Vishwa Nigam

How to Improve the Backup Success Rate of Data Centers?

Improve Backup Success Rate

According to industry analysts, a significant number of backup jobs (from 5 to 25%) are failing across various tiers of data centers. This impacts datacenters heavily on revenue loss, SLA-based penalties, and customer experience. Furthermore, loss of important data alienates customers and leads to low NPS (net promoter score).

Below image shows the current datacenter landscape with success rate and failure rate across various tiers. [Source: Gartner]

It is clear that the backup failure rate is least for tier 1 data centers at 5% and highest for tier 2 remote offices at 25%.

To reduce backup failure, it is necessary to find the root cause of the failure.

Following are a few frequent causes identified after monitoring some of the top datacenters.

Once the reason for failure is identified, the next step is to create a solution strategy to eliminate the problem and create successful backups. Below are some of the proven strategies to minimize backup failure.

Backup Audit tool

An audit tool helps data centers to conduct automated periodic audits of target systems and backup servers. Some of the features of audit tools are:

Firewall Configuration Audit: The audit tool runs a script at regular intervals to test the accessibility of the selected range of IP subnets, ports, target systems and backup servers. If any issue is found in the bidirectional communication, a ticket is raised automatically and sent to the concerned team.

Server Decommissioning: For servers that are consistently “inaccessible” and lead to backup failure, the audit tool automatically checks with the relevant internal application (Contract Management) for contractual information. If the application responds with a “contract expiry” message, the tool will initiate a ticket for decommissioning the server, disable further backups on it, and release the attached backup resources.

Backup Agent Audit: The audit tool runs a script on target systems and backup servers periodically to check if there is any mismatch in agent versions. If it finds a mismatch, a ticket gets generated automatically and sent to the admin team for customer approvals.

Database Permission Audit: Before running backup jobs, audit tool checks for access permission levels for the target system. If the tool identifies any gap, a ticket gets triggered to the admin team to get required permission from the customer for the associated target system.

Proactive capacity management

In order to effectively manage capacity and solve storage-related issues, data centers should focus on important areas to track and analyze various parameters such as location, libraries, media used, and storage-related information. Following measures would help data centers to reduce backup failures due to storage mismanagement.

Archival strategy for backup data

Primary storage is expensive as storage arrays are required to produce a sufficient level of input/output operations per second (IOPS) to meet operational requirements for user read/write activity. The data archives serve as an effective way for reducing primary storage consumption and related costs.

Cloud storage is a possible backup solution as it is cheaper and offers flexibility with an ongoing investment. Some solutions include Amazon Glacier, Microsoft StorSimple, and Google Drive.

Recommended archival strategy

Data classification: Classify the data with the most legal and regulatory exposure risk and update archival policy regularly for better compliance.

Well-defined retention policy: Align archival policy with different departments and BUs to retain information for varying time periods.

Tools for structuring the data: Use automation tools to structure data — indexing, auto-classification, text and content analytics helps to extract more value from data and store it efficiently.

Data centers can reduce approximately 25% of their storage costs by implementing an effective archival strategy. For instance, focusing on applications that accumulate unstructured data such as audiovisuals or images and archiving them into the cloud will reduce the load on the primary storage.

Parallel processing of backup jobs

The following diagrams illustrate the difference between serial backup processing and parallel backup processing. In parallel processing, the same storage is logically divided into different instances. This makes parallel processing of jobs more effective and faster than the serial backup processing.

Conclusion

This article provides insights on how having a customized backup tool can minimize failure, improve storage performance, and achieve 99% successful backup.

By Vishwa Nigam, Manager – Business Analysis & Insights, Prodapt

Vishwa is an experienced business manager with a demonstrated history of building & delivering actionable insights on DevOps, IoT, and robotic process automation. His areas of interest are analytics, process improvement, and business model innovation.

Cloud Syndicate

The 'Cloud Syndicate' is a mix of short term guest contributors, curated resources and syndication partners covering a variety of interesting technology related topics.

Contact us for syndication details on how to connect your technology article or news feed to our syndication network.

Quantum Computing opens new front in Cloud!

Quantum Computing opens new front in Cloud!

Quantum Computing As the amount of data in the world is rapidly increasing, so is the time required for machines to process it. Augmented Reality, Virtual Reality, Artificial Intelligence, Robotics, Real-Time Analytics, and Machine Learning algorithms are ...
The Shift from Monolithic to Microservices: What It Means for CTOs.

The Shift from Monolithic to Microservices: What It Means for CTOs.

The Shift to Microservices The shift in application development strategies is moving from monolithic design to isolated and resilient components known as microservices. As a result, applications that were designed with platform entanglements such as ...
How IoT and OT collaborate to usher in the data-driven factory of the future

How IoT and OT collaborate to usher in the data-driven factory of the future

The Data-driven Factory The next BriefingsDirect Internet of Things (IoT) technology trends interview explores how innovation is impacting modern factories and supply chains. We’ll now learn how a leading-edge manufacturer, Hirotec, in the global automotive industry, takes advantage of ...
Finding and Implementing Startup Tools

Finding and Implementing The Right Tools For Your Startup

Implementing Startup Tools Many startups believe implementing cloud tools help reduce operation costs as well as the time taken to enter a market, and so when taken along with the faster product development and increased ...
Bryan Doerr

Can You Afford the Risk of Not Going to the Cloud?

Risk of Not Going to the Cloud If you’re considering a migration to a public cloud environment, you’re most likely motivated by the potential to reduce costs, while increasing capital efficiency, productivity, agility, and overall ...
Star Wars IoT CES
The Lighter Side Of The Cloud - The Letter "G"
The Lighter Of The Cloud - Virtual Lunch Break
The Lighter Side Of The Cloud - Passwords
The Lighter Side Of The Cloud - Going Viral
The Lighter Side Of The Cloud - The Nanodegree
The Lighter Side Of The Cloud - Due Diligence
The Lighter Side Of The Cloud - Machine Learning
The Lighter Side Of The Cloud - Day 5

CLOUDBUZZ NEWS

Worldwide Services Revenue Posts Steady Year-Over-Year Growth in the Second Half of 2017, According to IDC

Worldwide Services Revenue Posts Steady Year-Over-Year Growth in the Second Half of 2017, According to IDC

FRAMINGHAM, Mass. May 15, 2018 – Worldwide revenues for IT Services and Business Services totaled $502 billion in the second half of 2017 (2H17), an increase of 3.6% year over year (in constant currency), according to ...
Security in the Cloud—A Little Known Advantage, Actually

Security in the Cloud—A Little Known Advantage, Actually

Okay, I’ll go ahead and say it: Public cloud infrastructures are more secure, and the security is more cost-effective, than the majority of on-premises data centers. That should get the blood flowing. With the word ...
Cambridge Analytica files for bankruptcy in U.S. following Facebook debacle

Cambridge Analytica files for bankruptcy in U.S. following Facebook debacle

(Reuters) - Cambridge Analytica, the political consultancy at the center of Facebook Inc’s (FB.O) privacy scandal, filed for Chapter 7 bankruptcy in the United States late on Thursday. This past March allegations surfaced that Cambridge ...