Resurrecting Speed

In a recent entry I pronounced <a href="http://www.informationweek.com/blog/main/archives/2008/05/speeds_dead.html">'speed is dead</a>' as it relates to solving the backup window problem. As the entry indicates, the NEED to reduce the backup window continues to be a desire. The ABILITY to reduce the backup window is the challenge. Due to the network infrastructure, the ability of the servers being protected to send that data fast enough, as well as a host of other issues, are the big limiters no

George Crump, President, Storage Switzerland

June 11, 2008

3 Min Read
Dark Reading logo in a gray background | Dark Reading

In a recent entry I pronounced 'speed is dead' as it relates to solving the backup window problem. As the entry indicates, the NEED to reduce the backup window continues to be a desire. The ABILITY to reduce the backup window is the challenge. Due to the network infrastructure, the ability of the servers being protected to send that data fast enough, as well as a host of other issues, are the big limiters now in backup window reduction.To collapse the backup window substantially, we need to understand that we are looking at the wrong end of the straw. The modern backup target, be it disk, tape, or something else, is plenty fast enough for most enterprises. To reduce the backup window is going to require either a substantial investment in upgrading the surrounding infrastructure, or reducing the amount of backup data that crosses the network, reducing the amount of data present on primary storage.

Improving the infrastructure is a budget issue for most customers we talk to. Most, if not all, are at 1 Gigabit Ethernet, so a dramatic move in infrastructure performance would have to come from moving most servers to the SAN or implementation of 10 GbE.

Reducing the amount of primary storage capacity can complement implementing a backup solution that moves less data across the network. To some extent, you probably have the ability to do this today with your current software. You could, for example, stop doing weekly full backups and just do incremental backups. This is less than desirable since it requires many more media mounts when doing a recovery. Many backup applications have the ability to do a consolidated full, where a baseline full is created and then subsequent incrementals can be consolidated to create a new master full.

The big shift in the network requirement for backup is using block-level incremental backups. NetApp does this with Open Systems SnapVault technology and Syncsort does this with Backup Express XRS. Block-level incremental backs up just the blocks that changed since the last backup. This is simpler than doing source-side data deduplication since the comparisons are on a volume-by-volume basis instead of across the whole enterprise. The result is almost no impact on the server during backup, and most backups typically completing in less than five minutes. This allows backups to be done repeatedly throughout the day with very little growth in backup storage.

Also, the backup target for these types of solutions is typically an active target, meaning that the backup target can be used as if it were any other file system, allowing for in-place recoveries, using the backup data set for test and development work or manual copy of data. Probably the most important component is that the move to secondary storage (tape, disk, deduplicated disk or VTL) is integrated into the process.

A complement to this is reducing the amount of primary storage capacity altogether. Implementing a Disk Archive System, like those available from Caringo, Copan Systems, Permabit and others, can do this. I have seen cases where, if-disk based archiving were implemented, as much as 80% of the data being backed up could be permanently archived. I detailed this a while back on in my entry on Data Keepage. An upcoming entry will examine some of the hardware platforms available to address this.

George Crump is founder of Storage Switzerland, an analyst firm focused on the virtualization and storage marketplaces. It provides strategic consulting and analysis to storage users, suppliers, and integrators. An industry veteran of more than 25 years, Crump has held engineering and sales positions at various IT industry manufacturers and integrators. Prior to Storage Switzerland, he was CTO at one of the nation's largest integrators.

About the Author

George Crump

President, Storage Switzerland

George Crump is president and founder of Storage Switzerland, an IT analyst firm focused on the storage and virtualization segments. With 25 years of experience designing storage solutions for datacenters across the US, he has seen the birth of such technologies as RAID, NAS, and SAN. Prior to founding Storage Switzerland, he was CTO at one the nation’s largest storage integrators, where he was in charge of technology testing, integration, and product selection. George is responsible for the storage blog on InformationWeek's website and is a regular contributor to publications such as Byte and Switch, SearchStorage, eWeek, SearchServerVirtualizaiton, and SearchDataBackup.

Keep up with the latest cybersecurity threats, newly discovered vulnerabilities, data breach information, and emerging trends. Delivered daily or weekly right to your email inbox.

You May Also Like


More Insights