Network Computing is part of the Informa Tech Division of Informa PLC

This site is operated by a business or businesses owned by Informa PLC and all copyright resides with them. Informa PLC's registered office is 5 Howick Place, London SW1P 1WG. Registered in England and Wales. Number 8860726.

A Data De-Duplication Survival Guide: Part 1: Page 3 of 7

General-purpose data de-duplication systems will typically have (or should have) the ability to do inline data de-duplication, since that is generally the most efficient process. Also, ideally, the data de-duplication system should have variable-length segment identification in order to provide the most aggressive data de-duplication effect. For example, it should be able to pick up and store only the changed segments within a database, as opposed to storing the entire file new on each backup.

Lastly, general-purpose data de-duplication systems that include replication provide the optimal way to replicate backup data to remote sites. By leveraging data de-duplication, the data de-duplication system only needs to replicate the net new segments of data across the network.

The most efficient systems will perform de-duplicated replication, in-line, across multiple sites. So far, Data Domain fits the bill. In addition, in-line de-duplication enables the replication process to begin the moment the system starts receiving data. This is unlike VTL systems that typically use post-process data de-duplication and therefore incur a time delay before the replication process can begin -- thus putting the disaster recovery data at risk.

To Page 3

VTL solutions

Suppliers of VTL solutions, such as FalconStor (which supplies EMC and Sun), NetApp, and Sepaton, typically qualify a range of backup applications, but they are not neutral in terms of data source or target.