For something that was supposed to become "just a feature" people sure do get passionate about de-duplication.
One of the areas I wanted to address was source side de-dupe, which a few posters have proclaimed the be-all and end-all of de-duplication. But before we handle that firecracker let's discuss the fragmentation of the de-dupe market.
First, de-duplication is showing up everywhere: primary storage, backup storage, archive storage, and even the wide-area network. Of course, there are different types of de-dupe implementations on each of these platforms, and each of the vendors thinks its solution is the best.
The guys that OEM technology seem to be at a disadvantage as the de-duplication use case expands. They end up with a de-dupe technology for primary storage, a different one for NAS storage, another one for backup (maybe two? or three?), and one for archives. This has to be confusing. Their advantage is they can move into this market faster -- but at what cost, a totally defragmented data reduction strategy?