Jasmine McTigue

Network Computing Blogger

Upcoming Events

Where the Cloud Touches Down: Simplifying Data Center Infrastructure Management

Thursday, July 25, 2013
10:00 AM PT/1:00 PM ET

In most data centers, DCIM rests on a shaky foundation of manual record keeping and scattered documentation. OpManager replaces data center documentation with a single repository for data, QRCodes for asset tracking, accurate 3D mapping of asset locations, and a configuration management database (CMDB). In this webcast, sponsored by ManageEngine, you will see how a real-world datacenter mapping stored in racktables gets imported into OpManager, which then provides a 3D visualization of where assets actually are. You'll also see how the QR Code generator helps you make the link between real assets and the monitoring world, and how the layered CMDB provides a single point of view for all your configuration data.

Register Now!

A Network Computing Webinar:
SDN First Steps

Thursday, August 8, 2013
11:00 AM PT / 2:00 PM ET

This webinar will help attendees understand the overall concept of SDN and its benefits, describe the different conceptual approaches to SDN, and examine the various technologies, both proprietary and open source, that are emerging. It will also help users decide whether SDN makes sense in their environment, and outline the first steps IT can take for testing SDN technologies.

Register Now!

More Events »

Subscribe to Newsletter

  • Keep up with all of the latest news and analysis on the fast-moving IT industry with Network Computing newsletters.
Sign Up

See more from this blogger

Microsoft ReFS and Oracle ZFS: How They Compare

Storage pros generally regard hardware-driven, RAID-based file systems as being highly reliable. The truth is, significant potential for error still exists, especially as systems scale for cloud architectures and big data applications. This potential for error is compounded by a lack of operating system or hypervisor awareness of low-level data error, as abstracted storage resources are passed up the stack and trust in the integrity of data is assumed.

It's also true that simple RAID parity schemes are inadequate to accommodate volumes in the thousands of disks. If you'd like to take a very deep dive on how storage methodologies are failing the cloud revolution, check out Stanford professor David S. H. Rosenthal's paper on data storage, "Keeping Bits Safe: How Hard Can It Be?"

More Insights


More >>

White Papers

More >>


More >>

To stay current, storage delivery engines need to get smarter and more transparent to the rest of the stack, and that means better file systems capable of detecting and remedying errors in real time for volumes that cannot simply be brought offline regularly for something as routine as error detection. Two options include Oracle's ZFS and Microsoft's ReFS. I'll compare both.

Sun started working on real-time error detection as early as 2005 with the initial development of Solaris 11, codenamed Nevada. The prototype product's open-source, next-generation file system was dubbed ZFS, and has continued to undergo development and improvement through the acquisition by Oracle and into the present day.

[SSDs can have an uneasy relationship with RAID. Howard Marks says RAID has to change. Find out why in "SSDs Vs. RAID? Fix RAID."]

Microsoft is attempting to create something similar in its new ReFS file system. ReFS is featured in Windows Server 2012 Storage Spaces. At its core, ReFS attempts to solve the same essential issues as ZFS while maintaining NTFS file system compatibility for legacy Windows applications, services and infrastructures. However, ReFS has some catching up to do compared to Oracle's mature product.

Both ReFS and ZFS offer several basic improvements over legacy file systems. First, both detect and repair data errors in real time, without dismounting or interfering with access to the volume. For ZFS, this is done at both the data level and metadata level through the use of hierarchical checksums that persist all the way up the file structure to the root node.

When a file is read, the checksum of the read file is compared to the stored checksum and errors are detected automatically. In ReFS, checksums are taken on file metadata. An additional optional feature called Integrity Streams makes sure that when changes are written, they are written to a different member volume, ensuring that an original is not compromised. This error detection fights against bit rot, the degradation of stored data on disk, as well as hardware failures, firmware glitches, cosmic radiation and other integrity damaging events without compromising the availability of the online file system.

While the two file systems share similar features, ZFS distinguishes itself in performance across parity striped volumes. According to Microsoft documentation, parity on ReFS comes from the paired Microsoft feature Storage Spaces, and requires that existing parity information "...be read and processed before a new write can occur."

While Microsoft advocates using dedicated SSDs for journals to mitigate poor write performance, ZFS combats this effect directly by writing parity information with a variable-sized stripe. This mitigates write speed penalties and also corrects the infamous "RAID 5 Hole," a limitation in conventional RAID systems that allows data to be lost in the event of a crash or power failure, which actually constitutes a whole new paradigm in the way that RAID sets implement parity.

The two volumes differ widely in terms of caching implementations as well. ReFS can accommodate SSD caching, but architectural details regarding ReFS caching functionality are sparse, though they do state that DRAM caching is not supported.

By contrast, ZFS caching is well documented and DRAM cache is a core feature. ZFS features three levels of caching: ARC, which is intelligent memory caching to RAM and uses as much free RAM as available; L2ARC, which is disposable (non-storage) SSD read caching; and ZIL, which is SSD write caching that buffers writes to the underlying volume.

ZFS supports using PCIe SSDs or battery-backed RAM disk devices with ultra-low latency as ZIL to dramatically reduce write latencies to ZFS volumes. Microsoft's only statement on this is that "customers can use third party solutions for this."

Both file systems support snapshots, encryption, very large disk sets, enormous numbers of files and a maximum volume size of 16 exabytes. ReFS, however, does not support compression or de-duplication, both of which are core features in ZFS, even though Windows Storage Spaces does support de-dupe on NTFS volumes.

The net result is a vastly dissimilar product offering that seems to offer similar features without closer inspection. On the surface, ReFS and ZFS may seem to be competitors, but Microsoft's first-generation ReFS offering is hardly in the same ballpark as the more mature ZFS. Still, compared with Oracle's 12+ year developmental lead, ReFS is a noble effort to keep Windows competitive.

Jasmine McTigue is the IT manager for Carwild Corp. She is responsible for IT infrastructure and has worked on numerous customer projects as well as ongoing network management and support throughout her 10-plus-year career.

Related Reading

Network Computing encourages readers to engage in spirited, healthy debate, including taking us to task. However, Network Computing moderates all comments posted to our site, and reserves the right to modify or remove any content that it determines to be derogatory, offensive, inflammatory, vulgar, irrelevant/off-topic, racist or obvious marketing/SPAM. Network Computing further reserves the right to disable the profile of any commenter participating in said activities.

Disqus Tips To upload an avatar photo, first complete your Disqus profile. | Please read our commenting policy.
Vendor Comparisons
Network Computing’s Vendor Comparisons provide extensive details on products and services, including downloadable feature matrices. Our categories include:

Research and Reports

Network Computing: April 2013

TechWeb Careers