Special Coverage Series

Network Computing

Special Coverage Series

Commentary

Tom Hollingsworth
Tom Hollingsworth

TRILL's Hidden Cost

TRILL allows for much needed advances over STP, but vendor licensing requirements make it hard to switch.

The world of today's data center is focused on moving bits back and forth as quickly as possible without interruption. We've taken every measure to ensure that data is flowing as freely as it can between 10, 40 or even 100 Gbit links connecting clusters of virtual servers. We want to crunch, analyze and query data instantly. All of this works as well as it can until we start looking under the hood. There we discover the root of all evil: spanning tree.

802.1d spanning tree (STP) and its progeny are elegant solutions for a bygone age. Radia Perlman created STP in a time when people had never heard of an Ethernet bridge. It prevented people from creating ignorance-based disasters. As time marched on, we created new additions to STP to decrease convergence time, improve recovery from catastrophic configuration events and even take into account changes in technology.

More Insights

Webcasts

More >>

White Papers

More >>

Reports

More >>

However, at the core, STP was still based on the idea that there should only be one link to a central bridge. STP effectively reduced the entire bandwidth of a data center to a single Layer 2 link between devices.

Perlman has gone on to create the true successor to her original design. With the help of some brilliant people inside the IETF, she created RFC 6326 -- Transparent Interconnection of Lots of Links (TRILL). TRILL allows for much-needed advances like mulitpathing and rapid failover between links in the event of failure. And TRILL can do it without any human interaction. So why aren't we using TRILL?

Network Computing contributor Ethan Banks recently posted a possible answer on Twitter: "If we want STP to die, replacement protocols can't continue to be license-only features." Ignoring the issues with hardware refreshes and suitability of TRILL to non-data center environments for the moment, I think Ethan has hit upon a very critical point in the development of TRILL and its vendor-specific kin.

Cisco has a TRILL-based solution in FabricPath. Brocades uses TRILL as the basis for VCS. Other vendors are developing products in much the same way: Use TRILL as a baseline to create the Layer 2 routing bridge (RBridge) underlay network, build proprietary functionality on top of it, then charge a license fee for the whole kit. Essentially, make people pay big bucks to get rid of spanning tree.

Ask anyone and they'll tell you STP is evil. Perlman resolved to replace it after reading the story of a hospital network core melting down due to STP misconfiguration and impacting caregivers. Every network engineer has a story of forgetting to configure it on a link and watching the ensuing bridging looping destroying a network. If STP is so bad, why are we being charged to replace it?

[TRILL can get us past STP, but will it have a place in a world moving toward software-defined networking? Ethan Banks digs into the issue in “Will SDN Kill TRILL?”]

I understand there are significant development costs involved in creating FabricPath, VCS and products just like them. Vendors want to ensure that their development work is adequately compensated. But if I have the choice of paying thousands of dollars and hundreds of hours rearchitecting my network or just enabling MSTP or RSTP and calling it a day, I'll pick the cheaper of two evils every time.

I've brought this up before with people who make these decisions. Why not reduce or outright eliminate the licensing requirement for TRILL or TRILL-like features to hasten the demise of STP? The response was nothing if not predictable: "We have demo licenses to allow customers to try these features out and see how they can benefit." Anyone ever redesigned an entire network for the sake of enabling a time-limited demo feature?

Spanning tree reached the end of its useful life many years ago. Newer, better protocols have been developed, and hardware has been created that can run both STP and TRILL without the need to change hardware modules on the fly.

The only thing that is still stuck in the past is the desire to license these features and derive revenue from them. If DEC had charged extra for STP in the beginning, I'm sure no one would have purchased it. It's only when a technology is included for everyone at no additional cost that engineers and architects begin to test it and find a way to use it to replace existing designs.

SDN and the coming wave of overlays and underlays won't change the need to ensure your network is built on a solid foundation. Tunnels fall apart when they are built on quicksand. Building your data center on the quagmire that is STP is going to result in pain and expensive reconfiguration down the road. I think we need to take this changing of the guard as a sign that we need to replace STP in our data centers as well.

What do you think? Is it past time to replace STP? Will you pay to replace it? Or should vendors make the superior protocols freely available to hasten the demise of STP and the rise of TRILL? Use the comments section below to voice your opinion.



Related Reading



Network Computing encourages readers to engage in spirited, healthy debate, including taking us to task. However, Network Computing moderates all comments posted to our site, and reserves the right to modify or remove any content that it determines to be derogatory, offensive, inflammatory, vulgar, irrelevant/off-topic, racist or obvious marketing/SPAM. Network Computing further reserves the right to disable the profile of any commenter participating in said activities.

 
Disqus Tips To upload an avatar photo, first complete your Disqus profile. | Please read our commenting policy.
 

Editor's Choice

2013 SDN Survey: Growing Pains

2013 SDN Survey: Growing Pains

In this report, we'll look at the state of Software-defined networking in the marketplace and examine the survey results. We'll also outline what a typical SDN infrastructure looks like in 2013, including a controller, programmable infrastructure, applications and network multitenancy. We'll dig into both OpenDaylight and OpenFlow, two open source initiatives that are shaping SDN from outside. Finally, We'll offer guidance for network professionals on how to approach bringing SDN into their own environments.
Get full survey results now! »

Vendor Turf Wars

Vendor Turf Wars

The enterprise tech market used to be an orderly place, where vendors had clearly defined markets. No more. Driven both by increasing complexity and Wall Street demands for growth, big vendors are duking it out for primacy -- and refusing to work together for IT's benefit. Must we now pick a side, or is neutrality an option?
Get the Digital Issue »

WEBCAST: Software Defined Networking (SDN) First Steps

WEBCAST: Software Defined Networking (SDN) First Steps


Software defined networking encompasses several emerging technologies that bring programmable interfaces to data center networks and promise to make networks more observable and automated, as well as better suited to the specific needs of large virtualized data centers. Attend this webcast to learn the overall concept of SDN and its benefits, describe the different conceptual approaches to SDN, and examine the various technologies, both proprietary and open source, that are emerging.
Register Today »

Related Content

From Our Sponsor

How Data Center Infrastructure Management Software Improves Planning and Cuts Operational Cost

How Data Center Infrastructure Management Software Improves Planning and Cuts Operational Cost

Business executives are challenging their IT staffs to convert data centers from cost centers into producers of business value. Data centers can make a significant impact to the bottom line by enabling the business to respond more quickly to market demands. This paper demonstrates, through a series of examples, how data center infrastructure management software tools can simplify operational processes, cut costs, and speed up information delivery.

Impact of Hot and Cold Aisle Containment on Data Center Temperature and Efficiency

Impact of Hot and Cold Aisle Containment on Data Center Temperature and Efficiency

Both hot-air and cold-air containment can improve the predictability and efficiency of traditional data center cooling systems. While both approaches minimize the mixing of hot and cold air, there are practical differences in implementation and operation that have significant consequences on work environment conditions, PUE, and economizer mode hours. The choice of hot-aisle containment over cold-aisle containment can save 43% in annual cooling system energy cost, corresponding to a 15% reduction in annualized PUE. This paper examines both methodologies and highlights the reasons why hot-aisle containment emerges as the preferred best practice for new data centers.

Monitoring Physical Threats in the Data Center

Monitoring Physical Threats in the Data Center

Traditional methodologies for monitoring the data center environment are no longer sufficient. With technologies such as blade servers driving up cooling demands and regulations such as Sarbanes-Oxley driving up data security requirements, the physical environment in the data center must be watched more closely. While well understood protocols exist for monitoring physical devices such as UPS systems, computer room air conditioners, and fire suppression systems, there is a class of distributed monitoring points that is often ignored. This paper describes this class of threats, suggests approaches to deploying monitoring devices, and provides best practices in leveraging the collected data to reduce downtime.

Cooling Strategies for Ultra-High Density Racks and Blade Servers

Cooling Strategies for Ultra-High Density Racks and Blade Servers

Rack power of 10 kW per rack or more can result from the deployment of high density information technology equipment such as blade servers. This creates difficult cooling challenges in a data center environment where the industry average rack power consumption is under 2 kW. Five strategies for deploying ultra-high power racks are described, covering practical solutions for both new and existing data centers.

Power and Cooling Capacity Management for Data Centers

Power and Cooling Capacity Management for Data Centers

High density IT equipment stresses the power density capability of modern data centers. Installation and unmanaged proliferation of this equipment can lead to unexpected problems with power and cooling infrastructure including overheating, overloads, and loss of redundancy. The ability to measure and predict power and cooling capability at the rack enclosure level is required to ensure predictable performance and optimize use of the physical infrastructure resource. This paper describes the principles for achieving power and cooling capacity management.