Special Coverage Series

Network Computing

Special Coverage Series

Commentary

Kurt Marko
Kurt Marko Contributing Editor

The IT Generation Gap

VMworld panel exposes a rift between innovative startups focused on cloud, mobile and social technologies and old IT that clings to its data centers and self-managed applications.

In its 10th year, VMworld has grown into perhaps the largest and most important enterprise IT event of the year. The conference venue, the Moscone Center, is in the heart of San Francisco's SoMa district, a tech oasis that's transformed a skid row full of warehouses and other structural detritus into a thriving area teeming with startups full of (largely) young software wizards with innovative ideas and the energy and moxie to pull them off. This new generation is steeped in the world of mobile, social and cloud, having little use for legacy, self-managed infrastructure, traditional ways of doing IT and, quite frankly, Windows and thick local applications.

There's a palpable generation gap between this new IT culture and old IT, which was on high-powered display during an executive roundtable at VMworld entitled "The Data Center is Dead, Long Live the Data Center."

More Insights

Webcasts

More >>

White Papers

More >>

Reports

More >>

The panel featured Web pioneer and Silicon Valley venture capitalist, Marc Andreessen, VMware CEO Pat Gelsinger, Sun co-founder and current chief development officer at Arista Andy Bechtolsheim and Graeme Hay, head of architecture and infrastructure for Credit Suisse. Despite his keynote touting the company's software-defined data center vision, roadmap and supporting products, Gelsinger clearly sees the cloud as a secondary adjunct to homegrown and operated enterprise data centers, and was supported in this view by one of his best customers at Credit Suisse. Andreessen was having none of it, and two exchanges highlighted the chasm between world views.

One of the self-congratulatory points Gelsinger made is that virtualization software like vSphere and vCloud have helped IT move the needle on the percentage of time it spends on operations and maintenance versus developing new services and innovating. From a ratio of about 70:30, he presented data showing that, at least among VMware customers, the split is now closer to 60:40, on its way to his stated goal of 50:50. Our survey, analyzed in this year's InformationWeek data center research report, puts the split at 64:36, with the level of innovation increasing four points in the past year.

Unfortunately for Gelsinger, his on-stage customer, Hay, refuted VMware's numbers. When asked, he said the ratio at Credit Suisse was more like 80:20, citing the firm's large and expensive legacy infrastructure and software, along with the overhead and red tape that come with running a financial enterprise.

Andreessen had heard enough, blurting out that at the companies he funds and works with at his VC firm, the numbers are more like 99% innovation, 1% operations. The IT old guard must have thought this was somewhere between naiveté and bravado since in their world, no one could conceivably outsource all their IT--save a few MacBooks and smartphones--to the cloud. But Andreessen was unyielding, reiterating the point that Gelsinger and his co-travelers are setting the bar far too low. Indeed, as one who almost exclusively lives in the cloud--where my primary work machine is a Chromebook, productivity suite is Google Apps for Business, and note and information database is Evernote--I can assure you he wasn't bluffing.

[Read how data centers are becoming the domain of service providers as smaller enterprises increasingly outsource their data center operations in "Data Center Study: The Big Get Bigger."]

The fissure between old and new IT was further illustrated when the discussion turned to the importance, or lack thereof, of internal enterprise data centers, now and in the future. Gelsinger contended that enterprises will be building private data centers for the next 40 or 50 years. The implication is clear: Forget the hype; the cloud will remain a secondary source of IT infrastructure for generations. This is completely at odds with the startup culture Andreessen lives in and ignores the increasing economies of scale accruing to mega center operators, whether commercial hosters like Switch or DuPont Fabros, or cloud services like Amazon, Google and Rackspace.

At this point, I joined the discussion. The moderator had asked the audience to Tweet questions. In light of Gelsinger's comment, I asked, "Why is building your own data center not like building your own power plants?" With only 140 characters I couldn't add any context, and the panelists proceeded to misinterpret my point, saying, well yes, they kind of are. Going to the microphone, I reprised author Nick Carr's argument. Just as companies stopped generating their own power decades ago as the cost of building and operating plants was undercut by the price at which utilities with massive, multitenant plants and intricate distribution grids could deliver the same energy, why would enterprises continue to build their own information plants (data centers) when utilities (cloud services, co-los) using the Internet (grid) can do it far more efficiently?

Once again, the old/new rift was evident. Everyone agreed that smaller data centers are increasingly untenable and that you need scale for facilities to make economic sense. But the old guard, represented by Gelsinger and Hay, clung to the notion that on-premise, self-managed data centers maintain sufficient--what I would contend are intangible or illusory--values of control, security and operational responsiveness and accountability to justify their existence. Aside from the fact that most enterprise data centers today are tiny--almost 72% of InformationWeek's survey respondents operate facilities smaller than 10,000 square feet--these "values" aren't demonstrable.

As Andreessen pointed out, most clouds are more secure and reliable than enterprise infrastructure. Gelsinger countered by citing the recent Amazon outage. Again, Andreessen had a response. It's the classic "man bites dog" news story: The Nasdaq exchange goes down for more than three hours and it's just the price of complexity in our technological era; a fraction of Gmail users see a glitch for a few minutes and it's proof of the cloud's fragility.

Like all generation gaps, there's no bridging old and new IT; the old just fades away. Old IT will cling to its data centers, on-premise infrastructure and self-managed applications until the weight of inefficiency and inflexibility swamps their businesses under a tide of nimble newcomers using utility resources, SaaS applications and rapidly adaptive strategies. SoMa and Silicon Valley are full of such companies run by brilliant risk takers and advised by visionaries like Andreessen and his Sand Hill Road colleagues. Big enterprise IT needs to get out of its bubble and see this rapidly growing new world before it's forced into an early retirement.



Related Reading



Network Computing encourages readers to engage in spirited, healthy debate, including taking us to task. However, Network Computing moderates all comments posted to our site, and reserves the right to modify or remove any content that it determines to be derogatory, offensive, inflammatory, vulgar, irrelevant/off-topic, racist or obvious marketing/SPAM. Network Computing further reserves the right to disable the profile of any commenter participating in said activities.

 
Disqus Tips To upload an avatar photo, first complete your Disqus profile. | Please read our commenting policy.
 

Editor's Choice

2013 SDN Survey: Growing Pains

2013 SDN Survey: Growing Pains

In this report, we'll look at the state of Software-defined networking in the marketplace and examine the survey results. We'll also outline what a typical SDN infrastructure looks like in 2013, including a controller, programmable infrastructure, applications and network multitenancy. We'll dig into both OpenDaylight and OpenFlow, two open source initiatives that are shaping SDN from outside. Finally, We'll offer guidance for network professionals on how to approach bringing SDN into their own environments.
Get full survey results now! »

Vendor Turf Wars

Vendor Turf Wars

The enterprise tech market used to be an orderly place, where vendors had clearly defined markets. No more. Driven both by increasing complexity and Wall Street demands for growth, big vendors are duking it out for primacy -- and refusing to work together for IT's benefit. Must we now pick a side, or is neutrality an option?
Get the Digital Issue »

WEBCAST: Software Defined Networking (SDN) First Steps

WEBCAST: Software Defined Networking (SDN) First Steps


Software defined networking encompasses several emerging technologies that bring programmable interfaces to data center networks and promise to make networks more observable and automated, as well as better suited to the specific needs of large virtualized data centers. Attend this webcast to learn the overall concept of SDN and its benefits, describe the different conceptual approaches to SDN, and examine the various technologies, both proprietary and open source, that are emerging.
Register Today »

Related Content

From Our Sponsor

How Data Center Infrastructure Management Software Improves Planning and Cuts Operational Cost

How Data Center Infrastructure Management Software Improves Planning and Cuts Operational Cost

Business executives are challenging their IT staffs to convert data centers from cost centers into producers of business value. Data centers can make a significant impact to the bottom line by enabling the business to respond more quickly to market demands. This paper demonstrates, through a series of examples, how data center infrastructure management software tools can simplify operational processes, cut costs, and speed up information delivery.

Impact of Hot and Cold Aisle Containment on Data Center Temperature and Efficiency

Impact of Hot and Cold Aisle Containment on Data Center Temperature and Efficiency

Both hot-air and cold-air containment can improve the predictability and efficiency of traditional data center cooling systems. While both approaches minimize the mixing of hot and cold air, there are practical differences in implementation and operation that have significant consequences on work environment conditions, PUE, and economizer mode hours. The choice of hot-aisle containment over cold-aisle containment can save 43% in annual cooling system energy cost, corresponding to a 15% reduction in annualized PUE. This paper examines both methodologies and highlights the reasons why hot-aisle containment emerges as the preferred best practice for new data centers.

Monitoring Physical Threats in the Data Center

Monitoring Physical Threats in the Data Center

Traditional methodologies for monitoring the data center environment are no longer sufficient. With technologies such as blade servers driving up cooling demands and regulations such as Sarbanes-Oxley driving up data security requirements, the physical environment in the data center must be watched more closely. While well understood protocols exist for monitoring physical devices such as UPS systems, computer room air conditioners, and fire suppression systems, there is a class of distributed monitoring points that is often ignored. This paper describes this class of threats, suggests approaches to deploying monitoring devices, and provides best practices in leveraging the collected data to reduce downtime.

Cooling Strategies for Ultra-High Density Racks and Blade Servers

Cooling Strategies for Ultra-High Density Racks and Blade Servers

Rack power of 10 kW per rack or more can result from the deployment of high density information technology equipment such as blade servers. This creates difficult cooling challenges in a data center environment where the industry average rack power consumption is under 2 kW. Five strategies for deploying ultra-high power racks are described, covering practical solutions for both new and existing data centers.

Power and Cooling Capacity Management for Data Centers

Power and Cooling Capacity Management for Data Centers

High density IT equipment stresses the power density capability of modern data centers. Installation and unmanaged proliferation of this equipment can lead to unexpected problems with power and cooling infrastructure including overheating, overloads, and loss of redundancy. The ability to measure and predict power and cooling capability at the rack enclosure level is required to ensure predictable performance and optimize use of the physical infrastructure resource. This paper describes the principles for achieving power and cooling capacity management.