Feeds

Why blade servers still don't cut it, and how they might

The blade manifesto

Security for virtualized datacentres

A more radical approach

I don't think that these projections and predictions by Gartner are all that unlikely, or even remotely radical. And I don't think these standardizations are sufficient, and I am advocating for a far more radical approach to future server designs that takes advantage of modular system components, the manufacturing scale and commoditization that comes from standardization, and an approach that allows these components to be used in tower, rack, or blade servers alike.

I have outlined many of these ideas before, and I am weaving them all together, plus a few new ones, into a single set of goals for the server industry, goals that lean heavily on the blade approach but allow for some of the peripheral expansion that is necessary and which still drives rack and tower server sales.

One socket to rule them all: The processor socket needs to be standardized across different processor architectures. I can envision a single, standardized interconnection - think of it as taking all of the best elements of Advanced Micro Devices' HyperTransport, Intel's QuickPath Interconnect, and IBM's GX interconnect at the heart of its Power and mainframe systems.

There is no reason why a socket cannot be created that allows X64, Power, Sparc, and Itanium processors to all plug into extra sockets and make use of a single standard interconnect. Instead of trying to standardize servers at the chip level, picking one architecture over another, such an approach would allow instruction set architectures to live on. Perhaps for longer than they might otherwise, in fact.

Modular motherboard designs: The genius of blade servers is that they take the elements of a rack server - servers, storage, networking switches, and such - and modularize them, shrink them, and connect them to each other over midplanes and internal wiring. It might be time to break down the motherboard into blade components and modularize them, too.

Imagine a system design that allowed chip sockets, chipsets, memory modules (meaning the actual memory slots), peripheral slots to be plugged together to make what would have been a single motherboard in the past. Imagine further that these motherboards could be created out of modular elements that allowed CPU sockets, memory, and I/O slots to be scaled up independently of each other by just plugging in extra modules. Think of a motherboard as being a 3D set of interconnected sub-blades instead of a single, 2D piece of motherboards with chips and slots mounted permanently and unchanging-ly on it.

Blade and sub-blade standards: To make the volume economics work in favor of the customers, I want blade server standards. And to be precise, I want innovation and standardization at the blade level and at the sub-blade level if that idea could come to pass. I want something like the set of standards for commercial blades, like those the telecommunications industry has had for a very long time now because it has power over the vendors thanks to the purse and long buying cycles.

It is apparent to me that commercial vendors do not like the telco blade standards, and for good reasons. First, the form factors are not consistent with data center form factors, and second, the lack of standards means a vendor who sells a half populated chassis to a customer can expect to sell the additional blades for the chassis as customers as computers for no sales costs. This account control is why server makers are in love with blades. But if they weren't so short-sighted concerning standards, the blade market might be five times as large already.

Common components for blades, racks, and towers: There is no reason why different server styles - blade, rack, and tower - could not be created from a variety of sub-blade components. One set of motherboard features makes many different kinds of servers, and there are volume economics that span server form factors and vendors - just like processors, memory chips, and I/O peripherals plugging in to PCI slots do today.

No more AC power inside the data center: The AC-to-DC conversions that go on inside the data center are idiotic. It is time for a set of DC standards for server gear, and AC should be the exception in the data center, not the rule. In tower servers and small blade chassis, obviously, AC wall power will prevail.

Integrated refrigerant cooling for key components: I said this back in early 2005, and I will say it again: as many watts of energy that might dissipate into the air as heat and then have to be recaptured by data center air conditioning should be trapped and moved out of the data center using water or other refrigerants linked directly into the external chillers outside of data centers. Air cooling is just too inefficient, particularly at the compute densities in modern data centers.

Yeah, I know. This is a lot of pie in the sky. But, you never get what you don't ask for in this world.

Copyright © 1996-2008 Guild Companies, Inc. All Rights Reserved. ®

Providing a secure and efficient Helpdesk

More from The Register

next story
It's Big, it's Blue... it's simply FABLESS! IBM's chip-free future
Or why the reversal of globalisation ain't gonna 'appen
IBM storage revenues sink: 'We are disappointed,' says CEO
Time to put the storage biz up for sale?
'Hmm, why CAN'T I run a water pipe through that rack of media servers?'
Leaving Las Vegas for Armenia kludging and Dubai dune bashing
Microsoft and Dell’s cloud in a box: Instant Azure for the data centre
A less painful way to run Microsoft’s private cloud
Facebook slurps 'paste sites' for STOLEN passwords, sprinkles on hash and salt
Zuck's ad empire DOESN'T see details in plain text. Phew!
Windows 10: Forget Cloudobile, put Security and Privacy First
But - dammit - It would be insane to say 'don't collect, because NSA'
prev story

Whitepapers

Forging a new future with identity relationship management
Learn about ForgeRock's next generation IRM platform and how it is designed to empower CEOS's and enterprises to engage with consumers.
Why and how to choose the right cloud vendor
The benefits of cloud-based storage in your processes. Eliminate onsite, disk-based backup and archiving in favor of cloud-based data protection.
Three 1TB solid state scorchers up for grabs
Big SSDs can be expensive but think big and think free because you could be the lucky winner of one of three 1TB Samsung SSD 840 EVO drives that we’re giving away worth over £300 apiece.
Reg Reader Research: SaaS based Email and Office Productivity Tools
Read this Reg reader report which provides advice and guidance for SMBs towards the use of SaaS based email and Office productivity tools.
Security for virtualized datacentres
Legacy security solutions are inefficient due to the architectural differences between physical and virtual environments.