POD giveth: Is factory-built HPC the future?
Mini data-centers getting big love
Hewlett-Packard is looking to steal a step on everyone in the industry with their POD-Works factory approach to building new data centers or adding capacity to old ones.
PODs are HP’s term for the now-ubiquitous shipping container-style mini data centers that started catching on (at least with vendors) a few years ago. HP is taking the concept a step further by applying large-scale industrialization to the process of producing PODs in volume.
Our pal TPM provides further detail on PODs here.
HP’s POD-Works is co-located in Houston with one of its major manufacturing and distribution centers. This way it has quick access to loads of hardware, as this is a primary receiving center for HP x86 gear. It has a new facility it has purpose-built for building out, testing, and qualifying the pods before they’re shipped to the customer. I think this has the potential to change the economics of buying at scale. Why?
The process of delivering, installing, and certifying a supercomputer is a lot like building a house. Materials in the form of racks, boards, processors, memory, cables, drives, cords, and everything else are delivered to the empty data center just like concrete, wood, sheetrock, and other supplies would be delivered to an empty lot. Then a group of skilled or semi-skilled workers start the building process from the ground up.
It’s pretty labor-intensive. Just testing all of the parts can consume quite a bit of time. If you’re building a modest system with, say, two thousand nodes, you might be looking at 4,000 processors and 32,000 individual DIMMs. A system this size might put you in the middle of the pack on the Top500 list.
When the numbers get that big (and a 2,000 node box isn’t all that big these days) you’re going to see problems simply due to the scale of the system. You’ll see memory speed mismatches, or the wrong model processors in the shipment, or some components that are DOA. Even if the vendor ships the right working parts 99.9 per cent of the time, it still means that you’ll see 32 bad DIMMS and three or four bad processors.
If you’re the one building this system (either as the end customer or an integrator/vendor), you have to inspect all the stuff to make sure it’s right and then test it to make sure it works. You’ll have to ship the wrong/defective/broken parts back and get new stuff shipped out to you. Delays ensue, voices are raised and, often, tears follow.
Contrast this to building the same system in a dedicated factory right next door to a massive system distribution facility. You received the wrong part or have a busted one? Walk next door and get a new one. Components not working as they should? The folks at the factory have built lots of these things in every imaginable configuration and have seen it all – they’ll get it working. Testing and certification can also take place at the factory with customer representatives flown in to participate in the process.
HP estimates that the POD factory will result in an 8x speed-up in system deployment. They also say that it is 37 per cent more efficient and 45 per cent less expensive than a typical data center. I’m not sure exactly how that’s measured, but it’s obvious to me that there would be significant savings from this model.
Of course, there are some trade-offs inherent in this too. The first is that you’re going to get your systems in POD-sized chunks – up to 1,100U at a time. Rather than row after row of servers and storage in a single large room, you’d have POD after POD arrayed in what could be just a shell of a warehouse providing physical security and protection from the elements but little else. (Well, a restroom and drinking fountain wouldn’t be too much to ask.)
In a lot of ways, it really comes down to aesthetic and logistic considerations. A house built in a factory with standardized parts can definitely be better from a quality standpoint and be less expensive to build and own. But the trade-off comes from the constraints arising from that model – if you want something highly customized, you’re out of luck.
I would also assume that with HP PODs, if you want them populated with, say, Dell servers, you are similarly out of luck. However, in HPC systems, the vast majority of the gear is already built from standard x86 system building blocks that are easily configurable to provide for variations in needs or budgets.
I think we may be looking at the beginnings of a significant trend here – an evolution in the HPC buying model at least, assuming that customers are willing to put speedy deployment and bang for the buck ahead of customization, logistics, and looks. If the savings and speed pan out, I expect that we’ll see other vendors respond by putting together factories of their own. HP said that it'll have a POD at SC10. I’m going to take a closer look at it, video camera in hand, and drill down into the economics of it a bit more too.
If you’re interested in seeing more, HP has some short videos up showing how a POD is constructed (“Birth of an HP POD”), a look at the tech inside a pod (“Engineer tour of the HP POD”), plus there are plenty of others that you can easily find on its site. ®
Sponsored: Benefits from the lessons learned in HPC