Feeds

King K super: does it refute hybrid HPC model?

GPUs are still GPU-riffic

Boost IT visibility and business value

ISC'11 It's been an eventful International Supercomputing Conference (ISC'11) in Hamburg. The Japanese sprang their K Computer on an unsuspecting HPC world, throwing down 8.126 Pflops on the table and raising the high-water performance mark by a factor of three.

Just as surprising was the fact that they did it the old-fashioned waywith semi-proprietary processors, a custom interconnect, and no fancy accelerators.

Was it only six months ago when the Chinese, with their 2.56 Pflop Tianhe system, appeared to have locked down the top spot for at least a year or more with heavy use of GPU accelerators?

This led many pundits (myself included) to say that the age of hybrid HPC was upon us, and that we probably wouldn’t see another non-hybrid system topping the chart anytime soon.

So is the K computer a signpost pointing to the resurgence of traditional CPU plus custom interconnect HPC? Or is it an aberration on the road to our hybrid future?

King K is hard to argue against. It has 93 per cent computational efficiency (RMax/RPeak), which is much more efficient than the rest of the top 10 systems, which sport numbers ranging from the low 40s to the low 80s.

Although it takes a staggering amount of power to run, almost 10 megawatts, it delivers 824 Mflops per watt, which makes it a close second on the top ten; it's barely edged out by the fifth largest system, the NEC/HP system at the Tokyo Institute of Technology.

And, as others have pointed out, this system isn't done growing - there are still roughly another 200 racks of room available. But I tend to think that the K computer is an superbly-executed aberration.

I think the biggest of the big supercomputers will ultimately be hybrid CPU + accelerator systems. In the final analysis, performance comes down to parallelism, cores, and core density.

Specialized cores, like those in GPUs, don't need all the trappings of general-purpose CPUs and can thus be crammed closer together. The NVIDIA Fermi GPU sports 512 cores running at 1.3 GHz, while Intel's Westmere has 6 cores running at 3.4GHz.

Clock for clock, that's an advantage of 32x in favor of the Fermi. One of the better CPU vs. GPU discussions is here, in the Top 10 Objections to GPU Computing Reconsidered.

It was authored by Dr. Vincent Natoli, a computational physicist who has spent 20 years working in HPC. In the article, he lays out the major arguments against GPU computing and responds to them with clear explanations and convincing logic.

It also serves as a primer on the value proposition behind the move toward hybrid HPC and is well worth a few minutes reading time. ®

The essential guide to IT transformation

More from The Register

next story
The Return of BSOD: Does ANYONE trust Microsoft patches?
Sysadmins, you're either fighting fires or seen as incompetents now
Microsoft: Azure isn't ready for biz-critical apps … yet
Microsoft will move its own IT to the cloud to avoid $200m server bill
Oracle reveals 32-core, 10 BEEELLION-transistor SPARC M7
New chip scales to 1024 cores, 8192 threads 64 TB RAM, at speeds over 3.6GHz
Docker kicks KVM's butt in IBM tests
Big Blue finds containers are speedy, but may not have much room to improve
US regulators OK sale of IBM's x86 server biz to Lenovo
Now all that remains is for gov't offices to ban the boxes
Gartner's Special Report: Should you believe the hype?
Enough hot air to carry a balloon to the Moon
Flash could be CHEAPER than SAS DISK? Come off it, NetApp
Stats analysis reckons we'll hit that point in just three years
Nimble's latest mutants GORGE themselves on unlucky forerunners
Crossing Sandy Bridges without stopping for breath
prev story

Whitepapers

5 things you didn’t know about cloud backup
IT departments are embracing cloud backup, but there’s a lot you need to know before choosing a service provider. Learn all the critical things you need to know.
Implementing global e-invoicing with guaranteed legal certainty
Explaining the role local tax compliance plays in successful supply chain management and e-business and how leading global brands are addressing this.
Build a business case: developing custom apps
Learn how to maximize the value of custom applications by accelerating and simplifying their development.
Rethinking backup and recovery in the modern data center
Combining intelligence, operational analytics, and automation to enable efficient, data-driven IT organizations using the HP ABR approach.
Next gen security for virtualised datacentres
Legacy security solutions are inefficient due to the architectural differences between physical and virtual environments.