Cluster cake pulled open: Oozes cores, GPU gravy, mineral oil

Recipe for Student Cluster Compo destruction

Choosing a cloud hosting partner with confidence

SC12 At last, I can reveal the final configurations of the systems the SC12 Student Cluster Competition (SCC) warriors brought to Salt Lake City last week.

Here’s the gear they were running…

First, kudos to all of the sponsors who generously provided so much equipment at no cost to the cluster crown competitors. (There isn’t a real crown.) How about next year, all of you chip in and we get an actual Cluster Crown that can go to the winning team? Maybe make it out of old motherboards and chips and stuff?

So, what’s noteworthy on the builds above? First, it’s too bad that I don’t have the total memory figures in the table above. Memory per node and per core is important and might explain some of performance differences that we’ll see in the final results. But that aside, there’s plenty of other interesting information in the table.

We see that at least five of the eight teams used graphics processing units (GPUs) this time around. Last year, about half the teams were using accelerators, and they dominated the field on both LINPACK and application performance. Team Longhorn used GPUs for the first time this year and was excited about the performance boost they’ve seen when compared to the traditional system they ran last year. Team Venus was thinking about using GPUs, but decided against it due to the team's relative inexperience with clusters and HPC in general. Purdue was also pondering the whole accelerator issue, and I think this will be their last non-hybrid SCC cluster.

Yum... not one but two flavours of GPU

Team Boston, sticking with their tradition of bringing anything that will run code, brought two flavours of GPUs. They also looked at running ARM processors and even talked about mixing some Phi (Intel Phi, that is) into their cluster cake. While their configuration sports 256 cores, I don’t think they were able to use them all during the competition due to power considerations. Teams are capped at 26 total amps (110 watts), and I’m pretty sure that if Boston fired up all that hardware at once, they’d be well past that limit.

The Chinese representatives, Teams NUDT and USTC, both brought six NVIDIA 2090s. NUDT had the lowest node count in the competition and yet the highest LINPACK score at just over 3 Teraflop/s. Their countrymen from USTC finished second with 2.73 Teraflop/s.

Laptops, a canister of mineral oil... and disappointment

And then there’s Team Red Raider from Texas Tech University. Every year at the SCC, there’s a team that seems almost cursed. Deliveries fall through, things break, software doesn’t perform like it did in the lab, etc. This year, that team is Team Red Raider.

They had big plans for SCC12. First, they designed their own immersion liquid cooling mechanism, using parts they scrounged or built themselves. They used an automobile transmission radiator for cooling and a boat bilge pump to move the liquid through flexible plastic piping to their homegrown server deep-fryer enclosure. They did a lot of custom engineering and shade tree fixes in order to make this design extremely efficient. (For more on the Texas Tech cluster, click here.

The only problem was that their system didn’t show up in Salt Lake City. It was somehow stuck in Mexico, of all places, and couldn’t be shipped in enough time to make the contest. At one point, the tech test system was going to be shipped to the show, but that fell through at the last minute. So the Red Raiders were left with laptops, a big canister full of mineral oil, and a huge sense of disappointment. But that’s not the end of the story…

Cluster kindness

When the rest of the field heard about the Red Raider problem, the teams from UT and Purdue stepped in and contributed some gear to get Texas Tech back in the game. Team Boston also contributed a couple of nodes to the effort. Meanwhile, both Dell and Mellanox came up with some servers to help fill out the rest of the cluster. The Texas Tech team busily incorporated all of this disparate gear into some sort of clustery order and managed to complete all of the scientific applications. Their attitude stayed upbeat, and they made the best of a bad situation.

Their configuration changed too quickly to keep up with, and I don’t think the final configuration is documented. But they ran what they had and put in a respectable showing. I hope they get the chance to compete next year.

Even though their cooling solution was homegrown, it was clear that they put an enormous amount of thought into it, and it would have consumed considerably less power than previous immersion solutions we’ve seen at SCC. According to the team, they were able to achieve a 4.0 teraflop/s LINPACK score in their lab under competition conditions. This would have landed them the LINPACK award for sure; who knows what they could have done on the scientific apps? But everyone’s system does better in the lab than in the real competition – so we need the Red Raiders to bring their stuff to SC13 in Denver and prove it.

More Student Cluster Competition coverage to follow. In upcoming blogs, we’ll talk to the teams on the final day of the competition, learn more about the LittleFe Traveling Salesman challenge, and attend the gala awards luncheon too. ®

Top 5 reasons to deploy VMware with Tegile

More from The Register

next story
Just don't blame Bono! Apple iTunes music sales PLUMMET
Cupertino revenue hit by cheapo downloads, says report
The DRUGSTORES DON'T WORK, CVS makes IT WORSE ... for Apple Pay
Goog Wallet apparently also spurned in NFC lockdown
IBM, backing away from hardware? NEVER!
Don't be so sure, so-surers
Hey - who wants 4.8 TERABYTES almost AS FAST AS MEMORY?
China's Memblaze says they've got it in PCIe. Yow
Microsoft brings the CLOUD that GOES ON FOREVER
Sky's the limit with unrestricted space in the cloud
This time it's SO REAL: Overcoming the open-source orgasm myth with TODO
If the web giants need it to work, hey, maybe it'll work
'ANYTHING BUT STABLE' Netflix suffers BIG Europe-wide outage
Friday night LIVE? Nope. The only thing streaming are tears down my face
Google roolz! Nest buys Revolv, KILLS new sales of home hub
Take my temperature, I'm feeling a little bit dizzy
Storage array giants can use Azure to evacuate their back ends
Site Recovery can help to move snapshots around
prev story


Why and how to choose the right cloud vendor
The benefits of cloud-based storage in your processes. Eliminate onsite, disk-based backup and archiving in favor of cloud-based data protection.
A strategic approach to identity relationship management
ForgeRock commissioned Forrester to evaluate companies’ IAM practices and requirements when it comes to customer-facing scenarios versus employee-facing ones.
Reg Reader Research: SaaS based Email and Office Productivity Tools
Read this Reg reader report which provides advice and guidance for SMBs towards the use of SaaS based email and Office productivity tools.
New hybrid storage solutions
Tackling data challenges through emerging hybrid storage solutions that enable optimum database performance whilst managing costs and increasingly large data stores.
The Heartbleed Bug: how to protect your business with Symantec
What happens when the next Heartbleed (or worse) comes along, and what can you do to weather another chapter in an all-too-familiar string of debilitating attacks?