Feeds

Intel doubles throughput, slashes power to stave off DATAPOCALYPSE

With 750 gigabytes/sec looming, strong measures must be taken

  • alert
  • submit to reddit

Remote control for virtualized desktops

Research@Intel Intel researchers have developed a prototype interconnect that they say will both increase bandwidth and lower power requirements, whether used for simple CPU-to-CPU connections or scaled up to connect tens of thousands of CPUs in a data center.

"Based on past trends, every four years the bandwidth requirements for systems is increasing by an order of magnitude," Intel senior research scientist Mozhgan Mansuri told The Reg at the annual Research@Intel dog and pony show in San Francisco on Tuesday.

If this trend continues, she said, by 2020 bandwidth needs will be in the range of 750 gigabytes per second. And since today's most efficient interconnects require about 10 or 15 picojoules per bit, the juice required to power that bandwidth will be about 50 per cent of the whole system in 2020, "which is unacceptable."

Motivations of the need for higher-performance, more power-efficient interconnects

Mansouri describing this slide showing the growing power needs of I/O: 'There's a lot of assumptions
there, but the point is that if we don't do anything about the I/O power, we're getting into trouble.'

The solution that Mansouri and her colleagues have come up with is called, rather straightforwardly, "Scalable Energy Efficient I/O". Their setup consists of 64 bidirectional lanes controlled by a global clock – and it's that clock that is key to their system's efficiency, Mansouri said.

The global clock takes advantage of the fact that although systems require more and more peak bandwidth, most of the time they're not running at that rate, but instead they're perhaps idling or running workloads that require only half that peak.

If you want to conserve power in current implementations, she said, you simply turn off lanes that aren't needed, but keep the clock at the same rate – if you only need half the bandwidth, you turn off half the lanes. "So in that case, you just get the linear power saving – your bandwidth goes down by half, your power goes down by half."

In Scalable Energy Efficient I/O, if you want to halve the bandwidth of the interconnect, instead of turning off half the lanes, you keep all the lanes running but drop the throughput of each of the lanes. By doing so, the power requirements drop in a curve that's notably steeper than mere linearity. "So potentially I get much better power saving than the linear bandwidth of scaling gives me," Mansuri said.

The power savings and the throughput can can both be considerable, she said, showing slides that gave examples such as an 8-link, four-socket blade using a 32-lane setup for her team's design consuming 22 watts and providing 8 terabits per second throughput, compared with a 32-lane PCIe 3.0 setup consuming 62 watts and providing 4 terabits per second throughput.

Intel's Scalable Energy Efficient I/O in a server blade

Twice the throughput at less than half the power; what's not to like except the price tag?

Another slide imagined a data center with 100,000 links, in which a prototype 32-lane Intel system consumed 270 kilowatts to provide an aggregate of 100 petabits per second, while a 32-lane PCIe 3.0 setup consumed 770 kilowatts to pump along 50 petabits per second.

Mansuri emphasized that the prototype which she showed us humming along nicely was just that – a prototype – and that much work needs to be done both in its design and its manufacturability at a reasonable cost. "So this will be more attractive to the high end," she said, "and that's why we're looking to the data center." ®

Beginner's guide to SSL certificates

More from The Register

next story
Azure TITSUP caused by INFINITE LOOP
Fat fingered geo-block kept Aussies in the dark
NASA launches new climate model at SC14
75 days of supercomputing later ...
Yahoo! blames! MONSTER! email! OUTAGE! on! CUT! CABLE! bungle!
Weekend woe for BT as telco struggles to restore service
Cloud unicorns are extinct so DiData cloud mess was YOUR fault
Applications need to be built to handle TITSUP incidents
BOFH: WHERE did this 'fax-enabled' printer UPGRADE come from?
Don't worry about that cable, it's part of the config
Stop the IoT revolution! We need to figure out packet sizes first
Researchers test 802.15.4 and find we know nuh-think! about large scale sensor network ops
DEATH by COMMENTS: WordPress XSS vuln is BIGGEST for YEARS
Trio of XSS turns attackers into admins
SanDisk vows: We'll have a 16TB SSD WHOPPER by 2016
Flash WORM has a serious use for archived photos and videos
prev story

Whitepapers

Why cloud backup?
Combining the latest advancements in disk-based backup with secure, integrated, cloud technologies offer organizations fast and assured recovery of their critical enterprise data.
Forging a new future with identity relationship management
Learn about ForgeRock's next generation IRM platform and how it is designed to empower CEOS's and enterprises to engage with consumers.
Designing and building an open ITOA architecture
Learn about a new IT data taxonomy defined by the four data sources of IT visibility: wire, machine, agent, and synthetic data sets.
How to determine if cloud backup is right for your servers
Two key factors, technical feasibility and TCO economics, that backup and IT operations managers should consider when assessing cloud backup.
High Performance for All
While HPC is not new, it has traditionally been seen as a specialist area – is it now geared up to meet more mainstream requirements?