Feeds

Open MPI hits milestone with FORTRAN-ready 1.7.4 release

A point release? FORTRAN? Yes this does matter, especially to HPC folks

Next gen security for virtualised datacentres

The decade-old Open MPI project, beloved of the HPC community, has shipped code for a major feature release that brings it close to a complete MPI 3 implementation.

Speaking to The Register, Cisco's Jeff Squyres (who along with Richard Graham, then of Los Alamos and now with Mellanox, was one of the instigators of the Open MPI project ten years ago) said that perhaps one of the most important characteristics of the new release is it wraps up much-needed FORTRAN support in the interface.

While that will probably raise giggles among the cool kids, FORTRAN is not only alive and well in the HPC community, it remains one of the key languages in the world of big number crunchers.

“This is the first set of MPI-FORTRAN bindings that are actually compliant with the language,” Squyers said.

MPI, he explained offers handles to languages and compilers that connect them to the underlying MPI objects – but “all of those handles used to be simple integers. With this release, the handles are now unique types.” With those types FORTRAN-compliant, he said, the user – the scientist writing software for a particular function – gets better assistance from the FORTRAN compiler.

“There are also strict prototypes for all the MPI function engines”, Squyres said, which again means users get better assistance in fixing errors.

The 1.7.4 Open MPI release also comes close to a full implementation of the MPI 3 standard, he said, with only one major item – support for one-way operations – missing.

A big item in the new release is support for “non-blocking collectives” in the communications model. While there are a host of communication types beyond simple point-to-point, Squyers chose non-blocking broadcasts as a good example of what this means for the HPC user.

“It can be really useful to ask MPI to start a broadcast, while the compute thread goes away to do computations because that's what it does, and come back later to see if the broadcast is finished.”

For example, he said, a much-used operation in fluid dynamics is an all-reduce operation: a vector of numbers is distributed across thousands of cores, and the user wants the calculations on each core to be tested against each other.

A non-blocking blocking broadcast means you can start an iteration of the calculation, send a broadcast to collect the result – and instead of waiting for responses to the broadcast, go ahead with the next iteration.

“I can overlap iteration and computation, without waiting for the communication,” Squyres said.

  • A new binding policy that makes “extensive” use of process affinity by default, so that a process can be forced to remain on a given core; and
  • Better CUDA support, courtesy of contributions from Nvidia.

What is Open MPI anyway?

So why would there be a fuss about a feature release of a message-passing standard? Haven't we been doing this, practically forever?

The Message Passing Interface, MPI – the standard that underlies both the open source efforts and proprietary implementations – dates back to the 1990s, and its longevity makes it the lingua franca of the HPC community.

Supercomputing, Squyers explained to The Register, has requirements that were never fulfilled by commercial middleware busses, nor by the world of TCP sockets.

In particular, HPC needs extremely low latency.

A very well-tuned TCP implementation might be able to get pings of 10 microseconds between a couple of servers in the same data centre, but the supercomputer user wants latency closer to a tenth of that. That's one of the aims of MPI, and both its proprietary and open source implementations.

The other, Squyers said, is that the middleware should present communications to the users as simply as possible.

“It's a very high-level network abstraction,” he explained. “The person writing the application is a chemical engineer, or a physicist or a biologist. They want to crunch some numbers, and they understand their own algorithms and math, but they don't understand the network.

“They want to tell their software 'take this gigabyte of double-precision numbers and send it to number 38'. Under MPI, a miracle occurs: the data gets sent, but the user doesn't have to know whether it's on a core a couple of racks away, or on the same server sent via shared memory.”

The MPI standard first got turned into Open MPI by Squyers and Graham in a project that began in 2004 and shipped its first code in 2005.

While there are alternatives to MPI – SGI's SHMEM (symmetric hierarchical memory access) which has an open implementation, and PGAS (partitioned global address space), for example – Squyers believes that MPI underlies as much as 90-95 per cent of HPC code out there.

The penetration of the Open MPI implementation is harder to assess, since it's included in Linux and some OpenBSD distributions. ®

Gartner critical capabilities for enterprise endpoint backup

More from The Register

next story
The Return of BSOD: Does ANYONE trust Microsoft patches?
Sysadmins, you're either fighting fires or seen as incompetents now
Microsoft: Azure isn't ready for biz-critical apps … yet
Microsoft will move its own IT to the cloud to avoid $200m server bill
Shoot-em-up: Sony Online Entertainment hit by 'large scale DDoS attack'
Games disrupted as firm struggles to control network
Cutting cancer rates: Data, models and a happy ending?
How surgery might be making cancer prognoses worse
Silicon Valley jolted by magnitude 6.1 quake – its biggest in 25 years
Did the earth move for you at VMworld – oh, OK. It just did. A lot
Forrester says it's time to give up on physical storage arrays
The physical/virtual storage tipping point may just have arrived
prev story

Whitepapers

Implementing global e-invoicing with guaranteed legal certainty
Explaining the role local tax compliance plays in successful supply chain management and e-business and how leading global brands are addressing this.
5 things you didn’t know about cloud backup
IT departments are embracing cloud backup, but there’s a lot you need to know before choosing a service provider. Learn all the critical things you need to know.
Why and how to choose the right cloud vendor
The benefits of cloud-based storage in your processes. Eliminate onsite, disk-based backup and archiving in favor of cloud-based data protection.
Top 8 considerations to enable and simplify mobility
In this whitepaper learn how to successfully add mobile capabilities simply and cost effectively.
High Performance for All
While HPC is not new, it has traditionally been seen as a specialist area – is it now geared up to meet more mainstream requirements?