Intel adds Lustre support to Hadoop
Chipzilla hopes HPC owners get busy with map reduce in their spare time
World+dog agrees that Hadoop is a very fine tool with which to tackle map reduce chores, but the software has a couple of constraints, especially its reliance on the Hadoop Distributed File System (HDFS).
There's nothing wrong with HDFS, but its integration with Hadoop means the software needs a dedicated cluster of computers on which to run.
That's not a bad thing for many reasons. But folks who run high performance computing clusters for other purposes often don't run HDFS, which leaves them with a bunch of computing power, tasks that could almost certainly benefit from a bit of map reduce and no way to put that power to work running Hadoop.
Intel's noticed this and, in version 2.5 of its Hadoop distribution that it quietly released last week, has added support for Lustre.
Girish Juneja, Intel's general manager for big data and software services, thinks Chipzilla's HPC customers are going to love that, and that the rest of us won't mid either given Intel's following all the open source rules with this contribution.
“Many customers do not want to deploy an entirely separate physical cluster just because we couldn't figure out how to run Hadoop on their file system,” Juneja told The Reg at Intel's Big Data and Cloud Summit in Ho Chi Minh City*. “HPC was a prime target for that. In the HPC segment a large part of the market runs on GPFS or Lustre and we had the fortune of having Lustre in my division.”
“We abstracted out an HDFS layer but underneath that it is actually talking to lustre.
“So if you look at Los Alamos lab and these research labs with humongous clusters that run HPC jobs 90 per cent of the time, but for ten per cent of the time they want to run a Hadoop job they can run it in exactly the same environment without moving data.”
Given such labs deal in vast quantities of data, the chance to leave it in place will be welcome.
Chipzilla's also turned is attention to encryption and an access control list for HBase.
“In this NoSQL environment the challenge becomes how to designate who has access to what data,” Juneja said. “We have added capabilities to allow for access control lists” that let administrators set policies for who can access what data in Hbase.
Janeja feels that addition, together with the introduction of encryption and data anonymisation, will mean financial services providers and users with heavy compliance requirements can now consider Hadoop. In the past, Janeja said, the absence of those security-oriented features meant Hadoop represented unacceptable risk.
Intel also sells its own management software to drive the access control lists, an arrangement Juneja feels customers won't mind too much.
Version 3.0 of Chipzilla's Hadoop distribution is also close to release, with September targeted for its emergence. Juneja said users can expect Intel's distribution to very closely resemble efforts of the wider Hadoop community. ®
*The author attended the summit as a guest of Intel, which paid for flights and accomodation.
Sponsored: Network DDoS protection