Facebook puts some brains in Open Vault JBOD storage

ARM or Atom, pick your embedded CPU and interconnect poison

Secure remote control for conventional and virtual desktops

Open Compute 2013 At last week's Open Compute Summit 2013, the people behind the open source hardware project were showing off some enhancements for the Open Vault JBOD storage array that Facebook has cooked up for its own use in its two newest data centers and presumably will be added to its existing data center.

The Open Vault array, known by its code-name "Knox," has been contributed by the social network to the Open Compute Project open source hardware design effort. Among other things, Open Vault is used for cold storage of the 240 billion photos on the site, which is growing at a rate of 350 million per day, or 7PB per month. (You can see the Open Vault specs here.)

Open Vault is a JBOD array, which means it is just a bunch of disks and is intended to hang off a SAS controller inside of a server. In Facebook's case, that is a custom Open Compute V2 server using Intel's custom "Windmill" two-socket Xeon E5 server node.

The Open Vault array has two 1U disk trays, which each hold fifteen 3.5-inch SAS drives and two SAS expander boards. The four SAS extender boards feed back to the server and make all of the 30 drives in the Open Vault look like they are connected directly to the server. Open Vault is designed so any disk or any one SAS expander can be changed without having to take the JBOD offline.

While Open Vault is great for what it does, it lacks brains. And so, through an extension of the Open Vault called "Knockout," ARM server chip upstart Calxeda and Intel are both working on variants that put some brains and internetworking into each JBOD to turn it into a smarter storage cluster.

Frank Frankovsky, vice president of hardware design and supply chain at Facebook and also chairman of the Open Compute effort, showed off two compute boards that slide into the Open Vault array where the SAS extenders currently fit and give them a bit of brains, like the Scarecrow in the Wizard of Oz.

The first one that Frankovsky showed off – and one that is close to being in production – is an ARM-based compute add-on card that is based on the 32-bit ECX-1000 processor from Calxeda:

A Calxeda ARM server node for the Open Vault JBOD

A Calxeda ARM server node for the Open Vault JBOD

The idea, Gina Longoria, product marketing manager at Calxeda, explains to El Reg is to allow for companies deploying Open Vault storage to possibly run Lustre or Gluster clustered file system code inside each Open Vault tray and maybe only use an x86 node in the rack to run a head node.

The additional computing power could also be used to run other storage software, such as the Ceph distributed object store that is being closely affiliated with OpenStack, or even the Cassandra NoSQL data store that was created by Facebook when it ran up against the limits of MySQL relational databases.

The precise software that can be run on an intelligent storage server is not the point. Giving Open Vault some cheap yet power brains is the point.

Intel wants a piece of this action, too, and is a good buddy of the Open Compute Project as well, and thus Frankovsky was careful to hold up a similar brain transplant card based on Intel's forthcoming "Avoton" Atom S Series processor, which is expected to have on-chip Ethernet links:

An Intel

An Intel "Avoton" Atom server node for Open Vault storage

The feeds and speeds of the Intel board were not divulged, but Calxeda was happy to talk about different configurations of its Knockout compute and networking cards for the Open Vault JBOD.

The Calxeda board has a single ECX-1000 processor with four Cortex-A9 cores running at 1.4GHz and 4GB of DDR3 main memory running at 1.33GHz. The board can have two RJ45 ports running at gigabit speeds and five SATA port mulitpliers, supporting all drives in a single Open Vault tray - you put one in each tray.

The card can be equipped to run software RAID or to run iSCSI target software, mapping from an x86 head node at the top of the rack. You can also have SFP+ or QSFP ports put on this card if you want to spend a little more money.

The current

The current "Knox" Open Vault and its computationally enhanced "Knockout" derivative

Or, if you want to go with the cheaper and better option, you could use CX4 connectors and use the on-chip distributed Layer 2 network on the ECX-1000 chips to be real clever. First, you could put a 24-port Gigabit Ethernet switch between the Windmill head node and the computationally enhanced Open Vault JBODs.

This switch would link the JBODs to other Windmill head nodes for redundancy, eliminating a single point of failure in the rack. Then you could add data compression, hashing, or other algorithms on the local ARM nodes, or Atom-based nodes, too.

By tucking the ECX-1000 server nodes into the Open Vault JBODs, however, you can do one other thing: cross-couple the arrays and their compute nodes across racks. Here's one example:

How you might use the interconnect on ECX-1000s to do a 2D torus between storage JBODs

How you might use the interconnect on ECX-1000s to do a 2D torus between storage JBODs

With the variant of the Knockout server board that has four 10GE ports coming off the ECX-1000 chip, you can turn on the integrated Fleet Services fabric and use the top of rack switch to handle north-south traffic out of the array and to the network, feeding applications, and use the Fleet Services interconnect to provide data replication and other services on an east-west network that spans multiple racks.

All of this can happen under the covers of the Open Vault and behind the scenes where the head node in a storage cluster is blissfully unaware. This would also mean that the x86 node could potentially be quite a bit less powerful in terms of memory and CPU capacity, and in fact, you could have an array of ARM servers acting as the head node if you wanted, according to Longoria.

This is, of course, the option that Calxeda is excited about. ®

Secure remote control for conventional and virtual desktops

More from The Register

next story
NSA SOURCE CODE LEAK: Information slurp tools to appear online
Now you can run your own intelligence agency
Fat fingered geo-block kept Aussies in the dark
Yahoo! blames! MONSTER! email! OUTAGE! on! CUT! CABLE! bungle!
Weekend woe for BT as telco struggles to restore service
Cloud unicorns are extinct so DiData cloud mess was YOUR fault
Applications need to be built to handle TITSUP incidents
Stop the IoT revolution! We need to figure out packet sizes first
Researchers test 802.15.4 and find we know nuh-think! about large scale sensor network ops
Turnbull should spare us all airline-magazine-grade cloud hype
Box-hugger is not a dirty word, Minister. Box-huggers make the cloud WORK
SanDisk vows: We'll have a 16TB SSD WHOPPER by 2016
Flash WORM has a serious use for archived photos and videos
Astro-boffins start opening universe simulation data
Got a supercomputer? Want to simulate a universe? Here you go
Microsoft adds video offering to Office 365. Oh NOES, you'll need Adobe Flash
Lovely presentations... but not on your Flash-hating mobe
prev story


Free virtual appliance for wire data analytics
The ExtraHop Discovery Edition is a free virtual appliance will help you to discover the performance of your applications across the network, web, VDI, database, and storage tiers.
Forging a new future with identity relationship management
Learn about ForgeRock's next generation IRM platform and how it is designed to empower CEOS's and enterprises to engage with consumers.
How to determine if cloud backup is right for your servers
Two key factors, technical feasibility and TCO economics, that backup and IT operations managers should consider when assessing cloud backup.
High Performance for All
While HPC is not new, it has traditionally been seen as a specialist area – is it now geared up to meet more mainstream requirements?
Business security measures using SSL
Examines the major types of threats to information security that businesses face today and the techniques for mitigating those threats.