You know what storage needs? More doughnuts to flatten us up

Flatter hierachies, fewer hops: Rockport and the Torus interconnect

Torus
A torus concept could flatten network equipment hierarchies

The hierarchy of adapters, switches, routers and directors involved in storage networking is unwieldy, complex and costly and needs replacing with a flatter scheme of direct connections between servers and storage devices. That’s the networking message from start-up Rockport Networks.

It says that current fat tree and spine-and-leaf network architectures tend to prevent economical scalability and hinder cloud computing and analytics. It’s better to rethink storage network architecture, indeed network architecture, than to try and software-define current networking hardware to gain relatively marginal scalability, flexibility and cost-savings.

This company has been operating under the radar since being founded in 2012, in Ottawa, by a trio of storage and networking vets: CEO Doug Carwardine, CTO Dan Oprea and COO Michael McLay.

Its idea is that storage and server nodes in a network should be directly interconnected using a Direct Interconnect (torus mesh) of optical cables. A torus mesh is a geometric shape analogy of their interconnection scheme, being a ring with a circular cross-section and a hole inside the middle of the ring – think doughnut.

This is not a new thing. Torus mesh connectivity has been used to interconnect high performance computing environments for years.

Rockport’s CTO Dan Oprea says its torus is a Direct Interconnect scheme but the two terms are not synonymous. You can have either without the other. Supercomputers use torus network schemes which are not Direct Interconnects.

Oprea says Rockport wanted to use a torus because one of its properties is minimising the number of hops between any source node and any destination node, “so you can jump much faster.”

It also wanted a characteristic of Direct Interconnect; the ability to add or drop traffic on every node, which you cannot do with a spine and leaf design.

Rockport says Direct Interconnect is widely recognised as the most effective way to connect; it is also recognised as a very complex way to cable things together. The company claims it has solved the cabling challenge, with high radix connectivity and simplified cabling with short runs. Direct interconnect is the most efficient way to interconnect data nodes in the data centre and provides dramatic linear scalability, it says.

Because of this, Rockport's software is going to bring the performance and scalability of its architecture to the larger enterprise market.

Torus

A torus. Imagine this is formed from a 3-dimensional grid of networked nodes with links between them.

The torus networking concept

How this torus mesh idea applies to networking is ingenious and this is how your storage correspondent, who is definitely not a network engineer, envisages it.

Imagine a grid of nodes, three to a side and with another grid of 3 x 3 nodes, below them. The upper grid nodes have orange tops and are interconnected by red lines. The bottom grid has yellow tops with the nodes interconnected by blue lines. The nodes vertically in line with one another are connected by green vertical lines, giving us this concept:

2_layer_3_x_3_Grid

2-layer 3x3 node grid with interconnections

Now imagine the lines are network paths between nodes in the grid and any node can communicate with any other node using these logical link lines. You can trace several paths from the top left node to the bottom right node using four intermediate nodes and five hops (links) between them. Got that?

Now imagine the nodes increase and the grid extends to two layers of 30 x 30 nodes. The same basic point applies, only now you’ll have the top left to bottom right node path involve some fifty or so intermediate nodes and links. Okay? Now extend it to a grid of 100 x 100 nodes and two layers. The paths get longer and longer, meaning 200 or so intermediate nodes and hops, which is clearly unacceptable. So; prepare for an imaginary leap here; suppose we wrap this grid around a flat column so that the far left set of nodes are now positioned so as to be underneath the far right set of nodes below them in our imagined 3D space;

Wrapped_node_grid

Wrapped node grid firms a quasi-tube

Imagine a cross-section of this tube is the circular cross-section of the torus doughnut tube. Now the network path from the top left node of the grid to the bottom right node has been considerable shortened to just over 100 intermediate nodes and hops.

Now we need another imaginary leap. Ready? The front and rear edges of our grid are now extended as well, so as to draw a circle with the front and rear edges meeting to that the nodes on the edges directly face each other; we now have a 3 dimensional torus mesh topology as shown in the blue torus doughnut image above.

Now imagine the 3 dimensional nodes all connected in another dimension of connectivity; a 4 dimensional torus mesh topology. And then go to 5D or 6D. This concept can scale to support literally billions of devices. As dimensions are added, there are much shorter links between previously far members of our grid, meaning fewer hops and lower latency. A typical 4D or 6D torus mesh will have more than 5,000 devices within single digit hops. Oprea says that the more equal the dimensions (node count) are on either edge of a grid, the better the torus Direct Interconnect network; 10 x 10 grids are better than 5 x 15 ones for example.

Here is a Rockport 2D configuration with a 6 x 7 grid (42 data nodes);

Rockport_2D

Rockport 2D config scheme with logical links, not physical cables.

Now we’ll extend it to 3D;

Rockport_3D

Three dimensions with logical links and 126 nodes

And to scale beyond that, 4D;

Rockport_4D

4 dimensions with logical links and 378 nodes

Finally, here is a Rockport Direct Interconnect concept diagram using IT racks;

Rockport_Torus

Rockport torus concept. Important point; the green lines are notional, indicative interconnect paths and not actual cables.

Sponsored: Minds Mastering Machines - Call for papers now open

Next page: Ditch the switch

Biting the hand that feeds IT © 1998–2018