This article is more than 1 year old

Compute-storage mashup: Great for Big Data, but what of my mobile?

Latency HELL as 'tops and 'slabs probe cloud

Blocks and Files Should compute and storage dance closer together or farther apart?

In the enterprise there is a strengthening under-current of thought that says data access latency is bad. You shouldn't have to wait for data. In the networked storage array field that is causing the import of flash storage into arrays to kick disk seeks into the long grass, and the provision of higher-speed networking, such as InfiniBand and 1Gbit/s Fibre Channel. But that's just tinkering with network latency.

We're seeing flattening SANs and virtualised server area networkers using server IO virtualisation, like the Oracle Xsigo technology, to provide SAN data access facilities at direct-attached storage (DAS) speeds. The purest exhibition of this is to re-invent DAS as flash storage memory, treated as an adjunct to a server's RAM accessed through the memory IO subsystem and not the disk IO subsystem.

These are examples of bringing storage to compute. A reverse trend is also visible though not as readily apparent, and that is bringing compute to storage by running application software inside storage arrays, as EMC keeps on saying its going to do with VMAX, VNX and Isilon storage arrays. DataDirect Networks is actually doing it by running file systems such as GPFS and Lustre inside its storage arrays via the ExaScaler and Gridscaler products. These are system-level applications though and not end-user apps. Running actual end-user apps in storage arrays is still a distant prospect.

In client, end-point computing-land the reverse is happening; to a large extent compute and storage are getting further apart. The smart end-points, like mobile phones, tablets, Google-ised Chromebooks and virtualised desktops have very little local storage and the cloud or enterprise data centre is where the data is stored. Like Big Data, client data is centripetal. Why is network latency acceptable for end-points but not in servers?

Frankly, sitting at the end of what my ISP or mobile phone service provider laughably calls broadband, having the remote servers access their data at microsecond speeds instead of milliseconds while I'm waiting multiple seconds leaves me cold. I could care less. I want my data - right now. If low latency is good for data centre servers then it's good for client devices too. But no shared network, no affordable shared network, is ever going to be as fast as local storage, or as dependable, as those who suffer ISP link outages know only too well.

Put another way: Big Data is a data centre thing. Here out in the far-flung reaches, the outer galactic spiral arms of end-point computing, Big Data doesn't exist. It's a case of little or no data. Curious isn't it? Latency sauce for the data centre server goose is not latency sauce for the end-point gander. ®

More about

More about

More about

TIP US OFF

Send us news


Other stories you might like