Feeds

Understanding what's going on in storage arrays is like doing MAGIC

Software defined storage? It's a load of bollocks

Combat fraud and increase customer satisfaction

Storagebod As the announcements and acquisitions which fall into the realms of Software Defined Storage – or "storage", as I like to call it – continue to come, one starts to ponder how this is all going to work in the real world.

I think it is extremely important to remember that you are going to need hardware to run this software on. While current trends show this moving towards a commodity model, there are going to be subtle differences that are going to need accounting for. And as we move down this track, there is going to be a real focus on understanding workloads and the impact of different infrastructure and infrastructure patterns on this.

I am seeing more and more products that enable DAS to work as a shared-storage resource, removing the SAN from the infrastructure and reducing the complexity. I am going to argue that this does not necessarily remove complexity but it shifts it. In fact, it doesn't remove the SAN at all – it just changes it.

It is not uncommon now to see storage vendor presentations that show Shared-Nothing-Cluster architectures in some form or another; often these are software and hardware "packaged" solutions. Yet as end-users start to demand the ability to deploy on their own hardware, this brings a whole new world of unknown behaviours into play.

Once vendors relinquish control of the underlying infrastructure, the software is going to have to become a lot more intelligent, and the end-user implementation teams are going to have to start thinking more like the hardware teams in vendors.

For example, the East-West traffic models in your data-centre become even more important. You might find yourself implementing low-latency storage networks; your new SAN is no longer a North-South model but Server-Server (East-West). This is something that the virtualisation guys have been dealing with for some time.

Then there's understanding performance and failure domains: do you protect the local DAS with RAID or move to a distributed RAIN (Redundant Array of Independent Nodes) model? If you do something like aggregate the storage on your compute farm into one big pool, what is the impact if one node in the compute farm starts to come under load? Can it impact the performance of the whole pool?

Anyone who has worked with any kind of distributed storage model will tell you that a slow performing node – or a failing node – can have impacts which far exceed what you'd have believed possible. At times, it can feel like the good old days of token ring, where a single misconfigured interface can kill the performance for everyone. Forget about the impact of a duplicate IP address – that's nothing compared to this.

So what is the impact of the failure of a single compute/storage node? What about if multiple compute/storage nodes go down?

In the past, this was all handled by the storage hardware vendor and, pretty much invisibly, at implementation phase by the local storage team. But you will need now to make decisions about how data is protected and understand the impact of replication.

In theory, you want your data as close to the processing as you can get it, but data has weight and persistence; it will have to move. Or do you come up with a method that allows a dynamic infrastructure that identifies where data is located and spins/moves the compute to it?

The vendors are going to have to improve their instrumentation as well. Let me tell you from experience that understanding what is going on in such environments is deep magic. Also the software's ability to cope with the differing capabilities and vagaries of a large-scale commodity infrastructure is going to be have to be a lot more robust than it is today.

Yet I see a lot of activity from vendors, open-source and closed-source, and I see a lot of interest from the large storage consumers. This all points towards a large prize to be won. But I'm expecting to see a lot of people fall by the road.

It's an interesting time to be in storage... ®

Combat fraud and increase customer satisfaction

More from The Register

next story
This time it's 'Personal': new Office 365 sub covers just two devices
Redmond also brings Office into Google's back yard
Kingston DataTraveler MicroDuo: Turn your phone into a 72GB beast
USB-usiness in the front, micro-USB party in the back
Dropbox defends fantastically badly timed Condoleezza Rice appointment
'Nothing is going to change with Dr. Rice's appointment,' file sharer promises
BOFH: Oh DO tell us what you think. *CLICK*
$%%&amp Oh dear, we've been cut *CLICK* Well hello *CLICK* You're breaking up...
AMD's 'Seattle' 64-bit ARM server chips now sampling, set to launch in late 2014
But they won't appear in SeaMicro Fabric Compute Systems anytime soon
Amazon reveals its Google-killing 'R3' server instances
A mega-memory instance that never forgets
Cisco reps flog Whiptail's Invicta arrays against EMC and Pure
Storage reseller report reveals who's selling what
prev story

Whitepapers

Securing web applications made simple and scalable
In this whitepaper learn how automated security testing can provide a simple and scalable way to protect your web applications.
3 Big data security analytics techniques
Applying these Big Data security analytics techniques can help you make your business safer by detecting attacks early, before significant damage is done.
The benefits of software based PBX
Why you should break free from your proprietary PBX and how to leverage your existing server hardware.
Top three mobile application threats
Learn about three of the top mobile application security threats facing businesses today and recommendations on how to mitigate the risk.
Combat fraud and increase customer satisfaction
Based on their experience using HP ArcSight Enterprise Security Manager for IT security operations, Finansbank moved to HP ArcSight ESM for fraud management.