Feeds

What makes a tier-1 enterprise flash array a superior beast?

Storage monoliths enter the fray

Next gen security for virtualised datacentres

Enterprise tier-1 storage arrays are a breed apart, focusing on providing fast access to vast amounts of data without either losing data or limiting access to it.

This is in contrast to the broad mass of storage arrays, which have dual controllers and thus limit the amount of data they can store and the number of concurrent accesses they allow.

There are four main tier-1 arrays on the market:

  • EMC's VMAX
  • IBM's DS8000
  • Hitachi Data Systems'(HDS) VSP
  • HP’s XP, an OEM version of the Hitachi VSP

Typically such arrays have more than two controllers and an internal backplane or fabric to link the controllers with the storage shelves and ensure there is enough internal bandwidth to service the high number of I/O requests coming in.

The VMAX 40K, for example, had a 4PB capacity with 2,000 2TB disk drives. There is a Virtual Matrix interconnecting the controllers, aka storage engines, and there can be up to eight of them.

These have high availability, with data access continuing if a controller fails and no loss of data if a disk drive fails. Updates to the systems software and firmware should be accomplished without hindering access to data.

This class of array also has a rich set of data-management features, such as replication to another array, either local or remote, and snapshots.

Suppliers assert that the array operating system needs two or more years of development and real-world testing to be truly reliable. Overall these arrays are highly available and highly reliable.

The key components are:

  • More than two controllers
  • Internal interconnect fabric
  • Scale-up rather than scale-out design
  • High performance for large number of concurrent accesses

The scale-up nature of their design, adding controllers and disk shelves as a customer's needs grow, is why some call these arrays monolithic. They are very large single systems, in contrast to dual-controller arrays which can scale out to add capacity and performance by linking separate boxes, using clustering for example.

The internal fabric is vital to support performance, although other things being equal if that is missing but the other tier-1 enterprise array features are present then the array’s status is not compromised.

Fine performance

On that basis the main features of a tier-1 enterprise class array are performance, scaling up past two controllers and reliability.

It is generally agreed that scale-out arrays do not have the claimed bullet-proof reliability of tier-1 enterprise-class arrays.

Suppose an all-flash array from a startup can equal a tier-1 array’s performance – does that make it a tier-1 enterprise class array?

We have to understand how performance is measured. It is not enough simply to say the system can run a certain number of random read or write IOPS, or that its bandwidth when writing or reading is a certain number of GBps.

We need to understand the mix of IOs in the real world, the percentage of reads and writes, the size of the data blocks and their distribution.

Unless a benchmark is run on your own workloads, any performance rating is going to be artificial, but you can at least get an industry-standard rating by looking at the Storage Performance Council (SPC) benchmarks.

The SPC-1 benchmark provides a rating of how well a storage array does at random IO, with SPC-2 doing the same for sequential IO. A check of results for various arrays is quite informative:

Hitachi VSP All-flash array SPC-1 ranking

SPC-1 rankings (click chart to enlarge)

This chart shows SPC-1 rankings for the Hitachi VSP array in disk (269,506.69) and all-flash (602,019.47) versions and IBM’s DS8870 (451,082.27 IOPS)

The top-scoring system is a Kaminario K2 all-flash array with 1,239,898 IOPS. Does that make this a tier-1 enterprise-class array?

Test of endurance

We would say not because despite out-performing mainstream vendors’ tier-1 arrays it lacks the enterprise-class array features mentioned above.

There are no other all-flash array or hybrid array startup systems in the SPC-1 rankings. That means we cannot compare Pure Storage or Nimble Storage, for example, with other systems in the SPC-1 rankings.

Also there is no independent and objective measure of their performance based on an industry-standard workload. Customers have to run their own internal tests if they wish to evaluate such an array.

How about the SPC-2 benchmark, the sequential IO one? Here is a chart of some SPC-2 results:

SPC_2_500_May_2014

SPC-2 rankings (click chart to enlarge)

This chart is a combination of Mbps throughput and price/performance. That is based on list price, so it is less relevant in the real world where discounts may apply, but we still get a worthwhile comparison.

The chart shows IBM's DS8870, HDS’s VSP and HP’s XP systems, which are soundly beaten in sequential throughput terms by the Kaminario array again. But the same reasoning applies: the K2 is not a tier-1 enterprise class array because it lacks the features we have already mentioned.

In general then, it seems there is no startup offering a scale-up all-flash array that scales past two controllers and has the reliability features needed.

The logic of this pretty much dictates that the only straightforward way to build a tier-1 all-flash enterprise array is to put the flash inside an existing one, like the HDS VSP.

You can also buy all-flash versions of IBM's DS8000 and EMC's VMAX, and we would assume that they are better performers than their disk-based versions, although we have no SPC lens, as it were, through which to verify this.

However, suppliers such as HP with its all-flash 3PAR system and NetApp with its all-flash FAS series, particularly the top-end 8080 EX, would say that not only do they have the performance of a VMAX, VSP or DS8000 monolithic array, they also have scale-up attributes as they can use more than two controllers through, for example, clustering of nodes.

The 8080 EX can scale up past 4.5PB of flash and HP's 7450 can have 460TB of raw flash and 1.3PB of effective capacity after deduplicating redundant data. The capacity scale obstacle goes away.

Risky disks

These vendors would assert that their array operating systems have many years of development behind them and are battle hardened. They also have features such as replication and snapshots to provide the same rich data-management services as VMAX and VSP-class systems.

The vendors would also say that the most unreliable element of a disk-drive array is the disk drive. When an array has a thousand or more drives then disk failure will be a regular occurrence. Using SSDs instead of disk drives removes that mechanical risk from the equation.

As flash storage responds to requests so much faster than disk you don't need so many storage controllers to handle the same number of access requests from users.

All in all, an enterprise array previously rated as a high-end mid-tier array, such as FAS and 3PAR arrays, can with flash storage inside satisfy tier-1 enterprise-class array requirements. Acquisition, power, cooling and operating costs are also lower than for a disk-drive legacy array.

Flash-based arrays take up far less data-centre space than disk-drive arrays of equivalent capacity and performance, need less power to operate and generate much less heat.

If nothing else, workloads could be selectively moved from a legacy enterprise array onto newer all-flash arrays. We might envisage workloads with low latency requirements or a high proportion of random IO requests being candidates for this.

Customers for these high-end arrays are famously conservative and require a lot of convincing that a new candidate array is as good as their existing ones.

It begins to look, though, as if high-end flash arrays from HP and NetApp can now compete at this level and provide the incumbents with a run for their money. ®

5 things you didn’t know about cloud backup

More from The Register

next story
The Return of BSOD: Does ANYONE trust Microsoft patches?
Sysadmins, you're either fighting fires or seen as incompetents now
Microsoft: Azure isn't ready for biz-critical apps … yet
Microsoft will move its own IT to the cloud to avoid $200m server bill
Oracle reveals 32-core, 10 BEEELLION-transistor SPARC M7
New chip scales to 1024 cores, 8192 threads 64 TB RAM, at speeds over 3.6GHz
Docker kicks KVM's butt in IBM tests
Big Blue finds containers are speedy, but may not have much room to improve
US regulators OK sale of IBM's x86 server biz to Lenovo
Now all that remains is for gov't offices to ban the boxes
Gartner's Special Report: Should you believe the hype?
Enough hot air to carry a balloon to the Moon
Flash could be CHEAPER than SAS DISK? Come off it, NetApp
Stats analysis reckons we'll hit that point in just three years
Dell The Man shrieks: 'We've got a Bitcoin order, we've got a Bitcoin order'
$50k of PowerEdge servers? That'll be 85 coins in digi-dosh
prev story

Whitepapers

Endpoint data privacy in the cloud is easier than you think
Innovations in encryption and storage resolve issues of data privacy and key requirements for companies to look for in a solution.
Implementing global e-invoicing with guaranteed legal certainty
Explaining the role local tax compliance plays in successful supply chain management and e-business and how leading global brands are addressing this.
Top 8 considerations to enable and simplify mobility
In this whitepaper learn how to successfully add mobile capabilities simply and cost effectively.
Solving today's distributed Big Data backup challenges
Enable IT efficiency and allow a firm to access and reuse corporate information for competitive advantage, ultimately changing business outcomes.
Reg Reader Research: SaaS based Email and Office Productivity Tools
Read this Reg reader report which provides advice and guidance for SMBs towards the use of SaaS based email and Office productivity tools.