Feeds

Virtualising your infrastructure? Get your numbers right

Avoid nightmares later

Intelligent flash storage arrays

Virtualisation can be a powerful tool for your IT department, making your infrastructure far more efficient. But without proper planning it is easy to trip yourself up by not scaling the system properly.

How can you plan the capacity needed for a virtualised system so that you don’t end up overspending or under-resourcing? Here are some key points to consider.

Stay single

One of the biggest mistakes companies can make is trying to run consolidated systems over non-consolidated networks. Discrete data centre networks for specific components such as storage and servers are no longer appropriate.

As companies move to private cloud computing, in which systems are consolidated onto fewer devices, networks must be simpler to manage while supporting large volumes of traffic.

Converged networks, with storage and server traffic running across a unified fabric, form the basis for next-generation data centres.

The converged network protocol wars are largely settling down and it is now possible to buy vendor-agnostic 10Gbit connections that will support your servers and storage well enough to help you avoid problems such as bootstorms.

Managing virtual systems over a single network fabric can make life far easier for administrators, decreasing management costs while maintaining system uptime.

Baseline everything

Before you can scale up a pilot it helps to understand what your current system is doing. Baselining system activity using management tools is a crucial step, but there are different levels of knowledge.

“Nine out of ten times your customers think they have the numbers, but it’s just the apps installed on a workstation they are looking at and not so much the utilisation of those apps,” warns Erwin Vollering, service director for virtualisation at Glasshouse Technologies.

Measuring that, with network probes or client-side agents, will give you an idea of what to expect when your desktop virtualisation system goes from 15 to 1,500 machines.

Don’t forget to sample your system to get a sense of how spiky your computing demand is. Then you can plan to hit a percentage of your peak load.

Talk to vendors

When planning for a scaled system, it makes sense to get a comprehensive picture of your hardware’s capability. Take as much information as possible from your storage vendors about how their systems will perform at certain loads.

In an application-centric world, you must be sure to factor in the software too. How capable is the software of working at scale? Is your relational database going to cut it, or do you need to move to a large, multi-node database?

In an age of Big Data, that becomes a critical decision. You may need to completely rethink the way you deliver an application, especially if it is a legacy app designed for a different infrastructure.

Integrate, integrate

Consolidated virtualised systems work far more effectively when all of your hardware is integrated effectively.

The next-generation data centre features a unified network fabric supporting an integrated stack, in which storage, servers, and networking equipment are all optimised to work together.

This provides better performance and also makes it easier for systems management tools to get an end-to-end view of the hardware infrastructure, and therefore of the software running on it. Consider a next-generation data centre architecture to keep consolidated systems running more smoothly.

Points of failure

Systems may be designed based on an expectation of a specific number of transactions in a given time window.

When a system’s transaction volume scales, single points of failure become a real issue. If a switch fails in your pilot system, the failover switch may be able to take the strain. But when you scale by an order of magnitude, will that still be the case?

Designing multiple failure zones that are segmented in different parts of the system is a good idea, so that no part of the infrastructure can be compromised at any one time.

Ideally, you need to also design those systems to cope with the system’s entire peak load so that they can pick up the slack in the event of a failure.

Amazon had to truck in more equipment to increase capacity

This can be difficult as the scale of the system increases. One reason that Amazon’s Dublin-based Elastic Block Store system took so long to get back online recently was that so many servers went down at once because of a transformer failure.

The ones that were left kept trying to replicate to anything they could find, which led to customers’ volumes being “stuck”. The firm had to truck in more equipment to increase capacity.

“There were delays as it was night time in Dublin and the logistics of trucking required mobilising transportation some distance from the data centre,” the firm said.

“Plan and test a restart,” says John Soanes, head of architecture at IT managed services firm Adapt.

“Because you have contention, you have successive failures. So it’s really important to try and bring up a large system in a controlled way.”

Conclusion

The axiom “garbage in, garbage out” pervades every aspect of IT, from coding through to systems design. It applies to virtualisation, too.

Get quality information at the start to avoid problems later. And consider designing everything from the bottom up as a converged architecture to increase performance and simplify management. ®

Beginner's guide to SSL certificates

More from The Register

next story
All aboard the Poo Bus! Ding ding, route Number Two departing
Only another three days of pooing and I can have a ride!
Ford's B-Max: Fiesta-based runaround that goes THUNK
... when you close the slidey doors, that is ...
Official: European members prefer to fondle Apple iPads
Only 7 of 50 parliamentarians plump for Samsung Galaxy S
Fujitsu CTO: We'll be 3D-printing tech execs in 15 years
Fleshy techie disses network neutrality, helmet-less motorcyclists
Space Commanders rebel as Elite:Dangerous kills offline mode
Frontier cops an epic kicking in its own forums ahead of December revival
Nexus 7 fandroids tell of salty taste after sucking on Google's Lollipop
Web giant looking into why version 5.0 of Android is crippling older slabs
prev story

Whitepapers

Choosing cloud Backup services
Demystify how you can address your data protection needs in your small- to medium-sized business and select the best online backup service to meet your needs.
A strategic approach to identity relationship management
ForgeRock commissioned Forrester to evaluate companies’ IAM practices and requirements when it comes to customer-facing scenarios versus employee-facing ones.
High Performance for All
While HPC is not new, it has traditionally been seen as a specialist area – is it now geared up to meet more mainstream requirements?
Website security in corporate America
Find out how you rank among other IT managers testing your website's vulnerabilities.
Top 5 reasons to deploy VMware with Tegile
Data demand and the rise of virtualization is challenging IT teams to deliver storage performance, scalability and capacity that can keep up, while maximizing efficiency.