Feeds

Is server virtualization ready for production?

Beyond the low hanging fruit

  • alert
  • submit to reddit

Internet Security Threat Report 2014

Lab The adoption of server virtualization technology follows several trajectories. We could consider the breadth of its penetration in terms of the number of organisations using it today, or we could consider the depth of its use in individual organisations, in terms of what they actually do with it.

The latter is of more relevance to IT departments that have already taken first steps down the server virtualization path. For other organizations, if we think beyond the workloads already being run in a virtualized environment, is there a ‘next’. And if there is, what is it?

Perhaps this is simplistic. When you think about virtualization do you think in terms of a proportion of the x86 server estate in your datacenter that ‘could’ be virtualized or do you think about the different types of workloads that need executing? Vendors with a vested interest in shifting virtualization technology tend to presume that something that could be a candidate for virtualization, automatically will be.

However, we know from Reg reader research that decision-making is typically focused on a simpler criterion: can it save money right now? As a result, virtualization tends to be employed for the more straightforward workloads that can be easily consolidated.

Admittedly, the notion of ‘straight forward’ is relative, although there are some commonly accepted candidates such as print servers, web servers and the like. Whether these are chosen because they are seen as cost-saving, low risk, ‘non-core’ or ‘non-critical’ areas, it’s where most organizations cut their teeth. So where do we go from here? The answer has to be into areas of higher potential risk, and less evident cost-benefit. So then: what is the rationale for making decisions?

Work up the list

To reiterate, the factors at play are: cost savings; virtualization benefit; business importance; and migration of risk. Does IT simply ‘work up the list’ from least risk / importance? Or are those with prior experience now applying virtualization to areas which would benefit specifically from it, regardless of their importance to the business?

Factors around migration-risk bring into question enough experience and confidence exists in the technology itself and on the periphery (availability, resilience and back-up and recovery systems), as well as the skills of the IT department itself to be able to consider higher-risk workloads as virtualization candidates.

One must also take into consideration the socio-political aspects of IT ownership. A line of business leader might have concerns about ‘his’ application running in a virtualized environment, even if he's perfectly happy with the service he gets from ‘lower value’ services. But if the technology is proven elsewhere, what’s the fuss?

Part of the answer could lie in how big the first step down the virtualization route was. Did the IT department have to fight to make it to happen, or did someone in the business make a request for it directly or indirectly – e.g., a demand that could only be fulfilled by employing technology in this way?

One argument suggests that had it not been for the economic crisis in 2008, many organizations would not have felt it necessary to virtualise any server infrastructure.

So, if you have moved virtualization beyond the pilot, how did you decide? Was it via the same process you employed the first time you decided to take advantage of this technology? Did it involve a complex risk management exercise or was it more about gut feel and trust in your collective abilities and the technology itself?

‘If you’ve already taken the ‘next step’, or are thinking about it, we’d like to hear about the decision-making processes you or your department have been working through and your experiences of migrating ‘next level’ workloads into the virtual environment so far.

Choosing a cloud hosting partner with confidence

More from The Register

next story
Preview redux: Microsoft ships new Windows 10 build with 7,000 changes
Latest bleeding-edge bits borrow Action Center from Windows Phone
Google opens Inbox – email for people too thick to handle email
Print this article out and give it to someone tech-y if you get stuck
Microsoft promises Windows 10 will mean two-factor auth for all
Sneak peek at security features Redmond's baking into new OS
FTDI yanks chip-bricking driver from Windows Update, vows to fight on
Next driver to battle fake chips with 'non-invasive' methods
UNIX greybeards threaten Debian fork over systemd plan
'Veteran Unix Admins' fear desktop emphasis is betraying open source
Entity Framework goes 'code first' as Microsoft pulls visual design tool
Visual Studio database diagramming's out the window
Google+ goes TITSUP. But WHO knew? How long? Anyone ... Hello ...
Wobbly Gmail, Contacts, Calendar on the other hand ...
prev story

Whitepapers

Why cloud backup?
Combining the latest advancements in disk-based backup with secure, integrated, cloud technologies offer organizations fast and assured recovery of their critical enterprise data.
A strategic approach to identity relationship management
ForgeRock commissioned Forrester to evaluate companies’ IAM practices and requirements when it comes to customer-facing scenarios versus employee-facing ones.
Security for virtualized datacentres
Legacy security solutions are inefficient due to the architectural differences between physical and virtual environments.
Reg Reader Research: SaaS based Email and Office Productivity Tools
Read this Reg reader report which provides advice and guidance for SMBs towards the use of SaaS based email and Office productivity tools.
New hybrid storage solutions
Tackling data challenges through emerging hybrid storage solutions that enable optimum database performance whilst managing costs and increasingly large data stores.