Technology -> Applications
By: Clive Longbottom, Head of Research, Quocirca
Published: 23rd November 2012
Copyright Quocirca © 2012
The growing use of virtualisation has really helped many organisations.
Not only have the average utilisation rates of servers and storage improved, but the use of applications and other software packaged ready for installation – commonly known as virtual images or virtual machines – has meant that systems can be implemented or recovered far faster than they used to be.
However, this can be a two-edged sword.
The good side of being able to implement a runtime application rapidly is seen in hosted systems, cloud computing and private datacentres; but the bad side is seen most in development and test departments, and is spreading out into the runtime.
The problem is that virtual machines (VMs) are just too easy to use. In the past, if you wanted to install a copy of an application, the first thing to do was order a server. Then wait to receive the server. Then get it up and running, install all the patches to the operating system that the supplier had neglected to put in place. Then install all the support software that is required – app server, database, whatever, followed finally by the software you want to run. Long-winded? Yes – and often enough to put a general developer off, and they would just re-use a single server time and time again, cleaning the server down after each test and building back up from a golden back-up image to then test the next iteration of their software. Maybe a couple of hours each time to get to a “clean” position.
Today it is possible to grab some spare resource from a virtualised hardware base, spin up a VM and then install your software. This takes just a few minutes, and as the resource pool can be pretty big, it is easy for the developer to “forget” that they have a live VM running and just start up another one. IT departments could experience greater problems with VM sprawl – with test groups growing the VM pool and users being able to self-service systems that they may only use a couple of times.
The move towards a development/operations (DevOps) model for organising IT, where the development and test employees can push new images directly into the runtime, will make it much harder for IT administrators to keep track of all VMs.
Effective management of software licences and VMs
The result is that not only are resources being locked down by VMs that are not doing anything useful, but there could also be licences tied up in these VMs that are doing absolutely nothing useful. For many, it may not appear to be an issue – unless someone from the Federation Against Software Theft (FAST) walks in through the door asking to carry out a licence inspection.
Managing licences is something that many organisations still do not do. Suppliers such as Flexera offer full-service licence management, which can not only track licence usage, but also manage them against suppliers’ licence agreements and, in most cases, against their tiering systems, ensuring that an organisation gets the best value from its licences. Others, such as Centrix Software, can track licences and advise on how they are being used so that an organisation can decide how licences should be allocated more effectively, although Centrix really is for dealing with virtual desktop systems. However, what a buyer really should be looking for is a system that not only manages licences, but also manages the lifecycle of the VM itself. Features to look out for include:
Optimising the virtual environment
Most of the incumbent systems management companies – IBM with Tivoli, CA, BMC – are moving in this direction in one way or another. However, others are doing more. Dell has been building on its Kace acquisition, and now that it has acquired Quest Software, expect to see a rapid move to a more full-service physical/ virtual systems management toolset.
Another company to watch is Serena Software. Under the umbrella of “orchestrated IT”, Serena is taking its existing application lifecycle management (ALM) approach and expanding it through to offer an organisation the choice of running as separate, but closely managed, development and test teams and a runtime team, or moving towards a more seamless DevOps approach where the various VMs are all fully managed according to a corporately and technically defined set of rules.
Outside of its Tivoli systems management capability, IBM also has its PureSystems and its z/Enterprise groups, with a universal resource manager that can ensure that a workload is placed on the best available resources – whether this be Windows, Linux or even a mainframe platform in the case of z/Enterprise, and also whether an Intel or Power chip is the best place for that workload to lie. This still needs the basic capabilities of Tivoli for other areas of managing the build and management of VMs, but gives good pointers as to the probable future of a fully managed virtual environment.
Virtualisation is a definite positive evolution in the use of available hardware resources.
However, organisations and technical teams have to understand that it is no silver bullet on its own. In fact, uncontrolled usage of virtualisation can lead to bigger problems where VM sprawl happens, at both the resource and the corporate responsibility levels. It is incumbent on those responsible for the IT function to ensure that the right systems are in place, to enable VMs to be managed at the right levels of granularity for full lifecycle management, with licence recovery and full audit capabilities in place to ensure that everything works to the best possible level.
We automatically stop accepting comments 180 days after a post is published. If you would like to know more about this subject, please contact us and we'll try to help.
Published by: electronicdawn Ltd.