Cookie policy: This site uses cookies (small files stored on your computer) to simplify and improve your experience of this website. Cookies are small text files stored on the device you are using to access this website. For more information on how we use and manage cookies please take a look at our privacy and cookie policies. Some parts of the site may not work properly if you choose not to accept cookies.


The future of virtualization

  • Print
  • Share
  • Comment
  • Save

Cloud computing and virtualization go hand in hand. Virtualization is cloud’s foundation and cloud computing software – such as OpenStack, for example – “is just the bit that sits on top of the house to manage this”, as Rackspace’s VP of technology and product, Nigel Beighton, will tell you.

But as the Cloud evolves, so too must virtualization to support more IO-intensive network and storage workloads, and to ensure that open standards being developed across the industry can also be applied to hypervisor designs. Beighton says most clouds today run on virtualization technology that is ten years old. But work is taking place behind the scenes to revolutionize the way virtualization is done.

The market leader
VMware has the largest market share when it comes to virtualization – 60%, according to research firm IDC. VMware global field CTO Paul Strong says that today most adoption is being driven by consolidation and the desire for private and hybrid cloud environments. But market share does not mean VMware can rest on its laurels. New demands, such as the software-defined data center (SDDC), are pushing virtualization technology forward.

“We have spent the last 15 years focused primarily on separating the application from the compute, or server, infrastructure. However, most applications are in fact network-distributed (think multi-tier apps). So, even though you can switch an instance (say a database server) on almost instantly, it may take hours, days or even weeks to re-program firewalls and load balancers, to assign new IP addresses, or to set up new VLANs for the complete application,” Strong says.

VMware purchased Nicira in July last year for its software-defined network (SDN) capabilities. “This (SDN) provides us with an opportunity to define a container, a virtual data center or virtual application for network-distributed applications,” Strong says.

He says this will eventually mean a container can be “manipulated” in much the same way that VMs are today, but for complete applications. And it is this that provides the foundation for the SDDC.

“SDDC is about extending virtualization from servers to networking and storage, separating the application from the infrastructure and encapsulating it in a container. Once these applications are in these containers, we can automate the lifecycle of the containers and thus, by proxy, the applications within them. Large enterprises have thousands of applications. It is nigh on impossible to automate provisioning and management of each of these individually. But when you place each of these applications within its own container, it looks more or less the same from the perspective of the daily operations, such as provisioning, moving to scale up or down, moving for availability, moving to and from the Cloud.”

Strong says the SDDC will place new demands on hypervisors, which will have to handle IO-intensive network and storage virtual appliances, as well as traditional applications. The hypervisor of tomorrow will need to be denser in order to handle increasing processor cores and threads.

New hypervisor designs
Some new players entering the hypervisor market are basing their initial product offerings around cloud. ZeroVM is one. It claims to have the “first hypervisor specially designed” for the model, with deployment speed, application isolation and efficiency built into its features. It claims a ZeroVM takes 5ms to create, making it possible to separate every single task into its own container. It only virtualizes parts of the server that are required to do the work. “Existing clouds are giant server farms that are spending precious resources virtualizing unneeded things,” ZeroVM’s website says.

The hypervisor uses Unix-style processes that communicate through pipes such as VMware, XEN and KVM, and claims to create a new VM for every single incoming request. It can also aggregate many physical servers and represent them as a single virtual system, or represent a number of virtual systems backed up by any number of physical servers. The end result for ZeroVM is a lightweight hypervisor that offers smaller footprint, allowing you to, hypothetically, divide the hypervisor into 10,000 processes instead of 10 servers, bringing virtualization down to another level.

Beighton says such developments are moving us toward a new age of virtualization. “It is the concept of moving away from just creating servers to deconstructing applications and virtualizing at the application and right down to the process and user level. It is at this point when you know cloud will become very different.”

Open cloud
Cloud has already made demands different, especially when it comes to open standards. OpenStack is just one example of the popularity of the open model. IBM CTO Matt Hogstrom, who is working on IBM’s new software-defined environment program, which is looking at policy-based management of workloads (read more about this in the FOCUS 31 iPad edition at DCDFOCUS), says one of the biggest challenges VMware will have in future will be removing vendor lock-in to create more fluid delivery options through the common use of programming languages.

Strong says VMware is working with a number of open standards and community initiatives from the Distributed Management Task Force (behind the open virtualization format for movement across different vendor’s hypervisors) and OpenStack, which it contributes to through Nicira. These will be another influential factor on the future generations of hypervisor, designed especially for the Cloud.


Over the next two days, FOCUS will be covering the latest announcements from VMworld Europe in Barcelona. Sign up to our daily newsletter to hear the latest news. Or you can read more about virtualization in FOCUS magazine, edition 31. Available as a digital edition here.


Related images

  • VMware's Paul STrong

Have your say

Please view our terms and conditions before submitting your comment.

  • Print
  • Share
  • Comment
  • Save


  • Overhead Power Distribution – Best Practice in Modular Design

    Tue, 10 Nov 2015 16:00:00

    Overhead power distribution in your data center offers many attractive possibilities but is not without its challenges. Join UE Corp’s Director of Marketing, Mark Swift, and CPI’s Senior Data Center Consultant, Steve Bornfield, for an exploration of the options and some of the pitfalls, supported by real-life examples from the field.

  • Overcoming the Challenges of High Power Density Deployments

    Wed, 4 Nov 2015 19:00:00

    Increasing rack power densities saves space and energy, and improves both OPEX and CAPEX. But it can also create unintended problems that could bring your data center to a screeching halt. Join Raritan’s VP of Products & Marketing, Henry Hsu, and DCD’s CTO Stephen Worn, as they reveal the three key challenges in deploying a high density cabinet, and explain how to: Reduce operating costs, Increase up-time , Improve mean time to repair, become more energy-efficient manage existing capacity and plan for growth.

  • Squeezing the Lemon - The Power to do More with Less

    Tue, 20 Oct 2015 08:00:00

    Energy costs rising, manpower resources falling – managing a data center is getting more stressful by the day. One cold night could be all it takes to tip your power supply over the edge. And let's not forget the never-ending demands from IT for additional space. More information on its own is not the answer. Join Rittal's webinar to understand how to: • Lower your power consumption and OPEX charges, with 'smart' power distribution • Identify issues before they become problems, with intelligent PDUs' monitoring capabilities • Expand your DC as your business grows, with modular PDUs • Profile your power requirements to help you plan and make better-informed decisions REGISTER NOW Note: All attendees will receive a free copy of the latest White Paper from Rittal.

  • Live Customer Roundtable: Optimizing Capacity (12:00 EST)

    Tue, 8 Sep 2015 16:00:00

    The biggest challenge facing many data centers today? Capacity. How to optimize what you have today. And when you need to expand, how to expand your capacity smarter. Learn from the experts about how Data Center Infrastructure Management (DCIM) and Prefabricated Modular Data Centers are driving best practices in how capacity is managed and optimized: - lower costs - improved efficiencies and performance - better IT services delivered to the business - accurate long-range planning Don;t miss out on our LIVE customer roundtable and your chance to pose questions to expert speakers from Commscope, VIRTUS and University of Montana. These enterprises are putting best practices to work today in the only place that counts – the real world.

  • Power Optimization – Can Your Business Survive an Unplanned Outage? (APAC)

    Wed, 26 Aug 2015 05:00:00

    Most outages are accidental; by adopting an intelligent power chain, you can help mitigate them and reduce your mean-time to repair. Join Anixter and DatacenterDynamics for a webinar on the five best practices and measurement techniques to help you obtain the performance data you need to optimize your power chain. Register today!

More link