Is it time to plan for the virtual future in our virtual designs? Happy New Year and welcome to 2013!! What a year 2012 turned out to be for virtualization and cloud computing in general. Microsoft Hyper-V, Red Hat, and VMware have all made quite a few enhancements to the hypervisor, and we have finally reached a point where we really have some good competition between hypervisors. Also, the competition boundaries are being expanded to include much more than just the hypervisor itself as we start to focus on the ecosystem as a whole. Toward the end of 2012 the industry had really begun presenting multi-hypervisor management capabilities and solutions. I see this area as something to really watch in 2013, and I propose this question.
Although multi-hypervisor management is still in its infancy, I think that it is time we start the planning process for the future. What I mean by that is, as we move forward with new deployments and/or infrastructure refreshes, we should be building the capability of running multiple hypervisors into our virtual designs. The trend of running multiple hypervisors, in my opinion, will continue to trend in the future. I am not saying the complete design for multiple platforms should be presented, but rather in our designs we should have a placeholder available and marked off so that if the decision is made to deploy a parallel hypervisor, we will have something ready for the future when called.
I have been talking with a few companies lately that are in the planning stages for their virtual refresh, and during these conversations I like to ask this question: What are the plans, if any, for the deployment of multiple hypervisors? The answers that I have been hearing lately are that the idea is being considered but for now we are going to concentrate on one hypervisor, and we will revisit this idea once the initial deployment has been completed. That answer is neither a “No” nor a “Yes”, and to help make life easier in the long run, consideration for this kind of deployment should be evaluated, but up front with a place to put things and at least a general idea of how we could manage this should we be called to task.
I understand that things are changing rapidly, and what we would consider doing now might not be what we would actually do tomorrow. For example, we could say that we have been tasked to deploy the VMware vCloud Suite onto a datacenter with two clusters (management and production). Once we have finished the virtual design, we could present Hotlink technologies as an integration engine that would give us the ability to branch different hypervisors off the main virtual design, which would give us a plan for the future. When the time comes to actually deploy the branched off-hypervisor, Hotlink might not be the best choice, and Microsoft System Center could be a different choice. We could easily replace Hotlink with System Center in the design and still be able to work from the initial plan.
My point is this: technology will change, but the time has come to start planning for what the future may bring and to at least have a plan in place that can be modified as technology changes over time. Why do we need to reinvent the wheel for each technology? We should be planning for heterogeneous environments from the start.
Share this Article:
Latest posts by Steve Beaver (see all)
- How Is Artificial Intelligence Development Going in Your Environment? - January 18, 2017
- Will 2017 Be the Year of the Self-Healing Data Center? - January 3, 2017
- The Next-Generation Data Center - December 19, 2016