The last mile of virtualization has multiple dimensions, based on where you are going with virtualization. When you ask about the last mile of virtualization—about what it will take to get to 100% virtualized either within your data center or within a cloud (hybrid, public, or private)—the “it depends” answer is the one you will get most often. So, what will it take to get to 100% virtualized? Continue reading Virtualizing the Last Mile
Recently I had the pleasure of speaking with Tim Mackey on the User Experience Podcast, Episode 8.
Tim has been with Citrix for ten years and has been working with XenServer exclusively since 2009. Our conversation revolves around what the XenServer density capabilities are. There has been an ongoing conflict between density stories about Citrix and VMware. The way Citrix defines it centers around how many VMs Citrix “supports” vs. how much VM density can actually be achieved. These are two very separate things in the Citrite dictionary. Tim goes into the actual numbers in our conversation, so if you are interested in hearing it straight from the horse’s mouth, I encourage you to listen to the podcast. Continue reading Open Sorcerer, XenServer Ambassador—Tim Mackey Speaks from the Citrix Mother Ship
Recently, I wrote an article about what Citrix has done around virtualizing GPUs and GPU sharing, based on a podcast with Derek Thorslund, director of product management for HDX at Citrix. When the story hit the social media sphere, I got clobbered with hits about Nutanix and its partnership with Citrix and NVIDIA, along with a handful of requests to lead a podcast and write about its involvement with the GRID vGPU tech in its virtual desktop infrastructure (VDI) business. So, just for the folks who reached out, here is my response. Continue reading VDI, a Beauty & a Beast: A Nutanix Story
In March 2013, Citrix announced they had GPU sharing working and available for XenApp (multi-user/RDS). In December 2013, they announced it was available for XenDesktop (multi-OS/VDI). This has been a major barrier to adoption for many companies that need the ability to deliver a high-end multimedia experience to their end users in order to gain acceptance for adoption. Continue reading Citrix Solved the vGPU Problem, So…Where Is the Parade?
As you can probably tell from the title, Citrix is leveraging their biggest advantages in the mobility/BYOD race: their understanding of ALL client operating systems, multimedia in both SBC and VDI environments, and their established partnerships with hardware and OS vendors. In a conversation I had with Chris Fleck, VP Mobility & Alliances at Citrix (@chrisfleck), we spent an hour talking about the various methods Citrix has decided to use to manage mobile devices in both multi-user and multi-OS virtual environments, while extending their function from consumption to productivity. Oh, yeah, they have also changed their product and technology names to reflect their commitment to mobility; shocking I know. Continue reading Mobility Bytes – Citrix’s Mobile Strategy Has Sharp Teeth & It’s Attacking from Every Side
While not a major version release (we will have to wait for 6.0 next year for that), the new 5.1 version of the VMware products contains some significant new functionality, in addition to the packing of all of the components into the vCloud Suite.
New Features in vSphere 5.1
- User Access – There is no longer a dependency on a shared root account. Local users assigned administrative privileges automatically get full shell access
- Auditing – All host activity from both the shell and the Direct Console User Interface is now logged under the account of the logged in user
- Monitoring – Support is added for SNMPv3. The SNMP agent has been unbundled from the VMkernel and can now be independently updated.
- vMotion – a vMotion and an Storage vMotion can be combined into one operation. This allows a VM to be moved between two hosts or clusters that do not have any shared storage.
- New Windows Support – Support for both the Desktop and Server Editions of Windows 8/2012
- Hardware Accelerated 3D Graphics – Teaming up with NVIDIA, vSphere can now map a vGPU to each VM on a system. Not only does this feature accelerate 3D graphics but provides a GPU for high performance computing needs
- Improvements in Virtual hardware virtualization support – This brings Intel-VT/AMD RVI features further into the virtual machine which will improve virtualization within virtualization. In addition, more low level CPU counters are exposed which can be further used for high performance computing and real time style applications.
- Agentless Antivirus and Antimalware – vShield Endpoint is now included in vSphere 5.1 and offloads anti-virus and antimalware processing inside virtual machines to a secure dedicated virtual appliance delivered by VMware partners. This change lowers the cost of entry for Agentless Angivirus and Malware.
- New 64-vCPU Support – Virtual machines running on a vSphere 5.1 host can be configured with up to 64 virtual CPU’s and 1TB of RAM.
- Auto-Deploy – Auto-Deploy is extended with two new modes, “stateless caching” and “stateful installs”. In addition the number of concurrent reboots per Auto-Deploy host has been increased to 80
- SR-IOV Support – Single Root I/O Virtualization allows certain Intel NIC’s to transfer data directly into the memory space of a virtual machine without any involvement from the hypervisor. See this Intel Video
- Space Reclaiming Thin Provisioned Disks – These types of disks add the ability to reclaim deleted blocks from existing thin provisioned disks while the VM is running. To reclaim space is a two-part function of first wiping the disk marking unused blocks as free, and then to shrink the disk. These two features have been a part of VMware Tools for a number of years but now do things differently for thin provisioned disks. The underlying hardware is not initially part a part of the reclamation process. Instead the vSCSI layer within ESX reorganizes unused blocks to keep the used part of the thin provisioned disk contiguous. Once the unused parts are at the end of the thin provisioned disk then the hardware is involved.
- Tunable Block Size – Normally thin provisioned disks use a 4KB block size that is unchanging, however, this block size can be tuned indirectly as it is now based on the requirements of the underlying storage array. There is no method to tune this by hand.
- All Paths Down Improvements – When there was an all paths down (APD) situation, the vSphere management service would hang waiting on disk IO, which would cause the vSphere host to inadvertently disconnect from vCenter and in effect become unmanageable. APD handling has been improved such that transient APD events will not cause the vSphere management service to hang waiting on disk IO, use of vSphere HA to move workloads around to other hosts if APD detects a permanent device lost (PDL) situation, and implement a way to detect PDL for iSCSI arrays that present only one LUN.
- Storage Hardware/Software improvements – These improvements include the ability to boot from software FCoE, additions of Jumbo frame support for all iSCSI adapters (software or hardware), and support for 16Gb FC
- VAAI Improvements – VAAI has added support to allow vCloud Director fast-provisioned vApps to make use of VAAI enabled NAS array-based snapshots.
- vSphere S.M.A.R.T. (Self-Monitoring, Analysis and Reporting Technology) Implementation – vSphere has implemented SMART reporting via the esxcli commands so that SSD and other disks can report back on their status. In addition, esxcli has been upgraded to include ways to reset specific FC adapters directly as well as methods to retrieve event caching information such as link-up and link-down.
- Storage IO Contral Statistics and Settings Improvements – Finding the proper value for SIOC has been problematic, now it is possible to set a percentage instead of a millisecond value to determine when SIOC should fire. In addition, SIOC will report stats immediately instead of waiting. This allows Storage DRS has statistics available immediately, which improve its decision process. In addition, the observed latency of a VM (a new metric) is available within the vSphere Client performance charts. The observed latency is latency within the host and not just latency after storage packets leave the host
- Storage DRS Improvements – Storage DRS has been improved for workloads using vCloud Directory. Linked clones can now be migrated between datastores if there exists either the base disk or a shadow copy of the base disk. Storage DRS is also used now for initial placement of workloads when using vCloud Director.
- Improvements in Datastore Correlation for Non-VASA enabled arrays – For storage devices that do not support VASA it is difficult to correlate datastores against disk spindles on an array. There are now improvements in the datastore correlation such that vSphere can now detect if spindles are shared by datastores on the array regardless of VASA support.
By exposing virtual hardware (Intel-VT/AMD RVI) virtualization as well as more CPU counters and components VMware has exposed more capability than ever before. Tie this with virtual graphics processing units and we now have the ability to implement virtualized high performance and real-time computing environments. Add to this the storage improvements and large scale big data applications as well as high performance computing environments can be virtualized. Both require low latency networking and storage.
Virtualization has long since been the bane of high performance applications whether that is 3D graphics, high performance computing, big data, or real-time applications. vSphere 5.1 provides a possible solution to use cases while improving integration with the VMware vCloud Suite.