VMware ESXi

VMware ESXi
Developer(s) VMware, Inc.
Stable release
6.0.0 Update 2 (Build 3620759) / 15 March 2016 (2016-03-15)
Platform i386 (discontinued in 4.0 onwards),[1] x86-64
Type Hypervisor
License Proprietary
Website vmware.com/products/esxi-and-esx/

VMware ESXi (formerly ESX) is an enterprise-class, type-1 hypervisor developed by VMware for deploying and serving virtual computers. As a type-1 hypervisor, ESXi is not a software application that one installs in an operating system (OS); instead, it includes and integrates vital OS components, such as a kernel.[2]

After version 4.1 (released in 2010), VMware renamed ESX to ESXi. ESXi replaces Service Console (a rudimentary operating system) with a more closely integrated OS. ESX/ESXi is the primary component in the VMware Infrastructure software suite.[3]

The name ESX originated as an abbreviation of Elastic Sky X.[4][5]

Architecture

ESX runs on bare metal (without running an operating system)[6] unlike other VMware products.[7] It includes its own kernel: A Linux kernel is started first,[8] and is then used to load a variety of specialized virtualization components, including ESX, which is otherwise known as the vmkernel component.[9] The Linux kernel is the primary virtual machine; it is invoked by the service console. At normal run-time, the vmkernel is running on the bare computer, and the Linux-based service console runs as the first virtual machine. VMWare dropped development of ESX at version 4.1, and now uses ESXi, which does not include a Linux kernel.[10]

The vmkernel is a microkernel[11] with three interfaces: hardware, guest systems, and the service console (Console OS).

Interface to hardware

The vmkernel handles CPU and memory directly, using scan-before-execution (SBE) to handle special or privileged CPU instructions[12][13] and the SRAT (system resource allocation table) to track allocated memory.[14]

Access to other hardware (such as network or storage devices) takes place using modules. At least some of the modules derive from modules used in the Linux kernel. To access these modules, an additional module called vmklinux implements the Linux module interface. According to the README file, "This module contains the Linux emulation layer used by the vmkernel."[15]

The vmkernel uses the device drivers:[15]

  1. net/e100
  2. net/e1000
  3. net/e1000e
  4. net/bnx2
  5. net/tg3
  6. net/forcedeth
  7. net/pcnet32
  8. block/cciss
  9. scsi/adp94xx
  10. scsi/aic7xxx
  11. scsi/aic79xx
  12. scsi/ips
  13. scsi/lpfcdd-v732
  14. scsi/megaraid2
  15. scsi/mptscsi_2xx
  16. scsi/qla2200-v7.07
  17. scsi/megaraid_sas
  18. scsi/qla4010
  19. scsi/qla4022
  20. scsi/vmkiscsi
  21. scsi/aacraid_esx30
  22. scsi/lpfcdd-v7xx
  23. scsi/qla2200-v7xx

These drivers mostly equate to those described in VMware's hardware compatibility list.[16] All these modules fall under the GPL. Programmers have adapted them to run with the vmkernel: VMware Inc has changed the module-loading and some other minor things.[15]

Service console

In ESX (and not ESXi), the Service Console is a vestigial general purpose operating system most significantly used as bootstrap for the VMware kernel, vmkernel, and secondarily used as a management interface. Both of these Console Operating System functions are being deprecated from version 5.0, as VMware migrates exclusively to the ESXi model, current version being ESXi.[17] The Service Console, for all intents and purposes, is the operating system used to interact with VMware ESX and the virtual machines that run on the server.

Linux dependencies

ESX uses a Linux kernel to load additional code: often referred to by VMware, Inc. as the "vmkernel". The dependencies between the "vmkernel" and the Linux part of the ESX server have changed drastically over different major versions of the software. The VMware FAQ[18] states: "ESX Server also incorporates a service console based on a Linux 2.4 kernel that is used to boot the ESX Server virtualization layer". The Linux kernel runs before any other software on an ESX host.[8] On ESX versions 1 and 2, no VMkernel processes run on the system during the boot process.[19] After the Linux kernel has loaded, the S90vmware script loads the vmkernel.[19] VMware Inc states that vmkernel does not derive from Linux, but acknowledges that it has adapted certain device-drivers from Linux device drivers. The Linux kernel continues running, under the control of the vmkernel, providing functions including the proc file system used by the ESX and an environment to run support applications.[19] ESX version 3 loads the VMkernel from the Linux initrd, thus much earlier in the boot-sequence than in previous ESX versions.

In traditional systems, a given operating system runs a single kernel. The VMware FAQ mentions that ESX has both a Linux 2.4 kernel and vmkernel – hence confusion over whether ESX has a Linux base. An ESX system starts a Linux kernel first, but it loads vmkernel (also described by VMware as a kernel), which according to VMware 'wraps around' the linux kernel, and which (according to VMware Inc) does not derive from Linux.

The ESX userspace environment, known as the "Service Console" (or as "COS" or as "vmnix"), derives from a modified version of Red Hat Linux, (Red Hat 7.2 for ESX 2.x and Red Hat Enterprise Linux 3 for ESX 3.x). In general, this Service Console provides management interfaces (CLI, webpage MUI, Remote Console).

As a further detail which differentiates the ESX from other VMware virtualization products: ESX supports the VMware proprietary cluster file system VMFS. VMFS enables multiple hosts to access the same SAN LUNs simultaneously, while file-level locking provides simple protection to file-system integrity.

Purple Screen of Death

A purple diagnostic screen as seen in VMware ESX Server 3.0
A purple diagnostic screen from VMware ESXi 4.1

In the event of a hardware error, the vmkernel can 'catch' a Machine Check Exception.[20] This results in an error message displayed on a purple diagnostic screen. This is colloquially known as a purple diagnostic screen, or purple screen of death (PSOD, cf. Blue Screen of Death (BSOD)).

Upon displaying a purple diagnostic screen, the vmkernel writes debug information to the core dump partition. This information, together with the error codes displayed on the purple diagnostic screen can be used by VMware support to determine the cause of the problem.

vMotion: live migration

Live migration (vMotion) in ESX allows a virtual machine to move between two different hosts. Live storage migration (Storage vMotion) enables live migration of virtual disks on the fly.[21] During vMotion Live Migration (vLM) of a running virtual machine (VM) the content of the (RAM) memory of the VM is sent from the running VM to the new VM (the instance on another host that will become the running VM after the vLM). The content of memory is by its nature changing all the time. ESX uses a system where the content is sent to the other VM and then it will check what data is changed and send that, each time smaller blocks. At the last moment it will very briefly 'freeze' the existing VM, transfer the last changes in the RAM content and then start the new VM. The intended effect of this process is to minimize the time during which the VM is suspended; in a best case this will be the time of the final transfer plus the time required to start the new VM.[22][23]

Versions

VMware ESX is available in two main types: ESX and ESXi, although since version 5 only ESXi is continued.

VMware ESX

Version release history:

VMware ESX 1.5

VMware ESX 2.0 (21 July 2003)

VMware ESX 2.5 (14 December 2004)

VMware Infrastructure 3.0 (VI3) (5 June 2006)

VMware vSphere 4.0 (20 May 2009)

ESX and ESXi before version 5.0 do not support Windows 8/Windows 2012. These Microsoft operating systems can only run on ESXi 5.x or later.[25]

18 July 2010 vSphere 4.1 and its subsequent update and patch releases are the last releases to include both ESX and ESXi hypervisor architectures. Future major releases of VMware vSphere will include only the VMware ESXi architecture. For this reason, VMware recommends that deployments of vSphere 4.x utilize the ESXi hypervisor architecture.

VMware ESXi

VMware ESXi, a smaller-footprint version of ESX, does not include the ESX Service Console. It is available - without the need to purchase a vCenter license - as a free download from VMware, with some features disabled.[26][27][28]

ESXi apparently stands for "ESX integrated".[29]

VMware ESXi originated as a compact version of VMware ESX that allowed for a smaller 32 MB disk footprint on the host. With a simple configuration console for mostly network configuration and remote based VMware Infrastructure Client Interface, this allows for more resources to be dedicated to the guest environments.

Two variations of ESXi exist:

The same installation media will install to either one or the other of these installation modes depending on the size of the target media.[30] One can upgrade ESXi to VMware Infrastructure 3[31] or to VMware vSphere 4.0 ESXi.

Originally named VMware ESX Server ESXi edition, through several revisions the ESXi product finally became VMware ESXi 3. New editions then followed: ESXi 3.5, ESXi 4 ESXi 5 and (as of 2015) ESXi 6.

To virtualize Windows 8 or Windows Server 2012 as a guest operating system, the ESXi version must be 5.0 update 1 or later.

Version history

Version release history:[28]

Lawsuit

VMware has been sued by Christoph Hellwig, a prominent Linux kernel developer, for GPL license violations. It is alleged that VMware has misappropriated portions of the Linux kernel,[32] and used them without permission.

The following products operate in conjunction with ESX:

Cisco Nexus 1000v

Network-connectivity between ESX hosts and the VMs running on it relies on virtual NICs (inside the VM) and virtual switches. The latter exists in two versions: the 'standard' vSwitch allowing several VMs on a single ESX host to share a physical NIC and the 'distributed vSwitch' where the vSwitches on different ESX hosts together form one logical switch. Cisco offers in their Cisco Nexus product-line the Nexus 1000v, an advanced version of the standard distributed vSwitch. A Nexus 1000v consists of two parts: a supervisor module (VSM) and on each ESX host a virtual ethernet module (VEM). The VSM runs as a virtual appliance within the ESX cluster or on dedicated hardware (Nexus 1010 series) and the VEM runs as module on each host and replaces a standard dvS (distributed virtual switch) from VMware. Configuration of the switch is done on the VSM using the standard NX-OS CLI. It offers capabilities to create standard port-profiles which can then be assigned to virtual machines using vCenter.

There are several differences between the standard dvS and the N1000v, one is that the Cisco switch generally has full support for network technologies such as LACP link aggregation or that the VMware switch supports new features such as routing based on physical NIC load. However the main difference lies in the architecture: Nexus 1000v is working in the same way as a physical Ethernet switch does while dvS is relying on information from ESX. This has consequences for example in scalability where the Kappa limit for a N1000v is 2048 virtual ports against 60000 for a dvS. The Nexus1000v is developed in co-operation between Cisco and VMware and uses the API of the dvS[35]

Third party management tools

Because VMware ESX is a leader in the server-virtualisation market,[36] software and hardware vendors offer a range of tools to integrate their products or services with ESX. Examples are the products from Veeam Software with backup and management applications[37] and a plugin to monitor and manage ESX using HP OpenView,[38] Quest Software with a range of management and backup-applications and most major backup-solution providers have plugins or modules for ESX. Using Microsoft Operations Manager (SCOM) 2007/2012 with a Bridgeways ESX management pack gives you a realtime ESX datacenter health view.

Also hardware-vendors such as HP and Dell include tools to support the use of ESX(i) on their hardware platforms. An example is the ESX module for Dell's OpenManage management platform.[39]
VMware have added a Web Client[40] since v5 but it will work on vCenter only and does not contain all features.[41] vEMan[42] is a Linux application which is trying to fill that gap. These are just a few examples: there are numerous 3rd party products to manage, monitor or backup ESX infrastructures and the VMs running on them.[43]

Known limitations

Known limitations of VMware ESXi, as of April 2015, include the following:

Infrastructure limitations

Some maximums in ESXi Server 6.0 may influence the design of data centers:[44]

Performance limitations

In terms of performance, virtualization imposes a cost in the additional work the CPU has to perform to virtualize the underlying hardware. Instructions that perform this extra work, and other activities that require virtualization, tend to lie in operating system calls. In an unmodified operating system, OS calls introduce the greatest portion of virtualization "overhead".

Paravirtualization or other virtualization techniques may help with these issues. VMware developed the Virtual Machine Interface for this purpose, and selected operating systems currently support this. A comparison between full virtualization and paravirtualization for the ESX Server[45] shows that in some cases paravirtualization is much faster.

Network limitations

When using the advanced and extended network capabilities by using the Cisco Nexus 1000v distributed virtual switch the following network-related limitations apply:[35]

  • 64 ESX/ESXi hosts per VSM (Virtual Supervisor Module)
  • 2048 virtual ethernet interfaces per VMWare vDS (virtual distributed switch)
  • and a maximum of 216 virtual interfaces per ESX/ESXi host
  • 2048 active VLAN's (one to be used for communication between VEM's and VSM)
  • 2048 port-profiles
  • 32 physical NIC's per ESX/ESXi (physical) host
  • 256 port-channels per VMWare vDS (virtual distributed switch)
  • and a maximum of 8 port-channels per ESX/ESXi host

Fibre Channel Fabric limitations

Regardless of the type of virtual SCSI adapter used, there are these limitations:[44]

See also

References

  1. "VMware ESX 4.0 only installs and runs on servers with 64bit x86 CPUs. 32bit systems are no longer supported.". VMware, Inc.
  2. "ESX Server Architecture". Vmware.com. Archived from the original on 2009-11-07. Retrieved 2009-10-22.
  3. VMWare:vSphere ESX and ESXi Info Center
  4. What does ESX stand for?
  5. "Glossary" (PDF). Developer’s Guide to Building vApps and Virtual Appliances: VMware Studio 2.5. Palo Alto: VMware. 2011. p. 153. Retrieved 2011-11-09.
  6. "ESX Server Datasheet"
  7. "ESX Server Architecture". Vmware.com. Archived from the original on 29 September 2007. Retrieved 2009-07-01.
  8. 1 2 "ESX machine boots". Video.google.com.au. 12 June 2006. Retrieved 2009-07-01.
  9. "VMKernel Scheduler". vmware.com. Retrieved 2016-03-10.
  10. Mike, Foley. "It's a Unix system, I know this!". VMWare Blogs. VMware.
  11. "Support for 64-bit Computing". Vmware.com. 19 April 2004. Retrieved 2009-07-01.
  12. Gerstel, Markus: "Virtualisierungsansätze mit Schwerpunkt Xen" Archived 15 August 2013 at the Wayback Machine.
  13. VMware ESX
  14. "VMware ESX Server 2: NUMA Support" (PDF). Palo Alto, California: VMware Inc. 2005. p. 7. Retrieved 2011-03-29. SRAT (system resource allocation table) – table that keeps track of memory allocated to a virtual machine.
  15. 1 2 3 "ESX Server Open Source". Vmware.com. Retrieved 2009-07-01.
  16. "ESX Hardware Compatibility List". Vmware.com. 10 December 2008. Retrieved 2009-07-01.
  17. "ESXi vs. ESX: A comparison of features". Vmware, Inc. Retrieved 2009-06-01.
  18. VMware FAQ Archived 11 February 2007 at the Wayback Machine.
  19. 1 2 3 ESX Server Advanced Technical Design Guide Archived 25 February 2007 at the Wayback Machine.
  20. "KB: Decoding Machine Check Exception (MCE) output after a purple diagnostic screen |publisher=VMware, Inc."
  21. The Design and Evolution of Live Storage Migration in VMware ESX
  22. 1 2 VMWare Blog by Kyle Gleed: vMotion: what's going on under the covers, 25 February 2011, visited: 2 February 2012
  23. 1 2 VMware website vMotion brochure. Retrieved 3 February 2012
  24. http://kb.vmware.com/kb/1003661
  25. VMWare KBArticle Windows 8/Windows 2012 doesn't boot on ESX, visited 12 September 2012
  26. "Download VMware vSphere Hypervisor (ESXi)". www.vmware.com. Retrieved 22 July 2014.
  27. "Getting Started with ESXi Installable" (PDF). VMware. Retrieved 22 July 2014.
  28. 1 2 "VMware ESX and ESXi 4.1 Comparison". Vmware.com. Retrieved 2011-06-09.
  29. "What do ESX and ESXi stand for?". VM.Blog. 2011-08-31. Retrieved 2016-06-21. Apparently, the 'i' in ESXi stands for Integrated, probably coming from the fact that this version of ESX can be embedded in a small bit of flash memory on the server hardware.
  30. Andreas Peetz. "ESXi embedded vs. ESXi installable FAQ". Retrieved 2014-08-11.
  31. "Free VMware ESXi: Bare Metal Hypervisor with Live Migration". Vmware.com. Retrieved 2009-07-01.
  32. "Conservancy Announces Funding for GPL Compliance Lawsuit". sfconservancy.org. 5 March 2015. Retrieved 2015-08-27.
  33. http://www.vmware.com/files/pdf/VMware-Storage-VMotion-DS-EN.pdf
  34. http://www.vmware.com/files/pdf/VMware-Fault-Tolerance-FT-DS-EN.pdf
  35. 1 2 Overview of the Nexus 1000v virtual switch, visited 9 July 2012
  36. VMware continues virtualization market romp, 18 April 2012. Visited: 9 July 2012
  37. About Veeam, visited 9 July 2012
  38. Veeam OpenView plugin for VMware, visited 9 July 2012
  39. OpenManage (omsa) support for ESXi 5.0, visited 9 July 2012
  40. VMware info about Web Client – VMware ESXi/ESX 4.1 and ESXi 5.0 Comparison
  41. Availability of vSphere Client for Linux systems – What the web client can do and what not
  42. vEMan website vEMan – Linux vSphere client
  43. Petri website 3rd party ESX tools, 23 December 2008. Visited: 11 september 2001
  44. 1 2 "Configuration Maximums" (PDF). VMware, Inc. 22 April 2015. Retrieved 2015-04-22.
  45. "Performance of VMware VMI" (PDF). VMware, Inc. 13 February 2008. Retrieved 2009-01-22.
This article is issued from Wikipedia - version of the 12/2/2016. The text is available under the Creative Commons Attribution/Share Alike but additional terms may apply for the media files.