Hypervisor Reference

How Nutanix AHV actually works, what changes when you move from ESXi, and what stays the same

AHV is the native hypervisor of the Nutanix Cloud Platform. It is a KVM-based hypervisor with Nutanix orchestration on top, included with the platform license at no separate hypervisor cost.

This guide is the engineer-grade reference for what runs on an AHV host, how it integrates with the rest of Nutanix, how it differs operationally from ESXi, and what a vSphere team should expect during the first 90 days of adoption.

⏱ 18 min read Engineering-led | Multi-vendor | Operations-focused

Key Takeaways

  • AHV is a Linux distribution running KVM, libvirt, QEMU, and Open vSwitch with Nutanix orchestration on top - not a proprietary hypervisor stack.
  • The operational shift from ESXi to AHV follows a predictable pattern: unfamiliar in the first 30 days, natural by 90 days, with operational parity thereafter.
  • AHV is included with Nutanix Cloud Platform at no separate hypervisor cost, eliminating per-CPU or per-core licensing that applies to ESXi deployments.
  • PowerCLI and vSphere automation does not translate to AHV - existing scripts require re-authoring against the Acropolis API surface and Nutanix tooling.
  • Nutanix Lifecycle Manager handles AOS, AHV, NCC, and firmware updates in a single coordinated workflow, consolidating upgrade complexity compared to vSphere environments.

What's Familiar About AHV, and What Isn't

AHV reads as unfamiliar to vSphere administrators because the names, the management interface, and the storage model are different. Most of the underlying technology is not unfamiliar at all. AHV is a Linux distribution running KVM, libvirt, QEMU, and Open vSwitch, with Nutanix orchestration on top. The operational shift is real, but it is smaller than it looks once the architecture is on the page.

What Runs On An AHV Host

The components below sit on every AHV host. The first three (KVM stack, Acropolis Agent, Controller VM) are the foundation. The remaining three (networking, VM lifecycle, storage path) are where most operational interaction happens. Reading this section is the difference between AHV feeling proprietary and AHV feeling like a familiar Linux virtualization stack with Nutanix orchestration on top.

The KVM Stack

AHV is a Linux distribution running KVM (the kernel virtualization module), libvirt (the management substrate via libvirtd), and QEMU (the userspace device model). A running VM on AHV is a QEMU process under libvirtd's management. This is standard Linux virtualization, the same foundation that underlies Red Hat Virtualization, oVirt, OpenStack KVM, and most cloud-provider compute platforms.

The Acropolis Agent

The Nutanix-specific agent on each host. It communicates with the cluster's Acropolis service, which is distributed across the CVMs. The agent handles VM lifecycle requests, host status reporting, live migration coordination, and integration with AOS for storage and networking. This is the orchestration layer that turns standard KVM into a coordinated cluster.

The Controller VM

A Linux VM that runs on every AHV host and hosts AOS services. The CVM owns the local NVMe drives and serves storage to other VMs through the Distributed Storage Fabric. It is allocated dedicated resources at boot and is not migrated like other VMs. CVM health is critical to cluster health, which is why Aegis Managed Nutanix monitors CVM status as a top-level concern.

AHV Networking

Open vSwitch is the SDN data plane. Linux bridges (br0 by default, with additional bridges for traffic segmentation) connect VMs to the network. Physical NIC bonds (LACP, active-backup, balance-slb) carry all traffic. VLANs are configured per VM NIC. Flow Network Security provides VM-level microsegmentation, the AHV equivalent of NSX-T's distributed firewall capability.

AHV VM Lifecycle

VMs are created via Prism, the Acropolis API, Terraform, or imported via Nutanix Move. Standard VM constructs apply: vCPUs, memory, vDisks, vNICs, BIOS or UEFI boot. Live migration is QEMU/libvirt migration with Nutanix coordination: memory state copies between hosts, and disk state stays in DSF. VM HA, affinity rules, and Acropolis Dynamic Scheduling (ADS) for cluster load balancing are platform-native.

The Storage Path

A VM writes to a vDisk in the Distributed Storage Fabric. The path is VM, then QEMU storage controller, then AOS volume group, then CVM, then local NVMe (or remote NVMe via the cluster fabric for non-local data). The CVM adds a layer compared to ESXi's VMkernel-direct model. It also adds capability: deduplication, compression, snapshots, and replication that VMware historically required separate licensing for.

What The Operational Shift Looks Like

For a team coming from vSphere, the AHV operational shift has predictable phases. The timeline below is an illustrative pattern, not a commitment for any specific environment. Engagement-specific assessment determines real timelines.

First 30 Days Post-Migration

Daily operations feel different. Prism's interface is unfamiliar. Common operations (create VM, modify VM, snapshot, migrate) are rediscovered in the new tool. Existing automation needs re-authoring against the AHV API surface. The team's vSphere muscle memory is briefly a friction point. This is the period where co-managed coverage matters most.

30 To 90 Days Post-Migration

Prism becomes natural. Common operations are faster than they were in vCenter for most tasks. Operational patterns settle: backup, monitoring, change management workflows stabilize. Edge cases and atypical operations require Nutanix knowledge that the team is still building, which is normal at this stage.

90 Days And Beyond

Operational parity with the prior vSphere environment. The platform's cost model and lifecycle simplicity become evident. Upgrade cycles execute through LCM with less coordination overhead than comparable vCenter-driven upgrades. The team is no longer in transition: AHV is the operational baseline.

What Accelerates The Shift

Hands-on Nutanix training before migration (Nutanix offers official training; IVI engagements typically include knowledge transfer components). Aegis Managed Nutanix during the early period, where IVI absorbs the operational expertise gap while the team builds it. Re-authoring the most-used automation early, before the team's vSphere-era automation atrophies and the requirements get harder to capture.

How Operational Dimensions Map From ESXi To AHV

The dimensions below are the ones that actually change for a team migrating off vSphere. This list is the working reference for planning the operational shift. VM concepts (CPU, memory, disk, network), guest OS support, NGT (Nutanix Guest Tools, the VMware Tools equivalent), and modern backup tooling do not change meaningfully and are not listed here.

Licensing Model

ESXi is licensed separately from the rest of the platform. AHV is included with Nutanix Cloud Platform at no separate hypervisor cost. There is no per-CPU or per-core hypervisor SKU.

Management Surface

ESXi is managed through vCenter. AHV is managed through Prism Element (per-cluster) and Prism Central (multi-cluster). Prism is web-based and mobile-responsive, designed for operational simplicity with fewer dials. The tradeoff is fewer ways to misconfigure, fewer ways to fine-tune.

CLI And Automation

ESXi has esxcli and PowerCLI. AHV has nuclei (Nutanix CLI), the Acropolis CLI, the v3 REST API, and the Nutanix PowerShell module. Both platforms are fully programmable; the API surfaces are different, so existing automation requires re-authoring rather than translation.

Lifecycle And Upgrades

ESXi upgrades go through vCenter Update Manager or vSphere Lifecycle Manager. AHV upgrades go through Nutanix Lifecycle Manager (LCM), which handles AOS, AHV, NCC, and firmware in a single coordinated workflow. LCM upgrades are non-disruptive at the cluster level: VMs migrate as hosts are updated in rolling fashion.

Networking And Storage Architecture

vSphere uses distributed switches and VMFS or VVOLs. AHV uses Open vSwitch with consistent per-host configuration (no distributed switch needed) and Nutanix DSF (no VMFS equivalent). vDisks are objects in DSF, not files on a filesystem in the traditional sense. Snapshot, clone, and replication are platform-native.

Support Model

ESXi support is from VMware (now Broadcom). AHV support is from Nutanix. Under Aegis, IVI is the first call for both operational issues and escalation management, with Nutanix TAC handling software defects and the customer team owning operational decisions.

Four architectural truths worth understanding before adopting AHV

AHV's architectural choices have consequences for operations, cost, and ecosystem fit. The four points below are the ones engineers and architects should understand before evaluating AHV against ESXi or other hypervisors. None of them is a marketing claim; each is a structural property of how the platform is built.

KVM-native and Linux-ecosystem aligned

Standard hypervisor stack: AHV is a Linux distribution running KVM, libvirt, QEMU, and Open vSwitch. The same technology stack underlies Red Hat Virtualization, oVirt, OpenStack KVM, and most cloud-provider compute platforms. AHV is not a fork; it is an operationally hardened distribution with Nutanix orchestration layered on top.

What this means in practice: Linux administrators recognize the stack. Standard KVM and libvirt tooling can inspect VMs at the host level if needed. The hypervisor follows Linux kernel direction rather than diverging from it. KVM skills built elsewhere transfer to AHV, which is part of why the operational shift is smaller than the unfamiliar interface initially suggests.

Hypervisor included with the platform license

No separate SKU: AHV is included with Nutanix Cloud Platform at no separate hypervisor cost. There is no per-CPU or per-core hypervisor licensing component, and the included hypervisor delivers the full feature set (HA, live migration, ADS, VM-level operations).

Cost implication in the platform comparison: For environments running ESXi on Nutanix hardware, the VMware license is a separate line item that disappears when AHV is the hypervisor. This is one of the structural cost differences in a Nutanix versus VMware total-cost comparison, and it is the reason AHV adoption is often paired with AIM cluster planning rather than treated as a standalone hypervisor decision.

Storage integrated with the hypervisor

No VMFS layer; data services are platform-native: AHV uses Nutanix DSF directly. There is no VMFS or VVOLs layer to manage. The CVM provides distributed storage services, including deduplication, compression, snapshots, and replication, as platform capabilities rather than separately licensed features.

Operational implication: vDisks are objects in DSF, served by the local CVM whenever possible (data locality). Storage operations such as snapshot, clone, and replicate use AHV-native primitives rather than VMware ecosystem tools. Backup vendor integration is well-supported; verify with each backup vendor for current AHV version coverage.

Coordinated, non-disruptive lifecycle via LCM

One tool for AOS, AHV, NCC, and firmware: Nutanix Lifecycle Manager handles the full software and firmware update cycle in a single coordinated workflow. Upgrades roll across hosts non-disruptively: VMs migrate as each host is updated in turn.

Comparison to vSphere upgrade coordination: Equivalent vCenter-driven upgrades require coordinated planning across vCenter Update Manager or vSphere Lifecycle Manager and firmware tooling, which is often vendor-specific. LCM consolidates the coordination overhead into one tool with one workflow, which is one of the largest day-to-day operational simplifications versus a comparable vSphere environment.

Who AHV Is For

AHV is the right hypervisor for environments running or moving to Nutanix where the operational shift is acceptable or covered by co-managed support. It is not the right hypervisor when VMware-specific dependencies dominate or when the platform direction is not Nutanix.

Related Resources

FAQs

Frequently Asked Questions

Is AHV a fork of KVM?

No. AHV is an operationally hardened distribution of KVM with Nutanix orchestration on top, not a fork. The underlying KVM, libvirt, QEMU, and Open vSwitch are standard Linux virtualization components maintained in their upstream projects. AHV follows Linux kernel direction rather than diverging from it, which is what allows AHV to remain aligned with the broader Linux virtualization ecosystem.

Does AHV run only on Nutanix-branded hardware?

No. AHV runs on Nutanix NX, Cisco UCS, HPE, Dell, Lenovo, and other supported platforms. The Nutanix hardware compatibility list is the authoritative source for current support. For AIM clients, the most common platform pairing is AHV on Cisco UCS, but the hypervisor itself is platform-portable within the supported HCL.

How does AHV live migration compare to vMotion?

Operationally similar. AHV live migration is QEMU and libvirt migration with Nutanix coordination. Memory state copies between hosts; disk state is preserved by the Distributed Storage Fabric because the disk does not need to migrate. It is already accessible from any host in the cluster. Migration time is bandwidth-bound and scales with VM memory size, comparable to vMotion under similar conditions.

Can we keep using ESXi on Nutanix hardware?

Yes. Nutanix supports ESXi as the hypervisor on Nutanix Cloud Platform. The cost tradeoff is that VMware licensing applies separately when ESXi is the hypervisor, while AHV is included with the Nutanix platform license at no additional hypervisor cost. For VMware-dependent workloads or VMware-skilled teams without migration appetite, ESXi on Nutanix remains a supported configuration.

How long does the operational shift from ESXi to AHV take?

As an illustrative pattern, daily operations feel unfamiliar in the first 30 days, become natural between 30 and 90 days, and reach operational parity at 90 days and beyond. This is not a commitment for any specific environment. Hands-on Nutanix training, Aegis Managed Nutanix coverage during the early period, and early re-authoring of the most-used automation all accelerate the shift.

What happens to our PowerCLI and vSphere automation?

Existing PowerCLI and vSphere API automation does not translate directly to AHV. The AHV equivalents are nuclei (Nutanix CLI), the Acropolis CLI, the v3 REST API, and the Nutanix PowerShell module. Automation requires re-authoring against the AHV API surface. The most-used scripts should be re-authored early, while the team's understanding of what they need to do is fresh: vSphere-era automation atrophies once the team is on AHV, and the window to capture the requirements is short.

Ready to evaluate AHV for your environment?

IVI's data center team works with organizations through AHV evaluation, migration planning, and operational transition - including Aegis Managed Nutanix coverage during the shift from vSphere.

Start a Conversation