Hypervisor-to-Hypervisor Migration Pattern¶
Scope¶
This file covers hypervisor-to-hypervisor migration: moving virtual machine workloads from one hypervisor platform to another (e.g., VMware ESXi to Nutanix AHV, VMware to Hyper-V, VMware to KVM/Proxmox). It addresses source assessment, target selection, migration tooling, driver conversion, network mapping, management plane transition, and guest OS compatibility. For cloud migration (on-prem to IaaS/PaaS), see general/workload-migration.md. For cutover execution, see patterns/migration-cutover.md.
Checklist¶
- [Critical] Complete VM inventory with OS versions, virtual hardware versions, disk formats, and resource consumption
- [Critical] Identify VMs with hardware dependencies that block automated migration (GPU passthrough, PCIe devices, USB passthrough, RDM/physical-mode disks, independent/persistent disks)
- [Critical] Verify guest OS compatibility with target hypervisor — check vendor support matrices for every OS version in the estate
- [Critical] Plan and test driver injection strategy (VirtIO for KVM/AHV, Hyper-V Integration Services for Hyper-V) before production migration
- [Critical] Map source network constructs (vDS port groups, VLANs, security policies, traffic shaping) to target equivalents — document every VLAN and port group mapping
- [Critical] Validate licensing — confirm target hypervisor licensing covers all hosts and that guest OS licenses (especially Windows Server) are valid under the new hypervisor
- [Critical] Define rollback procedure: can source VMs be restarted if migration fails? Are snapshots preserved?
- [Recommended] Capture performance baselines on source hypervisor (CPU, memory, IOPS, network throughput per VM) for post-migration comparison
- [Recommended] Run a pilot wave of 3-5 non-critical VMs through the full migration process before committing to production waves
- [Recommended] Identify and plan for management tool replacement: vCenter alternatives (Prism, SCVMM, Proxmox web UI), monitoring agents, backup agents
- [Recommended] Audit VMware-specific features in use (DRS, HA, vMotion, Storage DRS, fault tolerance) and map to target equivalents or alternatives
- [Recommended] Plan vCenter decommission timeline — do not decommission until all VMs are migrated, validated, and rollback window has closed
- [Recommended] Test backup and restore workflows on target hypervisor before migrating production workloads
- [Optional] Evaluate Broadcom/VMware contract renewal timeline to establish migration deadline and negotiate short-term bridge licensing if needed
- [Optional] Assess VMware NSX dependencies — micro-segmentation rules, distributed firewall policies, load balancers that must be replicated on the target platform
- [Optional] Plan physical host re-purposing or replacement if target hypervisor has different hardware requirements or certification lists
Why This Matters¶
Hypervisor migration is fundamentally different from cloud migration. Both source and target are on-premises infrastructure under your control, but the VM format, driver model, network abstraction, and management plane all change simultaneously. A VM that runs perfectly on ESXi may fail to boot on KVM because it lacks VirtIO drivers, or it may boot but lose network connectivity because the virtual NIC type changed.
The Broadcom acquisition of VMware (completed November 2023) has made hypervisor migration an urgent reality for many organizations. Broadcom eliminated perpetual licensing in favor of subscription-only bundles (VMware Cloud Foundation or VMware vSphere Foundation), shifted to per-core pricing with a 72-core minimum per CPU (effective April 2025), and terminated the VMware Cloud Service Provider (VCSP) program in favor of an invite-only model (January 2026). Many organizations have seen 3-5x cost increases, with some smaller deployments experiencing even higher multipliers. This has driven a wave of migrations to Nutanix AHV, Microsoft Hyper-V, Proxmox VE, and KVM-based platforms.
The most common hypervisor migration failures are: VMs that fail to boot because drivers were not injected before conversion, network connectivity loss because VLAN mappings were incomplete, performance degradation because storage I/O paths changed without tuning, and management gaps because teams decommissioned vCenter before establishing equivalent monitoring and automation on the target platform. All of these are preventable with proper assessment and pilot testing.
Common Decisions (ADR Triggers)¶
ADR: Target Hypervisor Selection¶
Context: Organization needs to migrate off VMware ESXi. Must select a replacement hypervisor platform. Options: - Nutanix AHV — Included with Nutanix HCI licensing (no separate hypervisor cost). Tightly integrated with Nutanix Prism management. Supports live migration, HA, and ADS (Acropolis Dynamic Scheduling, equivalent to DRS). Migration tooling (Nutanix Move) purpose-built for VMware-to-AHV. Requires Nutanix hardware or certified OEM hardware. - Microsoft Hyper-V / Azure Stack HCI — Included with Windows Server Datacenter licensing. Familiar to Windows-centric shops. SCVMM for management. Strong integration with Azure Arc and Azure hybrid services. Weaker Linux guest support historically, though improved significantly. - Proxmox VE — Open-source (AGPL), no licensing cost. KVM-based with LXC container support. Web UI included. Community and enterprise support subscriptions available. Strong fit for SMBs and cost-sensitive environments. Lacks enterprise features like distributed switching and centralized multi-cluster management at scale. - Red Hat OpenShift Virtualization (KubeVirt) — Runs VMs as Kubernetes pods alongside containers. Strategic for organizations already committed to OpenShift. Higher operational complexity. Not a drop-in hypervisor replacement. - Oracle Linux Virtualization Manager (oVirt) — Open-source KVM management. Free but smaller ecosystem. Suitable for Oracle-heavy environments.
Decision criteria: If already running Nutanix HCI or willing to adopt it, AHV is the lowest-friction path with purpose-built migration tooling. If Windows-centric with existing Microsoft EA, Hyper-V/Azure Stack HCI leverages existing licenses. If cost is the primary driver and the team is Linux-capable, Proxmox VE eliminates hypervisor licensing entirely. If the strategic direction is containers and Kubernetes, OpenShift Virtualization consolidates VMs and containers on one platform but is not a like-for-like hypervisor swap.
ADR: Migration Tooling Strategy¶
Context: Need to convert and migrate VMs from source hypervisor to target hypervisor format. Options: - Vendor-native tools — Nutanix Move (VMware to AHV), Microsoft MVMC/Azure Migrate (VMware to Hyper-V), SCVMM (import VMware VMs). Purpose-built, supported, handle driver injection automatically. Limited to vendor-specific target. - virt-v2v (Red Hat/open-source) — Converts VMs from VMware, Xen, or Hyper-V to KVM. Handles driver injection (VirtIO) for both Linux and Windows guests. Supports output to libvirt, oVirt, OpenStack, and local disk. Industry standard for KVM targets. - Veeam / NAKIVO / Zerto — Backup-based migration: back up from source hypervisor, restore to target. Works across hypervisors. Adds a tested backup/restore path as a side benefit. Slower than streaming replication tools. - Manual conversion — Export OVA/VMDK from source, convert disk format (qemu-img convert), import to target, manually install drivers. Full control but labor-intensive and error-prone at scale.
Decision criteria: Always prefer vendor-native tools when migrating to their platform — they handle driver injection, hardware abstraction, and network mapping automatically. Use virt-v2v for KVM/Proxmox targets. Consider backup-based migration for small estates (< 50 VMs) or when you want the migration to also validate your backup/restore pipeline. Manual conversion is a last resort for VMs that automated tools cannot handle (e.g., unusual OS, non-standard disk configurations).
ADR: Network Architecture Transition¶
Context: VMware vDS (Distributed Switch) provides centralized network management, port mirroring, LACP, traffic shaping, and network I/O control. Target hypervisor uses a different networking model. Options: - Pre-migration to vSS — Migrate VMware VMs from vDS to vSS (Standard Switch) before hypervisor migration. Simplifies the hypervisor migration by removing vDS dependency. Extra step but reduces variables during conversion. - Direct vDS-to-target mapping — Map vDS port groups directly to target network constructs (AHV managed networks, Hyper-V virtual switches, OVS bridges). Requires complete port group inventory and VLAN mapping documentation. - Network redesign — Use the hypervisor migration as an opportunity to simplify or redesign the network architecture. Higher risk, higher reward. Only appropriate if the current network architecture has known deficiencies.
Decision criteria: Direct mapping is safest for most migrations — preserve existing VLAN assignments and security zones. Pre-migration to vSS is warranted if the vDS configuration is complex (LACP, NetFlow, port mirroring) and those features are not needed on the target. Network redesign should be a separate project, not combined with hypervisor migration.
ADR: Management Plane Transition Strategy¶
Context: vCenter provides VM lifecycle management, monitoring, alerting, RBAC, automation (vRealize/Aria), and third-party integrations. Decommissioning vCenter requires replacing all of these functions. Options: - Parallel management — Run both vCenter and target management (Prism, SCVMM, Proxmox UI) during migration. Decommission vCenter only after all VMs are migrated and validated. Safest but requires maintaining vCenter licensing during transition. - Immediate vCenter decommission — Decommission vCenter as soon as VMs are migrated off ESXi hosts. Fastest license cost reduction. Risk: lose visibility into any remaining VMware infrastructure. - Phased feature migration — Migrate management functions in phases: VM lifecycle first, then monitoring, then automation, then RBAC. Ensures no management gaps.
Decision criteria: Always use parallel management during migration. Decommission vCenter only after the last VMware VM is migrated, validated, and the rollback window has closed. Budget for short-term vCenter licensing overlap — the cost of maintaining vCenter for 3-6 months during migration is far less than the cost of a management gap during a production incident.
Source Assessment¶
VM Inventory and Classification¶
Catalog every VM in the VMware environment with the following attributes:
| Attribute | Why It Matters |
|---|---|
| Guest OS and version | Target hypervisor may not support EOL operating systems (Windows Server 2008, RHEL 5, etc.) |
| Virtual hardware version | Newer hardware versions may use features not available on the target hypervisor |
| Disk type and format | RDM (Raw Device Mapping), independent disks, eager/lazy thick provisioning affect conversion |
| Disk size | Some migration tools have maximum disk size limits per migration task |
| CPU/memory allocation | Right-sizing opportunity during migration |
| Network adapters and port groups | Must map to target network constructs |
| VMware Tools version | Must be replaced with target guest agents (Nutanix Guest Tools, Hyper-V Integration Services) |
| Snapshots | Must be consolidated before migration — do not migrate VMs with active snapshot chains |
| GPU/PCIe passthrough | Cannot be automatically migrated — requires manual reconfiguration on target |
| Affinity/anti-affinity rules | Must be recreated on target management plane |
| Custom attributes and tags | Used for automation and organization — must be replicated or mapped |
Workloads That Require Special Handling¶
- VMs with RDM (Raw Device Mapping) — Convert to VMDK before migration, or re-attach raw storage on target
- VMs with independent/persistent disks — Nutanix Move and most automated tools do not support these; migrate data separately
- VMs with PCIe passthrough (GPU, NIC SR-IOV) — Requires manual passthrough reconfiguration on target hypervisor
- VMs with active snapshot trees — Consolidate all snapshots before migration
- VMs running unsupported guest OS — May need to retain on VMware or perform P2V-style conversion
- VMs with VMware-specific drivers or agents — NSX agents, vRealize agents, third-party agents tied to VMware APIs
Target Selection Considerations¶
Feature Parity Matrix¶
| VMware Feature | Nutanix AHV | Hyper-V / Azure Stack HCI | Proxmox VE (KVM) |
|---|---|---|---|
| Live migration (vMotion) | AHV live migration | Live Migration | QEMU live migration |
| HA (auto-restart on failure) | AHV HA | Failover Clustering | HA (Corosync/QEMU) |
| DRS (dynamic load balancing) | ADS (Acropolis Dynamic Scheduling) | SCVMM dynamic optimization | Manual / third-party |
| Distributed switching (vDS) | AHV managed networking (OVS-based) | Logical switch (SCVMM) | OVS bridges / Linux bridges |
| Storage vMotion | AHV storage live migration | Storage Migration | Storage live migration |
| Fault tolerance | Not available (use HA + application-level clustering) | Not available (use HA) | Not available |
| NSX micro-segmentation | Flow Network Security (Nutanix Flow) | Azure Network Security Groups / Windows Firewall | OVN / firewall rules |
| vRealize / Aria automation | Nutanix Calm / Prism Pro | Azure Arc / SCVMM / WAC | Ansible / Terraform |
| Content library (templates) | Image service | SCVMM library | Template system |
Hardware Compatibility¶
- Nutanix AHV — Runs on Nutanix-branded hardware or certified OEM nodes (Dell, Lenovo, HPE, Fujitsu). Cannot run on arbitrary servers.
- Hyper-V — Runs on any x86-64 server meeting Windows Server hardware requirements. Broad compatibility.
- Proxmox VE — Runs on any x86-64 server with VT-x/VT-d support. Widest hardware compatibility. No vendor lock-in.
- General — Verify CPU, NIC, and storage controller compatibility with the target hypervisor's HCL (Hardware Compatibility List) before purchasing or repurposing hardware.
Migration Tooling¶
Nutanix Move¶
Nutanix Move is a purpose-built cross-hypervisor migration tool for migrating VMs to AHV.
Capabilities: - Agentless migration from VMware ESXi (connects to vCenter) - Seeding (initial full copy) followed by incremental replication of changed blocks - Automatic driver injection (VirtIO drivers for AHV) - Network mapping (source port groups to AHV managed networks) - Automated VMware Tools uninstall and Nutanix Guest Tools (NGT) install - Supports both Linux and Windows guests - Bulk migration with configurable bandwidth throttling - Cutover scheduling with minimal downtime (final delta sync + VM power-on)
Limitations: - Does not support VMs with independent/persistent disks — must detach before migration - Does not support VMs with PCIe passthrough devices — must remove before migration - Does not support VMs with active snapshot chains — consolidate first - Does not support physical-mode RDMs — convert to virtual-mode or VMDK first - Does not migrate VMware-level constructs (resource pools, DRS rules, custom attributes) - Does not handle program management (wave scheduling, stakeholder communication, executive reporting) - UEFI-to-UEFI migration support depends on Move version and guest OS — verify against release notes
virt-v2v (KVM/Proxmox Target)¶
Red Hat's open-source VM conversion tool, the industry standard for converting VMs to KVM.
Capabilities: - Converts VMs from VMware (via vCenter or direct ESXi), Hyper-V, and Xen to KVM - Injects VirtIO drivers (block, network, balloon) for both Linux and Windows guests - Windows VirtIO driver injection requires the virtio-win package installed on the conversion host - Outputs to libvirt, oVirt, OpenStack, or local QCOW2/raw disk files - Handles boot loader reconfiguration (GRUB updates for Linux guests) - Companion tool virt-p2v for physical-to-virtual conversion
Limitations: - Offline conversion only — VM must be shut down during conversion - Windows driver injection requires virtio-win ISO/drivers to be available on the conversion host - Complex VMware configurations (vApps, multi-NIC with specific port group mappings) require manual post-conversion configuration - No incremental replication — each conversion is a full copy
Microsoft MVMC / SCVMM (Hyper-V Target)¶
Capabilities: - Converts VMware VMs to Hyper-V (VHD/VHDX format) - Installs Hyper-V Integration Services during conversion - SCVMM can import VMware VMs directly by connecting to vCenter
Limitations: - MVMC has been deprecated — SCVMM or Azure Migrate is the current supported path - Some Linux guest OS versions may require manual installation of Hyper-V Integration Services
Driver Injection and Conversion¶
The Driver Problem¶
Every hypervisor presents different virtual hardware to guest operating systems. VMware guests use PVSCSI (storage) and VMXNET3 (network) paravirtual drivers. KVM/AHV guests use VirtIO drivers. Hyper-V guests use Hyper-V Integration Services (VSC drivers). If the target drivers are not present when the VM boots on the new hypervisor, the VM either fails to boot (cannot find boot disk) or boots without network connectivity.
Driver Injection Strategies¶
| Strategy | When to Use | Risk |
|---|---|---|
| Automated tool injection | Migration tool handles it (Nutanix Move, virt-v2v) — always prefer this | Low — tool is tested against supported OS matrix |
| Pre-install drivers on source | Install target drivers (VirtIO, Integration Services) on the VM while it still runs on VMware | Low — drivers coexist, activate when hardware is detected |
| Offline injection | Mount VM disk on conversion host, inject drivers into offline OS | Medium — requires knowledge of OS driver store and boot configuration |
| Recovery mode injection | Boot VM on target hypervisor with IDE/emulated devices, install drivers, switch to paravirtual | Medium — requires guest OS recovery/safe mode boot |
Windows Guest Driver Injection¶
- Preferred: Use migration tool's automated injection (Nutanix Move, virt-v2v with virtio-win)
- Fallback — pre-install on source: Install VirtIO drivers (Red Hat signed) or NGT on the Windows VM while it runs on VMware. Drivers register but remain inactive until matching hardware is detected.
- Emergency — offline injection: Mount the VMDK/QCOW2 on a Linux host, use
virt-customizeor manual registry edits to inject drivers into the Windows driver store - Last resort: Boot VM on target with IDE emulation (slow but functional), install paravirtual drivers, shut down, reconfigure to paravirtual, restart
Linux Guest Driver Injection¶
Linux kernels 2.6.25+ include VirtIO drivers in-tree. Most modern Linux distributions (RHEL 6+, Ubuntu 12.04+, SLES 11+, Debian 7+) have VirtIO modules compiled into the kernel or initramfs. Driver injection is rarely needed for Linux guests — the kernel detects the new virtual hardware and loads appropriate modules automatically. Verify that the initramfs/initrd includes VirtIO modules (lsinitrd on RHEL, lsinitramfs on Debian/Ubuntu).
Network Mapping¶
VMware vDS to Target Hypervisor Networking¶
VMware Distributed Switch (vDS) provides centralized network configuration across an ESXi cluster. When migrating to a different hypervisor, every vDS port group must be mapped to an equivalent construct on the target.
Network Mapping Checklist¶
- Export vDS configuration — Document every port group, VLAN ID, security policy (promiscuous mode, MAC changes, forged transmits), traffic shaping policy, and LACP/LAG configuration
- Map VLANs — VLANs are hypervisor-independent; the same VLAN IDs work on any hypervisor. Map each vDS port group to a target virtual switch port with the same VLAN tag
- Recreate security policies — Promiscuous mode, MAC spoofing settings must be configured on target virtual switches
- LACP/LAG — If using LACP on vDS, configure equivalent bonding on target (OVS bond on AHV/Proxmox, NIC teaming on Hyper-V)
- Traffic shaping — If using vDS traffic shaping or Network I/O Control, configure QoS on target (OVS QoS, Hyper-V bandwidth management)
- Port mirroring — If using vDS port mirroring (SPAN), configure equivalent on target (OVS mirror ports, Hyper-V port mirroring)
Target Networking Equivalents¶
| VMware vDS Feature | Nutanix AHV (OVS) | Hyper-V | Proxmox VE (OVS/Linux Bridge) |
|---|---|---|---|
| Port group with VLAN | AHV managed network with VLAN | Virtual switch with VLAN ID | OVS port with VLAN tag |
| LACP / LAG | OVS bond (LACP mode) | NIC teaming (LACP) | OVS bond (LACP mode) |
| Promiscuous mode | OVS port mirror or promisc flag | Hyper-V port mirroring | OVS or bridge promisc |
| Traffic shaping | OVS QoS (ingress policing, egress shaping) | Hyper-V bandwidth management | OVS QoS / tc |
| Private VLAN | Nutanix Flow micro-segmentation | Hyper-V isolation mode | OVS ACL rules |
| NetFlow/IPFIX | OVS sFlow/IPFIX | Not native (third-party) | OVS sFlow/IPFIX |
VM Hardware Compatibility¶
Virtual Hardware Differences¶
| Component | VMware ESXi | KVM/AHV | Hyper-V |
|---|---|---|---|
| Storage controller (paravirtual) | PVSCSI | VirtIO-blk / VirtIO-SCSI | StorVSC |
| Network adapter (paravirtual) | VMXNET3 | VirtIO-net | NetVSC |
| Display adapter | SVGA | QXL / VirtIO-GPU | Synthetic video |
| Memory ballooning | VMware balloon driver | VirtIO-balloon | Dynamic Memory |
| Guest agent | VMware Tools | QEMU Guest Agent / NGT | Hyper-V Integration Services |
| Disk format | VMDK | QCOW2 (KVM) / raw (AHV) | VHD / VHDX |
| UEFI firmware | VMware EFI | OVMF (TianoCore) | Hyper-V Gen 2 |
| TPM | vTPM | swtpm | vTPM (Gen 2) |
| Secure Boot | Supported (HW v13+) | Supported (OVMF + signed shim) | Supported (Gen 2) |
Conversion Considerations¶
- BIOS vs UEFI — Ensure the target hypervisor supports the same firmware type. BIOS VMs on VMware should remain BIOS on the target. Converting BIOS to UEFI (or vice versa) during migration adds unnecessary risk.
- SCSI controller type — Migration tools typically handle this, but verify that the boot disk is attached to a paravirtual controller on the target (VirtIO-SCSI for KVM/AHV, not IDE emulation which is significantly slower).
- vCPU topology — VMware sockets/cores-per-socket settings should be mapped to equivalent topology on the target, especially for applications with socket-based licensing (SQL Server, Oracle).
- Memory reservation — VMware memory reservations do not exist on most target hypervisors. Evaluate whether the application requires guaranteed memory or can use overcommitted memory with ballooning.
Guest OS Compatibility¶
Support Matrix Considerations¶
Each hypervisor vendor publishes a supported guest OS list. VMs running unsupported guest operating systems may work but will not receive vendor support if issues arise.
| Guest OS | Nutanix AHV | Hyper-V (WS 2022) | Proxmox VE / KVM |
|---|---|---|---|
| Windows Server 2022 | Supported | Supported | Supported |
| Windows Server 2019 | Supported | Supported | Supported |
| Windows Server 2016 | Supported | Supported | Supported |
| Windows Server 2012 R2 | Supported (check EOL) | Supported | Supported |
| Windows 10/11 | Supported | Supported | Supported |
| RHEL 8.x / 9.x | Supported | Supported | Supported |
| Ubuntu 20.04 / 22.04 / 24.04 | Supported | Supported | Supported |
| SLES 12 / 15 | Supported | Supported | Supported |
| Debian 11 / 12 | Community (check AHV matrix) | Supported | Supported |
| CentOS 7 (EOL June 2024) | Supported (may drop) | Supported | Supported |
| Windows Server 2008 R2 (EOL) | Not supported | Legacy support | Works (unsupported) |
| RHEL 5/6 (EOL) | Not supported | Not supported | Works (unsupported) |
EOL Operating Systems¶
VMs running end-of-life operating systems are the highest-risk workloads in a hypervisor migration:
- Option 1: Upgrade OS before migration — Safest. Upgrade to a supported OS version, then migrate. Two changes, but each is well-understood.
- Option 2: Migrate as-is, accept risk — The VM may work on the target hypervisor even if not officially supported. No vendor support if issues arise.
- Option 3: Retain on VMware — Keep the EOL VM on a minimal VMware cluster until the application is decommissioned. Requires maintaining VMware licensing for that cluster.
- Option 4: P2V to target — If the EOL OS cannot be migrated by automated tools, use physical-to-virtual techniques (disk-level clone and conversion).
Management Plane Transition¶
vCenter Replacement Checklist¶
vCenter is not just a VM manager — it is the integration point for monitoring, backup, automation, compliance, and operations tooling. Decommissioning vCenter requires replacing every integration.
| vCenter Function | Replacement Required |
|---|---|
| VM lifecycle (create, clone, delete, snapshot) | Target management plane (Prism, SCVMM, Proxmox UI) |
| Performance monitoring | Target monitoring + external tools (Prometheus, Grafana, Datadog) |
| Alerting | Target alerting + external tools |
| RBAC and permissions | Target RBAC (Prism roles, AD integration on all platforms) |
| Automation (vRealize/Aria) | Nutanix Calm, Ansible, Terraform, Rundeck |
| Backup integration (Veeam, Commvault) | Reconfigure backup agents/proxies for target hypervisor |
| Compliance scanning (vRealize SaltStack, Tanzu) | Target-compatible compliance tools |
| API consumers (custom scripts, CMDB sync) | Rewrite against target API (Nutanix Prism v3 API, Hyper-V WMI/PowerShell, Proxmox API) |
| Log forwarding (syslog from ESXi/vCenter) | Configure syslog on target hypervisor hosts |
vCenter Decommission Timeline¶
Migration start: vCenter remains fully operational
Target management plane deployed in parallel
All new VMs created on target platform
Migration waves: VMs migrate wave by wave
vCenter visibility decreases as VMs leave
Maintain vCenter for rollback capability
Last VM migrated: Begin rollback timer (typically 2-4 weeks)
vCenter still running for emergency rollback
Rollback window Decommission vCenter, ESXi hosts
closed: Reclaim VMware licenses
Archive vCenter database for historical reference
Migration Wave Planning¶
Wave Sequencing for Hypervisor Migration¶
- Wave 0 — Infrastructure preparation: Deploy target hypervisor, configure networking (VLANs, bonds, virtual switches), configure storage, deploy management plane, validate backup integration
- Wave 1 — Pilot (3-5 VMs): Non-critical VMs, mix of Linux and Windows, test full migration lifecycle including rollback
- Wave 2 — Development/test environments: Lower-risk workloads, larger batch, validate automation and wave process
- Wave 3-N — Production waves: Group by application dependency clusters. Migrate tightly coupled VMs together. Schedule during maintenance windows.
- Final wave — Stragglers and special cases: VMs requiring manual conversion, EOL OS workloads, GPU passthrough VMs
VMs Per Wave Sizing¶
- Pilot: 3-5 VMs (manual, high-touch)
- Early waves: 10-20 VMs (building confidence)
- Steady state: 20-50 VMs per wave (depends on network bandwidth for replication and team capacity for validation)
- Constraint: Total migration bandwidth — each VM's disk must be fully replicated. A 500 GB VM over a 1 Gbps link takes ~70 minutes for initial seed. Plan wave sizes against available bandwidth and cutover window duration.
Reference Links¶
See Also¶
general/workload-migration.md— Cloud migration strategy and the 6 Rs frameworkpatterns/migration-cutover.md— Cutover runbook pattern, rollback procedures, communication templatesgeneral/disaster-recovery.md— DR planning for migrated workloadsgeneral/networking.md— Network architecture patterns