KubeVirt

From David's Wiki
Revision as of 20:41, 13 February 2022 by David (talk | contribs) (→‎vGPU)
\( \newcommand{\P}[]{\unicode{xB6}} \newcommand{\AA}[]{\unicode{x212B}} \newcommand{\empty}[]{\emptyset} \newcommand{\O}[]{\emptyset} \newcommand{\Alpha}[]{Α} \newcommand{\Beta}[]{Β} \newcommand{\Epsilon}[]{Ε} \newcommand{\Iota}[]{Ι} \newcommand{\Kappa}[]{Κ} \newcommand{\Rho}[]{Ρ} \newcommand{\Tau}[]{Τ} \newcommand{\Zeta}[]{Ζ} \newcommand{\Mu}[]{\unicode{x039C}} \newcommand{\Chi}[]{Χ} \newcommand{\Eta}[]{\unicode{x0397}} \newcommand{\Nu}[]{\unicode{x039D}} \newcommand{\Omicron}[]{\unicode{x039F}} \DeclareMathOperator{\sgn}{sgn} \def\oiint{\mathop{\vcenter{\mathchoice{\huge\unicode{x222F}\,}{\unicode{x222F}}{\unicode{x222F}}{\unicode{x222F}}}\,}\nolimits} \def\oiiint{\mathop{\vcenter{\mathchoice{\huge\unicode{x2230}\,}{\unicode{x2230}}{\unicode{x2230}}{\unicode{x2230}}}\,}\nolimits} \)

KubeVirt lets you setup and manage virtual machines on your kubernetes cluster.

Getting Started

Background

KubeVirt creates two new types of resources on your cluster: VirtualMachine (vm) and VirtualMachineInstance (vmi). VirtualMachine defines how to create VMIs. VirtualMachineInstance represent a running virtual machine.

Similar to deployments-pods, you will typically not create VirtualMachineInstance manually. Instead you define VirtualMachine in your manifests and control them using virtctl. Then KubeVirt will automatically create VirtualMachineInstance.

Requirements

See requirements

  • You need a kubernetes cluster with kubectl set up.
  • You do not need to install qemu-kvm libvirt-daemon-system on the nodes.

Install KubeVirt

See installation

Install commands
# Get the latest version string
export VERSION=$(curl -s https://api.github.com/repos/kubevirt/kubevirt/releases | grep tag_name | grep -v -- '-rc' | sort -r | head -1 | awk -F': ' '{print $2}' | sed 's/,//' | xargs)
echo $VERSION

# Deploy operator
kubectl create -f https://github.com/kubevirt/kubevirt/releases/download/${VERSION}/kubevirt-operator.yaml
# Deploy custom resources
kubectl create -f https://github.com/kubevirt/kubevirt/releases/download/${VERSION}/kubevirt-cr.yaml

# Install virtctl
VERSION=$(kubectl get kubevirt.kubevirt.io/kubevirt -n kubevirt -o=jsonpath="{.status.observedKubeVirtVersion}")
ARCH=$(uname -s | tr A-Z a-z)-$(uname -m | sed 's/x86_64/amd64/') || windows-amd64.exe
echo ${ARCH}
curl -L -o virtctl https://github.com/kubevirt/kubevirt/releases/download/${VERSION}/virtctl-${VERSION}-${ARCH}
chmod +x virtctl
sudo install virtctl /usr/local/bin

Creating a VM

Loading ISOs into the cluster

Windows

  • Deploy the example manifest below.
  • During the install, you will need to load the `viostor` driver.
  • After install, load the remaining drivers using device manager.
  • Then you can remove the cdrom and windows-guest-tools disks.
Example manifest
apiVersion: v1
kind: PersistentVolumeClaim
metadata:
  name: windows10vm-pvc
spec:
  storageClassName: ""
  volumeName: windows10vm-pv
  accessModes:
    - ReadWriteMany
  resources:
    requests:
      storage: 100Gi
---
apiVersion: kubevirt.io/v1alpha3
kind: VirtualMachine
metadata:
  name: windows10vm
spec:
  runStrategy: Manual
  template:
    metadata:
      labels:
        kubevirt.io/domain: win10vm1
    spec:
      terminationGracePeriodSeconds: 300
      domain:
        cpu:
          cores: 4
        resources:
          requests:
            memory: 8G
        devices:
          disks:
            - bootOrder: 1
              cdrom:
                bus: sata
              name: cdrom
            - name: maindrive
              bootOrder: 1
              disk:
                bus: virtio
            - name: windows-guest-tools
              bootOrder: 2
              cdrom:
                bus: sata
          interfaces:
            - masquerade: {}
              model: e1000e
              name: nic-0
          sound:
            name: audio
            model: ich9
      networks:
        - name: nic-0
          pod: {}
      volumes:
        - name: cdrom
          persistentVolumeClaim:
            claimName: iso-win10
        - name: maindrive
          persistentVolumeClaim:
            claimName: windows10vm-pvc
        - containerDisk:
            image: quay.io/kubevirt/virtio-container-disk
          name: windows-guest-tools

vGPU

Intel vGPU

See intel-vgpu-kubevirt

Instructions
Setup the nodes

Run the following on each node with an Intel GPU.

# Enable kvmgt and iommu
sudo sh -c "echo kvmgt > /etc/modules-load.d/gpu-kvmgt.conf"
sudo sed -i 's/^GRUB_CMDLINE_LINUX_DEFAULT="/&intel_iommu=on i915.enable_gvt=1/'  /etc/default/grub
sudo update-grub
sudo reboot

# Check that kvmgt modules are loaded
sudo lsmod | grep kvmgt

# Create two vGPUs
pci_id=$(sudo lspci | grep -oP '([\d:\.]+)(?=\sVGA)')
uuid1=$(uuidgen)
uuid2=$(uuidgen)

cat > ~/gvtg-enable.service << EOF
[Unit]
Description=Create Intel GVT-g vGPU

[Service]
Type=oneshot
ExecStart=/bin/sh -c "echo '${uuid1}' > /sys/devices/pci0000:00/0000:${pci_id}/mdev_supported_types/i915-GVTg_V5_8/create"
ExecStart=/bin/sh -c "echo '${uuid2}' > /sys/devices/pci0000:00/0000:${pci_id}/mdev_supported_types/i915-GVTg_V5_8/create"
ExecStop=/bin/sh -c "echo '1' > /sys/devices/pci0000:00/0000:${pci_id}/${uuid1}/remove"
ExecStop=/bin/sh -c "echo '1' > /sys/devices/pci0000:00/0000:${pci_id}/${uuid2}/remove"
RemainAfterExit=yes

[Install]
WantedBy=multi-user.target
EOF
sudo mv ~/gvtg-enable.service /etc/systemd/system/gvtg-enable.service
sudo systemctl enable gvtg-enable --now

Resources