Implement new provisioner setup
This commit is contained in:
		
							
								
								
									
										425
									
								
								api-daemon/provisioner/examples/1-noinstall.py
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										425
									
								
								api-daemon/provisioner/examples/1-noinstall.py
									
									
									
									
									
										Normal file
									
								
							@@ -0,0 +1,425 @@
 | 
			
		||||
#!/usr/bin/env python3
 | 
			
		||||
 | 
			
		||||
# 1-noinstall.py - PVC Provisioner example script for noop install
 | 
			
		||||
# Part of the Parallel Virtual Cluster (PVC) system
 | 
			
		||||
#
 | 
			
		||||
#    Copyright (C) 2018-2022 Joshua M. Boniface <joshua@boniface.me>
 | 
			
		||||
#
 | 
			
		||||
#    This program is free software: you can redistribute it and/or modify
 | 
			
		||||
#    it under the terms of the GNU General Public License as published by
 | 
			
		||||
#    the Free Software Foundation, version 3.
 | 
			
		||||
#
 | 
			
		||||
#    This program is distributed in the hope that it will be useful,
 | 
			
		||||
#    but WITHOUT ANY WARRANTY; without even the implied warranty of
 | 
			
		||||
#    MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
 | 
			
		||||
#    GNU General Public License for more details.
 | 
			
		||||
#
 | 
			
		||||
#    You should have received a copy of the GNU General Public License
 | 
			
		||||
#    along with this program.  If not, see <https://www.gnu.org/licenses/>.
 | 
			
		||||
#
 | 
			
		||||
###############################################################################
 | 
			
		||||
 | 
			
		||||
# This script provides an example of a PVC provisioner script. It will create a
 | 
			
		||||
# standard VM config but do no preparation/installation/cleanup (noop).
 | 
			
		||||
 | 
			
		||||
# This script can thus be used as an example or reference implementation of a
 | 
			
		||||
# PVC provisioner script and expanded upon as required.
 | 
			
		||||
 | 
			
		||||
# The script must implement the class "VMBuilderScript" which extens "VMBuilder",
 | 
			
		||||
# providing the 5 functions indicated. Detailed explanation of the role of each
 | 
			
		||||
# function is provided.
 | 
			
		||||
 | 
			
		||||
# Within the VMBuilderScript class, several common variables are exposed:
 | 
			
		||||
#  self.vm_name: The name of the VM from PVC's perspective
 | 
			
		||||
#  self.vm_id: The VM ID (numerical component of the vm_name) from PVC's perspective
 | 
			
		||||
#  self.vm_uuid: An automatically-generated UUID for the VM
 | 
			
		||||
#  self.vm_profile: The PVC provisioner profile name used for the VM
 | 
			
		||||
#  self.vm-data: A dictionary of VM data collected by the provisioner; an example:
 | 
			
		||||
#    {
 | 
			
		||||
#      "ceph_monitor_list": [
 | 
			
		||||
#        "hv1.pvcstorage.tld",
 | 
			
		||||
#        "hv2.pvcstorage.tld",
 | 
			
		||||
#        "hv3.pvcstorage.tld"
 | 
			
		||||
#      ],
 | 
			
		||||
#      "ceph_monitor_port": "6789",
 | 
			
		||||
#      "ceph_monitor_secret": "96721723-8650-4a72-b8f6-a93cd1a20f0c",
 | 
			
		||||
#      "mac_template": null,
 | 
			
		||||
#      "networks": [
 | 
			
		||||
#        {
 | 
			
		||||
#          "eth_bridge": "vmbr1001",
 | 
			
		||||
#          "id": 72,
 | 
			
		||||
#          "network_template": 69,
 | 
			
		||||
#          "vni": "1001"
 | 
			
		||||
#        },
 | 
			
		||||
#        {
 | 
			
		||||
#          "eth_bridge": "vmbr101",
 | 
			
		||||
#          "id": 73,
 | 
			
		||||
#          "network_template": 69,
 | 
			
		||||
#          "vni": "101"
 | 
			
		||||
#        }
 | 
			
		||||
#      ],
 | 
			
		||||
#      "script": [contents of this file]
 | 
			
		||||
#      "script_arguments": {
 | 
			
		||||
#        "deb_mirror": "http://ftp.debian.org/debian",
 | 
			
		||||
#        "deb_release": "bullseye"
 | 
			
		||||
#      },
 | 
			
		||||
#      "system_architecture": "x86_64",
 | 
			
		||||
#      "system_details": {
 | 
			
		||||
#        "id": 78,
 | 
			
		||||
#        "migration_method": "live",
 | 
			
		||||
#        "name": "small",
 | 
			
		||||
#        "node_autostart": false,
 | 
			
		||||
#        "node_limit": null,
 | 
			
		||||
#        "node_selector": null,
 | 
			
		||||
#        "ova": null,
 | 
			
		||||
#        "serial": true,
 | 
			
		||||
#        "vcpu_count": 2,
 | 
			
		||||
#        "vnc": false,
 | 
			
		||||
#        "vnc_bind": null,
 | 
			
		||||
#        "vram_mb": 2048
 | 
			
		||||
#      },
 | 
			
		||||
#      "volumes": [
 | 
			
		||||
#        {
 | 
			
		||||
#          "disk_id": "sda",
 | 
			
		||||
#          "disk_size_gb": 4,
 | 
			
		||||
#          "filesystem": "ext4",
 | 
			
		||||
#          "filesystem_args": "-L=root",
 | 
			
		||||
#          "id": 9,
 | 
			
		||||
#          "mountpoint": "/",
 | 
			
		||||
#          "pool": "vms",
 | 
			
		||||
#          "source_volume": null,
 | 
			
		||||
#          "storage_template": 67
 | 
			
		||||
#        },
 | 
			
		||||
#        {
 | 
			
		||||
#          "disk_id": "sdb",
 | 
			
		||||
#          "disk_size_gb": 4,
 | 
			
		||||
#          "filesystem": "ext4",
 | 
			
		||||
#          "filesystem_args": "-L=var",
 | 
			
		||||
#          "id": 10,
 | 
			
		||||
#          "mountpoint": "/var",
 | 
			
		||||
#          "pool": "vms",
 | 
			
		||||
#          "source_volume": null,
 | 
			
		||||
#          "storage_template": 67
 | 
			
		||||
#        },
 | 
			
		||||
#        {
 | 
			
		||||
#          "disk_id": "sdc",
 | 
			
		||||
#          "disk_size_gb": 4,
 | 
			
		||||
#          "filesystem": "ext4",
 | 
			
		||||
#          "filesystem_args": "-L=log",
 | 
			
		||||
#          "id": 11,
 | 
			
		||||
#          "mountpoint": "/var/log",
 | 
			
		||||
#          "pool": "vms",
 | 
			
		||||
#          "source_volume": null,
 | 
			
		||||
#          "storage_template": 67
 | 
			
		||||
#        }
 | 
			
		||||
#      ]
 | 
			
		||||
#    }
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
class VMBuilderScript(VMBuilder):
 | 
			
		||||
    def setup(self):
 | 
			
		||||
        """
 | 
			
		||||
        setup(): Perform special setup steps or validation before proceeding
 | 
			
		||||
 | 
			
		||||
        Since we do no install in this example, it does nothing.
 | 
			
		||||
        """
 | 
			
		||||
 | 
			
		||||
        pass
 | 
			
		||||
 | 
			
		||||
    def create(self):
 | 
			
		||||
        """
 | 
			
		||||
        create(): Create the VM libvirt schema definition
 | 
			
		||||
 | 
			
		||||
        This step *must* return a fully-formed Libvirt XML document as a string.
 | 
			
		||||
 | 
			
		||||
        This example leverages the built-in libvirt_schema objects provided by PVC; these
 | 
			
		||||
        can be used as-is, or replaced with your own schema(s) on a per-script basis.
 | 
			
		||||
        """
 | 
			
		||||
 | 
			
		||||
        # Run any imports first
 | 
			
		||||
        import datetime
 | 
			
		||||
        import random
 | 
			
		||||
        import pvcapid.libvirt_schema as libvirt_schema
 | 
			
		||||
 | 
			
		||||
        schema = ""
 | 
			
		||||
 | 
			
		||||
        # Prepare a description based on the VM profile
 | 
			
		||||
        description = (
 | 
			
		||||
            f"PVC provisioner @ {datetime.datetime.now()}, profile '{self.vm_profile}'"
 | 
			
		||||
        )
 | 
			
		||||
 | 
			
		||||
        # Format the header
 | 
			
		||||
        schema += libvirt_schema.libvirt_header.format(
 | 
			
		||||
            vm_name=self.vm_name,
 | 
			
		||||
            vm_uuid=self.vm_uuid,
 | 
			
		||||
            vm_description=description,
 | 
			
		||||
            vm_memory=self.vm_data["system_details"]["vram_mb"],
 | 
			
		||||
            vm_vcpus=self.vm_data["system_details"]["vcpu_count"],
 | 
			
		||||
            vm_architecture=self.vm_data["system_architecture"],
 | 
			
		||||
        )
 | 
			
		||||
 | 
			
		||||
        # Add the disk devices
 | 
			
		||||
        monitor_list = self.vm_data["ceph_monitor_list"]
 | 
			
		||||
        monitor_port = self.vm_data["ceph_monitor_port"]
 | 
			
		||||
        monitor_secret = self.vm_data["ceph_monitor_secret"]
 | 
			
		||||
 | 
			
		||||
        for volume in self.vm_data["volumes"]:
 | 
			
		||||
            schema += libvirt_schema.devices_disk_header.format(
 | 
			
		||||
                ceph_storage_secret=monitor_secret,
 | 
			
		||||
                disk_pool=volume["pool"],
 | 
			
		||||
                vm_name=self.vm_name,
 | 
			
		||||
                disk_id=volume["disk_id"],
 | 
			
		||||
            )
 | 
			
		||||
            for monitor in monitor_list:
 | 
			
		||||
                schema += libvirt_schema.devices_disk_coordinator.format(
 | 
			
		||||
                    coordinator_name=monitor,
 | 
			
		||||
                    coordinator_ceph_mon_port=monitor_port,
 | 
			
		||||
                )
 | 
			
		||||
            schema += libvirt_schema.devices_disk_footer
 | 
			
		||||
 | 
			
		||||
        # Add the special vhostmd device for hypervisor information inside the VM
 | 
			
		||||
        schema += libvirt_schema.devices_vhostmd
 | 
			
		||||
 | 
			
		||||
        # Add the network devices
 | 
			
		||||
        network_id = 0
 | 
			
		||||
        for network in self.vm_data["networks"]:
 | 
			
		||||
            vm_id_hex = "{:x}".format(int(vm_id % 16))
 | 
			
		||||
            net_id_hex = "{:x}".format(int(network_id % 16))
 | 
			
		||||
 | 
			
		||||
            if self.vm_data.get("mac_template") is not None:
 | 
			
		||||
                mac_prefix = "52:54:01"
 | 
			
		||||
                macgen_template = self.vm_data["mac_template"]
 | 
			
		||||
                eth_macaddr = macgen_template.format(
 | 
			
		||||
                    prefix=mac_prefix, vmid=vm_id_hex, netid=net_id_hex
 | 
			
		||||
                )
 | 
			
		||||
            else:
 | 
			
		||||
                mac_prefix = "52:54:00"
 | 
			
		||||
                random_octet_A = "{:x}".format(random.randint(16, 238))
 | 
			
		||||
                random_octet_B = "{:x}".format(random.randint(16, 238))
 | 
			
		||||
                random_octet_C = "{:x}".format(random.randint(16, 238))
 | 
			
		||||
 | 
			
		||||
                macgen_template = "{prefix}:{octetA}:{octetB}:{octetC}"
 | 
			
		||||
                eth_macaddr = macgen_template.format(
 | 
			
		||||
                    prefix=mac_prefix,
 | 
			
		||||
                    octetA=random_octet_A,
 | 
			
		||||
                    octetB=random_octet_B,
 | 
			
		||||
                    octetC=random_octet_C,
 | 
			
		||||
                )
 | 
			
		||||
 | 
			
		||||
            schema += libvirt_schema.devices_net_interface.format(
 | 
			
		||||
                eth_macaddr=eth_macaddr,
 | 
			
		||||
                eth_bridge=eth_bridge,
 | 
			
		||||
            )
 | 
			
		||||
 | 
			
		||||
            network_id += 1
 | 
			
		||||
 | 
			
		||||
        # Add default devices
 | 
			
		||||
        schema += libvirt_schema.devices_default
 | 
			
		||||
 | 
			
		||||
        # Add serial device
 | 
			
		||||
        if self.vm_data["system_details"]["serial"]:
 | 
			
		||||
            schema += libvirt_schema.devices_serial.format(vm_name=self.vm_name)
 | 
			
		||||
 | 
			
		||||
        # Add VNC device
 | 
			
		||||
        if self.vm_data["system_details"]["vnc"]:
 | 
			
		||||
            if self.vm_data["system_details"]["vnc_bind"]:
 | 
			
		||||
                vm_vnc_bind = self.vm_data["system_details"]["vnc_bind"]
 | 
			
		||||
            else:
 | 
			
		||||
                vm_vnc_bind = "127.0.0.1"
 | 
			
		||||
 | 
			
		||||
            vm_vncport = 5900
 | 
			
		||||
            vm_vnc_autoport = "yes"
 | 
			
		||||
 | 
			
		||||
            schema += libvirt_schema.devices_vnc.format(
 | 
			
		||||
                vm_vncport=vm_vncport,
 | 
			
		||||
                vm_vnc_autoport=vm_vnc_autoport,
 | 
			
		||||
                vm_vnc_bind=vm_vnc_bind,
 | 
			
		||||
            )
 | 
			
		||||
 | 
			
		||||
        # Add SCSI controller
 | 
			
		||||
        schema += libvirt_schema.devices_scsi_controller
 | 
			
		||||
 | 
			
		||||
        # Add footer
 | 
			
		||||
        schema += libvirt_schema.libvirt_footer
 | 
			
		||||
 | 
			
		||||
        return schema
 | 
			
		||||
 | 
			
		||||
    def prepare(self):
 | 
			
		||||
        """
 | 
			
		||||
        prepare(): Prepare any disks/volumes for the install() step
 | 
			
		||||
 | 
			
		||||
        This function should use the various exposed PVC commands as indicated to create
 | 
			
		||||
        block devices and map them to the host.
 | 
			
		||||
        """
 | 
			
		||||
 | 
			
		||||
        # First loop: Create the disks, either by cloning (pvc_ceph.clone_volume), or by
 | 
			
		||||
        # new creation (pvc_ceph.add_volume).
 | 
			
		||||
        for volume in self.vm_data["volumes"]:
 | 
			
		||||
            if volume.get("source_volume") is not None:
 | 
			
		||||
                with open_zk(config) as zkhandler:
 | 
			
		||||
                    success, message = pvc_ceph.clone_volume(
 | 
			
		||||
                        zkhandler,
 | 
			
		||||
                        volume["pool"],
 | 
			
		||||
                        volume["source_volume"],
 | 
			
		||||
                        f"{self.vm_name}_{volume['disk_id']}",
 | 
			
		||||
                    )
 | 
			
		||||
                    print(message)
 | 
			
		||||
                    if not success:
 | 
			
		||||
                        raise ProvisioningError(
 | 
			
		||||
                            f"Failed to clone volume '{volume['source_volume']}' to '{volume['disk_id']}'."
 | 
			
		||||
                        )
 | 
			
		||||
            else:
 | 
			
		||||
                with open_zk(config) as zkhandler:
 | 
			
		||||
                    success, message = pvc_ceph.add_volume(
 | 
			
		||||
                        zkhandler,
 | 
			
		||||
                        volume["pool"],
 | 
			
		||||
                        f"{self.vm_name}_{volume['disk_id']}",
 | 
			
		||||
                        f"{volume['disk_size_gb']}G",
 | 
			
		||||
                    )
 | 
			
		||||
                    print(message)
 | 
			
		||||
                    if not success:
 | 
			
		||||
                        raise ProvisioningError(
 | 
			
		||||
                            f"Failed to create volume '{volume['disk_id']}'."
 | 
			
		||||
                        )
 | 
			
		||||
 | 
			
		||||
        # Second loop: Map the disks to the local system
 | 
			
		||||
        for volume in self.vm_data["volumes"]:
 | 
			
		||||
            dst_volume_name = f"{self.vm_name}_{volume['disk_id']}"
 | 
			
		||||
            dst_volume = f"{volume['pool']}/{dst_volume_name}"
 | 
			
		||||
 | 
			
		||||
            with open_zk(config) as zkhandler:
 | 
			
		||||
                success, message = pvc_ceph.map_volume(
 | 
			
		||||
                    zkhandler,
 | 
			
		||||
                    volume["pool"],
 | 
			
		||||
                    dst_volume_name,
 | 
			
		||||
                )
 | 
			
		||||
                print(message)
 | 
			
		||||
                if not retcode:
 | 
			
		||||
                    raise ProvisioningError(f"Failed to map volume '{dst_volume}'.")
 | 
			
		||||
 | 
			
		||||
        # Third loop: Create filesystems on the volumes
 | 
			
		||||
        for volume in self.vm_data["volumes"]:
 | 
			
		||||
            dst_volume_name = f"{self.vm_name}_{volume['disk_id']}"
 | 
			
		||||
            dst_volume = f"{volume['pool']}/{dst_volume_name}"
 | 
			
		||||
 | 
			
		||||
            if volume.get("source_volume") is not None:
 | 
			
		||||
                continue
 | 
			
		||||
 | 
			
		||||
            if volume.get("filesystem") is None:
 | 
			
		||||
                continue
 | 
			
		||||
 | 
			
		||||
            filesystem_args_list = list()
 | 
			
		||||
            for arg in volume["filesystem_args"].split():
 | 
			
		||||
                arg_entry, *arg_data = arg.split("=")
 | 
			
		||||
                arg_data = "=".join(arg_data)
 | 
			
		||||
                filesystem_args_list.append(arg_entry)
 | 
			
		||||
                filesystem_args_list.append(arg_data)
 | 
			
		||||
            filesystem_args = " ".join(filesystem_args_list)
 | 
			
		||||
 | 
			
		||||
            if volume["filesystem"] == "swap":
 | 
			
		||||
                retcode, stdout, stderr = pvc_common.run_os_command(
 | 
			
		||||
                    f"mkswap -f /dev/rbd/{dst_volume}"
 | 
			
		||||
                )
 | 
			
		||||
                if retcode:
 | 
			
		||||
                    raise ProvisioningError(
 | 
			
		||||
                        f"Failed to create swap on '{dst_volume}': {stderr}"
 | 
			
		||||
                    )
 | 
			
		||||
            else:
 | 
			
		||||
                retcode, stdout, stderr = pvc_common.run_os_command(
 | 
			
		||||
                    f"mkfs.{volume['filesystem']} {filesystem_args} /dev/rbd/{dst_volume}"
 | 
			
		||||
                )
 | 
			
		||||
                if retcode:
 | 
			
		||||
                    raise ProvisioningError(
 | 
			
		||||
                        f"Faield to create {volume['filesystem']} file on '{dst_volume}': {stderr}"
 | 
			
		||||
                    )
 | 
			
		||||
 | 
			
		||||
            print(stdout)
 | 
			
		||||
 | 
			
		||||
        # Create a temporary directory to use during install
 | 
			
		||||
        temp_dir = "/tmp/target"
 | 
			
		||||
        if not os.exists(temp_dir):
 | 
			
		||||
            os.mkdir(temp_dir)
 | 
			
		||||
 | 
			
		||||
        # Fourth loop: Mount the volumes to a set of temporary directories
 | 
			
		||||
        for volume in self.vm_data["volumes"]:
 | 
			
		||||
            dst_volume_name = f"{self.vm_name}_{volume['disk_id']}"
 | 
			
		||||
            dst_volume = f"{volume['pool']}/{dst_volume_name}"
 | 
			
		||||
 | 
			
		||||
            if volume.get("source_volume") is not None:
 | 
			
		||||
                continue
 | 
			
		||||
 | 
			
		||||
            if volume.get("filesystem") is None:
 | 
			
		||||
                continue
 | 
			
		||||
 | 
			
		||||
            mapped_dst_volume = f"/dev/rbd/{dst_volume}"
 | 
			
		||||
 | 
			
		||||
            mount_path = f"{temp_dir}/{volume['mountpoint']}"
 | 
			
		||||
 | 
			
		||||
            if not os.exists(mount_path):
 | 
			
		||||
                os.mkdir(mount_path)
 | 
			
		||||
 | 
			
		||||
            # Mount filesystem
 | 
			
		||||
            retcode, stdout, stderr = pvc_common.run_os_command(
 | 
			
		||||
                f"mount {mapped_dst_volume} {mount_path}"
 | 
			
		||||
            )
 | 
			
		||||
            if retcode:
 | 
			
		||||
                raise ProvisioningError(
 | 
			
		||||
                    f"Failed to mount '{mapped_dst_volume}' on '{mount_path}': {stderr}"
 | 
			
		||||
                )
 | 
			
		||||
 | 
			
		||||
    def install(self):
 | 
			
		||||
        """
 | 
			
		||||
        install(): Perform the installation
 | 
			
		||||
 | 
			
		||||
        Since this is a noop example, this step does nothing, aside from getting some
 | 
			
		||||
        arguments for demonstration.
 | 
			
		||||
        """
 | 
			
		||||
 | 
			
		||||
        arguments = self.vm_data["script_arguments"]
 | 
			
		||||
        if arguments.get("vm_fqdn"):
 | 
			
		||||
            vm_fqdn = arguments.get("vm_fqdn")
 | 
			
		||||
        else:
 | 
			
		||||
            vm_fqdn = self.vm_name
 | 
			
		||||
 | 
			
		||||
        pass
 | 
			
		||||
 | 
			
		||||
    def cleanup(self):
 | 
			
		||||
        """
 | 
			
		||||
        cleanup(): Perform any cleanup required due to prepare()/install()
 | 
			
		||||
 | 
			
		||||
        It is important to now reverse *all* steps taken in those functions that might
 | 
			
		||||
        need cleanup before teardown of the overlay chroot environment.
 | 
			
		||||
        """
 | 
			
		||||
 | 
			
		||||
        temp_dir = "/tmp/target"
 | 
			
		||||
 | 
			
		||||
        for volume in list(reversed(self.vm_data["volumes"])):
 | 
			
		||||
            dst_volume_name = f"{self.vm_name}_{volume['disk_id']}"
 | 
			
		||||
            dst_volume = f"{volume['pool']}/{dst_volume_name}"
 | 
			
		||||
            mapped_dst_volume = f"/dev/rbd/{dst_volume}"
 | 
			
		||||
            mount_path = f"{temp_dir}/{volume['mountpoint']}"
 | 
			
		||||
 | 
			
		||||
            if (
 | 
			
		||||
                volume.get("source_volume") is None
 | 
			
		||||
                and volume.get("filesystem") is not None
 | 
			
		||||
            ):
 | 
			
		||||
                # Unmount filesystem
 | 
			
		||||
                retcode, stdout, stderr = pvc_common.run_os_command(
 | 
			
		||||
                    f"umount {mount_path}"
 | 
			
		||||
                )
 | 
			
		||||
                if retcode:
 | 
			
		||||
                    raise ProvisioningError(
 | 
			
		||||
                        f"Failed to unmount '{mapped_dst_volume}' on '{mount_path}': {stderr}"
 | 
			
		||||
                    )
 | 
			
		||||
 | 
			
		||||
                # Unmap volume
 | 
			
		||||
                with open_zk(config) as zkhandler:
 | 
			
		||||
                    success, message = pvc_ceph.unmap_volume(
 | 
			
		||||
                        zkhandler,
 | 
			
		||||
                        volume["pool"],
 | 
			
		||||
                        dst_volume_name,
 | 
			
		||||
                    )
 | 
			
		||||
                    if not success:
 | 
			
		||||
                        raise ProvisioningError(
 | 
			
		||||
                            f"Failed to unmap '{mapped_dst_volume}': {stderr}"
 | 
			
		||||
                        )
 | 
			
		||||
							
								
								
									
										602
									
								
								api-daemon/provisioner/examples/2-debootstrap.py
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										602
									
								
								api-daemon/provisioner/examples/2-debootstrap.py
									
									
									
									
									
										Normal file
									
								
							@@ -0,0 +1,602 @@
 | 
			
		||||
#!/usr/bin/env python3
 | 
			
		||||
 | 
			
		||||
# 2-debootstrap.py - PVC Provisioner example script for debootstrap install
 | 
			
		||||
# Part of the Parallel Virtual Cluster (PVC) system
 | 
			
		||||
#
 | 
			
		||||
#    Copyright (C) 2018-2022 Joshua M. Boniface <joshua@boniface.me>
 | 
			
		||||
#
 | 
			
		||||
#    This program is free software: you can redistribute it and/or modify
 | 
			
		||||
#    it under the terms of the GNU General Public License as published by
 | 
			
		||||
#    the Free Software Foundation, version 3.
 | 
			
		||||
#
 | 
			
		||||
#    This program is distributed in the hope that it will be useful,
 | 
			
		||||
#    but WITHOUT ANY WARRANTY; without even the implied warranty of
 | 
			
		||||
#    MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
 | 
			
		||||
#    GNU General Public License for more details.
 | 
			
		||||
#
 | 
			
		||||
#    You should have received a copy of the GNU General Public License
 | 
			
		||||
#    along with this program.  If not, see <https://www.gnu.org/licenses/>.
 | 
			
		||||
#
 | 
			
		||||
###############################################################################
 | 
			
		||||
 | 
			
		||||
# This script provides an example of a PVC provisioner script. It will create a
 | 
			
		||||
# standard VM config and install a Debian-like OS using debootstrap.
 | 
			
		||||
 | 
			
		||||
# This script can thus be used as an example or reference implementation of a
 | 
			
		||||
# PVC provisioner script and expanded upon as required.
 | 
			
		||||
 | 
			
		||||
# The script must implement the class "VMBuilderScript" which extens "VMBuilder",
 | 
			
		||||
# providing the 5 functions indicated. Detailed explanation of the role of each
 | 
			
		||||
# function is provided.
 | 
			
		||||
 | 
			
		||||
# Within the VMBuilderScript class, several common variables are exposed:
 | 
			
		||||
#  self.vm_name: The name of the VM from PVC's perspective
 | 
			
		||||
#  self.vm_id: The VM ID (numerical component of the vm_name) from PVC's perspective
 | 
			
		||||
#  self.vm_uuid: An automatically-generated UUID for the VM
 | 
			
		||||
#  self.vm_profile: The PVC provisioner profile name used for the VM
 | 
			
		||||
#  self.vm-data: A dictionary of VM data collected by the provisioner; an example:
 | 
			
		||||
#    {
 | 
			
		||||
#      "ceph_monitor_list": [
 | 
			
		||||
#        "hv1.pvcstorage.tld",
 | 
			
		||||
#        "hv2.pvcstorage.tld",
 | 
			
		||||
#        "hv3.pvcstorage.tld"
 | 
			
		||||
#      ],
 | 
			
		||||
#      "ceph_monitor_port": "6789",
 | 
			
		||||
#      "ceph_monitor_secret": "96721723-8650-4a72-b8f6-a93cd1a20f0c",
 | 
			
		||||
#      "mac_template": null,
 | 
			
		||||
#      "networks": [
 | 
			
		||||
#        {
 | 
			
		||||
#          "eth_bridge": "vmbr1001",
 | 
			
		||||
#          "id": 72,
 | 
			
		||||
#          "network_template": 69,
 | 
			
		||||
#          "vni": "1001"
 | 
			
		||||
#        },
 | 
			
		||||
#        {
 | 
			
		||||
#          "eth_bridge": "vmbr101",
 | 
			
		||||
#          "id": 73,
 | 
			
		||||
#          "network_template": 69,
 | 
			
		||||
#          "vni": "101"
 | 
			
		||||
#        }
 | 
			
		||||
#      ],
 | 
			
		||||
#      "script": [contents of this file]
 | 
			
		||||
#      "script_arguments": {
 | 
			
		||||
#        "deb_mirror": "http://ftp.debian.org/debian",
 | 
			
		||||
#        "deb_release": "bullseye"
 | 
			
		||||
#      },
 | 
			
		||||
#      "system_architecture": "x86_64",
 | 
			
		||||
#      "system_details": {
 | 
			
		||||
#        "id": 78,
 | 
			
		||||
#        "migration_method": "live",
 | 
			
		||||
#        "name": "small",
 | 
			
		||||
#        "node_autostart": false,
 | 
			
		||||
#        "node_limit": null,
 | 
			
		||||
#        "node_selector": null,
 | 
			
		||||
#        "ova": null,
 | 
			
		||||
#        "serial": true,
 | 
			
		||||
#        "vcpu_count": 2,
 | 
			
		||||
#        "vnc": false,
 | 
			
		||||
#        "vnc_bind": null,
 | 
			
		||||
#        "vram_mb": 2048
 | 
			
		||||
#      },
 | 
			
		||||
#      "volumes": [
 | 
			
		||||
#        {
 | 
			
		||||
#          "disk_id": "sda",
 | 
			
		||||
#          "disk_size_gb": 4,
 | 
			
		||||
#          "filesystem": "ext4",
 | 
			
		||||
#          "filesystem_args": "-L=root",
 | 
			
		||||
#          "id": 9,
 | 
			
		||||
#          "mountpoint": "/",
 | 
			
		||||
#          "pool": "vms",
 | 
			
		||||
#          "source_volume": null,
 | 
			
		||||
#          "storage_template": 67
 | 
			
		||||
#        },
 | 
			
		||||
#        {
 | 
			
		||||
#          "disk_id": "sdb",
 | 
			
		||||
#          "disk_size_gb": 4,
 | 
			
		||||
#          "filesystem": "ext4",
 | 
			
		||||
#          "filesystem_args": "-L=var",
 | 
			
		||||
#          "id": 10,
 | 
			
		||||
#          "mountpoint": "/var",
 | 
			
		||||
#          "pool": "vms",
 | 
			
		||||
#          "source_volume": null,
 | 
			
		||||
#          "storage_template": 67
 | 
			
		||||
#        },
 | 
			
		||||
#        {
 | 
			
		||||
#          "disk_id": "sdc",
 | 
			
		||||
#          "disk_size_gb": 4,
 | 
			
		||||
#          "filesystem": "ext4",
 | 
			
		||||
#          "filesystem_args": "-L=log",
 | 
			
		||||
#          "id": 11,
 | 
			
		||||
#          "mountpoint": "/var/log",
 | 
			
		||||
#          "pool": "vms",
 | 
			
		||||
#          "source_volume": null,
 | 
			
		||||
#          "storage_template": 67
 | 
			
		||||
#        }
 | 
			
		||||
#      ]
 | 
			
		||||
#    }
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
class VMBuilderScript(VMBuilder):
 | 
			
		||||
    def setup(self):
 | 
			
		||||
        """
 | 
			
		||||
        setup(): Perform special setup steps or validation before proceeding
 | 
			
		||||
        """
 | 
			
		||||
 | 
			
		||||
        # Ensure we have debootstrap intalled on the provisioner system
 | 
			
		||||
        retcode, stdout, stderr = pvc_common.run_os_command(f"which debootstrap")
 | 
			
		||||
        if retcode:
 | 
			
		||||
            raise ProvisioningError("Failed to find critical dependency: debootstrap")
 | 
			
		||||
 | 
			
		||||
    def create(self):
 | 
			
		||||
        """
 | 
			
		||||
        create(): Create the VM libvirt schema definition
 | 
			
		||||
 | 
			
		||||
        This step *must* return a fully-formed Libvirt XML document as a string.
 | 
			
		||||
 | 
			
		||||
        This example leverages the built-in libvirt_schema objects provided by PVC; these
 | 
			
		||||
        can be used as-is, or replaced with your own schema(s) on a per-script basis.
 | 
			
		||||
        """
 | 
			
		||||
 | 
			
		||||
        # Run any imports first
 | 
			
		||||
        import datetime
 | 
			
		||||
        import random
 | 
			
		||||
        import pvcapid.libvirt_schema as libvirt_schema
 | 
			
		||||
 | 
			
		||||
        schema = ""
 | 
			
		||||
 | 
			
		||||
        # Prepare a description based on the VM profile
 | 
			
		||||
        description = (
 | 
			
		||||
            f"PVC provisioner @ {datetime.datetime.now()}, profile '{self.vm_profile}'"
 | 
			
		||||
        )
 | 
			
		||||
 | 
			
		||||
        # Format the header
 | 
			
		||||
        schema += libvirt_schema.libvirt_header.format(
 | 
			
		||||
            vm_name=self.vm_name,
 | 
			
		||||
            vm_uuid=self.vm_uuid,
 | 
			
		||||
            vm_description=description,
 | 
			
		||||
            vm_memory=self.vm_data["system_details"]["vram_mb"],
 | 
			
		||||
            vm_vcpus=self.vm_data["system_details"]["vcpu_count"],
 | 
			
		||||
            vm_architecture=self.vm_data["system_architecture"],
 | 
			
		||||
        )
 | 
			
		||||
 | 
			
		||||
        # Add the disk devices
 | 
			
		||||
        monitor_list = self.vm_data["ceph_monitor_list"]
 | 
			
		||||
        monitor_port = self.vm_data["ceph_monitor_port"]
 | 
			
		||||
        monitor_secret = self.vm_data["ceph_monitor_secret"]
 | 
			
		||||
 | 
			
		||||
        for volume in self.vm_data["volumes"]:
 | 
			
		||||
            schema += libvirt_schema.devices_disk_header.format(
 | 
			
		||||
                ceph_storage_secret=monitor_secret,
 | 
			
		||||
                disk_pool=volume["pool"],
 | 
			
		||||
                vm_name=self.vm_name,
 | 
			
		||||
                disk_id=volume["disk_id"],
 | 
			
		||||
            )
 | 
			
		||||
            for monitor in monitor_list:
 | 
			
		||||
                schema += libvirt_schema.devices_disk_coordinator.format(
 | 
			
		||||
                    coordinator_name=monitor,
 | 
			
		||||
                    coordinator_ceph_mon_port=monitor_port,
 | 
			
		||||
                )
 | 
			
		||||
            schema += libvirt_schema.devices_disk_footer
 | 
			
		||||
 | 
			
		||||
        # Add the special vhostmd device for hypervisor information inside the VM
 | 
			
		||||
        schema += libvirt_schema.devices_vhostmd
 | 
			
		||||
 | 
			
		||||
        # Add the network devices
 | 
			
		||||
        network_id = 0
 | 
			
		||||
        for network in self.vm_data["networks"]:
 | 
			
		||||
            vm_id_hex = "{:x}".format(int(vm_id % 16))
 | 
			
		||||
            net_id_hex = "{:x}".format(int(network_id % 16))
 | 
			
		||||
 | 
			
		||||
            if self.vm_data.get("mac_template") is not None:
 | 
			
		||||
                mac_prefix = "52:54:01"
 | 
			
		||||
                macgen_template = self.vm_data["mac_template"]
 | 
			
		||||
                eth_macaddr = macgen_template.format(
 | 
			
		||||
                    prefix=mac_prefix, vmid=vm_id_hex, netid=net_id_hex
 | 
			
		||||
                )
 | 
			
		||||
            else:
 | 
			
		||||
                mac_prefix = "52:54:00"
 | 
			
		||||
                random_octet_A = "{:x}".format(random.randint(16, 238))
 | 
			
		||||
                random_octet_B = "{:x}".format(random.randint(16, 238))
 | 
			
		||||
                random_octet_C = "{:x}".format(random.randint(16, 238))
 | 
			
		||||
 | 
			
		||||
                macgen_template = "{prefix}:{octetA}:{octetB}:{octetC}"
 | 
			
		||||
                eth_macaddr = macgen_template.format(
 | 
			
		||||
                    prefix=mac_prefix,
 | 
			
		||||
                    octetA=random_octet_A,
 | 
			
		||||
                    octetB=random_octet_B,
 | 
			
		||||
                    octetC=random_octet_C,
 | 
			
		||||
                )
 | 
			
		||||
 | 
			
		||||
            schema += libvirt_schema.devices_net_interface.format(
 | 
			
		||||
                eth_macaddr=eth_macaddr,
 | 
			
		||||
                eth_bridge=eth_bridge,
 | 
			
		||||
            )
 | 
			
		||||
 | 
			
		||||
            network_id += 1
 | 
			
		||||
 | 
			
		||||
        # Add default devices
 | 
			
		||||
        schema += libvirt_schema.devices_default
 | 
			
		||||
 | 
			
		||||
        # Add serial device
 | 
			
		||||
        if self.vm_data["system_details"]["serial"]:
 | 
			
		||||
            schema += libvirt_schema.devices_serial.format(vm_name=self.vm_name)
 | 
			
		||||
 | 
			
		||||
        # Add VNC device
 | 
			
		||||
        if self.vm_data["system_details"]["vnc"]:
 | 
			
		||||
            if self.vm_data["system_details"]["vnc_bind"]:
 | 
			
		||||
                vm_vnc_bind = self.vm_data["system_details"]["vnc_bind"]
 | 
			
		||||
            else:
 | 
			
		||||
                vm_vnc_bind = "127.0.0.1"
 | 
			
		||||
 | 
			
		||||
            vm_vncport = 5900
 | 
			
		||||
            vm_vnc_autoport = "yes"
 | 
			
		||||
 | 
			
		||||
            schema += libvirt_schema.devices_vnc.format(
 | 
			
		||||
                vm_vncport=vm_vncport,
 | 
			
		||||
                vm_vnc_autoport=vm_vnc_autoport,
 | 
			
		||||
                vm_vnc_bind=vm_vnc_bind,
 | 
			
		||||
            )
 | 
			
		||||
 | 
			
		||||
        # Add SCSI controller
 | 
			
		||||
        schema += libvirt_schema.devices_scsi_controller
 | 
			
		||||
 | 
			
		||||
        # Add footer
 | 
			
		||||
        schema += libvirt_schema.libvirt_footer
 | 
			
		||||
 | 
			
		||||
        return schema
 | 
			
		||||
 | 
			
		||||
    def prepare(self):
 | 
			
		||||
        """
 | 
			
		||||
        prepare(): Prepare any disks/volumes for the install() step
 | 
			
		||||
 | 
			
		||||
        This function should use the various exposed PVC commands as indicated to create
 | 
			
		||||
        block devices and map them to the host.
 | 
			
		||||
        """
 | 
			
		||||
 | 
			
		||||
        # First loop: Create the disks, either by cloning (pvc_ceph.clone_volume), or by
 | 
			
		||||
        # new creation (pvc_ceph.add_volume).
 | 
			
		||||
        for volume in self.vm_data["volumes"]:
 | 
			
		||||
            if volume.get("source_volume") is not None:
 | 
			
		||||
                with open_zk(config) as zkhandler:
 | 
			
		||||
                    success, message = pvc_ceph.clone_volume(
 | 
			
		||||
                        zkhandler,
 | 
			
		||||
                        volume["pool"],
 | 
			
		||||
                        volume["source_volume"],
 | 
			
		||||
                        f"{self.vm_name}_{volume['disk_id']}",
 | 
			
		||||
                    )
 | 
			
		||||
                    print(message)
 | 
			
		||||
                    if not success:
 | 
			
		||||
                        raise ProvisioningError(
 | 
			
		||||
                            f"Failed to clone volume '{volume['source_volume']}' to '{volume['disk_id']}'."
 | 
			
		||||
                        )
 | 
			
		||||
            else:
 | 
			
		||||
                with open_zk(config) as zkhandler:
 | 
			
		||||
                    success, message = pvc_ceph.add_volume(
 | 
			
		||||
                        zkhandler,
 | 
			
		||||
                        volume["pool"],
 | 
			
		||||
                        f"{self.vm_name}_{volume['disk_id']}",
 | 
			
		||||
                        f"{volume['disk_size_gb']}G",
 | 
			
		||||
                    )
 | 
			
		||||
                    print(message)
 | 
			
		||||
                    if not success:
 | 
			
		||||
                        raise ProvisioningError(
 | 
			
		||||
                            f"Failed to create volume '{volume['disk_id']}'."
 | 
			
		||||
                        )
 | 
			
		||||
 | 
			
		||||
        # Second loop: Map the disks to the local system
 | 
			
		||||
        for volume in self.vm_data["volumes"]:
 | 
			
		||||
            dst_volume_name = f"{self.vm_name}_{volume['disk_id']}"
 | 
			
		||||
            dst_volume = f"{volume['pool']}/{dst_volume_name}"
 | 
			
		||||
 | 
			
		||||
            with open_zk(config) as zkhandler:
 | 
			
		||||
                success, message = pvc_ceph.map_volume(
 | 
			
		||||
                    zkhandler,
 | 
			
		||||
                    volume["pool"],
 | 
			
		||||
                    dst_volume_name,
 | 
			
		||||
                )
 | 
			
		||||
                print(message)
 | 
			
		||||
                if not retcode:
 | 
			
		||||
                    raise ProvisioningError(f"Failed to map volume '{dst_volume}'.")
 | 
			
		||||
 | 
			
		||||
        # Third loop: Create filesystems on the volumes
 | 
			
		||||
        for volume in self.vm_data["volumes"]:
 | 
			
		||||
            dst_volume_name = f"{self.vm_name}_{volume['disk_id']}"
 | 
			
		||||
            dst_volume = f"{volume['pool']}/{dst_volume_name}"
 | 
			
		||||
 | 
			
		||||
            if volume.get("source_volume") is not None:
 | 
			
		||||
                continue
 | 
			
		||||
 | 
			
		||||
            if volume.get("filesystem") is None:
 | 
			
		||||
                continue
 | 
			
		||||
 | 
			
		||||
            filesystem_args_list = list()
 | 
			
		||||
            for arg in volume["filesystem_args"].split():
 | 
			
		||||
                arg_entry, *arg_data = arg.split("=")
 | 
			
		||||
                arg_data = "=".join(arg_data)
 | 
			
		||||
                filesystem_args_list.append(arg_entry)
 | 
			
		||||
                filesystem_args_list.append(arg_data)
 | 
			
		||||
            filesystem_args = " ".join(filesystem_args_list)
 | 
			
		||||
 | 
			
		||||
            if volume["filesystem"] == "swap":
 | 
			
		||||
                retcode, stdout, stderr = pvc_common.run_os_command(
 | 
			
		||||
                    f"mkswap -f /dev/rbd/{dst_volume}"
 | 
			
		||||
                )
 | 
			
		||||
                if retcode:
 | 
			
		||||
                    raise ProvisioningError(
 | 
			
		||||
                        f"Failed to create swap on '{dst_volume}': {stderr}"
 | 
			
		||||
                    )
 | 
			
		||||
            else:
 | 
			
		||||
                retcode, stdout, stderr = pvc_common.run_os_command(
 | 
			
		||||
                    f"mkfs.{volume['filesystem']} {filesystem_args} /dev/rbd/{dst_volume}"
 | 
			
		||||
                )
 | 
			
		||||
                if retcode:
 | 
			
		||||
                    raise ProvisioningError(
 | 
			
		||||
                        f"Faield to create {volume['filesystem']} file on '{dst_volume}': {stderr}"
 | 
			
		||||
                    )
 | 
			
		||||
 | 
			
		||||
            print(stdout)
 | 
			
		||||
 | 
			
		||||
        # Create a temporary directory to use during install
 | 
			
		||||
        temp_dir = "/tmp/target"
 | 
			
		||||
        if not os.exists(temp_dir):
 | 
			
		||||
            os.mkdir(temp_dir)
 | 
			
		||||
 | 
			
		||||
        # Fourth loop: Mount the volumes to a set of temporary directories
 | 
			
		||||
        for volume in self.vm_data["volumes"]:
 | 
			
		||||
            dst_volume_name = f"{self.vm_name}_{volume['disk_id']}"
 | 
			
		||||
            dst_volume = f"{volume['pool']}/{dst_volume_name}"
 | 
			
		||||
 | 
			
		||||
            if volume.get("source_volume") is not None:
 | 
			
		||||
                continue
 | 
			
		||||
 | 
			
		||||
            if volume.get("filesystem") is None:
 | 
			
		||||
                continue
 | 
			
		||||
 | 
			
		||||
            mapped_dst_volume = f"/dev/rbd/{dst_volume}"
 | 
			
		||||
 | 
			
		||||
            mount_path = f"{temp_dir}/{volume['mountpoint']}"
 | 
			
		||||
 | 
			
		||||
            if not os.exists(mount_path):
 | 
			
		||||
                os.mkdir(mount_path)
 | 
			
		||||
 | 
			
		||||
            # Mount filesystem
 | 
			
		||||
            retcode, stdout, stderr = pvc_common.run_os_command(
 | 
			
		||||
                f"mount {mapped_dst_volume} {mount_path}"
 | 
			
		||||
            )
 | 
			
		||||
            if retcode:
 | 
			
		||||
                raise ProvisioningError(
 | 
			
		||||
                    f"Failed to mount '{mapped_dst_volume}' on '{mount_path}': {stderr}"
 | 
			
		||||
                )
 | 
			
		||||
 | 
			
		||||
    def install(self):
 | 
			
		||||
        """
 | 
			
		||||
        install(): Perform the installation
 | 
			
		||||
 | 
			
		||||
        Since this is a noop example, this step does nothing, aside from getting some
 | 
			
		||||
        arguments for demonstration.
 | 
			
		||||
        """
 | 
			
		||||
 | 
			
		||||
        # The directory we mounted things on earlier during prepare()
 | 
			
		||||
        temporary_directory = "/tmp/target"
 | 
			
		||||
 | 
			
		||||
        # Use these convenient aliases for later (avoiding lots of "self.vm_data" everywhere)
 | 
			
		||||
        vm_name = self.vm_name
 | 
			
		||||
        disks = self.vm_data["disks"]
 | 
			
		||||
        networks = self.vm_data["networks"]
 | 
			
		||||
 | 
			
		||||
        # Parse these arguments out of self.vm_data["script_arguments"]
 | 
			
		||||
        if self.vm_data["script_arguments"].get("deb_release") is not None:
 | 
			
		||||
            deb_release = self.vm_data["script_arguments"].get("deb_release")
 | 
			
		||||
        else:
 | 
			
		||||
            deb_release = "stable"
 | 
			
		||||
 | 
			
		||||
        if self.vm_data["script_arguments"].get("deb_mirror") is not None:
 | 
			
		||||
            deb_mirror = self.vm_data["script_arguments"].get("deb_mirror")
 | 
			
		||||
        else:
 | 
			
		||||
            deb_mirror = "http://ftp.debian.org/debian"
 | 
			
		||||
 | 
			
		||||
        if self.vm_data["script_arguments"].get("deb_packages") is not None:
 | 
			
		||||
            deb_packages = (
 | 
			
		||||
                self.vm_data["script_arguments"].get("deb_packages").split(",")
 | 
			
		||||
            )
 | 
			
		||||
        else:
 | 
			
		||||
            deb_packages = [
 | 
			
		||||
                "linux-image-amd64",
 | 
			
		||||
                "grub-pc",
 | 
			
		||||
                "cloud-init",
 | 
			
		||||
                "python3-cffi-backend",
 | 
			
		||||
                "wget",
 | 
			
		||||
            ]
 | 
			
		||||
 | 
			
		||||
        # We need to know our root disk
 | 
			
		||||
        root_disk = None
 | 
			
		||||
        for disk in disks:
 | 
			
		||||
            if disk["mountpoint"] == "/":
 | 
			
		||||
                root_disk = disk
 | 
			
		||||
        if not root_disk:
 | 
			
		||||
            raise ProvisioningError("Failed to find root disk in disks list")
 | 
			
		||||
 | 
			
		||||
        # Perform a deboostrap installation
 | 
			
		||||
        os.system(
 | 
			
		||||
            "debootstrap --include={pkgs} {suite} {target} {mirror}".format(
 | 
			
		||||
                suite=deb_release,
 | 
			
		||||
                target=temporary_directory,
 | 
			
		||||
                mirror=deb_mirror,
 | 
			
		||||
                pkgs=",".join(deb_packages),
 | 
			
		||||
            )
 | 
			
		||||
        )
 | 
			
		||||
 | 
			
		||||
        # Bind mount the devfs
 | 
			
		||||
        os.system("mount --bind /dev {}/dev".format(temporary_directory))
 | 
			
		||||
 | 
			
		||||
        # Create an fstab entry for each disk
 | 
			
		||||
        fstab_file = "{}/etc/fstab".format(temporary_directory)
 | 
			
		||||
        # The disk ID starts at zero and increments by one for each disk in the fixed-order
 | 
			
		||||
        # disk list. This lets us work around the insanity of Libvirt IDs not matching guest IDs,
 | 
			
		||||
        # while still letting us have some semblance of control here without enforcing things
 | 
			
		||||
        # like labels. It increments in the for loop below at the end of each iteration, and is
 | 
			
		||||
        # used to craft a /dev/disk/by-id/scsi-0QEMU_QEMU_HARDDISK_drive-scsi0-0-0-X device ID
 | 
			
		||||
        # which will always match the correct order from Libvirt (unlike sdX/vdX names).
 | 
			
		||||
        disk_id = 0
 | 
			
		||||
        for disk in disks:
 | 
			
		||||
            # We assume SSD-based/-like storage, and dislike atimes
 | 
			
		||||
            options = "defaults,discard,noatime,nodiratime"
 | 
			
		||||
 | 
			
		||||
            # The root, var, and log volumes have specific values
 | 
			
		||||
            if disk["mountpoint"] == "/":
 | 
			
		||||
                root_disk["scsi_id"] = disk_id
 | 
			
		||||
                dump = 0
 | 
			
		||||
                cpass = 1
 | 
			
		||||
            elif disk["mountpoint"] == "/var" or disk["mountpoint"] == "/var/log":
 | 
			
		||||
                dump = 0
 | 
			
		||||
                cpass = 2
 | 
			
		||||
            else:
 | 
			
		||||
                dump = 0
 | 
			
		||||
                cpass = 0
 | 
			
		||||
 | 
			
		||||
            # Append the fstab line
 | 
			
		||||
            with open(fstab_file, "a") as fh:
 | 
			
		||||
                data = "/dev/disk/by-id/scsi-0QEMU_QEMU_HARDDISK_drive-scsi0-0-0-{disk} {mountpoint} {filesystem} {options} {dump} {cpass}\n".format(
 | 
			
		||||
                    disk=disk_id,
 | 
			
		||||
                    mountpoint=disk["mountpoint"],
 | 
			
		||||
                    filesystem=disk["filesystem"],
 | 
			
		||||
                    options=options,
 | 
			
		||||
                    dump=dump,
 | 
			
		||||
                    cpass=cpass,
 | 
			
		||||
                )
 | 
			
		||||
                fh.write(data)
 | 
			
		||||
 | 
			
		||||
            # Increment the disk_id
 | 
			
		||||
            disk_id += 1
 | 
			
		||||
 | 
			
		||||
        # Write the hostname
 | 
			
		||||
        hostname_file = "{}/etc/hostname".format(temporary_directory)
 | 
			
		||||
        with open(hostname_file, "w") as fh:
 | 
			
		||||
            fh.write("{}".format(vm_name))
 | 
			
		||||
 | 
			
		||||
        # Fix the cloud-init.target since it's broken
 | 
			
		||||
        cloudinit_target_file = "{}/etc/systemd/system/cloud-init.target".format(
 | 
			
		||||
            temporary_directory
 | 
			
		||||
        )
 | 
			
		||||
        with open(cloudinit_target_file, "w") as fh:
 | 
			
		||||
            data = """[Install]
 | 
			
		||||
WantedBy=multi-user.target
 | 
			
		||||
[Unit]
 | 
			
		||||
Description=Cloud-init target
 | 
			
		||||
After=multi-user.target
 | 
			
		||||
"""
 | 
			
		||||
            fh.write(data)
 | 
			
		||||
 | 
			
		||||
        # NOTE: Due to device ordering within the Libvirt XML configuration, the first Ethernet interface
 | 
			
		||||
        #       will always be on PCI bus ID 2, hence the name "ens2".
 | 
			
		||||
        # Write a DHCP stanza for ens2
 | 
			
		||||
        ens2_network_file = "{}/etc/network/interfaces.d/ens2".format(
 | 
			
		||||
            temporary_directory
 | 
			
		||||
        )
 | 
			
		||||
        with open(ens2_network_file, "w") as fh:
 | 
			
		||||
            data = """auto ens2
 | 
			
		||||
iface ens2 inet dhcp
 | 
			
		||||
"""
 | 
			
		||||
            fh.write(data)
 | 
			
		||||
 | 
			
		||||
        # Write the DHCP config for ens2
 | 
			
		||||
        dhclient_file = "{}/etc/dhcp/dhclient.conf".format(temporary_directory)
 | 
			
		||||
        with open(dhclient_file, "w") as fh:
 | 
			
		||||
            data = (
 | 
			
		||||
                """# DHCP client configuration
 | 
			
		||||
# Written by the PVC provisioner
 | 
			
		||||
option rfc3442-classless-static-routes code 121 = array of unsigned integer 8;
 | 
			
		||||
interface "ens2" {
 | 
			
		||||
"""
 | 
			
		||||
                + """        send fqdn.fqdn = "{hostname}";
 | 
			
		||||
        send host-name = "{hostname}";
 | 
			
		||||
    """.format(
 | 
			
		||||
                    hostname=vm_name
 | 
			
		||||
                )
 | 
			
		||||
                + """        request subnet-mask, broadcast-address, time-offset, routers,
 | 
			
		||||
                domain-name, domain-name-servers, domain-search, host-name,
 | 
			
		||||
                dhcp6.name-servers, dhcp6.domain-search, dhcp6.fqdn, dhcp6.sntp-servers,
 | 
			
		||||
                netbios-name-servers, netbios-scope, interface-mtu,
 | 
			
		||||
                rfc3442-classless-static-routes, ntp-servers;
 | 
			
		||||
}
 | 
			
		||||
"""
 | 
			
		||||
            )
 | 
			
		||||
            fh.write(data)
 | 
			
		||||
 | 
			
		||||
        # Write the GRUB configuration
 | 
			
		||||
        grubcfg_file = "{}/etc/default/grub".format(temporary_directory)
 | 
			
		||||
        with open(grubcfg_file, "w") as fh:
 | 
			
		||||
            data = """# Written by the PVC provisioner
 | 
			
		||||
GRUB_DEFAULT=0
 | 
			
		||||
GRUB_TIMEOUT=1
 | 
			
		||||
GRUB_DISTRIBUTOR="PVC Virtual Machine"
 | 
			
		||||
GRUB_CMDLINE_LINUX_DEFAULT="root=/dev/disk/by-id/scsi-0QEMU_QEMU_HARDDISK_drive-scsi0-0-0-{root_disk} console=tty0 console=ttyS0,115200n8"
 | 
			
		||||
GRUB_CMDLINE_LINUX=""
 | 
			
		||||
GRUB_TERMINAL=console
 | 
			
		||||
GRUB_SERIAL_COMMAND="serial --speed=115200 --unit=0 --word=8 --parity=no --stop=1"
 | 
			
		||||
GRUB_DISABLE_LINUX_UUID=false
 | 
			
		||||
""".format(
 | 
			
		||||
                root_disk=root_disk["scsi_id"]
 | 
			
		||||
            )
 | 
			
		||||
            fh.write(data)
 | 
			
		||||
 | 
			
		||||
        # Chroot, do some in-root tasks, then exit the chroot
 | 
			
		||||
        with chroot_target(temporary_directory):
 | 
			
		||||
            # Install and update GRUB
 | 
			
		||||
            os.system(
 | 
			
		||||
                "grub-install --force /dev/rbd/{}/{}_{}".format(
 | 
			
		||||
                    root_disk["pool"], vm_name, root_disk["disk_id"]
 | 
			
		||||
                )
 | 
			
		||||
            )
 | 
			
		||||
            os.system("update-grub")
 | 
			
		||||
            # Set a really dumb root password [TEMPORARY]
 | 
			
		||||
            os.system("echo root:test123 | chpasswd")
 | 
			
		||||
            # Enable cloud-init target on (first) boot
 | 
			
		||||
            # NOTE: Your user-data should handle this and disable it once done, or things get messy.
 | 
			
		||||
            #       That cloud-init won't run without this hack seems like a bug... but even the official
 | 
			
		||||
            #       Debian cloud images are affected, so who knows.
 | 
			
		||||
            os.system("systemctl enable cloud-init.target")
 | 
			
		||||
 | 
			
		||||
        # Unmount the bound devfs
 | 
			
		||||
        os.system("umount {}/dev".format(temporary_directory))
 | 
			
		||||
 | 
			
		||||
    def cleanup(self):
 | 
			
		||||
        """
 | 
			
		||||
        cleanup(): Perform any cleanup required due to prepare()/install()
 | 
			
		||||
 | 
			
		||||
        It is important to now reverse *all* steps taken in those functions that might
 | 
			
		||||
        need cleanup before teardown of the overlay chroot environment.
 | 
			
		||||
        """
 | 
			
		||||
 | 
			
		||||
        temp_dir = "/tmp/target"
 | 
			
		||||
 | 
			
		||||
        for volume in list(reversed(self.vm_data["volumes"])):
 | 
			
		||||
            dst_volume_name = f"{self.vm_name}_{volume['disk_id']}"
 | 
			
		||||
            dst_volume = f"{volume['pool']}/{dst_volume_name}"
 | 
			
		||||
            mapped_dst_volume = f"/dev/rbd/{dst_volume}"
 | 
			
		||||
            mount_path = f"{temp_dir}/{volume['mountpoint']}"
 | 
			
		||||
 | 
			
		||||
            if (
 | 
			
		||||
                volume.get("source_volume") is None
 | 
			
		||||
                and volume.get("filesystem") is not None
 | 
			
		||||
            ):
 | 
			
		||||
                # Unmount filesystem
 | 
			
		||||
                retcode, stdout, stderr = pvc_common.run_os_command(
 | 
			
		||||
                    f"umount {mount_path}"
 | 
			
		||||
                )
 | 
			
		||||
                if retcode:
 | 
			
		||||
                    raise ProvisioningError(
 | 
			
		||||
                        f"Failed to unmount '{mapped_dst_volume}' on '{mount_path}': {stderr}"
 | 
			
		||||
                    )
 | 
			
		||||
 | 
			
		||||
                # Unmap volume
 | 
			
		||||
                with open_zk(config) as zkhandler:
 | 
			
		||||
                    success, message = pvc_ceph.unmap_volume(
 | 
			
		||||
                        zkhandler,
 | 
			
		||||
                        volume["pool"],
 | 
			
		||||
                        dst_volume_name,
 | 
			
		||||
                    )
 | 
			
		||||
                    if not success:
 | 
			
		||||
                        raise ProvisioningError(
 | 
			
		||||
                            f"Failed to unmap '{mapped_dst_volume}': {stderr}"
 | 
			
		||||
                        )
 | 
			
		||||
							
								
								
									
										604
									
								
								api-daemon/provisioner/examples/3-ova.py
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										604
									
								
								api-daemon/provisioner/examples/3-ova.py
									
									
									
									
									
										Normal file
									
								
							@@ -0,0 +1,604 @@
 | 
			
		||||
#!/usr/bin/env python3
 | 
			
		||||
 | 
			
		||||
# 3-ova.py - PVC Provisioner example script for OVA install
 | 
			
		||||
# Part of the Parallel Virtual Cluster (PVC) system
 | 
			
		||||
#
 | 
			
		||||
#    Copyright (C) 2018-2022 Joshua M. Boniface <joshua@boniface.me>
 | 
			
		||||
#
 | 
			
		||||
#    This program is free software: you can redistribute it and/or modify
 | 
			
		||||
#    it under the terms of the GNU General Public License as published by
 | 
			
		||||
#    the Free Software Foundation, version 3.
 | 
			
		||||
#
 | 
			
		||||
#    This program is distributed in the hope that it will be useful,
 | 
			
		||||
#    but WITHOUT ANY WARRANTY; without even the implied warranty of
 | 
			
		||||
#    MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
 | 
			
		||||
#    GNU General Public License for more details.
 | 
			
		||||
#
 | 
			
		||||
#    You should have received a copy of the GNU General Public License
 | 
			
		||||
#    along with this program.  If not, see <https://www.gnu.org/licenses/>.
 | 
			
		||||
#
 | 
			
		||||
###############################################################################
 | 
			
		||||
 | 
			
		||||
# This script provides an example of a PVC provisioner script. It will create a
 | 
			
		||||
# custom VM config based on an OVA profile.
 | 
			
		||||
 | 
			
		||||
# This script can thus be used as an example or reference implementation of a
 | 
			
		||||
# PVC provisioner script and expanded upon as required; this specific script is
 | 
			
		||||
# also hard-coded into a normal PVC provisioner system.
 | 
			
		||||
 | 
			
		||||
# The script must implement the class "VMBuilderScript" which extens "VMBuilder",
 | 
			
		||||
# providing the 5 functions indicated. Detailed explanation of the role of each
 | 
			
		||||
# function is provided.
 | 
			
		||||
 | 
			
		||||
# Within the VMBuilderScript class, several common variables are exposed:
 | 
			
		||||
#  self.vm_name: The name of the VM from PVC's perspective
 | 
			
		||||
#  self.vm_id: The VM ID (numerical component of the vm_name) from PVC's perspective
 | 
			
		||||
#  self.vm_uuid: An automatically-generated UUID for the VM
 | 
			
		||||
#  self.vm_profile: The PVC provisioner profile name used for the VM
 | 
			
		||||
#  self.vm-data: A dictionary of VM data collected by the provisioner; an example:
 | 
			
		||||
#    {
 | 
			
		||||
#      "ceph_monitor_list": [
 | 
			
		||||
#        "hv1.pvcstorage.tld",
 | 
			
		||||
#        "hv2.pvcstorage.tld",
 | 
			
		||||
#        "hv3.pvcstorage.tld"
 | 
			
		||||
#      ],
 | 
			
		||||
#      "ceph_monitor_port": "6789",
 | 
			
		||||
#      "ceph_monitor_secret": "96721723-8650-4a72-b8f6-a93cd1a20f0c",
 | 
			
		||||
#      "mac_template": null,
 | 
			
		||||
#      "networks": [
 | 
			
		||||
#        {
 | 
			
		||||
#          "eth_bridge": "vmbr1001",
 | 
			
		||||
#          "id": 72,
 | 
			
		||||
#          "network_template": 69,
 | 
			
		||||
#          "vni": "1001"
 | 
			
		||||
#        },
 | 
			
		||||
#        {
 | 
			
		||||
#          "eth_bridge": "vmbr101",
 | 
			
		||||
#          "id": 73,
 | 
			
		||||
#          "network_template": 69,
 | 
			
		||||
#          "vni": "101"
 | 
			
		||||
#        }
 | 
			
		||||
#      ],
 | 
			
		||||
#      "script": [contents of this file]
 | 
			
		||||
#      "script_arguments": {
 | 
			
		||||
#        "deb_mirror": "http://ftp.debian.org/debian",
 | 
			
		||||
#        "deb_release": "bullseye"
 | 
			
		||||
#      },
 | 
			
		||||
#      "system_architecture": "x86_64",
 | 
			
		||||
#      "system_details": {
 | 
			
		||||
#        "id": 78,
 | 
			
		||||
#        "migration_method": "live",
 | 
			
		||||
#        "name": "small",
 | 
			
		||||
#        "node_autostart": false,
 | 
			
		||||
#        "node_limit": null,
 | 
			
		||||
#        "node_selector": null,
 | 
			
		||||
#        "ova": null,
 | 
			
		||||
#        "serial": true,
 | 
			
		||||
#        "vcpu_count": 2,
 | 
			
		||||
#        "vnc": false,
 | 
			
		||||
#        "vnc_bind": null,
 | 
			
		||||
#        "vram_mb": 2048
 | 
			
		||||
#      },
 | 
			
		||||
#      "volumes": [
 | 
			
		||||
#        {
 | 
			
		||||
#          "disk_id": "sda",
 | 
			
		||||
#          "disk_size_gb": 4,
 | 
			
		||||
#          "filesystem": "ext4",
 | 
			
		||||
#          "filesystem_args": "-L=root",
 | 
			
		||||
#          "id": 9,
 | 
			
		||||
#          "mountpoint": "/",
 | 
			
		||||
#          "pool": "vms",
 | 
			
		||||
#          "source_volume": null,
 | 
			
		||||
#          "storage_template": 67
 | 
			
		||||
#        },
 | 
			
		||||
#        {
 | 
			
		||||
#          "disk_id": "sdb",
 | 
			
		||||
#          "disk_size_gb": 4,
 | 
			
		||||
#          "filesystem": "ext4",
 | 
			
		||||
#          "filesystem_args": "-L=var",
 | 
			
		||||
#          "id": 10,
 | 
			
		||||
#          "mountpoint": "/var",
 | 
			
		||||
#          "pool": "vms",
 | 
			
		||||
#          "source_volume": null,
 | 
			
		||||
#          "storage_template": 67
 | 
			
		||||
#        },
 | 
			
		||||
#        {
 | 
			
		||||
#          "disk_id": "sdc",
 | 
			
		||||
#          "disk_size_gb": 4,
 | 
			
		||||
#          "filesystem": "ext4",
 | 
			
		||||
#          "filesystem_args": "-L=log",
 | 
			
		||||
#          "id": 11,
 | 
			
		||||
#          "mountpoint": "/var/log",
 | 
			
		||||
#          "pool": "vms",
 | 
			
		||||
#          "source_volume": null,
 | 
			
		||||
#          "storage_template": 67
 | 
			
		||||
#        }
 | 
			
		||||
#      ]
 | 
			
		||||
#    }
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
# Run any imports first
 | 
			
		||||
import datetime
 | 
			
		||||
import random
 | 
			
		||||
import pvcapid.libvirt_schema as libvirt_schema
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
class VMBuilderScript(VMBuilder):
 | 
			
		||||
    def setup(self):
 | 
			
		||||
        """
 | 
			
		||||
        setup(): Perform special setup steps or validation before proceeding
 | 
			
		||||
        """
 | 
			
		||||
 | 
			
		||||
        # Ensure we have debootstrap intalled on the provisioner system
 | 
			
		||||
        retcode, stdout, stderr = pvc_common.run_os_command(f"which debootstrap")
 | 
			
		||||
        if retcode:
 | 
			
		||||
            raise ProvisioningError("Failed to find critical dependency: debootstrap")
 | 
			
		||||
 | 
			
		||||
    def create(self):
 | 
			
		||||
        """
 | 
			
		||||
        create(): Create the VM libvirt schema definition
 | 
			
		||||
 | 
			
		||||
        This step *must* return a fully-formed Libvirt XML document as a string.
 | 
			
		||||
 | 
			
		||||
        This example leverages the built-in libvirt_schema objects provided by PVC; these
 | 
			
		||||
        can be used as-is, or replaced with your own schema(s) on a per-script basis.
 | 
			
		||||
        """
 | 
			
		||||
 | 
			
		||||
        schema = ""
 | 
			
		||||
 | 
			
		||||
        # Prepare a description based on the VM profile
 | 
			
		||||
        description = (
 | 
			
		||||
            f"PVC provisioner @ {datetime.datetime.now()}, profile '{self.vm_profile}'"
 | 
			
		||||
        )
 | 
			
		||||
 | 
			
		||||
        # Format the header
 | 
			
		||||
        schema += libvirt_schema.libvirt_header.format(
 | 
			
		||||
            vm_name=self.vm_name,
 | 
			
		||||
            vm_uuid=self.vm_uuid,
 | 
			
		||||
            vm_description=description,
 | 
			
		||||
            vm_memory=self.vm_data["system_details"]["vram_mb"],
 | 
			
		||||
            vm_vcpus=self.vm_data["system_details"]["vcpu_count"],
 | 
			
		||||
            vm_architecture=self.vm_data["system_architecture"],
 | 
			
		||||
        )
 | 
			
		||||
 | 
			
		||||
        # Add the disk devices
 | 
			
		||||
        monitor_list = self.vm_data["ceph_monitor_list"]
 | 
			
		||||
        monitor_port = self.vm_data["ceph_monitor_port"]
 | 
			
		||||
        monitor_secret = self.vm_data["ceph_monitor_secret"]
 | 
			
		||||
 | 
			
		||||
        for volume in self.vm_data["volumes"]:
 | 
			
		||||
            schema += libvirt_schema.devices_disk_header.format(
 | 
			
		||||
                ceph_storage_secret=monitor_secret,
 | 
			
		||||
                disk_pool=volume["pool"],
 | 
			
		||||
                vm_name=self.vm_name,
 | 
			
		||||
                disk_id=volume["disk_id"],
 | 
			
		||||
            )
 | 
			
		||||
            for monitor in monitor_list:
 | 
			
		||||
                schema += libvirt_schema.devices_disk_coordinator.format(
 | 
			
		||||
                    coordinator_name=monitor,
 | 
			
		||||
                    coordinator_ceph_mon_port=monitor_port,
 | 
			
		||||
                )
 | 
			
		||||
            schema += libvirt_schema.devices_disk_footer
 | 
			
		||||
 | 
			
		||||
        # Add the special vhostmd device for hypervisor information inside the VM
 | 
			
		||||
        schema += libvirt_schema.devices_vhostmd
 | 
			
		||||
 | 
			
		||||
        # Add the network devices
 | 
			
		||||
        network_id = 0
 | 
			
		||||
        for network in self.vm_data["networks"]:
 | 
			
		||||
            vm_id_hex = "{:x}".format(int(vm_id % 16))
 | 
			
		||||
            net_id_hex = "{:x}".format(int(network_id % 16))
 | 
			
		||||
 | 
			
		||||
            if self.vm_data.get("mac_template") is not None:
 | 
			
		||||
                mac_prefix = "52:54:01"
 | 
			
		||||
                macgen_template = self.vm_data["mac_template"]
 | 
			
		||||
                eth_macaddr = macgen_template.format(
 | 
			
		||||
                    prefix=mac_prefix, vmid=vm_id_hex, netid=net_id_hex
 | 
			
		||||
                )
 | 
			
		||||
            else:
 | 
			
		||||
                mac_prefix = "52:54:00"
 | 
			
		||||
                random_octet_A = "{:x}".format(random.randint(16, 238))
 | 
			
		||||
                random_octet_B = "{:x}".format(random.randint(16, 238))
 | 
			
		||||
                random_octet_C = "{:x}".format(random.randint(16, 238))
 | 
			
		||||
 | 
			
		||||
                macgen_template = "{prefix}:{octetA}:{octetB}:{octetC}"
 | 
			
		||||
                eth_macaddr = macgen_template.format(
 | 
			
		||||
                    prefix=mac_prefix,
 | 
			
		||||
                    octetA=random_octet_A,
 | 
			
		||||
                    octetB=random_octet_B,
 | 
			
		||||
                    octetC=random_octet_C,
 | 
			
		||||
                )
 | 
			
		||||
 | 
			
		||||
            schema += libvirt_schema.devices_net_interface.format(
 | 
			
		||||
                eth_macaddr=eth_macaddr,
 | 
			
		||||
                eth_bridge=eth_bridge,
 | 
			
		||||
            )
 | 
			
		||||
 | 
			
		||||
            network_id += 1
 | 
			
		||||
 | 
			
		||||
        # Add default devices
 | 
			
		||||
        schema += libvirt_schema.devices_default
 | 
			
		||||
 | 
			
		||||
        # Add serial device
 | 
			
		||||
        if self.vm_data["system_details"]["serial"]:
 | 
			
		||||
            schema += libvirt_schema.devices_serial.format(vm_name=self.vm_name)
 | 
			
		||||
 | 
			
		||||
        # Add VNC device
 | 
			
		||||
        if self.vm_data["system_details"]["vnc"]:
 | 
			
		||||
            if self.vm_data["system_details"]["vnc_bind"]:
 | 
			
		||||
                vm_vnc_bind = self.vm_data["system_details"]["vnc_bind"]
 | 
			
		||||
            else:
 | 
			
		||||
                vm_vnc_bind = "127.0.0.1"
 | 
			
		||||
 | 
			
		||||
            vm_vncport = 5900
 | 
			
		||||
            vm_vnc_autoport = "yes"
 | 
			
		||||
 | 
			
		||||
            schema += libvirt_schema.devices_vnc.format(
 | 
			
		||||
                vm_vncport=vm_vncport,
 | 
			
		||||
                vm_vnc_autoport=vm_vnc_autoport,
 | 
			
		||||
                vm_vnc_bind=vm_vnc_bind,
 | 
			
		||||
            )
 | 
			
		||||
 | 
			
		||||
        # Add SCSI controller
 | 
			
		||||
        schema += libvirt_schema.devices_scsi_controller
 | 
			
		||||
 | 
			
		||||
        # Add footer
 | 
			
		||||
        schema += libvirt_schema.libvirt_footer
 | 
			
		||||
 | 
			
		||||
        return schema
 | 
			
		||||
 | 
			
		||||
    def prepare(self):
 | 
			
		||||
        """
 | 
			
		||||
        prepare(): Prepare any disks/volumes for the install() step
 | 
			
		||||
 | 
			
		||||
        This function should use the various exposed PVC commands as indicated to create
 | 
			
		||||
        block devices and map them to the host.
 | 
			
		||||
        """
 | 
			
		||||
 | 
			
		||||
        # First loop: Create the disks, either by cloning (pvc_ceph.clone_volume), or by
 | 
			
		||||
        # new creation (pvc_ceph.add_volume).
 | 
			
		||||
        for volume in self.vm_data["volumes"]:
 | 
			
		||||
            if volume.get("source_volume") is not None:
 | 
			
		||||
                with open_zk(config) as zkhandler:
 | 
			
		||||
                    success, message = pvc_ceph.clone_volume(
 | 
			
		||||
                        zkhandler,
 | 
			
		||||
                        volume["pool"],
 | 
			
		||||
                        volume["source_volume"],
 | 
			
		||||
                        f"{self.vm_name}_{volume['disk_id']}",
 | 
			
		||||
                    )
 | 
			
		||||
                    print(message)
 | 
			
		||||
                    if not success:
 | 
			
		||||
                        raise ProvisioningError(
 | 
			
		||||
                            f"Failed to clone volume '{volume['source_volume']}' to '{volume['disk_id']}'."
 | 
			
		||||
                        )
 | 
			
		||||
            else:
 | 
			
		||||
                with open_zk(config) as zkhandler:
 | 
			
		||||
                    success, message = pvc_ceph.add_volume(
 | 
			
		||||
                        zkhandler,
 | 
			
		||||
                        volume["pool"],
 | 
			
		||||
                        f"{self.vm_name}_{volume['disk_id']}",
 | 
			
		||||
                        f"{volume['disk_size_gb']}G",
 | 
			
		||||
                    )
 | 
			
		||||
                    print(message)
 | 
			
		||||
                    if not success:
 | 
			
		||||
                        raise ProvisioningError(
 | 
			
		||||
                            f"Failed to create volume '{volume['disk_id']}'."
 | 
			
		||||
                        )
 | 
			
		||||
 | 
			
		||||
        # Second loop: Map the disks to the local system
 | 
			
		||||
        for volume in self.vm_data["volumes"]:
 | 
			
		||||
            dst_volume_name = f"{self.vm_name}_{volume['disk_id']}"
 | 
			
		||||
            dst_volume = f"{volume['pool']}/{dst_volume_name}"
 | 
			
		||||
 | 
			
		||||
            with open_zk(config) as zkhandler:
 | 
			
		||||
                success, message = pvc_ceph.map_volume(
 | 
			
		||||
                    zkhandler,
 | 
			
		||||
                    volume["pool"],
 | 
			
		||||
                    dst_volume_name,
 | 
			
		||||
                )
 | 
			
		||||
                print(message)
 | 
			
		||||
                if not retcode:
 | 
			
		||||
                    raise ProvisioningError(f"Failed to map volume '{dst_volume}'.")
 | 
			
		||||
 | 
			
		||||
        # Third loop: Create filesystems on the volumes
 | 
			
		||||
        for volume in self.vm_data["volumes"]:
 | 
			
		||||
            dst_volume_name = f"{self.vm_name}_{volume['disk_id']}"
 | 
			
		||||
            dst_volume = f"{volume['pool']}/{dst_volume_name}"
 | 
			
		||||
 | 
			
		||||
            if volume.get("source_volume") is not None:
 | 
			
		||||
                continue
 | 
			
		||||
 | 
			
		||||
            if volume.get("filesystem") is None:
 | 
			
		||||
                continue
 | 
			
		||||
 | 
			
		||||
            filesystem_args_list = list()
 | 
			
		||||
            for arg in volume["filesystem_args"].split():
 | 
			
		||||
                arg_entry, *arg_data = arg.split("=")
 | 
			
		||||
                arg_data = "=".join(arg_data)
 | 
			
		||||
                filesystem_args_list.append(arg_entry)
 | 
			
		||||
                filesystem_args_list.append(arg_data)
 | 
			
		||||
            filesystem_args = " ".join(filesystem_args_list)
 | 
			
		||||
 | 
			
		||||
            if volume["filesystem"] == "swap":
 | 
			
		||||
                retcode, stdout, stderr = pvc_common.run_os_command(
 | 
			
		||||
                    f"mkswap -f /dev/rbd/{dst_volume}"
 | 
			
		||||
                )
 | 
			
		||||
                if retcode:
 | 
			
		||||
                    raise ProvisioningError(
 | 
			
		||||
                        f"Failed to create swap on '{dst_volume}': {stderr}"
 | 
			
		||||
                    )
 | 
			
		||||
            else:
 | 
			
		||||
                retcode, stdout, stderr = pvc_common.run_os_command(
 | 
			
		||||
                    f"mkfs.{volume['filesystem']} {filesystem_args} /dev/rbd/{dst_volume}"
 | 
			
		||||
                )
 | 
			
		||||
                if retcode:
 | 
			
		||||
                    raise ProvisioningError(
 | 
			
		||||
                        f"Faield to create {volume['filesystem']} file on '{dst_volume}': {stderr}"
 | 
			
		||||
                    )
 | 
			
		||||
 | 
			
		||||
            print(stdout)
 | 
			
		||||
 | 
			
		||||
        # Create a temporary directory to use during install
 | 
			
		||||
        temp_dir = "/tmp/target"
 | 
			
		||||
        if not os.exists(temp_dir):
 | 
			
		||||
            os.mkdir(temp_dir)
 | 
			
		||||
 | 
			
		||||
        # Fourth loop: Mount the volumes to a set of temporary directories
 | 
			
		||||
        for volume in self.vm_data["volumes"]:
 | 
			
		||||
            dst_volume_name = f"{self.vm_name}_{volume['disk_id']}"
 | 
			
		||||
            dst_volume = f"{volume['pool']}/{dst_volume_name}"
 | 
			
		||||
 | 
			
		||||
            if volume.get("source_volume") is not None:
 | 
			
		||||
                continue
 | 
			
		||||
 | 
			
		||||
            if volume.get("filesystem") is None:
 | 
			
		||||
                continue
 | 
			
		||||
 | 
			
		||||
            mapped_dst_volume = f"/dev/rbd/{dst_volume}"
 | 
			
		||||
 | 
			
		||||
            mount_path = f"{temp_dir}/{volume['mountpoint']}"
 | 
			
		||||
 | 
			
		||||
            if not os.exists(mount_path):
 | 
			
		||||
                os.mkdir(mount_path)
 | 
			
		||||
 | 
			
		||||
            # Mount filesystem
 | 
			
		||||
            retcode, stdout, stderr = pvc_common.run_os_command(
 | 
			
		||||
                f"mount {mapped_dst_volume} {mount_path}"
 | 
			
		||||
            )
 | 
			
		||||
            if retcode:
 | 
			
		||||
                raise ProvisioningError(
 | 
			
		||||
                    f"Failed to mount '{mapped_dst_volume}' on '{mount_path}': {stderr}"
 | 
			
		||||
                )
 | 
			
		||||
 | 
			
		||||
    def install(self):
 | 
			
		||||
        """
 | 
			
		||||
        install(): Perform the installation
 | 
			
		||||
 | 
			
		||||
        Since this is a noop example, this step does nothing, aside from getting some
 | 
			
		||||
        arguments for demonstration.
 | 
			
		||||
        """
 | 
			
		||||
 | 
			
		||||
        # The directory we mounted things on earlier during prepare()
 | 
			
		||||
        temporary_directory = "/tmp/target"
 | 
			
		||||
 | 
			
		||||
        # Use these convenient aliases for later (avoiding lots of "self.vm_data" everywhere)
 | 
			
		||||
        vm_name = self.vm_name
 | 
			
		||||
        disks = self.vm_data["disks"]
 | 
			
		||||
        networks = self.vm_data["networks"]
 | 
			
		||||
 | 
			
		||||
        # Parse these arguments out of self.vm_data["script_arguments"]
 | 
			
		||||
        if self.vm_data["script_arguments"].get("deb_release") is not None:
 | 
			
		||||
            deb_release = self.vm_data["script_arguments"].get("deb_release")
 | 
			
		||||
        else:
 | 
			
		||||
            deb_release = "stable"
 | 
			
		||||
 | 
			
		||||
        if self.vm_data["script_arguments"].get("deb_mirror") is not None:
 | 
			
		||||
            deb_mirror = self.vm_data["script_arguments"].get("deb_mirror")
 | 
			
		||||
        else:
 | 
			
		||||
            deb_mirror = "http://ftp.debian.org/debian"
 | 
			
		||||
 | 
			
		||||
        if self.vm_data["script_arguments"].get("deb_packages") is not None:
 | 
			
		||||
            deb_packages = (
 | 
			
		||||
                self.vm_data["script_arguments"].get("deb_packages").split(",")
 | 
			
		||||
            )
 | 
			
		||||
        else:
 | 
			
		||||
            deb_packages = [
 | 
			
		||||
                "linux-image-amd64",
 | 
			
		||||
                "grub-pc",
 | 
			
		||||
                "cloud-init",
 | 
			
		||||
                "python3-cffi-backend",
 | 
			
		||||
                "wget",
 | 
			
		||||
            ]
 | 
			
		||||
 | 
			
		||||
        # We need to know our root disk
 | 
			
		||||
        root_disk = None
 | 
			
		||||
        for disk in disks:
 | 
			
		||||
            if disk["mountpoint"] == "/":
 | 
			
		||||
                root_disk = disk
 | 
			
		||||
        if not root_disk:
 | 
			
		||||
            raise ProvisioningError("Failed to find root disk in disks list")
 | 
			
		||||
 | 
			
		||||
        # Perform a deboostrap installation
 | 
			
		||||
        os.system(
 | 
			
		||||
            "debootstrap --include={pkgs} {suite} {target} {mirror}".format(
 | 
			
		||||
                suite=deb_release,
 | 
			
		||||
                target=temporary_directory,
 | 
			
		||||
                mirror=deb_mirror,
 | 
			
		||||
                pkgs=",".join(deb_packages),
 | 
			
		||||
            )
 | 
			
		||||
        )
 | 
			
		||||
 | 
			
		||||
        # Bind mount the devfs
 | 
			
		||||
        os.system("mount --bind /dev {}/dev".format(temporary_directory))
 | 
			
		||||
 | 
			
		||||
        # Create an fstab entry for each disk
 | 
			
		||||
        fstab_file = "{}/etc/fstab".format(temporary_directory)
 | 
			
		||||
        # The disk ID starts at zero and increments by one for each disk in the fixed-order
 | 
			
		||||
        # disk list. This lets us work around the insanity of Libvirt IDs not matching guest IDs,
 | 
			
		||||
        # while still letting us have some semblance of control here without enforcing things
 | 
			
		||||
        # like labels. It increments in the for loop below at the end of each iteration, and is
 | 
			
		||||
        # used to craft a /dev/disk/by-id/scsi-0QEMU_QEMU_HARDDISK_drive-scsi0-0-0-X device ID
 | 
			
		||||
        # which will always match the correct order from Libvirt (unlike sdX/vdX names).
 | 
			
		||||
        disk_id = 0
 | 
			
		||||
        for disk in disks:
 | 
			
		||||
            # We assume SSD-based/-like storage, and dislike atimes
 | 
			
		||||
            options = "defaults,discard,noatime,nodiratime"
 | 
			
		||||
 | 
			
		||||
            # The root, var, and log volumes have specific values
 | 
			
		||||
            if disk["mountpoint"] == "/":
 | 
			
		||||
                root_disk["scsi_id"] = disk_id
 | 
			
		||||
                dump = 0
 | 
			
		||||
                cpass = 1
 | 
			
		||||
            elif disk["mountpoint"] == "/var" or disk["mountpoint"] == "/var/log":
 | 
			
		||||
                dump = 0
 | 
			
		||||
                cpass = 2
 | 
			
		||||
            else:
 | 
			
		||||
                dump = 0
 | 
			
		||||
                cpass = 0
 | 
			
		||||
 | 
			
		||||
            # Append the fstab line
 | 
			
		||||
            with open(fstab_file, "a") as fh:
 | 
			
		||||
                data = "/dev/disk/by-id/scsi-0QEMU_QEMU_HARDDISK_drive-scsi0-0-0-{disk} {mountpoint} {filesystem} {options} {dump} {cpass}\n".format(
 | 
			
		||||
                    disk=disk_id,
 | 
			
		||||
                    mountpoint=disk["mountpoint"],
 | 
			
		||||
                    filesystem=disk["filesystem"],
 | 
			
		||||
                    options=options,
 | 
			
		||||
                    dump=dump,
 | 
			
		||||
                    cpass=cpass,
 | 
			
		||||
                )
 | 
			
		||||
                fh.write(data)
 | 
			
		||||
 | 
			
		||||
            # Increment the disk_id
 | 
			
		||||
            disk_id += 1
 | 
			
		||||
 | 
			
		||||
        # Write the hostname
 | 
			
		||||
        hostname_file = "{}/etc/hostname".format(temporary_directory)
 | 
			
		||||
        with open(hostname_file, "w") as fh:
 | 
			
		||||
            fh.write("{}".format(vm_name))
 | 
			
		||||
 | 
			
		||||
        # Fix the cloud-init.target since it's broken
 | 
			
		||||
        cloudinit_target_file = "{}/etc/systemd/system/cloud-init.target".format(
 | 
			
		||||
            temporary_directory
 | 
			
		||||
        )
 | 
			
		||||
        with open(cloudinit_target_file, "w") as fh:
 | 
			
		||||
            data = """[Install]
 | 
			
		||||
WantedBy=multi-user.target
 | 
			
		||||
[Unit]
 | 
			
		||||
Description=Cloud-init target
 | 
			
		||||
After=multi-user.target
 | 
			
		||||
"""
 | 
			
		||||
            fh.write(data)
 | 
			
		||||
 | 
			
		||||
        # NOTE: Due to device ordering within the Libvirt XML configuration, the first Ethernet interface
 | 
			
		||||
        #       will always be on PCI bus ID 2, hence the name "ens2".
 | 
			
		||||
        # Write a DHCP stanza for ens2
 | 
			
		||||
        ens2_network_file = "{}/etc/network/interfaces.d/ens2".format(
 | 
			
		||||
            temporary_directory
 | 
			
		||||
        )
 | 
			
		||||
        with open(ens2_network_file, "w") as fh:
 | 
			
		||||
            data = """auto ens2
 | 
			
		||||
iface ens2 inet dhcp
 | 
			
		||||
"""
 | 
			
		||||
            fh.write(data)
 | 
			
		||||
 | 
			
		||||
        # Write the DHCP config for ens2
 | 
			
		||||
        dhclient_file = "{}/etc/dhcp/dhclient.conf".format(temporary_directory)
 | 
			
		||||
        with open(dhclient_file, "w") as fh:
 | 
			
		||||
            data = (
 | 
			
		||||
                """# DHCP client configuration
 | 
			
		||||
# Written by the PVC provisioner
 | 
			
		||||
option rfc3442-classless-static-routes code 121 = array of unsigned integer 8;
 | 
			
		||||
interface "ens2" {
 | 
			
		||||
"""
 | 
			
		||||
                + """        send fqdn.fqdn = "{hostname}";
 | 
			
		||||
        send host-name = "{hostname}";
 | 
			
		||||
    """.format(
 | 
			
		||||
                    hostname=vm_name
 | 
			
		||||
                )
 | 
			
		||||
                + """        request subnet-mask, broadcast-address, time-offset, routers,
 | 
			
		||||
                domain-name, domain-name-servers, domain-search, host-name,
 | 
			
		||||
                dhcp6.name-servers, dhcp6.domain-search, dhcp6.fqdn, dhcp6.sntp-servers,
 | 
			
		||||
                netbios-name-servers, netbios-scope, interface-mtu,
 | 
			
		||||
                rfc3442-classless-static-routes, ntp-servers;
 | 
			
		||||
}
 | 
			
		||||
"""
 | 
			
		||||
            )
 | 
			
		||||
            fh.write(data)
 | 
			
		||||
 | 
			
		||||
        # Write the GRUB configuration
 | 
			
		||||
        grubcfg_file = "{}/etc/default/grub".format(temporary_directory)
 | 
			
		||||
        with open(grubcfg_file, "w") as fh:
 | 
			
		||||
            data = """# Written by the PVC provisioner
 | 
			
		||||
GRUB_DEFAULT=0
 | 
			
		||||
GRUB_TIMEOUT=1
 | 
			
		||||
GRUB_DISTRIBUTOR="PVC Virtual Machine"
 | 
			
		||||
GRUB_CMDLINE_LINUX_DEFAULT="root=/dev/disk/by-id/scsi-0QEMU_QEMU_HARDDISK_drive-scsi0-0-0-{root_disk} console=tty0 console=ttyS0,115200n8"
 | 
			
		||||
GRUB_CMDLINE_LINUX=""
 | 
			
		||||
GRUB_TERMINAL=console
 | 
			
		||||
GRUB_SERIAL_COMMAND="serial --speed=115200 --unit=0 --word=8 --parity=no --stop=1"
 | 
			
		||||
GRUB_DISABLE_LINUX_UUID=false
 | 
			
		||||
""".format(
 | 
			
		||||
                root_disk=root_disk["scsi_id"]
 | 
			
		||||
            )
 | 
			
		||||
            fh.write(data)
 | 
			
		||||
 | 
			
		||||
        # Chroot, do some in-root tasks, then exit the chroot
 | 
			
		||||
        with chroot_target(temporary_directory):
 | 
			
		||||
            # Install and update GRUB
 | 
			
		||||
            os.system(
 | 
			
		||||
                "grub-install --force /dev/rbd/{}/{}_{}".format(
 | 
			
		||||
                    root_disk["pool"], vm_name, root_disk["disk_id"]
 | 
			
		||||
                )
 | 
			
		||||
            )
 | 
			
		||||
            os.system("update-grub")
 | 
			
		||||
            # Set a really dumb root password [TEMPORARY]
 | 
			
		||||
            os.system("echo root:test123 | chpasswd")
 | 
			
		||||
            # Enable cloud-init target on (first) boot
 | 
			
		||||
            # NOTE: Your user-data should handle this and disable it once done, or things get messy.
 | 
			
		||||
            #       That cloud-init won't run without this hack seems like a bug... but even the official
 | 
			
		||||
            #       Debian cloud images are affected, so who knows.
 | 
			
		||||
            os.system("systemctl enable cloud-init.target")
 | 
			
		||||
 | 
			
		||||
        # Unmount the bound devfs
 | 
			
		||||
        os.system("umount {}/dev".format(temporary_directory))
 | 
			
		||||
 | 
			
		||||
    def cleanup(self):
 | 
			
		||||
        """
 | 
			
		||||
        cleanup(): Perform any cleanup required due to prepare()/install()
 | 
			
		||||
 | 
			
		||||
        It is important to now reverse *all* steps taken in those functions that might
 | 
			
		||||
        need cleanup before teardown of the overlay chroot environment.
 | 
			
		||||
        """
 | 
			
		||||
 | 
			
		||||
        temp_dir = "/tmp/target"
 | 
			
		||||
 | 
			
		||||
        for volume in list(reversed(self.vm_data["volumes"])):
 | 
			
		||||
            dst_volume_name = f"{self.vm_name}_{volume['disk_id']}"
 | 
			
		||||
            dst_volume = f"{volume['pool']}/{dst_volume_name}"
 | 
			
		||||
            mapped_dst_volume = f"/dev/rbd/{dst_volume}"
 | 
			
		||||
            mount_path = f"{temp_dir}/{volume['mountpoint']}"
 | 
			
		||||
 | 
			
		||||
            if (
 | 
			
		||||
                volume.get("source_volume") is None
 | 
			
		||||
                and volume.get("filesystem") is not None
 | 
			
		||||
            ):
 | 
			
		||||
                # Unmount filesystem
 | 
			
		||||
                retcode, stdout, stderr = pvc_common.run_os_command(
 | 
			
		||||
                    f"umount {mount_path}"
 | 
			
		||||
                )
 | 
			
		||||
                if retcode:
 | 
			
		||||
                    raise ProvisioningError(
 | 
			
		||||
                        f"Failed to unmount '{mapped_dst_volume}' on '{mount_path}': {stderr}"
 | 
			
		||||
                    )
 | 
			
		||||
 | 
			
		||||
                # Unmap volume
 | 
			
		||||
                with open_zk(config) as zkhandler:
 | 
			
		||||
                    success, message = pvc_ceph.unmap_volume(
 | 
			
		||||
                        zkhandler,
 | 
			
		||||
                        volume["pool"],
 | 
			
		||||
                        dst_volume_name,
 | 
			
		||||
                    )
 | 
			
		||||
                    if not success:
 | 
			
		||||
                        raise ProvisioningError(
 | 
			
		||||
                            f"Failed to unmap '{mapped_dst_volume}': {stderr}"
 | 
			
		||||
                        )
 | 
			
		||||
@@ -1,248 +0,0 @@
 | 
			
		||||
#!/usr/bin/env python3
 | 
			
		||||
 | 
			
		||||
# debootstrap_script.py - PVC Provisioner example script for Debootstrap
 | 
			
		||||
# Part of the Parallel Virtual Cluster (PVC) system
 | 
			
		||||
#
 | 
			
		||||
#    Copyright (C) 2018-2021 Joshua M. Boniface <joshua@boniface.me>
 | 
			
		||||
#
 | 
			
		||||
#    This program is free software: you can redistribute it and/or modify
 | 
			
		||||
#    it under the terms of the GNU General Public License as published by
 | 
			
		||||
#    the Free Software Foundation, version 3.
 | 
			
		||||
#
 | 
			
		||||
#    This program is distributed in the hope that it will be useful,
 | 
			
		||||
#    but WITHOUT ANY WARRANTY; without even the implied warranty of
 | 
			
		||||
#    MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
 | 
			
		||||
#    GNU General Public License for more details.
 | 
			
		||||
#
 | 
			
		||||
#    You should have received a copy of the GNU General Public License
 | 
			
		||||
#    along with this program.  If not, see <https://www.gnu.org/licenses/>.
 | 
			
		||||
#
 | 
			
		||||
###############################################################################
 | 
			
		||||
 | 
			
		||||
# This script provides an example of a PVC provisioner script. It will install
 | 
			
		||||
# a Debian system, of the release specified in the keyword argument `deb_release`
 | 
			
		||||
# and from the mirror specified in the keyword argument `deb_mirror`, and
 | 
			
		||||
# including the packages specified in the keyword argument `deb_packages` (a list
 | 
			
		||||
# of strings, which is then joined together as a CSV and passed to debootstrap),
 | 
			
		||||
# to the configured disks, configure fstab, and install GRUB. Any later config
 | 
			
		||||
# should be done within the VM, for instance via cloud-init.
 | 
			
		||||
 | 
			
		||||
# This script can thus be used as an example or reference implementation of a
 | 
			
		||||
# PVC provisioner script and expanded upon as required.
 | 
			
		||||
 | 
			
		||||
# This script will run under root privileges as the provisioner does. Be careful
 | 
			
		||||
# with that.
 | 
			
		||||
 | 
			
		||||
import os
 | 
			
		||||
from contextlib import contextmanager
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
# Create a chroot context manager
 | 
			
		||||
# This can be used later in the script to chroot to the destination directory
 | 
			
		||||
# for instance to run commands within the target.
 | 
			
		||||
@contextmanager
 | 
			
		||||
def chroot_target(destination):
 | 
			
		||||
    try:
 | 
			
		||||
        real_root = os.open("/", os.O_RDONLY)
 | 
			
		||||
        os.chroot(destination)
 | 
			
		||||
        fake_root = os.open("/", os.O_RDONLY)
 | 
			
		||||
        os.fchdir(fake_root)
 | 
			
		||||
        yield
 | 
			
		||||
    finally:
 | 
			
		||||
        os.fchdir(real_root)
 | 
			
		||||
        os.chroot(".")
 | 
			
		||||
        os.fchdir(real_root)
 | 
			
		||||
        os.close(fake_root)
 | 
			
		||||
        os.close(real_root)
 | 
			
		||||
        del fake_root
 | 
			
		||||
        del real_root
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
# Installation function - performs a debootstrap install of a Debian system
 | 
			
		||||
# Note that the only arguments are keyword arguments.
 | 
			
		||||
def install(**kwargs):
 | 
			
		||||
    # The provisioner has already mounted the disks on kwargs['temporary_directory'].
 | 
			
		||||
    # by this point, so we can get right to running the debootstrap after setting
 | 
			
		||||
    # some nicer variable names; you don't necessarily have to do this.
 | 
			
		||||
    vm_name = kwargs["vm_name"]
 | 
			
		||||
    temporary_directory = kwargs["temporary_directory"]
 | 
			
		||||
    disks = kwargs["disks"]
 | 
			
		||||
    networks = kwargs["networks"]
 | 
			
		||||
    # Our own required arguments. We should, though are not required to, handle
 | 
			
		||||
    # failures of these gracefully, should administrators forget to specify them.
 | 
			
		||||
    try:
 | 
			
		||||
        deb_release = kwargs["deb_release"]
 | 
			
		||||
    except Exception:
 | 
			
		||||
        deb_release = "stable"
 | 
			
		||||
    try:
 | 
			
		||||
        deb_mirror = kwargs["deb_mirror"]
 | 
			
		||||
    except Exception:
 | 
			
		||||
        deb_mirror = "http://ftp.debian.org/debian"
 | 
			
		||||
    try:
 | 
			
		||||
        deb_packages = kwargs["deb_packages"].split(",")
 | 
			
		||||
    except Exception:
 | 
			
		||||
        deb_packages = [
 | 
			
		||||
            "linux-image-amd64",
 | 
			
		||||
            "grub-pc",
 | 
			
		||||
            "cloud-init",
 | 
			
		||||
            "python3-cffi-backend",
 | 
			
		||||
            "wget",
 | 
			
		||||
        ]
 | 
			
		||||
 | 
			
		||||
    # We need to know our root disk
 | 
			
		||||
    root_disk = None
 | 
			
		||||
    for disk in disks:
 | 
			
		||||
        if disk["mountpoint"] == "/":
 | 
			
		||||
            root_disk = disk
 | 
			
		||||
    if not root_disk:
 | 
			
		||||
        return
 | 
			
		||||
 | 
			
		||||
    # Ensure we have debootstrap intalled on the provisioner system; this is a
 | 
			
		||||
    # good idea to include if you plan to use anything that is not part of the
 | 
			
		||||
    # base Debian host system, just in case the provisioner host is not properly
 | 
			
		||||
    # configured already.
 | 
			
		||||
    os.system("apt-get install -y debootstrap")
 | 
			
		||||
 | 
			
		||||
    # Perform a deboostrap installation
 | 
			
		||||
    os.system(
 | 
			
		||||
        "debootstrap --include={pkgs} {suite} {target} {mirror}".format(
 | 
			
		||||
            suite=deb_release,
 | 
			
		||||
            target=temporary_directory,
 | 
			
		||||
            mirror=deb_mirror,
 | 
			
		||||
            pkgs=",".join(deb_packages),
 | 
			
		||||
        )
 | 
			
		||||
    )
 | 
			
		||||
 | 
			
		||||
    # Bind mount the devfs
 | 
			
		||||
    os.system("mount --bind /dev {}/dev".format(temporary_directory))
 | 
			
		||||
 | 
			
		||||
    # Create an fstab entry for each disk
 | 
			
		||||
    fstab_file = "{}/etc/fstab".format(temporary_directory)
 | 
			
		||||
    # The disk ID starts at zero and increments by one for each disk in the fixed-order
 | 
			
		||||
    # disk list. This lets us work around the insanity of Libvirt IDs not matching guest IDs,
 | 
			
		||||
    # while still letting us have some semblance of control here without enforcing things
 | 
			
		||||
    # like labels. It increments in the for loop below at the end of each iteration, and is
 | 
			
		||||
    # used to craft a /dev/disk/by-id/scsi-0QEMU_QEMU_HARDDISK_drive-scsi0-0-0-X device ID
 | 
			
		||||
    # which will always match the correct order from Libvirt (unlike sdX/vdX names).
 | 
			
		||||
    disk_id = 0
 | 
			
		||||
    for disk in disks:
 | 
			
		||||
        # We assume SSD-based/-like storage, and dislike atimes
 | 
			
		||||
        options = "defaults,discard,noatime,nodiratime"
 | 
			
		||||
 | 
			
		||||
        # The root, var, and log volumes have specific values
 | 
			
		||||
        if disk["mountpoint"] == "/":
 | 
			
		||||
            root_disk["scsi_id"] = disk_id
 | 
			
		||||
            dump = 0
 | 
			
		||||
            cpass = 1
 | 
			
		||||
        elif disk["mountpoint"] == "/var" or disk["mountpoint"] == "/var/log":
 | 
			
		||||
            dump = 0
 | 
			
		||||
            cpass = 2
 | 
			
		||||
        else:
 | 
			
		||||
            dump = 0
 | 
			
		||||
            cpass = 0
 | 
			
		||||
 | 
			
		||||
        # Append the fstab line
 | 
			
		||||
        with open(fstab_file, "a") as fh:
 | 
			
		||||
            data = "/dev/disk/by-id/scsi-0QEMU_QEMU_HARDDISK_drive-scsi0-0-0-{disk} {mountpoint} {filesystem} {options} {dump} {cpass}\n".format(
 | 
			
		||||
                disk=disk_id,
 | 
			
		||||
                mountpoint=disk["mountpoint"],
 | 
			
		||||
                filesystem=disk["filesystem"],
 | 
			
		||||
                options=options,
 | 
			
		||||
                dump=dump,
 | 
			
		||||
                cpass=cpass,
 | 
			
		||||
            )
 | 
			
		||||
            fh.write(data)
 | 
			
		||||
 | 
			
		||||
        # Increment the disk_id
 | 
			
		||||
        disk_id += 1
 | 
			
		||||
 | 
			
		||||
    # Write the hostname
 | 
			
		||||
    hostname_file = "{}/etc/hostname".format(temporary_directory)
 | 
			
		||||
    with open(hostname_file, "w") as fh:
 | 
			
		||||
        fh.write("{}".format(vm_name))
 | 
			
		||||
 | 
			
		||||
    # Fix the cloud-init.target since it's broken
 | 
			
		||||
    cloudinit_target_file = "{}/etc/systemd/system/cloud-init.target".format(
 | 
			
		||||
        temporary_directory
 | 
			
		||||
    )
 | 
			
		||||
    with open(cloudinit_target_file, "w") as fh:
 | 
			
		||||
        data = """[Install]
 | 
			
		||||
WantedBy=multi-user.target
 | 
			
		||||
[Unit]
 | 
			
		||||
Description=Cloud-init target
 | 
			
		||||
After=multi-user.target
 | 
			
		||||
"""
 | 
			
		||||
        fh.write(data)
 | 
			
		||||
 | 
			
		||||
    # NOTE: Due to device ordering within the Libvirt XML configuration, the first Ethernet interface
 | 
			
		||||
    #       will always be on PCI bus ID 2, hence the name "ens2".
 | 
			
		||||
    # Write a DHCP stanza for ens2
 | 
			
		||||
    ens2_network_file = "{}/etc/network/interfaces.d/ens2".format(temporary_directory)
 | 
			
		||||
    with open(ens2_network_file, "w") as fh:
 | 
			
		||||
        data = """auto ens2
 | 
			
		||||
iface ens2 inet dhcp
 | 
			
		||||
"""
 | 
			
		||||
        fh.write(data)
 | 
			
		||||
 | 
			
		||||
    # Write the DHCP config for ens2
 | 
			
		||||
    dhclient_file = "{}/etc/dhcp/dhclient.conf".format(temporary_directory)
 | 
			
		||||
    with open(dhclient_file, "w") as fh:
 | 
			
		||||
        data = (
 | 
			
		||||
            """# DHCP client configuration
 | 
			
		||||
# Written by the PVC provisioner
 | 
			
		||||
option rfc3442-classless-static-routes code 121 = array of unsigned integer 8;
 | 
			
		||||
interface "ens2" {
 | 
			
		||||
"""
 | 
			
		||||
            + """        send fqdn.fqdn = "{hostname}";
 | 
			
		||||
        send host-name = "{hostname}";
 | 
			
		||||
""".format(
 | 
			
		||||
                hostname=vm_name
 | 
			
		||||
            )
 | 
			
		||||
            + """        request subnet-mask, broadcast-address, time-offset, routers,
 | 
			
		||||
                domain-name, domain-name-servers, domain-search, host-name,
 | 
			
		||||
                dhcp6.name-servers, dhcp6.domain-search, dhcp6.fqdn, dhcp6.sntp-servers,
 | 
			
		||||
                netbios-name-servers, netbios-scope, interface-mtu,
 | 
			
		||||
                rfc3442-classless-static-routes, ntp-servers;
 | 
			
		||||
}
 | 
			
		||||
"""
 | 
			
		||||
        )
 | 
			
		||||
        fh.write(data)
 | 
			
		||||
 | 
			
		||||
    # Write the GRUB configuration
 | 
			
		||||
    grubcfg_file = "{}/etc/default/grub".format(temporary_directory)
 | 
			
		||||
    with open(grubcfg_file, "w") as fh:
 | 
			
		||||
        data = """# Written by the PVC provisioner
 | 
			
		||||
GRUB_DEFAULT=0
 | 
			
		||||
GRUB_TIMEOUT=1
 | 
			
		||||
GRUB_DISTRIBUTOR="PVC Virtual Machine"
 | 
			
		||||
GRUB_CMDLINE_LINUX_DEFAULT="root=/dev/disk/by-id/scsi-0QEMU_QEMU_HARDDISK_drive-scsi0-0-0-{root_disk} console=tty0 console=ttyS0,115200n8"
 | 
			
		||||
GRUB_CMDLINE_LINUX=""
 | 
			
		||||
GRUB_TERMINAL=console
 | 
			
		||||
GRUB_SERIAL_COMMAND="serial --speed=115200 --unit=0 --word=8 --parity=no --stop=1"
 | 
			
		||||
GRUB_DISABLE_LINUX_UUID=false
 | 
			
		||||
""".format(
 | 
			
		||||
            root_disk=root_disk["scsi_id"]
 | 
			
		||||
        )
 | 
			
		||||
        fh.write(data)
 | 
			
		||||
 | 
			
		||||
    # Chroot, do some in-root tasks, then exit the chroot
 | 
			
		||||
    with chroot_target(temporary_directory):
 | 
			
		||||
        # Install and update GRUB
 | 
			
		||||
        os.system(
 | 
			
		||||
            "grub-install --force /dev/rbd/{}/{}_{}".format(
 | 
			
		||||
                root_disk["pool"], vm_name, root_disk["disk_id"]
 | 
			
		||||
            )
 | 
			
		||||
        )
 | 
			
		||||
        os.system("update-grub")
 | 
			
		||||
        # Set a really dumb root password [TEMPORARY]
 | 
			
		||||
        os.system("echo root:test123 | chpasswd")
 | 
			
		||||
        # Enable cloud-init target on (first) boot
 | 
			
		||||
        # NOTE: Your user-data should handle this and disable it once done, or things get messy.
 | 
			
		||||
        #       That cloud-init won't run without this hack seems like a bug... but even the official
 | 
			
		||||
        #       Debian cloud images are affected, so who knows.
 | 
			
		||||
        os.system("systemctl enable cloud-init.target")
 | 
			
		||||
 | 
			
		||||
    # Unmount the bound devfs
 | 
			
		||||
    os.system("umount {}/dev".format(temporary_directory))
 | 
			
		||||
 | 
			
		||||
    # Everything else is done via cloud-init user-data
 | 
			
		||||
@@ -1,43 +0,0 @@
 | 
			
		||||
#!/usr/bin/env python3
 | 
			
		||||
 | 
			
		||||
# dummy_script.py - PVC Provisioner example script for noop
 | 
			
		||||
# Part of the Parallel Virtual Cluster (PVC) system
 | 
			
		||||
#
 | 
			
		||||
#    Copyright (C) 2018-2021 Joshua M. Boniface <joshua@boniface.me>
 | 
			
		||||
#
 | 
			
		||||
#    This program is free software: you can redistribute it and/or modify
 | 
			
		||||
#    it under the terms of the GNU General Public License as published by
 | 
			
		||||
#    the Free Software Foundation, version 3.
 | 
			
		||||
#
 | 
			
		||||
#    This program is distributed in the hope that it will be useful,
 | 
			
		||||
#    but WITHOUT ANY WARRANTY; without even the implied warranty of
 | 
			
		||||
#    MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
 | 
			
		||||
#    GNU General Public License for more details.
 | 
			
		||||
#
 | 
			
		||||
#    You should have received a copy of the GNU General Public License
 | 
			
		||||
#    along with this program.  If not, see <https://www.gnu.org/licenses/>.
 | 
			
		||||
#
 | 
			
		||||
###############################################################################
 | 
			
		||||
 | 
			
		||||
# This script provides an example of a PVC provisioner script. It will do
 | 
			
		||||
# nothing and return back to the provisioner without taking any action, and
 | 
			
		||||
# expecting no special arguments.
 | 
			
		||||
 | 
			
		||||
# This script can thus be used as an example or reference implementation of a
 | 
			
		||||
# PVC provisioner script and expanded upon as required.
 | 
			
		||||
 | 
			
		||||
# This script will run under root privileges as the provisioner does. Be careful
 | 
			
		||||
# with that.
 | 
			
		||||
 | 
			
		||||
# Installation function - performs no actions then returns
 | 
			
		||||
# Note that the only arguments are keyword arguments.
 | 
			
		||||
def install(**kwargs):
 | 
			
		||||
    # The provisioner has already mounted the disks on kwargs['temporary_directory'].
 | 
			
		||||
    # by this point, so we can get right to running the debootstrap after setting
 | 
			
		||||
    # some nicer variable names; you don't necessarily have to do this.
 | 
			
		||||
    vm_name = kwargs["vm_name"]
 | 
			
		||||
    temporary_directory = kwargs["temporary_directory"]
 | 
			
		||||
    disks = kwargs["disks"]
 | 
			
		||||
    networks = kwargs["networks"]
 | 
			
		||||
    # No operation - this script just returns
 | 
			
		||||
    pass
 | 
			
		||||
@@ -29,6 +29,7 @@ from pvcapid.Daemon import config, strtobool, API_VERSION
 | 
			
		||||
 | 
			
		||||
import pvcapid.helper as api_helper
 | 
			
		||||
import pvcapid.provisioner as api_provisioner
 | 
			
		||||
import pvcapid.vmbuilder as api_vmbuilder
 | 
			
		||||
import pvcapid.benchmark as api_benchmark
 | 
			
		||||
import pvcapid.ova as api_ova
 | 
			
		||||
 | 
			
		||||
@@ -144,7 +145,7 @@ def Authenticator(function):
 | 
			
		||||
def create_vm(
 | 
			
		||||
    self, vm_name, profile_name, define_vm=True, start_vm=True, script_run_args=[]
 | 
			
		||||
):
 | 
			
		||||
    return api_provisioner.create_vm(
 | 
			
		||||
    return api_vmbuilder.create_vm(
 | 
			
		||||
        self,
 | 
			
		||||
        vm_name,
 | 
			
		||||
        profile_name,
 | 
			
		||||
 
 | 
			
		||||
@@ -19,23 +19,12 @@
 | 
			
		||||
#
 | 
			
		||||
###############################################################################
 | 
			
		||||
 | 
			
		||||
import json
 | 
			
		||||
import psycopg2
 | 
			
		||||
import psycopg2.extras
 | 
			
		||||
import re
 | 
			
		||||
 | 
			
		||||
from pvcapid.Daemon import config, strtobool
 | 
			
		||||
 | 
			
		||||
from daemon_lib.zkhandler import ZKHandler
 | 
			
		||||
 | 
			
		||||
import daemon_lib.common as pvc_common
 | 
			
		||||
import daemon_lib.node as pvc_node
 | 
			
		||||
import daemon_lib.vm as pvc_vm
 | 
			
		||||
import daemon_lib.network as pvc_network
 | 
			
		||||
import daemon_lib.ceph as pvc_ceph
 | 
			
		||||
 | 
			
		||||
import pvcapid.libvirt_schema as libvirt_schema
 | 
			
		||||
 | 
			
		||||
from pvcapid.ova import list_ova
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
@@ -1229,866 +1218,3 @@ def delete_profile(name):
 | 
			
		||||
        retcode = 400
 | 
			
		||||
    close_database(conn, cur)
 | 
			
		||||
    return retmsg, retcode
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
#
 | 
			
		||||
# Main VM provisioning function - executed by the Celery worker
 | 
			
		||||
#
 | 
			
		||||
def create_vm(
 | 
			
		||||
    self, vm_name, vm_profile, define_vm=True, start_vm=True, script_run_args=[]
 | 
			
		||||
):
 | 
			
		||||
    # Runtime imports
 | 
			
		||||
    import time
 | 
			
		||||
    import importlib
 | 
			
		||||
    import uuid
 | 
			
		||||
    import datetime
 | 
			
		||||
    import random
 | 
			
		||||
 | 
			
		||||
    temp_dir = None
 | 
			
		||||
 | 
			
		||||
    time.sleep(2)
 | 
			
		||||
 | 
			
		||||
    print(
 | 
			
		||||
        "Starting provisioning of VM '{}' with profile '{}'".format(vm_name, vm_profile)
 | 
			
		||||
    )
 | 
			
		||||
 | 
			
		||||
    # Phase 0 - connect to databases
 | 
			
		||||
    try:
 | 
			
		||||
        db_conn, db_cur = open_database(config)
 | 
			
		||||
    except Exception:
 | 
			
		||||
        raise ClusterError("Failed to connect to Postgres")
 | 
			
		||||
 | 
			
		||||
    try:
 | 
			
		||||
        zkhandler = ZKHandler(config)
 | 
			
		||||
        zkhandler.connect()
 | 
			
		||||
    except Exception:
 | 
			
		||||
        raise ClusterError("Failed to connect to Zookeeper")
 | 
			
		||||
 | 
			
		||||
    # Phase 1 - setup
 | 
			
		||||
    #  * Get the profile elements
 | 
			
		||||
    #  * Get the details from these elements
 | 
			
		||||
    #  * Assemble a VM configuration dictionary
 | 
			
		||||
    self.update_state(
 | 
			
		||||
        state="RUNNING",
 | 
			
		||||
        meta={"current": 1, "total": 10, "status": "Collecting configuration"},
 | 
			
		||||
    )
 | 
			
		||||
    time.sleep(1)
 | 
			
		||||
 | 
			
		||||
    vm_id = re.findall(r"/(\d+)$/", vm_name)
 | 
			
		||||
    if not vm_id:
 | 
			
		||||
        vm_id = 0
 | 
			
		||||
    else:
 | 
			
		||||
        vm_id = vm_id[0]
 | 
			
		||||
 | 
			
		||||
    vm_data = dict()
 | 
			
		||||
 | 
			
		||||
    # Get the profile information
 | 
			
		||||
    query = "SELECT * FROM profile WHERE name = %s"
 | 
			
		||||
    args = (vm_profile,)
 | 
			
		||||
    db_cur.execute(query, args)
 | 
			
		||||
    profile_data = db_cur.fetchone()
 | 
			
		||||
    if profile_data.get("arguments"):
 | 
			
		||||
        vm_data["script_arguments"] = profile_data.get("arguments").split("|")
 | 
			
		||||
    else:
 | 
			
		||||
        vm_data["script_arguments"] = []
 | 
			
		||||
 | 
			
		||||
    if profile_data.get("profile_type") == "ova":
 | 
			
		||||
        is_ova_install = True
 | 
			
		||||
        is_script_install = False  # By definition
 | 
			
		||||
    else:
 | 
			
		||||
        is_ova_install = False
 | 
			
		||||
 | 
			
		||||
    # Get the system details
 | 
			
		||||
    query = "SELECT * FROM system_template WHERE id = %s"
 | 
			
		||||
    args = (profile_data["system_template"],)
 | 
			
		||||
    db_cur.execute(query, args)
 | 
			
		||||
    vm_data["system_details"] = db_cur.fetchone()
 | 
			
		||||
 | 
			
		||||
    # Get the MAC template
 | 
			
		||||
    query = "SELECT mac_template FROM network_template WHERE id = %s"
 | 
			
		||||
    args = (profile_data["network_template"],)
 | 
			
		||||
    db_cur.execute(query, args)
 | 
			
		||||
    db_row = db_cur.fetchone()
 | 
			
		||||
    if db_row:
 | 
			
		||||
        vm_data["mac_template"] = db_row.get("mac_template")
 | 
			
		||||
    else:
 | 
			
		||||
        vm_data["mac_template"] = None
 | 
			
		||||
 | 
			
		||||
    # Get the networks
 | 
			
		||||
    query = "SELECT * FROM network WHERE network_template = %s"
 | 
			
		||||
    args = (profile_data["network_template"],)
 | 
			
		||||
    db_cur.execute(query, args)
 | 
			
		||||
    vm_data["networks"] = db_cur.fetchall()
 | 
			
		||||
 | 
			
		||||
    # Get the storage volumes
 | 
			
		||||
    # ORDER BY ensures disks are always in the sdX/vdX order, regardless of add order
 | 
			
		||||
    query = "SELECT * FROM storage WHERE storage_template = %s ORDER BY disk_id"
 | 
			
		||||
    args = (profile_data["storage_template"],)
 | 
			
		||||
    db_cur.execute(query, args)
 | 
			
		||||
    vm_data["volumes"] = db_cur.fetchall()
 | 
			
		||||
 | 
			
		||||
    # Get the script
 | 
			
		||||
    query = "SELECT script FROM script WHERE id = %s"
 | 
			
		||||
    args = (profile_data["script"],)
 | 
			
		||||
    db_cur.execute(query, args)
 | 
			
		||||
    db_row = db_cur.fetchone()
 | 
			
		||||
    if db_row:
 | 
			
		||||
        vm_data["script"] = db_row.get("script")
 | 
			
		||||
    else:
 | 
			
		||||
        vm_data["script"] = None
 | 
			
		||||
 | 
			
		||||
    if vm_data["script"] and not is_ova_install:
 | 
			
		||||
        is_script_install = True
 | 
			
		||||
    else:
 | 
			
		||||
        is_script_install = False
 | 
			
		||||
 | 
			
		||||
    # Get the OVA details
 | 
			
		||||
    if is_ova_install:
 | 
			
		||||
        query = "SELECT * FROM ova WHERE id = %s"
 | 
			
		||||
        args = (profile_data["ova"],)
 | 
			
		||||
        db_cur.execute(query, args)
 | 
			
		||||
        vm_data["ova_details"] = db_cur.fetchone()
 | 
			
		||||
 | 
			
		||||
        query = "SELECT * FROM ova_volume WHERE ova = %s"
 | 
			
		||||
        args = (profile_data["ova"],)
 | 
			
		||||
        db_cur.execute(query, args)
 | 
			
		||||
        vm_data["volumes"] = db_cur.fetchall()
 | 
			
		||||
 | 
			
		||||
    close_database(db_conn, db_cur)
 | 
			
		||||
 | 
			
		||||
    print(
 | 
			
		||||
        "VM configuration data:\n{}".format(
 | 
			
		||||
            json.dumps(vm_data, sort_keys=True, indent=2)
 | 
			
		||||
        )
 | 
			
		||||
    )
 | 
			
		||||
 | 
			
		||||
    # Phase 2 - verification
 | 
			
		||||
    #  * Ensure that at least one node has enough free RAM to hold the VM (becomes main host)
 | 
			
		||||
    #  * Ensure that all networks are valid
 | 
			
		||||
    #  * Ensure that there is enough disk space in the Ceph cluster for the disks
 | 
			
		||||
    # This is the "safe fail" step when an invalid configuration will be caught
 | 
			
		||||
    self.update_state(
 | 
			
		||||
        state="RUNNING",
 | 
			
		||||
        meta={
 | 
			
		||||
            "current": 2,
 | 
			
		||||
            "total": 10,
 | 
			
		||||
            "status": "Verifying configuration against cluster",
 | 
			
		||||
        },
 | 
			
		||||
    )
 | 
			
		||||
    time.sleep(1)
 | 
			
		||||
 | 
			
		||||
    # Verify that a VM with this name does not already exist
 | 
			
		||||
    if pvc_vm.searchClusterByName(zkhandler, vm_name):
 | 
			
		||||
        raise ClusterError(
 | 
			
		||||
            "A VM with the name '{}' already exists in the cluster.".format(vm_name)
 | 
			
		||||
        )
 | 
			
		||||
 | 
			
		||||
    # Verify that at least one host has enough free RAM to run the VM
 | 
			
		||||
    _discard, nodes = pvc_node.get_list(zkhandler, None)
 | 
			
		||||
    target_node = None
 | 
			
		||||
    last_free = 0
 | 
			
		||||
    for node in nodes:
 | 
			
		||||
        # Skip the node if it is not ready to run VMs
 | 
			
		||||
        if node["daemon_state"] != "run" or node["domain_state"] != "ready":
 | 
			
		||||
            continue
 | 
			
		||||
        # Skip the node if its free memory is less than the new VM's size, plus a 512MB buffer
 | 
			
		||||
        if node["memory"]["free"] < (vm_data["system_details"]["vram_mb"] + 512):
 | 
			
		||||
            continue
 | 
			
		||||
        # If this node has the most free, use it
 | 
			
		||||
        if node["memory"]["free"] > last_free:
 | 
			
		||||
            last_free = node["memory"]["free"]
 | 
			
		||||
            target_node = node["name"]
 | 
			
		||||
    # Raise if no node was found
 | 
			
		||||
    if not target_node:
 | 
			
		||||
        raise ClusterError(
 | 
			
		||||
            "No ready cluster node contains at least {}+512 MB of free RAM.".format(
 | 
			
		||||
                vm_data["system_details"]["vram_mb"]
 | 
			
		||||
            )
 | 
			
		||||
        )
 | 
			
		||||
 | 
			
		||||
    print(
 | 
			
		||||
        'Selecting target node "{}" with "{}" MB free RAM'.format(
 | 
			
		||||
            target_node, last_free
 | 
			
		||||
        )
 | 
			
		||||
    )
 | 
			
		||||
 | 
			
		||||
    # Verify that all configured networks are present on the cluster
 | 
			
		||||
    cluster_networks, _discard = pvc_network.getClusterNetworkList(zkhandler)
 | 
			
		||||
    for network in vm_data["networks"]:
 | 
			
		||||
        vni = str(network["vni"])
 | 
			
		||||
        if vni not in cluster_networks and vni not in [
 | 
			
		||||
            "upstream",
 | 
			
		||||
            "cluster",
 | 
			
		||||
            "storage",
 | 
			
		||||
        ]:
 | 
			
		||||
            raise ClusterError(
 | 
			
		||||
                'The network VNI "{}" is not present on the cluster.'.format(vni)
 | 
			
		||||
            )
 | 
			
		||||
 | 
			
		||||
    print("All configured networks for VM are valid")
 | 
			
		||||
 | 
			
		||||
    # Verify that there is enough disk space free to provision all VM disks
 | 
			
		||||
    pools = dict()
 | 
			
		||||
    for volume in vm_data["volumes"]:
 | 
			
		||||
        if volume.get("source_volume") is not None:
 | 
			
		||||
            volume_data = pvc_ceph.getVolumeInformation(
 | 
			
		||||
                zkhandler, volume["pool"], volume["source_volume"]
 | 
			
		||||
            )
 | 
			
		||||
            if not volume_data:
 | 
			
		||||
                raise ClusterError(
 | 
			
		||||
                    "The source volume {}/{} could not be found.".format(
 | 
			
		||||
                        volume["pool"], volume["source_volume"]
 | 
			
		||||
                    )
 | 
			
		||||
                )
 | 
			
		||||
            if not volume["pool"] in pools:
 | 
			
		||||
                pools[volume["pool"]] = int(
 | 
			
		||||
                    pvc_ceph.format_bytes_fromhuman(volume_data["stats"]["size"])
 | 
			
		||||
                    / 1024
 | 
			
		||||
                    / 1024
 | 
			
		||||
                    / 1024
 | 
			
		||||
                )
 | 
			
		||||
            else:
 | 
			
		||||
                pools[volume["pool"]] += int(
 | 
			
		||||
                    pvc_ceph.format_bytes_fromhuman(volume_data["stats"]["size"])
 | 
			
		||||
                    / 1024
 | 
			
		||||
                    / 1024
 | 
			
		||||
                    / 1024
 | 
			
		||||
                )
 | 
			
		||||
        else:
 | 
			
		||||
            if not volume["pool"] in pools:
 | 
			
		||||
                pools[volume["pool"]] = volume["disk_size_gb"]
 | 
			
		||||
            else:
 | 
			
		||||
                pools[volume["pool"]] += volume["disk_size_gb"]
 | 
			
		||||
 | 
			
		||||
    for pool in pools:
 | 
			
		||||
        try:
 | 
			
		||||
            pool_information = pvc_ceph.getPoolInformation(zkhandler, pool)
 | 
			
		||||
            if not pool_information:
 | 
			
		||||
                raise
 | 
			
		||||
        except Exception:
 | 
			
		||||
            raise ClusterError('Pool "{}" is not present on the cluster.'.format(pool))
 | 
			
		||||
        pool_free_space_gb = int(
 | 
			
		||||
            pool_information["stats"]["free_bytes"] / 1024 / 1024 / 1024
 | 
			
		||||
        )
 | 
			
		||||
        pool_vm_usage_gb = int(pools[pool])
 | 
			
		||||
 | 
			
		||||
        if pool_vm_usage_gb >= pool_free_space_gb:
 | 
			
		||||
            raise ClusterError(
 | 
			
		||||
                'Pool "{}" has only {} GB free and VM requires {} GB.'.format(
 | 
			
		||||
                    pool, pool_free_space_gb, pool_vm_usage_gb
 | 
			
		||||
                )
 | 
			
		||||
            )
 | 
			
		||||
 | 
			
		||||
    print("There is enough space on cluster to store VM volumes")
 | 
			
		||||
 | 
			
		||||
    if not is_ova_install:
 | 
			
		||||
        # Verify that every specified filesystem is valid
 | 
			
		||||
        used_filesystems = list()
 | 
			
		||||
        for volume in vm_data["volumes"]:
 | 
			
		||||
            if volume["source_volume"] is not None:
 | 
			
		||||
                continue
 | 
			
		||||
            if volume["filesystem"] and volume["filesystem"] not in used_filesystems:
 | 
			
		||||
                used_filesystems.append(volume["filesystem"])
 | 
			
		||||
 | 
			
		||||
        for filesystem in used_filesystems:
 | 
			
		||||
            if filesystem == "swap":
 | 
			
		||||
                retcode, stdout, stderr = pvc_common.run_os_command("which mkswap")
 | 
			
		||||
                if retcode:
 | 
			
		||||
                    raise ProvisioningError(
 | 
			
		||||
                        "Failed to find binary for mkswap: {}".format(stderr)
 | 
			
		||||
                    )
 | 
			
		||||
            else:
 | 
			
		||||
                retcode, stdout, stderr = pvc_common.run_os_command(
 | 
			
		||||
                    "which mkfs.{}".format(filesystem)
 | 
			
		||||
                )
 | 
			
		||||
                if retcode:
 | 
			
		||||
                    raise ProvisioningError(
 | 
			
		||||
                        "Failed to find binary for mkfs.{}: {}".format(
 | 
			
		||||
                            filesystem, stderr
 | 
			
		||||
                        )
 | 
			
		||||
                    )
 | 
			
		||||
 | 
			
		||||
        print("All selected filesystems are valid")
 | 
			
		||||
 | 
			
		||||
    # Phase 3 - provisioning script preparation
 | 
			
		||||
    #  * Import the provisioning script as a library with importlib
 | 
			
		||||
    #  * Ensure the required function(s) are present
 | 
			
		||||
    self.update_state(
 | 
			
		||||
        state="RUNNING",
 | 
			
		||||
        meta={"current": 3, "total": 10, "status": "Preparing provisioning script"},
 | 
			
		||||
    )
 | 
			
		||||
    time.sleep(1)
 | 
			
		||||
 | 
			
		||||
    if is_script_install:
 | 
			
		||||
        # Write the script out to a temporary file
 | 
			
		||||
        retcode, stdout, stderr = pvc_common.run_os_command("mktemp")
 | 
			
		||||
        if retcode:
 | 
			
		||||
            raise ProvisioningError(
 | 
			
		||||
                "Failed to create a temporary file: {}".format(stderr)
 | 
			
		||||
            )
 | 
			
		||||
        script_file = stdout.strip()
 | 
			
		||||
        with open(script_file, "w") as fh:
 | 
			
		||||
            fh.write(vm_data["script"])
 | 
			
		||||
            fh.write("\n")
 | 
			
		||||
 | 
			
		||||
        # Import the script file
 | 
			
		||||
        loader = importlib.machinery.SourceFileLoader("installer_script", script_file)
 | 
			
		||||
        spec = importlib.util.spec_from_loader(loader.name, loader)
 | 
			
		||||
        installer_script = importlib.util.module_from_spec(spec)
 | 
			
		||||
        loader.exec_module(installer_script)
 | 
			
		||||
 | 
			
		||||
        # Verify that the install() function is valid
 | 
			
		||||
        if "install" not in dir(installer_script):
 | 
			
		||||
            raise ProvisioningError(
 | 
			
		||||
                "Specified script does not contain an install() function."
 | 
			
		||||
            )
 | 
			
		||||
 | 
			
		||||
        print("Provisioning script imported successfully")
 | 
			
		||||
 | 
			
		||||
    # Phase 4 - configuration creation
 | 
			
		||||
    #  * Create the libvirt XML configuration
 | 
			
		||||
    self.update_state(
 | 
			
		||||
        state="RUNNING",
 | 
			
		||||
        meta={
 | 
			
		||||
            "current": 4,
 | 
			
		||||
            "total": 10,
 | 
			
		||||
            "status": "Preparing Libvirt XML configuration",
 | 
			
		||||
        },
 | 
			
		||||
    )
 | 
			
		||||
    time.sleep(1)
 | 
			
		||||
 | 
			
		||||
    print("Creating Libvirt configuration")
 | 
			
		||||
 | 
			
		||||
    # Get information about VM
 | 
			
		||||
    vm_uuid = uuid.uuid4()
 | 
			
		||||
    vm_description = "PVC provisioner @ {}, profile '{}'".format(
 | 
			
		||||
        datetime.datetime.now(), vm_profile
 | 
			
		||||
    )
 | 
			
		||||
 | 
			
		||||
    retcode, stdout, stderr = pvc_common.run_os_command("uname -m")
 | 
			
		||||
    system_architecture = stdout.strip()
 | 
			
		||||
 | 
			
		||||
    # Begin assembling libvirt schema
 | 
			
		||||
    vm_schema = ""
 | 
			
		||||
 | 
			
		||||
    vm_schema += libvirt_schema.libvirt_header.format(
 | 
			
		||||
        vm_name=vm_name,
 | 
			
		||||
        vm_uuid=vm_uuid,
 | 
			
		||||
        vm_description=vm_description,
 | 
			
		||||
        vm_memory=vm_data["system_details"]["vram_mb"],
 | 
			
		||||
        vm_vcpus=vm_data["system_details"]["vcpu_count"],
 | 
			
		||||
        vm_architecture=system_architecture,
 | 
			
		||||
    )
 | 
			
		||||
 | 
			
		||||
    # Add disk devices
 | 
			
		||||
    monitor_list = list()
 | 
			
		||||
    coordinator_names = config["storage_hosts"]
 | 
			
		||||
    for coordinator in coordinator_names:
 | 
			
		||||
        monitor_list.append("{}.{}".format(coordinator, config["storage_domain"]))
 | 
			
		||||
 | 
			
		||||
    ceph_storage_secret = config["ceph_storage_secret_uuid"]
 | 
			
		||||
 | 
			
		||||
    for volume in vm_data["volumes"]:
 | 
			
		||||
        vm_schema += libvirt_schema.devices_disk_header.format(
 | 
			
		||||
            ceph_storage_secret=ceph_storage_secret,
 | 
			
		||||
            disk_pool=volume["pool"],
 | 
			
		||||
            vm_name=vm_name,
 | 
			
		||||
            disk_id=volume["disk_id"],
 | 
			
		||||
        )
 | 
			
		||||
        for monitor in monitor_list:
 | 
			
		||||
            vm_schema += libvirt_schema.devices_disk_coordinator.format(
 | 
			
		||||
                coordinator_name=monitor,
 | 
			
		||||
                coordinator_ceph_mon_port=config["ceph_monitor_port"],
 | 
			
		||||
            )
 | 
			
		||||
        vm_schema += libvirt_schema.devices_disk_footer
 | 
			
		||||
 | 
			
		||||
    vm_schema += libvirt_schema.devices_vhostmd
 | 
			
		||||
 | 
			
		||||
    # Add network devices
 | 
			
		||||
    network_id = 0
 | 
			
		||||
    for network in vm_data["networks"]:
 | 
			
		||||
        vni = network["vni"]
 | 
			
		||||
        if vni in ["upstream", "cluster", "storage"]:
 | 
			
		||||
            eth_bridge = "br{}".format(vni)
 | 
			
		||||
        else:
 | 
			
		||||
            eth_bridge = "vmbr{}".format(vni)
 | 
			
		||||
 | 
			
		||||
        vm_id_hex = "{:x}".format(int(vm_id % 16))
 | 
			
		||||
        net_id_hex = "{:x}".format(int(network_id % 16))
 | 
			
		||||
 | 
			
		||||
        if vm_data.get("mac_template") is not None:
 | 
			
		||||
            mac_prefix = "52:54:01"
 | 
			
		||||
            macgen_template = vm_data["mac_template"]
 | 
			
		||||
            eth_macaddr = macgen_template.format(
 | 
			
		||||
                prefix=mac_prefix,
 | 
			
		||||
                vmid=vm_id_hex,
 | 
			
		||||
                netid=net_id_hex,
 | 
			
		||||
            )
 | 
			
		||||
        else:
 | 
			
		||||
            mac_prefix = "52:54:00"
 | 
			
		||||
            random_octet_A = "{:x}".format(random.randint(16, 238))
 | 
			
		||||
            random_octet_B = "{:x}".format(random.randint(16, 238))
 | 
			
		||||
            random_octet_C = "{:x}".format(random.randint(16, 238))
 | 
			
		||||
 | 
			
		||||
            macgen_template = "{prefix}:{octetA}:{octetB}:{octetC}"
 | 
			
		||||
            eth_macaddr = macgen_template.format(
 | 
			
		||||
                prefix=mac_prefix,
 | 
			
		||||
                octetA=random_octet_A,
 | 
			
		||||
                octetB=random_octet_B,
 | 
			
		||||
                octetC=random_octet_C,
 | 
			
		||||
            )
 | 
			
		||||
 | 
			
		||||
        vm_schema += libvirt_schema.devices_net_interface.format(
 | 
			
		||||
            eth_macaddr=eth_macaddr, eth_bridge=eth_bridge
 | 
			
		||||
        )
 | 
			
		||||
 | 
			
		||||
        network_id += 1
 | 
			
		||||
 | 
			
		||||
    # Add default devices
 | 
			
		||||
    vm_schema += libvirt_schema.devices_default
 | 
			
		||||
 | 
			
		||||
    # Add serial device
 | 
			
		||||
    if vm_data["system_details"]["serial"]:
 | 
			
		||||
        vm_schema += libvirt_schema.devices_serial.format(vm_name=vm_name)
 | 
			
		||||
 | 
			
		||||
    # Add VNC device
 | 
			
		||||
    if vm_data["system_details"]["vnc"]:
 | 
			
		||||
        if vm_data["system_details"]["vnc_bind"]:
 | 
			
		||||
            vm_vnc_bind = vm_data["system_details"]["vnc_bind"]
 | 
			
		||||
        else:
 | 
			
		||||
            vm_vnc_bind = "127.0.0.1"
 | 
			
		||||
 | 
			
		||||
        vm_vncport = 5900
 | 
			
		||||
        vm_vnc_autoport = "yes"
 | 
			
		||||
 | 
			
		||||
        vm_schema += libvirt_schema.devices_vnc.format(
 | 
			
		||||
            vm_vncport=vm_vncport,
 | 
			
		||||
            vm_vnc_autoport=vm_vnc_autoport,
 | 
			
		||||
            vm_vnc_bind=vm_vnc_bind,
 | 
			
		||||
        )
 | 
			
		||||
 | 
			
		||||
    # Add SCSI controller
 | 
			
		||||
    vm_schema += libvirt_schema.devices_scsi_controller
 | 
			
		||||
 | 
			
		||||
    # Add footer
 | 
			
		||||
    vm_schema += libvirt_schema.libvirt_footer
 | 
			
		||||
 | 
			
		||||
    print("Final VM schema:\n{}\n".format(vm_schema))
 | 
			
		||||
 | 
			
		||||
    # All the following steps may require cleanup later on, so catch them here and do cleanup in a Finally block
 | 
			
		||||
    try:
 | 
			
		||||
        # Phase 5 - definition
 | 
			
		||||
        #  * Create the VM in the PVC cluster
 | 
			
		||||
        self.update_state(
 | 
			
		||||
            state="RUNNING",
 | 
			
		||||
            meta={"current": 5, "total": 10, "status": "Defining VM on the cluster"},
 | 
			
		||||
        )
 | 
			
		||||
        time.sleep(1)
 | 
			
		||||
 | 
			
		||||
        if define_vm:
 | 
			
		||||
            print("Defining VM on cluster")
 | 
			
		||||
            node_limit = vm_data["system_details"]["node_limit"]
 | 
			
		||||
            if node_limit:
 | 
			
		||||
                node_limit = node_limit.split(",")
 | 
			
		||||
            node_selector = vm_data["system_details"]["node_selector"]
 | 
			
		||||
            node_autostart = vm_data["system_details"]["node_autostart"]
 | 
			
		||||
            migration_method = vm_data["system_details"]["migration_method"]
 | 
			
		||||
            retcode, retmsg = pvc_vm.define_vm(
 | 
			
		||||
                zkhandler,
 | 
			
		||||
                vm_schema.strip(),
 | 
			
		||||
                target_node,
 | 
			
		||||
                node_limit,
 | 
			
		||||
                node_selector,
 | 
			
		||||
                node_autostart,
 | 
			
		||||
                migration_method,
 | 
			
		||||
                vm_profile,
 | 
			
		||||
                initial_state="provision",
 | 
			
		||||
            )
 | 
			
		||||
            print(retmsg)
 | 
			
		||||
        else:
 | 
			
		||||
            print("Skipping VM definition")
 | 
			
		||||
 | 
			
		||||
        # Phase 6 - disk creation
 | 
			
		||||
        #  * Create each Ceph storage volume for the disks
 | 
			
		||||
        self.update_state(
 | 
			
		||||
            state="RUNNING",
 | 
			
		||||
            meta={"current": 6, "total": 10, "status": "Creating storage volumes"},
 | 
			
		||||
        )
 | 
			
		||||
        time.sleep(1)
 | 
			
		||||
 | 
			
		||||
        for volume in vm_data["volumes"]:
 | 
			
		||||
            if volume.get("source_volume") is not None:
 | 
			
		||||
                success, message = pvc_ceph.clone_volume(
 | 
			
		||||
                    zkhandler,
 | 
			
		||||
                    volume["pool"],
 | 
			
		||||
                    volume["source_volume"],
 | 
			
		||||
                    "{}_{}".format(vm_name, volume["disk_id"]),
 | 
			
		||||
                )
 | 
			
		||||
                print(message)
 | 
			
		||||
                if not success:
 | 
			
		||||
                    raise ProvisioningError(
 | 
			
		||||
                        'Failed to clone volume "{}" to "{}".'.format(
 | 
			
		||||
                            volume["source_volume"], volume["disk_id"]
 | 
			
		||||
                        )
 | 
			
		||||
                    )
 | 
			
		||||
            else:
 | 
			
		||||
                success, message = pvc_ceph.add_volume(
 | 
			
		||||
                    zkhandler,
 | 
			
		||||
                    volume["pool"],
 | 
			
		||||
                    "{}_{}".format(vm_name, volume["disk_id"]),
 | 
			
		||||
                    "{}G".format(volume["disk_size_gb"]),
 | 
			
		||||
                )
 | 
			
		||||
                print(message)
 | 
			
		||||
                if not success:
 | 
			
		||||
                    raise ProvisioningError(
 | 
			
		||||
                        'Failed to create volume "{}".'.format(volume["disk_id"])
 | 
			
		||||
                    )
 | 
			
		||||
 | 
			
		||||
        # Phase 7 - disk mapping
 | 
			
		||||
        #  * Map each volume to the local host in order
 | 
			
		||||
        #  * Format each volume with any specified filesystems
 | 
			
		||||
        #  * If any mountpoints are specified, create a temporary mount directory
 | 
			
		||||
        #  * Mount any volumes to their respective mountpoints
 | 
			
		||||
        self.update_state(
 | 
			
		||||
            state="RUNNING",
 | 
			
		||||
            meta={
 | 
			
		||||
                "current": 7,
 | 
			
		||||
                "total": 10,
 | 
			
		||||
                "status": "Mapping, formatting, and mounting storage volumes locally",
 | 
			
		||||
            },
 | 
			
		||||
        )
 | 
			
		||||
        time.sleep(1)
 | 
			
		||||
 | 
			
		||||
        for volume in vm_data["volumes"]:
 | 
			
		||||
            dst_volume_name = "{}_{}".format(vm_name, volume["disk_id"])
 | 
			
		||||
            dst_volume = "{}/{}".format(volume["pool"], dst_volume_name)
 | 
			
		||||
 | 
			
		||||
            if is_ova_install:
 | 
			
		||||
                src_volume_name = volume["volume_name"]
 | 
			
		||||
                src_volume = "{}/{}".format(volume["pool"], src_volume_name)
 | 
			
		||||
 | 
			
		||||
                print(
 | 
			
		||||
                    "Converting {} source volume {} to raw format on {}".format(
 | 
			
		||||
                        volume["volume_format"], src_volume, dst_volume
 | 
			
		||||
                    )
 | 
			
		||||
                )
 | 
			
		||||
 | 
			
		||||
                # Map the target RBD device
 | 
			
		||||
                retcode, retmsg = pvc_ceph.map_volume(
 | 
			
		||||
                    zkhandler, volume["pool"], dst_volume_name
 | 
			
		||||
                )
 | 
			
		||||
                if not retcode:
 | 
			
		||||
                    raise ProvisioningError(
 | 
			
		||||
                        'Failed to map destination volume "{}": {}'.format(
 | 
			
		||||
                            dst_volume_name, retmsg
 | 
			
		||||
                        )
 | 
			
		||||
                    )
 | 
			
		||||
                # Map the source RBD device
 | 
			
		||||
                retcode, retmsg = pvc_ceph.map_volume(
 | 
			
		||||
                    zkhandler, volume["pool"], src_volume_name
 | 
			
		||||
                )
 | 
			
		||||
                if not retcode:
 | 
			
		||||
                    raise ProvisioningError(
 | 
			
		||||
                        'Failed to map source volume "{}": {}'.format(
 | 
			
		||||
                            src_volume_name, retmsg
 | 
			
		||||
                        )
 | 
			
		||||
                    )
 | 
			
		||||
                # Convert from source to target
 | 
			
		||||
                retcode, stdout, stderr = pvc_common.run_os_command(
 | 
			
		||||
                    "qemu-img convert -C -f {} -O raw {} {}".format(
 | 
			
		||||
                        volume["volume_format"],
 | 
			
		||||
                        "/dev/rbd/{}".format(src_volume),
 | 
			
		||||
                        "/dev/rbd/{}".format(dst_volume),
 | 
			
		||||
                    )
 | 
			
		||||
                )
 | 
			
		||||
                if retcode:
 | 
			
		||||
                    raise ProvisioningError(
 | 
			
		||||
                        'Failed to convert {} volume "{}" to raw volume "{}": {}'.format(
 | 
			
		||||
                            volume["volume_format"], src_volume, dst_volume, stderr
 | 
			
		||||
                        )
 | 
			
		||||
                    )
 | 
			
		||||
 | 
			
		||||
                # Unmap the source RBD device (don't bother later)
 | 
			
		||||
                retcode, retmsg = pvc_ceph.unmap_volume(
 | 
			
		||||
                    zkhandler, volume["pool"], src_volume_name
 | 
			
		||||
                )
 | 
			
		||||
                if not retcode:
 | 
			
		||||
                    raise ProvisioningError(
 | 
			
		||||
                        'Failed to unmap source volume "{}": {}'.format(
 | 
			
		||||
                            src_volume_name, retmsg
 | 
			
		||||
                        )
 | 
			
		||||
                    )
 | 
			
		||||
                # Unmap the target RBD device (don't bother later)
 | 
			
		||||
                retcode, retmsg = pvc_ceph.unmap_volume(
 | 
			
		||||
                    zkhandler, volume["pool"], dst_volume_name
 | 
			
		||||
                )
 | 
			
		||||
                if not retcode:
 | 
			
		||||
                    raise ProvisioningError(
 | 
			
		||||
                        'Failed to unmap destination volume "{}": {}'.format(
 | 
			
		||||
                            dst_volume_name, retmsg
 | 
			
		||||
                        )
 | 
			
		||||
                    )
 | 
			
		||||
            else:
 | 
			
		||||
                if volume.get("source_volume") is not None:
 | 
			
		||||
                    continue
 | 
			
		||||
 | 
			
		||||
                if volume.get("filesystem") is None:
 | 
			
		||||
                    continue
 | 
			
		||||
 | 
			
		||||
                filesystem_args_list = list()
 | 
			
		||||
                for arg in volume["filesystem_args"].split():
 | 
			
		||||
                    arg_entry, *arg_data = arg.split("=")
 | 
			
		||||
                    arg_data = "=".join(arg_data)
 | 
			
		||||
                    filesystem_args_list.append(arg_entry)
 | 
			
		||||
                    filesystem_args_list.append(arg_data)
 | 
			
		||||
                filesystem_args = " ".join(filesystem_args_list)
 | 
			
		||||
 | 
			
		||||
                print(
 | 
			
		||||
                    "Creating {} filesystem on {}".format(
 | 
			
		||||
                        volume["filesystem"], dst_volume
 | 
			
		||||
                    )
 | 
			
		||||
                )
 | 
			
		||||
                print("Args: {}".format(filesystem_args))
 | 
			
		||||
 | 
			
		||||
                # Map the RBD device
 | 
			
		||||
                retcode, retmsg = pvc_ceph.map_volume(
 | 
			
		||||
                    zkhandler, volume["pool"], dst_volume_name
 | 
			
		||||
                )
 | 
			
		||||
                if not retcode:
 | 
			
		||||
                    raise ProvisioningError(
 | 
			
		||||
                        'Failed to map volume "{}": {}'.format(dst_volume, retmsg)
 | 
			
		||||
                    )
 | 
			
		||||
 | 
			
		||||
                # Create the filesystem
 | 
			
		||||
                if volume["filesystem"] == "swap":
 | 
			
		||||
                    retcode, stdout, stderr = pvc_common.run_os_command(
 | 
			
		||||
                        "mkswap -f /dev/rbd/{}".format(dst_volume)
 | 
			
		||||
                    )
 | 
			
		||||
                    if retcode:
 | 
			
		||||
                        raise ProvisioningError(
 | 
			
		||||
                            'Failed to create swap on "{}": {}'.format(
 | 
			
		||||
                                dst_volume, stderr
 | 
			
		||||
                            )
 | 
			
		||||
                        )
 | 
			
		||||
                else:
 | 
			
		||||
                    retcode, stdout, stderr = pvc_common.run_os_command(
 | 
			
		||||
                        "mkfs.{} {} /dev/rbd/{}".format(
 | 
			
		||||
                            volume["filesystem"], filesystem_args, dst_volume
 | 
			
		||||
                        )
 | 
			
		||||
                    )
 | 
			
		||||
                    if retcode:
 | 
			
		||||
                        raise ProvisioningError(
 | 
			
		||||
                            'Failed to create {} filesystem on "{}": {}'.format(
 | 
			
		||||
                                volume["filesystem"], dst_volume, stderr
 | 
			
		||||
                            )
 | 
			
		||||
                        )
 | 
			
		||||
 | 
			
		||||
                print(stdout)
 | 
			
		||||
 | 
			
		||||
        if is_script_install:
 | 
			
		||||
            # Create temporary directory
 | 
			
		||||
            retcode, stdout, stderr = pvc_common.run_os_command("mktemp -d")
 | 
			
		||||
            if retcode:
 | 
			
		||||
                raise ProvisioningError(
 | 
			
		||||
                    "Failed to create a temporary directory: {}".format(stderr)
 | 
			
		||||
                )
 | 
			
		||||
            temp_dir = stdout.strip()
 | 
			
		||||
 | 
			
		||||
            for volume in vm_data["volumes"]:
 | 
			
		||||
                if volume["source_volume"] is not None:
 | 
			
		||||
                    continue
 | 
			
		||||
 | 
			
		||||
                if not volume["mountpoint"] or volume["mountpoint"] == "swap":
 | 
			
		||||
                    continue
 | 
			
		||||
 | 
			
		||||
                mapped_dst_volume = "/dev/rbd/{}/{}_{}".format(
 | 
			
		||||
                    volume["pool"], vm_name, volume["disk_id"]
 | 
			
		||||
                )
 | 
			
		||||
                mount_path = "{}{}".format(temp_dir, volume["mountpoint"])
 | 
			
		||||
 | 
			
		||||
                # Ensure the mount path exists (within the filesystems)
 | 
			
		||||
                retcode, stdout, stderr = pvc_common.run_os_command(
 | 
			
		||||
                    "mkdir -p {}".format(mount_path)
 | 
			
		||||
                )
 | 
			
		||||
                if retcode:
 | 
			
		||||
                    raise ProvisioningError(
 | 
			
		||||
                        'Failed to create mountpoint "{}": {}'.format(
 | 
			
		||||
                            mount_path, stderr
 | 
			
		||||
                        )
 | 
			
		||||
                    )
 | 
			
		||||
 | 
			
		||||
                # Mount filesystems to temporary directory
 | 
			
		||||
                retcode, stdout, stderr = pvc_common.run_os_command(
 | 
			
		||||
                    "mount {} {}".format(mapped_dst_volume, mount_path)
 | 
			
		||||
                )
 | 
			
		||||
                if retcode:
 | 
			
		||||
                    raise ProvisioningError(
 | 
			
		||||
                        'Failed to mount "{}" on "{}": {}'.format(
 | 
			
		||||
                            mapped_dst_volume, mount_path, stderr
 | 
			
		||||
                        )
 | 
			
		||||
                    )
 | 
			
		||||
 | 
			
		||||
                print(
 | 
			
		||||
                    "Successfully mounted {} on {}".format(
 | 
			
		||||
                        mapped_dst_volume, mount_path
 | 
			
		||||
                    )
 | 
			
		||||
                )
 | 
			
		||||
 | 
			
		||||
        # Phase 8 - provisioning script execution
 | 
			
		||||
        #  * Execute the provisioning script main function ("install") passing any custom arguments
 | 
			
		||||
        self.update_state(
 | 
			
		||||
            state="RUNNING",
 | 
			
		||||
            meta={"current": 8, "total": 10, "status": "Executing provisioning script"},
 | 
			
		||||
        )
 | 
			
		||||
        time.sleep(1)
 | 
			
		||||
 | 
			
		||||
        if is_script_install:
 | 
			
		||||
            print("Running installer script")
 | 
			
		||||
 | 
			
		||||
            # Parse the script arguments
 | 
			
		||||
            script_arguments = dict()
 | 
			
		||||
            for argument in vm_data["script_arguments"]:
 | 
			
		||||
                argument_name, argument_data = argument.split("=")
 | 
			
		||||
                script_arguments[argument_name] = argument_data
 | 
			
		||||
 | 
			
		||||
            # Parse the runtime arguments
 | 
			
		||||
            if script_run_args is not None:
 | 
			
		||||
                for argument in script_run_args:
 | 
			
		||||
                    argument_name, argument_data = argument.split("=")
 | 
			
		||||
                    script_arguments[argument_name] = argument_data
 | 
			
		||||
 | 
			
		||||
            print("Script arguments: {}".format(script_arguments))
 | 
			
		||||
 | 
			
		||||
            # Run the script
 | 
			
		||||
            try:
 | 
			
		||||
                installer_script.install(
 | 
			
		||||
                    vm_name=vm_name,
 | 
			
		||||
                    vm_id=vm_id,
 | 
			
		||||
                    temporary_directory=temp_dir,
 | 
			
		||||
                    disks=vm_data["volumes"],
 | 
			
		||||
                    networks=vm_data["networks"],
 | 
			
		||||
                    **script_arguments
 | 
			
		||||
                )
 | 
			
		||||
            except Exception as e:
 | 
			
		||||
                raise ProvisioningError("Failed to run install script: {}".format(e))
 | 
			
		||||
 | 
			
		||||
    except Exception as e:
 | 
			
		||||
        start_vm = False
 | 
			
		||||
        raise e
 | 
			
		||||
 | 
			
		||||
    # Always perform the cleanup steps
 | 
			
		||||
    finally:
 | 
			
		||||
        # Phase 9 - install cleanup
 | 
			
		||||
        #  * Unmount any mounted volumes
 | 
			
		||||
        #  * Remove any temporary directories
 | 
			
		||||
        self.update_state(
 | 
			
		||||
            state="RUNNING",
 | 
			
		||||
            meta={
 | 
			
		||||
                "current": 9,
 | 
			
		||||
                "total": 10,
 | 
			
		||||
                "status": "Cleaning up local mounts and directories",
 | 
			
		||||
            },
 | 
			
		||||
        )
 | 
			
		||||
        time.sleep(1)
 | 
			
		||||
 | 
			
		||||
        if not is_ova_install:
 | 
			
		||||
            for volume in list(reversed(vm_data["volumes"])):
 | 
			
		||||
                if volume.get("source_volume") is not None:
 | 
			
		||||
                    continue
 | 
			
		||||
 | 
			
		||||
                if is_script_install:
 | 
			
		||||
                    # Unmount the volume
 | 
			
		||||
                    if (
 | 
			
		||||
                        volume.get("mountpoint") is not None
 | 
			
		||||
                        and volume.get("mountpoint") != "swap"
 | 
			
		||||
                    ):
 | 
			
		||||
                        print(
 | 
			
		||||
                            "Cleaning up mount {}{}".format(
 | 
			
		||||
                                temp_dir, volume["mountpoint"]
 | 
			
		||||
                            )
 | 
			
		||||
                        )
 | 
			
		||||
 | 
			
		||||
                        mount_path = "{}{}".format(temp_dir, volume["mountpoint"])
 | 
			
		||||
 | 
			
		||||
                        # Make sure any bind mounts or submounts are unmounted first
 | 
			
		||||
                        if volume["mountpoint"] == "/":
 | 
			
		||||
                            retcode, stdout, stderr = pvc_common.run_os_command(
 | 
			
		||||
                                "umount {}/**/**".format(mount_path)
 | 
			
		||||
                            )
 | 
			
		||||
                            retcode, stdout, stderr = pvc_common.run_os_command(
 | 
			
		||||
                                "umount {}/**".format(mount_path)
 | 
			
		||||
                            )
 | 
			
		||||
 | 
			
		||||
                        retcode, stdout, stderr = pvc_common.run_os_command(
 | 
			
		||||
                            "umount {}".format(mount_path)
 | 
			
		||||
                        )
 | 
			
		||||
                        if retcode:
 | 
			
		||||
                            print(
 | 
			
		||||
                                'Failed to unmount "{}": {}'.format(mount_path, stderr)
 | 
			
		||||
                            )
 | 
			
		||||
 | 
			
		||||
                # Unmap the RBD device
 | 
			
		||||
                if volume["filesystem"]:
 | 
			
		||||
                    print(
 | 
			
		||||
                        "Cleaning up RBD mapping /dev/rbd/{}/{}_{}".format(
 | 
			
		||||
                            volume["pool"], vm_name, volume["disk_id"]
 | 
			
		||||
                        )
 | 
			
		||||
                    )
 | 
			
		||||
 | 
			
		||||
                    rbd_volume = "/dev/rbd/{}/{}_{}".format(
 | 
			
		||||
                        volume["pool"], vm_name, volume["disk_id"]
 | 
			
		||||
                    )
 | 
			
		||||
                    retcode, stdout, stderr = pvc_common.run_os_command(
 | 
			
		||||
                        "rbd unmap {}".format(rbd_volume)
 | 
			
		||||
                    )
 | 
			
		||||
                    if retcode:
 | 
			
		||||
                        print(
 | 
			
		||||
                            'Failed to unmap volume "{}": {}'.format(rbd_volume, stderr)
 | 
			
		||||
                        )
 | 
			
		||||
 | 
			
		||||
        print("Cleaning up temporary directories and files")
 | 
			
		||||
 | 
			
		||||
        if is_script_install:
 | 
			
		||||
            # Remove temporary mount directory (don't fail if not removed)
 | 
			
		||||
            retcode, stdout, stderr = pvc_common.run_os_command(
 | 
			
		||||
                "rmdir {}".format(temp_dir)
 | 
			
		||||
            )
 | 
			
		||||
            if retcode:
 | 
			
		||||
                print(
 | 
			
		||||
                    'Failed to delete temporary directory "{}": {}'.format(
 | 
			
		||||
                        temp_dir, stderr
 | 
			
		||||
                    )
 | 
			
		||||
                )
 | 
			
		||||
 | 
			
		||||
            # Remote temporary script (don't fail if not removed)
 | 
			
		||||
            retcode, stdout, stderr = pvc_common.run_os_command(
 | 
			
		||||
                "rm -f {}".format(script_file)
 | 
			
		||||
            )
 | 
			
		||||
            if retcode:
 | 
			
		||||
                print(
 | 
			
		||||
                    'Failed to delete temporary script file "{}": {}'.format(
 | 
			
		||||
                        script_file, stderr
 | 
			
		||||
                    )
 | 
			
		||||
                )
 | 
			
		||||
 | 
			
		||||
        # Phase 10 - startup
 | 
			
		||||
        #  * Start the VM in the PVC cluster
 | 
			
		||||
        if start_vm:
 | 
			
		||||
            self.update_state(
 | 
			
		||||
                state="RUNNING",
 | 
			
		||||
                meta={"current": 10, "total": 10, "status": "Starting VM"},
 | 
			
		||||
            )
 | 
			
		||||
            time.sleep(1)
 | 
			
		||||
            retcode, retmsg = pvc_vm.start_vm(zkhandler, vm_name)
 | 
			
		||||
            print(retmsg)
 | 
			
		||||
 | 
			
		||||
    zkhandler.disconnect()
 | 
			
		||||
    del zkhandler
 | 
			
		||||
 | 
			
		||||
    return {
 | 
			
		||||
        "status": 'VM "{}" with profile "{}" has been provisioned successfully'.format(
 | 
			
		||||
            vm_name, vm_profile
 | 
			
		||||
        ),
 | 
			
		||||
        "current": 10,
 | 
			
		||||
        "total": 10,
 | 
			
		||||
    }
 | 
			
		||||
 
 | 
			
		||||
							
								
								
									
										730
									
								
								api-daemon/pvcapid/vmbuilder.py
									
									
									
									
									
										Executable file
									
								
							
							
						
						
									
										730
									
								
								api-daemon/pvcapid/vmbuilder.py
									
									
									
									
									
										Executable file
									
								
							@@ -0,0 +1,730 @@
 | 
			
		||||
#!/usr/bin/env python3
 | 
			
		||||
 | 
			
		||||
# vmbuilder.py - pvc api vm builder (provisioner) functions
 | 
			
		||||
# part of the parallel virtual cluster (pvc) system
 | 
			
		||||
#
 | 
			
		||||
#    copyright (c) 2018-2022 joshua m. boniface <joshua@boniface.me>
 | 
			
		||||
#
 | 
			
		||||
#    this program is free software: you can redistribute it and/or modify
 | 
			
		||||
#    it under the terms of the gnu general public license as published by
 | 
			
		||||
#    the free software foundation, version 3.
 | 
			
		||||
#
 | 
			
		||||
#    this program is distributed in the hope that it will be useful,
 | 
			
		||||
#    but without any warranty; without even the implied warranty of
 | 
			
		||||
#    MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
 | 
			
		||||
#    GNU General Public License for more details.
 | 
			
		||||
#
 | 
			
		||||
#    You should have received a copy of the GNU General Public License
 | 
			
		||||
#    along with this program.  If not, see <https://www.gnu.org/licenses/>.
 | 
			
		||||
#
 | 
			
		||||
###############################################################################
 | 
			
		||||
 | 
			
		||||
import json
 | 
			
		||||
import psycopg2
 | 
			
		||||
import psycopg2.extras
 | 
			
		||||
import re
 | 
			
		||||
import os
 | 
			
		||||
 | 
			
		||||
# import sys
 | 
			
		||||
import time
 | 
			
		||||
import importlib.util
 | 
			
		||||
import uuid
 | 
			
		||||
 | 
			
		||||
from contextlib import contextmanager
 | 
			
		||||
 | 
			
		||||
from pvcapid.Daemon import config
 | 
			
		||||
 | 
			
		||||
from daemon_lib.zkhandler import ZKHandler
 | 
			
		||||
 | 
			
		||||
import daemon_lib.common as pvc_common
 | 
			
		||||
import daemon_lib.node as pvc_node
 | 
			
		||||
import daemon_lib.vm as pvc_vm
 | 
			
		||||
import daemon_lib.network as pvc_network
 | 
			
		||||
import daemon_lib.ceph as pvc_ceph
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
#
 | 
			
		||||
# Exceptions (used by Celery tasks)
 | 
			
		||||
#
 | 
			
		||||
class ValidationError(Exception):
 | 
			
		||||
    """
 | 
			
		||||
    An exception that results from some value being un- or mis-defined.
 | 
			
		||||
    """
 | 
			
		||||
 | 
			
		||||
    pass
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
class ClusterError(Exception):
 | 
			
		||||
    """
 | 
			
		||||
    An exception that results from the PVC cluster being out of alignment with the action.
 | 
			
		||||
    """
 | 
			
		||||
 | 
			
		||||
    pass
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
class ProvisioningError(Exception):
 | 
			
		||||
    """
 | 
			
		||||
    An exception that results from a failure of a provisioning command.
 | 
			
		||||
    """
 | 
			
		||||
 | 
			
		||||
    pass
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
#
 | 
			
		||||
# VMBuilder class - subclassed by install scripts
 | 
			
		||||
#
 | 
			
		||||
class VMBuilder(object):
 | 
			
		||||
    def __init__(
 | 
			
		||||
        self,
 | 
			
		||||
        vm_name,
 | 
			
		||||
        vm_id,
 | 
			
		||||
        vm_profile,
 | 
			
		||||
        vm_data,
 | 
			
		||||
    ):
 | 
			
		||||
        self.vm_name = vm_name
 | 
			
		||||
        self.vm_id = vm_id
 | 
			
		||||
        self.vm_uuid = uuid.uuid4()
 | 
			
		||||
        self.vm_profile = vm_profile
 | 
			
		||||
        self.vm_data = vm_data
 | 
			
		||||
 | 
			
		||||
    #
 | 
			
		||||
    # Primary class functions; implemented by the individual scripts
 | 
			
		||||
    #
 | 
			
		||||
    def setup(self):
 | 
			
		||||
        """
 | 
			
		||||
        setup(): Perform special setup steps before proceeding
 | 
			
		||||
        OPTIONAL
 | 
			
		||||
        """
 | 
			
		||||
        pass
 | 
			
		||||
 | 
			
		||||
    def create(self):
 | 
			
		||||
        """
 | 
			
		||||
        create(): Create the VM libvirt schema definition which is defined afterwards
 | 
			
		||||
        """
 | 
			
		||||
        pass
 | 
			
		||||
 | 
			
		||||
    def prepare(self):
 | 
			
		||||
        """
 | 
			
		||||
        prepare(): Prepare any disks/volumes for the install step
 | 
			
		||||
        """
 | 
			
		||||
        pass
 | 
			
		||||
 | 
			
		||||
    def install(self):
 | 
			
		||||
        """
 | 
			
		||||
        install(): Perform the installation
 | 
			
		||||
        """
 | 
			
		||||
        pass
 | 
			
		||||
 | 
			
		||||
    def cleanup(self):
 | 
			
		||||
        """
 | 
			
		||||
        cleanup(): Perform any cleanup required after the prepare() step or on failure of the install() step
 | 
			
		||||
        """
 | 
			
		||||
        pass
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
#
 | 
			
		||||
# Helper functions (as context managers)
 | 
			
		||||
#
 | 
			
		||||
@contextmanager
 | 
			
		||||
def chroot(destination):
 | 
			
		||||
    """
 | 
			
		||||
    Change root directory to a given destination
 | 
			
		||||
    """
 | 
			
		||||
    try:
 | 
			
		||||
        real_root = os.open("/", os.O_RDONLY)
 | 
			
		||||
        os.chroot(destination)
 | 
			
		||||
        fake_root = os.open("/", os.O_RDONLY)
 | 
			
		||||
        os.fchdir(fake_root)
 | 
			
		||||
        yield
 | 
			
		||||
    except Exception:
 | 
			
		||||
        pass
 | 
			
		||||
    finally:
 | 
			
		||||
        os.fchdir(real_root)
 | 
			
		||||
        os.chroot(".")
 | 
			
		||||
        os.fchdir(fake_root)
 | 
			
		||||
        os.close(fake_root)
 | 
			
		||||
        os.close(real_root)
 | 
			
		||||
        del fake_root
 | 
			
		||||
        del real_root
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
@contextmanager
 | 
			
		||||
def open_db(config):
 | 
			
		||||
    try:
 | 
			
		||||
        conn = psycopg2.connect(
 | 
			
		||||
            host=config["database_host"],
 | 
			
		||||
            port=config["database_port"],
 | 
			
		||||
            dbname=config["database_name"],
 | 
			
		||||
            user=config["database_user"],
 | 
			
		||||
            password=config["database_password"],
 | 
			
		||||
        )
 | 
			
		||||
        cur = conn.cursor(cursor_factory=psycopg2.extras.RealDictCursor)
 | 
			
		||||
        yield cur
 | 
			
		||||
    except Exception:
 | 
			
		||||
        raise ClusterError("Failed to connect to Postgres")
 | 
			
		||||
    finally:
 | 
			
		||||
        conn.commit()
 | 
			
		||||
        cur.close()
 | 
			
		||||
        conn.close()
 | 
			
		||||
        del conn
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
@contextmanager
 | 
			
		||||
def open_zk(config):
 | 
			
		||||
    try:
 | 
			
		||||
        zkhandler = ZKHandler(config)
 | 
			
		||||
        zkhandler.connect()
 | 
			
		||||
        yield zkhandler
 | 
			
		||||
    except Exception:
 | 
			
		||||
        raise ClusterError("Failed to connect to Zookeeper")
 | 
			
		||||
    finally:
 | 
			
		||||
        zkhandler.disconnect()
 | 
			
		||||
        del zkhandler
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
#
 | 
			
		||||
# Main VM provisioning function - executed by the Celery worker
 | 
			
		||||
#
 | 
			
		||||
def create_vm(
 | 
			
		||||
    self, vm_name, vm_profile, define_vm=True, start_vm=True, script_run_args=[]
 | 
			
		||||
):
 | 
			
		||||
    print(f"Starting provisioning of VM '{vm_name}' with profile '{vm_profile}'")
 | 
			
		||||
 | 
			
		||||
    # Phase 1 - setup
 | 
			
		||||
    #  * Get the profile elements
 | 
			
		||||
    #  * Get the details from these elements
 | 
			
		||||
    #  * Assemble a VM configuration dictionary
 | 
			
		||||
    self.update_state(
 | 
			
		||||
        state="RUNNING",
 | 
			
		||||
        meta={"current": 1, "total": 10, "status": "Collecting configuration"},
 | 
			
		||||
    )
 | 
			
		||||
    time.sleep(1)
 | 
			
		||||
 | 
			
		||||
    vm_id = re.findall(r"/(\d+)$/", vm_name)
 | 
			
		||||
    if not vm_id:
 | 
			
		||||
        vm_id = 0
 | 
			
		||||
    else:
 | 
			
		||||
        vm_id = vm_id[0]
 | 
			
		||||
 | 
			
		||||
    vm_data = dict()
 | 
			
		||||
 | 
			
		||||
    with open_db(config) as db_cur:
 | 
			
		||||
        # Get the profile information
 | 
			
		||||
        query = "SELECT * FROM profile WHERE name = %s"
 | 
			
		||||
        args = (vm_profile,)
 | 
			
		||||
        db_cur.execute(query, args)
 | 
			
		||||
        profile_data = db_cur.fetchone()
 | 
			
		||||
        if profile_data.get("arguments"):
 | 
			
		||||
            vm_data["script_arguments"] = profile_data.get("arguments").split("|")
 | 
			
		||||
        else:
 | 
			
		||||
            vm_data["script_arguments"] = []
 | 
			
		||||
 | 
			
		||||
        if profile_data.get("profile_type") == "ova":
 | 
			
		||||
            is_ova_install = True
 | 
			
		||||
        else:
 | 
			
		||||
            is_ova_install = False
 | 
			
		||||
 | 
			
		||||
        # Get the system details
 | 
			
		||||
        query = "SELECT * FROM system_template WHERE id = %s"
 | 
			
		||||
        args = (profile_data["system_template"],)
 | 
			
		||||
        db_cur.execute(query, args)
 | 
			
		||||
        vm_data["system_details"] = db_cur.fetchone()
 | 
			
		||||
 | 
			
		||||
        # Get the MAC template
 | 
			
		||||
        query = "SELECT mac_template FROM network_template WHERE id = %s"
 | 
			
		||||
        args = (profile_data["network_template"],)
 | 
			
		||||
        db_cur.execute(query, args)
 | 
			
		||||
        db_row = db_cur.fetchone()
 | 
			
		||||
        if db_row:
 | 
			
		||||
            vm_data["mac_template"] = db_row.get("mac_template")
 | 
			
		||||
        else:
 | 
			
		||||
            vm_data["mac_template"] = None
 | 
			
		||||
 | 
			
		||||
        # Get the networks
 | 
			
		||||
        query = "SELECT * FROM network WHERE network_template = %s"
 | 
			
		||||
        args = (profile_data["network_template"],)
 | 
			
		||||
        db_cur.execute(query, args)
 | 
			
		||||
        _vm_networks = db_cur.fetchall()
 | 
			
		||||
        vm_networks = list()
 | 
			
		||||
 | 
			
		||||
        # Set the eth_bridge for each network
 | 
			
		||||
        for network in _vm_networks:
 | 
			
		||||
            vni = network["vni"]
 | 
			
		||||
            if vni in ["upstream", "cluster", "storage"]:
 | 
			
		||||
                eth_bridge = "br{}".format(vni)
 | 
			
		||||
            else:
 | 
			
		||||
                eth_bridge = "vmbr{}".format(vni)
 | 
			
		||||
            network["eth_bridge"] = eth_bridge
 | 
			
		||||
            vm_networks.append(network)
 | 
			
		||||
        vm_data["networks"] = vm_networks
 | 
			
		||||
 | 
			
		||||
        # Get the storage volumes
 | 
			
		||||
        # ORDER BY ensures disks are always in the sdX/vdX order, regardless of add order
 | 
			
		||||
        query = "SELECT * FROM storage WHERE storage_template = %s ORDER BY disk_id"
 | 
			
		||||
        args = (profile_data["storage_template"],)
 | 
			
		||||
        db_cur.execute(query, args)
 | 
			
		||||
        vm_data["volumes"] = db_cur.fetchall()
 | 
			
		||||
 | 
			
		||||
        # Get the script
 | 
			
		||||
        query = "SELECT script FROM script WHERE id = %s"
 | 
			
		||||
        args = (profile_data["script"],)
 | 
			
		||||
        db_cur.execute(query, args)
 | 
			
		||||
        db_row = db_cur.fetchone()
 | 
			
		||||
        if db_row:
 | 
			
		||||
            vm_data["script"] = db_row.get("script")
 | 
			
		||||
        else:
 | 
			
		||||
            vm_data["script"] = None
 | 
			
		||||
            query = "SELECT * FROM ova WHERE id = %s"
 | 
			
		||||
            args = (profile_data["ova"],)
 | 
			
		||||
            db_cur.execute(query, args)
 | 
			
		||||
            vm_data["ova_details"] = db_cur.fetchone()
 | 
			
		||||
 | 
			
		||||
            query = "SELECT * FROM ova_volume WHERE ova = %s"
 | 
			
		||||
            args = (profile_data["ova"],)
 | 
			
		||||
            db_cur.execute(query, args)
 | 
			
		||||
            vm_data["volumes"] = db_cur.fetchall()
 | 
			
		||||
 | 
			
		||||
    retcode, stdout, stderr = pvc_common.run_os_command("uname -m")
 | 
			
		||||
    vm_data["system_architecture"] = stdout.strip()
 | 
			
		||||
 | 
			
		||||
    monitor_list = list()
 | 
			
		||||
    coordinator_names = config["storage_hosts"]
 | 
			
		||||
    for coordinator in coordinator_names:
 | 
			
		||||
        monitor_list.append("{}.{}".format(coordinator, config["storage_domain"]))
 | 
			
		||||
    vm_data["ceph_monitor_list"] = monitor_list
 | 
			
		||||
    vm_data["ceph_monitor_port"] = config["ceph_monitor_port"]
 | 
			
		||||
    vm_data["ceph_monitor_secret"] = config["ceph_storage_secret_uuid"]
 | 
			
		||||
 | 
			
		||||
    # Parse the script arguments
 | 
			
		||||
    script_arguments = dict()
 | 
			
		||||
    for argument in vm_data["script_arguments"]:
 | 
			
		||||
        argument_name, argument_data = argument.split("=")
 | 
			
		||||
        script_arguments[argument_name] = argument_data
 | 
			
		||||
 | 
			
		||||
    # Parse the runtime arguments
 | 
			
		||||
    if script_run_args is not None:
 | 
			
		||||
        for argument in script_run_args:
 | 
			
		||||
            argument_name, argument_data = argument.split("=")
 | 
			
		||||
            script_arguments[argument_name] = argument_data
 | 
			
		||||
 | 
			
		||||
    print("Script arguments: {}".format(script_arguments))
 | 
			
		||||
    vm_data["script_arguments"] = script_arguments
 | 
			
		||||
 | 
			
		||||
    print(
 | 
			
		||||
        "VM configuration data:\n{}".format(
 | 
			
		||||
            json.dumps(vm_data, sort_keys=True, indent=2)
 | 
			
		||||
        )
 | 
			
		||||
    )
 | 
			
		||||
 | 
			
		||||
    # Phase 2 - verification
 | 
			
		||||
    #  * Ensure that at least one node has enough free RAM to hold the VM (becomes main host)
 | 
			
		||||
    #  * Ensure that all networks are valid
 | 
			
		||||
    #  * Ensure that there is enough disk space in the Ceph cluster for the disks
 | 
			
		||||
    # This is the "safe fail" step when an invalid configuration will be caught
 | 
			
		||||
    self.update_state(
 | 
			
		||||
        state="RUNNING",
 | 
			
		||||
        meta={
 | 
			
		||||
            "current": 2,
 | 
			
		||||
            "total": 10,
 | 
			
		||||
            "status": "Verifying configuration against cluster",
 | 
			
		||||
        },
 | 
			
		||||
    )
 | 
			
		||||
    time.sleep(1)
 | 
			
		||||
 | 
			
		||||
    with open_zk(config) as zkhandler:
 | 
			
		||||
        # Verify that a VM with this name does not already exist
 | 
			
		||||
        if pvc_vm.searchClusterByName(zkhandler, vm_name):
 | 
			
		||||
            raise ClusterError(
 | 
			
		||||
                "A VM with the name '{}' already exists in the cluster.".format(vm_name)
 | 
			
		||||
            )
 | 
			
		||||
 | 
			
		||||
        # Verify that at least one host has enough free RAM to run the VM
 | 
			
		||||
        _discard, nodes = pvc_node.get_list(zkhandler, None)
 | 
			
		||||
        target_node = None
 | 
			
		||||
        last_free = 0
 | 
			
		||||
        for node in nodes:
 | 
			
		||||
            # Skip the node if it is not ready to run VMs
 | 
			
		||||
            if node["daemon_state"] != "run" or node["domain_state"] != "ready":
 | 
			
		||||
                continue
 | 
			
		||||
            # Skip the node if its free memory is less than the new VM's size, plus a 512MB buffer
 | 
			
		||||
            if node["memory"]["free"] < (vm_data["system_details"]["vram_mb"] + 512):
 | 
			
		||||
                continue
 | 
			
		||||
            # If this node has the most free, use it
 | 
			
		||||
            if node["memory"]["free"] > last_free:
 | 
			
		||||
                last_free = node["memory"]["free"]
 | 
			
		||||
                target_node = node["name"]
 | 
			
		||||
        # Raise if no node was found
 | 
			
		||||
        if not target_node:
 | 
			
		||||
            raise ClusterError(
 | 
			
		||||
                "No ready cluster node contains at least {}+512 MB of free RAM.".format(
 | 
			
		||||
                    vm_data["system_details"]["vram_mb"]
 | 
			
		||||
                )
 | 
			
		||||
            )
 | 
			
		||||
 | 
			
		||||
        print(
 | 
			
		||||
            'Selecting target node "{}" with "{}" MB free RAM'.format(
 | 
			
		||||
                target_node, last_free
 | 
			
		||||
            )
 | 
			
		||||
        )
 | 
			
		||||
 | 
			
		||||
        # Verify that all configured networks are present on the cluster
 | 
			
		||||
        cluster_networks, _discard = pvc_network.getClusterNetworkList(zkhandler)
 | 
			
		||||
        for network in vm_data["networks"]:
 | 
			
		||||
            vni = str(network["vni"])
 | 
			
		||||
            if vni not in cluster_networks and vni not in [
 | 
			
		||||
                "upstream",
 | 
			
		||||
                "cluster",
 | 
			
		||||
                "storage",
 | 
			
		||||
            ]:
 | 
			
		||||
                raise ClusterError(
 | 
			
		||||
                    'The network VNI "{}" is not present on the cluster.'.format(vni)
 | 
			
		||||
                )
 | 
			
		||||
 | 
			
		||||
        print("All configured networks for VM are valid")
 | 
			
		||||
 | 
			
		||||
        # Verify that there is enough disk space free to provision all VM disks
 | 
			
		||||
        pools = dict()
 | 
			
		||||
        for volume in vm_data["volumes"]:
 | 
			
		||||
            if volume.get("source_volume") is not None:
 | 
			
		||||
                volume_data = pvc_ceph.getVolumeInformation(
 | 
			
		||||
                    zkhandler, volume["pool"], volume["source_volume"]
 | 
			
		||||
                )
 | 
			
		||||
                if not volume_data:
 | 
			
		||||
                    raise ClusterError(
 | 
			
		||||
                        "The source volume {}/{} could not be found.".format(
 | 
			
		||||
                            volume["pool"], volume["source_volume"]
 | 
			
		||||
                        )
 | 
			
		||||
                    )
 | 
			
		||||
                if not volume["pool"] in pools:
 | 
			
		||||
                    pools[volume["pool"]] = int(
 | 
			
		||||
                        pvc_ceph.format_bytes_fromhuman(volume_data["stats"]["size"])
 | 
			
		||||
                        / 1024
 | 
			
		||||
                        / 1024
 | 
			
		||||
                        / 1024
 | 
			
		||||
                    )
 | 
			
		||||
                else:
 | 
			
		||||
                    pools[volume["pool"]] += int(
 | 
			
		||||
                        pvc_ceph.format_bytes_fromhuman(volume_data["stats"]["size"])
 | 
			
		||||
                        / 1024
 | 
			
		||||
                        / 1024
 | 
			
		||||
                        / 1024
 | 
			
		||||
                    )
 | 
			
		||||
            else:
 | 
			
		||||
                if not volume["pool"] in pools:
 | 
			
		||||
                    pools[volume["pool"]] = volume["disk_size_gb"]
 | 
			
		||||
                else:
 | 
			
		||||
                    pools[volume["pool"]] += volume["disk_size_gb"]
 | 
			
		||||
 | 
			
		||||
        for pool in pools:
 | 
			
		||||
            try:
 | 
			
		||||
                pool_information = pvc_ceph.getPoolInformation(zkhandler, pool)
 | 
			
		||||
                if not pool_information:
 | 
			
		||||
                    raise
 | 
			
		||||
            except Exception:
 | 
			
		||||
                raise ClusterError(
 | 
			
		||||
                    'Pool "{}" is not present on the cluster.'.format(pool)
 | 
			
		||||
                )
 | 
			
		||||
            pool_free_space_gb = int(
 | 
			
		||||
                pool_information["stats"]["free_bytes"] / 1024 / 1024 / 1024
 | 
			
		||||
            )
 | 
			
		||||
            pool_vm_usage_gb = int(pools[pool])
 | 
			
		||||
 | 
			
		||||
            if pool_vm_usage_gb >= pool_free_space_gb:
 | 
			
		||||
                raise ClusterError(
 | 
			
		||||
                    'Pool "{}" has only {} GB free and VM requires {} GB.'.format(
 | 
			
		||||
                        pool, pool_free_space_gb, pool_vm_usage_gb
 | 
			
		||||
                    )
 | 
			
		||||
                )
 | 
			
		||||
 | 
			
		||||
        print("There is enough space on cluster to store VM volumes")
 | 
			
		||||
 | 
			
		||||
    if not is_ova_install:
 | 
			
		||||
        # Verify that every specified filesystem is valid
 | 
			
		||||
        used_filesystems = list()
 | 
			
		||||
        for volume in vm_data["volumes"]:
 | 
			
		||||
            if volume["source_volume"] is not None:
 | 
			
		||||
                continue
 | 
			
		||||
            if volume["filesystem"] and volume["filesystem"] not in used_filesystems:
 | 
			
		||||
                used_filesystems.append(volume["filesystem"])
 | 
			
		||||
 | 
			
		||||
        for filesystem in used_filesystems:
 | 
			
		||||
            if filesystem == "swap":
 | 
			
		||||
                retcode, stdout, stderr = pvc_common.run_os_command("which mkswap")
 | 
			
		||||
                if retcode:
 | 
			
		||||
                    raise ProvisioningError(
 | 
			
		||||
                        "Failed to find binary for mkswap: {}".format(stderr)
 | 
			
		||||
                    )
 | 
			
		||||
            else:
 | 
			
		||||
                retcode, stdout, stderr = pvc_common.run_os_command(
 | 
			
		||||
                    "which mkfs.{}".format(filesystem)
 | 
			
		||||
                )
 | 
			
		||||
                if retcode:
 | 
			
		||||
                    raise ProvisioningError(
 | 
			
		||||
                        "Failed to find binary for mkfs.{}: {}".format(
 | 
			
		||||
                            filesystem, stderr
 | 
			
		||||
                        )
 | 
			
		||||
                    )
 | 
			
		||||
 | 
			
		||||
        print("All selected filesystems are valid")
 | 
			
		||||
 | 
			
		||||
    # Phase 3 - provisioning script preparation
 | 
			
		||||
    #  * Import the provisioning script as a library with importlib
 | 
			
		||||
    #  * Ensure the required function(s) are present
 | 
			
		||||
    self.update_state(
 | 
			
		||||
        state="RUNNING",
 | 
			
		||||
        meta={"current": 3, "total": 10, "status": "Preparing provisioning script"},
 | 
			
		||||
    )
 | 
			
		||||
    time.sleep(1)
 | 
			
		||||
 | 
			
		||||
    # Write the script out to a temporary file
 | 
			
		||||
    retcode, stdout, stderr = pvc_common.run_os_command("mktemp")
 | 
			
		||||
    if retcode:
 | 
			
		||||
        raise ProvisioningError("Failed to create a temporary file: {}".format(stderr))
 | 
			
		||||
    script_file = stdout.strip()
 | 
			
		||||
    with open(script_file, "w") as fh:
 | 
			
		||||
        fh.write(vm_data["script"])
 | 
			
		||||
        fh.write("\n")
 | 
			
		||||
 | 
			
		||||
    # Import the script file
 | 
			
		||||
    loader = importlib.machinery.SourceFileLoader("installer_script", script_file)
 | 
			
		||||
    spec = importlib.util.spec_from_loader(loader.name, loader)
 | 
			
		||||
    installer_script = importlib.util.module_from_spec(spec)
 | 
			
		||||
    spec.loader.exec_module(installer_script)
 | 
			
		||||
 | 
			
		||||
    # Set up the VMBuilderScript object
 | 
			
		||||
    vm_builder = installer_script.VMBuilderScript(
 | 
			
		||||
        vm_name=vm_name,
 | 
			
		||||
        vm_id=vm_id,
 | 
			
		||||
        vm_profile=vm_profile,
 | 
			
		||||
        vm_data=vm_data,
 | 
			
		||||
    )
 | 
			
		||||
 | 
			
		||||
    print("Provisioning script imported successfully")
 | 
			
		||||
 | 
			
		||||
    # Create temporary directory for external chroot
 | 
			
		||||
    retcode, stdout, stderr = pvc_common.run_os_command("mktemp -d")
 | 
			
		||||
    if retcode:
 | 
			
		||||
        raise ProvisioningError(f"Failed to create a temporary directory: {stderr}")
 | 
			
		||||
    temp_dir = stdout.strip()
 | 
			
		||||
 | 
			
		||||
    # Bind mount / to the chroot location /
 | 
			
		||||
    retcode, stdout, stderr = pvc_common.run_os_command(
 | 
			
		||||
        f"mount --bind --options ro / {temp_dir}"
 | 
			
		||||
    )
 | 
			
		||||
    if retcode:
 | 
			
		||||
        raise ProvisioningError(
 | 
			
		||||
            f"Failed to mount rootfs onto {temp_dir} for chroot: {stderr}"
 | 
			
		||||
        )
 | 
			
		||||
 | 
			
		||||
    # Mount tmpfs to the chroot location /tmp
 | 
			
		||||
    retcode, stdout, stderr = pvc_common.run_os_command(
 | 
			
		||||
        f"mount --type tmpfs tmpfs {temp_dir}/tmp"
 | 
			
		||||
    )
 | 
			
		||||
    if retcode:
 | 
			
		||||
        raise ProvisioningError(
 | 
			
		||||
            f"Failed to mount tmpfs onto {temp_dir}/tmp for chroot: {stderr}"
 | 
			
		||||
        )
 | 
			
		||||
 | 
			
		||||
    # Bind mount /dev to the chroot location /dev
 | 
			
		||||
    retcode, stdout, stderr = pvc_common.run_os_command(
 | 
			
		||||
        f"mount --bind --options ro /dev {temp_dir}/dev"
 | 
			
		||||
    )
 | 
			
		||||
    if retcode:
 | 
			
		||||
        raise ProvisioningError(
 | 
			
		||||
            f"Failed to mount devfs onto {temp_dir}/dev for chroot: {stderr}"
 | 
			
		||||
        )
 | 
			
		||||
 | 
			
		||||
    # Bind mount /run to the chroot location /run
 | 
			
		||||
    retcode, stdout, stderr = pvc_common.run_os_command(
 | 
			
		||||
        f"mount --bind --options rw /run {temp_dir}/run"
 | 
			
		||||
    )
 | 
			
		||||
    if retcode:
 | 
			
		||||
        raise ProvisioningError(
 | 
			
		||||
            f"Failed to mount runfs onto {temp_dir}/run for chroot: {stderr}"
 | 
			
		||||
        )
 | 
			
		||||
 | 
			
		||||
    # Bind mount /sys to the chroot location /sys
 | 
			
		||||
    retcode, stdout, stderr = pvc_common.run_os_command(
 | 
			
		||||
        f"mount --bind --options rw /sys {temp_dir}/sys"
 | 
			
		||||
    )
 | 
			
		||||
    if retcode:
 | 
			
		||||
        raise ProvisioningError(
 | 
			
		||||
            f"Failed to mount sysfs onto {temp_dir}/sys for chroot: {stderr}"
 | 
			
		||||
        )
 | 
			
		||||
 | 
			
		||||
    print("Chroot environment prepared successfully")
 | 
			
		||||
 | 
			
		||||
    # Phase 4 - script: setup()
 | 
			
		||||
    #  * Run pre-setup steps
 | 
			
		||||
    self.update_state(
 | 
			
		||||
        state="RUNNING",
 | 
			
		||||
        meta={
 | 
			
		||||
            "current": 4,
 | 
			
		||||
            "total": 10,
 | 
			
		||||
            "status": "Running script setup() step",
 | 
			
		||||
        },
 | 
			
		||||
    )
 | 
			
		||||
    time.sleep(1)
 | 
			
		||||
 | 
			
		||||
    print("Running script setup() step")
 | 
			
		||||
 | 
			
		||||
    with chroot(temp_dir):
 | 
			
		||||
        vm_builder.setup()
 | 
			
		||||
 | 
			
		||||
    # Phase 5 - script: create()
 | 
			
		||||
    #  * Prepare the libvirt XML defintion for the VM
 | 
			
		||||
    self.update_state(
 | 
			
		||||
        state="RUNNING",
 | 
			
		||||
        meta={
 | 
			
		||||
            "current": 5,
 | 
			
		||||
            "total": 10,
 | 
			
		||||
            "status": "Running script create() step",
 | 
			
		||||
        },
 | 
			
		||||
    )
 | 
			
		||||
    time.sleep(1)
 | 
			
		||||
 | 
			
		||||
    if define_vm:
 | 
			
		||||
        print("Running script create() step")
 | 
			
		||||
 | 
			
		||||
        with chroot(temp_dir):
 | 
			
		||||
            vm_schema = vm_builder.create()
 | 
			
		||||
            print("Generated VM schema:\n{}\n".format(vm_schema))
 | 
			
		||||
 | 
			
		||||
        print("Defining VM on cluster")
 | 
			
		||||
        node_limit = vm_data["system_details"]["node_limit"]
 | 
			
		||||
        if node_limit:
 | 
			
		||||
            node_limit = node_limit.split(",")
 | 
			
		||||
        node_selector = vm_data["system_details"]["node_selector"]
 | 
			
		||||
        node_autostart = vm_data["system_details"]["node_autostart"]
 | 
			
		||||
        migration_method = vm_data["system_details"]["migration_method"]
 | 
			
		||||
        with open_zk(config) as zkhandler:
 | 
			
		||||
            retcode, retmsg = pvc_vm.define_vm(
 | 
			
		||||
                zkhandler,
 | 
			
		||||
                vm_schema.strip(),
 | 
			
		||||
                target_node,
 | 
			
		||||
                node_limit,
 | 
			
		||||
                node_selector,
 | 
			
		||||
                node_autostart,
 | 
			
		||||
                migration_method,
 | 
			
		||||
                vm_profile,
 | 
			
		||||
                initial_state="provision",
 | 
			
		||||
            )
 | 
			
		||||
        print(retmsg)
 | 
			
		||||
    else:
 | 
			
		||||
        print("Skipping VM definition due to define_vm=False")
 | 
			
		||||
 | 
			
		||||
    # Phase 6 - script: prepare()
 | 
			
		||||
    #  * Run preparation steps (e.g. disk creation and mapping, filesystem creation, etc.)
 | 
			
		||||
    self.update_state(
 | 
			
		||||
        state="RUNNING",
 | 
			
		||||
        meta={
 | 
			
		||||
            "current": 6,
 | 
			
		||||
            "total": 10,
 | 
			
		||||
            "status": "Running script prepare() step",
 | 
			
		||||
        },
 | 
			
		||||
    )
 | 
			
		||||
    time.sleep(1)
 | 
			
		||||
 | 
			
		||||
    print("Running script prepare() step")
 | 
			
		||||
 | 
			
		||||
    with chroot(temp_dir):
 | 
			
		||||
        vm_builder.prepare()
 | 
			
		||||
 | 
			
		||||
    # Phase 7 - script: install()
 | 
			
		||||
    #  * Run installation with arguments
 | 
			
		||||
    self.update_state(
 | 
			
		||||
        state="RUNNING",
 | 
			
		||||
        meta={
 | 
			
		||||
            "current": 7,
 | 
			
		||||
            "total": 10,
 | 
			
		||||
            "status": "Running script install() step",
 | 
			
		||||
        },
 | 
			
		||||
    )
 | 
			
		||||
    time.sleep(1)
 | 
			
		||||
 | 
			
		||||
    print("Running script install() step")
 | 
			
		||||
 | 
			
		||||
    with chroot(temp_dir):
 | 
			
		||||
        vm_builder.install()
 | 
			
		||||
 | 
			
		||||
    # Phase 8 - script: cleanup()
 | 
			
		||||
    #  * Run cleanup steps
 | 
			
		||||
    self.update_state(
 | 
			
		||||
        state="RUNNING",
 | 
			
		||||
        meta={
 | 
			
		||||
            "current": 8,
 | 
			
		||||
            "total": 10,
 | 
			
		||||
            "status": "Running script cleanup() step",
 | 
			
		||||
        },
 | 
			
		||||
    )
 | 
			
		||||
    time.sleep(1)
 | 
			
		||||
 | 
			
		||||
    print("Running script cleanup() step")
 | 
			
		||||
 | 
			
		||||
    with chroot(temp_dir):
 | 
			
		||||
        vm_builder.cleanup()
 | 
			
		||||
 | 
			
		||||
    # Phase 9 - general cleanup
 | 
			
		||||
    #  * Clean up the chroot from earlier
 | 
			
		||||
    self.update_state(
 | 
			
		||||
        state="RUNNING",
 | 
			
		||||
        meta={
 | 
			
		||||
            "current": 9,
 | 
			
		||||
            "total": 10,
 | 
			
		||||
            "status": "Running upper cleanup steps",
 | 
			
		||||
        },
 | 
			
		||||
    )
 | 
			
		||||
    time.sleep(1)
 | 
			
		||||
 | 
			
		||||
    print("Running upper cleanup steps")
 | 
			
		||||
 | 
			
		||||
    # Remote temporary script (don't fail if not removed)
 | 
			
		||||
    if not os.remove(script_file):
 | 
			
		||||
        print(f"Failed to delete temporary script file '{script_file}'.")
 | 
			
		||||
 | 
			
		||||
    # Unmount bind-mounted devfs on the chroot
 | 
			
		||||
    retcode, stdout, stderr = pvc_common.run_os_command(f"umount {temp_dir}/dev")
 | 
			
		||||
    if retcode:
 | 
			
		||||
        raise ProvisioningError(f"Failed to unmount devfs from chroot: {stderr}")
 | 
			
		||||
 | 
			
		||||
    # Unmount bind-mounted runfs on the chroot
 | 
			
		||||
    retcode, stdout, stderr = pvc_common.run_os_command(f"umount {temp_dir}/run")
 | 
			
		||||
    if retcode:
 | 
			
		||||
        raise ProvisioningError(f"Failed to unmount runfs from chroot: {stderr}")
 | 
			
		||||
 | 
			
		||||
    # Unmount bind-mounted sysfs on the chroot
 | 
			
		||||
    retcode, stdout, stderr = pvc_common.run_os_command(f"umount {temp_dir}/sys")
 | 
			
		||||
    if retcode:
 | 
			
		||||
        raise ProvisioningError(f"Failed to unmount sysfs from chroot: {stderr}")
 | 
			
		||||
 | 
			
		||||
    # Unmount bind-mounted tmpfs on the chroot
 | 
			
		||||
    retcode, stdout, stderr = pvc_common.run_os_command(f"umount {temp_dir}/tmp")
 | 
			
		||||
    if retcode:
 | 
			
		||||
        raise ProvisioningError(f"Failed to unmount tmpfs from chroot: {stderr}")
 | 
			
		||||
 | 
			
		||||
    # Unmount bind-mounted rootfs on the chroot
 | 
			
		||||
    retcode, stdout, stderr = pvc_common.run_os_command(f"umount {temp_dir}")
 | 
			
		||||
    if retcode:
 | 
			
		||||
        raise ProvisioningError(f"Failed to unmount rootfs from chroot: {stderr}")
 | 
			
		||||
 | 
			
		||||
    # Remove the temp_dir
 | 
			
		||||
    if not os.rmdir(temp_dir):
 | 
			
		||||
        print(f"Failed to delete temporary chroot directory '{temp_dir}'.")
 | 
			
		||||
 | 
			
		||||
    # Phase 10 - startup
 | 
			
		||||
    #  * Start the VM in the PVC cluster
 | 
			
		||||
    self.update_state(
 | 
			
		||||
        state="RUNNING",
 | 
			
		||||
        meta={
 | 
			
		||||
            "current": 10,
 | 
			
		||||
            "total": 10,
 | 
			
		||||
            "status": "Starting VM",
 | 
			
		||||
        },
 | 
			
		||||
    )
 | 
			
		||||
    time.sleep(1)
 | 
			
		||||
 | 
			
		||||
    if start_vm:
 | 
			
		||||
        print("Starting VM")
 | 
			
		||||
        with open_zk(config) as zkhandler:
 | 
			
		||||
            success, message = pvc_vm.start_vm(zkhandler, vm_name)
 | 
			
		||||
        print(message)
 | 
			
		||||
		Reference in New Issue
	
	Block a user