Add rinse example configuration
Provisions Rocky Linux 8 and 9 systems, and potentially older CentOS/Fedora/Scientific Linux/SuSE systems. Depends on a custom build of rinse (3.7.1) with Rocky 9 support.
This commit is contained in:
		
							
								
								
									
										741
									
								
								api-daemon/provisioner/examples/script/4-rinse.py
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										741
									
								
								api-daemon/provisioner/examples/script/4-rinse.py
									
									
									
									
									
										Normal file
									
								
							@@ -0,0 +1,741 @@
 | 
				
			|||||||
 | 
					#!/usr/bin/env python3
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					# 4-rinse.py - PVC Provisioner example script for rinse install
 | 
				
			||||||
 | 
					# Part of the Parallel Virtual Cluster (PVC) system
 | 
				
			||||||
 | 
					#
 | 
				
			||||||
 | 
					#    Copyright (C) 2018-2022 Joshua M. Boniface <joshua@boniface.me>
 | 
				
			||||||
 | 
					#
 | 
				
			||||||
 | 
					#    This program is free software: you can redistribute it and/or modify
 | 
				
			||||||
 | 
					#    it under the terms of the GNU General Public License as published by
 | 
				
			||||||
 | 
					#    the Free Software Foundation, version 3.
 | 
				
			||||||
 | 
					#
 | 
				
			||||||
 | 
					#    This program is distributed in the hope that it will be useful,
 | 
				
			||||||
 | 
					#    but WITHOUT ANY WARRANTY; without even the implied warranty of
 | 
				
			||||||
 | 
					#    MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
 | 
				
			||||||
 | 
					#    GNU General Public License for more details.
 | 
				
			||||||
 | 
					#
 | 
				
			||||||
 | 
					#    You should have received a copy of the GNU General Public License
 | 
				
			||||||
 | 
					#    along with this program.  If not, see <https://www.gnu.org/licenses/>.
 | 
				
			||||||
 | 
					#
 | 
				
			||||||
 | 
					###############################################################################
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					# This script provides an example of a PVC provisioner script. It will create a
 | 
				
			||||||
 | 
					# standard VM config and install a RHEL-like OS using rinse.
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					# This script can thus be used as an example or reference implementation of a
 | 
				
			||||||
 | 
					# PVC provisioner script and expanded upon as required.
 | 
				
			||||||
 | 
					# *** READ THIS SCRIPT THOROUGHLY BEFORE USING TO UNDERSTAND HOW IT WORKS. ***
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					# A script must implement the class "VMBuilderScript" which extends "VMBuilder",
 | 
				
			||||||
 | 
					# providing the 5 functions indicated. Detailed explanation of the role of each
 | 
				
			||||||
 | 
					# function is provided in context of the example; see the other examples for
 | 
				
			||||||
 | 
					# more potential uses.
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					# Within the VMBuilderScript class, several common variables are exposed through
 | 
				
			||||||
 | 
					# the parent VMBuilder class:
 | 
				
			||||||
 | 
					#  self.vm_name: The name of the VM from PVC's perspective
 | 
				
			||||||
 | 
					#  self.vm_id: The VM ID (numerical component of the vm_name) from PVC's perspective
 | 
				
			||||||
 | 
					#  self.vm_uuid: An automatically-generated UUID for the VM
 | 
				
			||||||
 | 
					#  self.vm_profile: The PVC provisioner profile name used for the VM
 | 
				
			||||||
 | 
					#  self.vm_data: A dictionary of VM data collected by the provisioner; as an example:
 | 
				
			||||||
 | 
					#    {
 | 
				
			||||||
 | 
					#      "ceph_monitor_list": [
 | 
				
			||||||
 | 
					#        "hv1.pvcstorage.tld",
 | 
				
			||||||
 | 
					#        "hv2.pvcstorage.tld",
 | 
				
			||||||
 | 
					#        "hv3.pvcstorage.tld"
 | 
				
			||||||
 | 
					#      ],
 | 
				
			||||||
 | 
					#      "ceph_monitor_port": "6789",
 | 
				
			||||||
 | 
					#      "ceph_monitor_secret": "96721723-8650-4a72-b8f6-a93cd1a20f0c",
 | 
				
			||||||
 | 
					#      "mac_template": null,
 | 
				
			||||||
 | 
					#      "networks": [
 | 
				
			||||||
 | 
					#        {
 | 
				
			||||||
 | 
					#          "eth_bridge": "vmbr1001",
 | 
				
			||||||
 | 
					#          "id": 72,
 | 
				
			||||||
 | 
					#          "network_template": 69,
 | 
				
			||||||
 | 
					#          "vni": "1001"
 | 
				
			||||||
 | 
					#        },
 | 
				
			||||||
 | 
					#        {
 | 
				
			||||||
 | 
					#          "eth_bridge": "vmbr101",
 | 
				
			||||||
 | 
					#          "id": 73,
 | 
				
			||||||
 | 
					#          "network_template": 69,
 | 
				
			||||||
 | 
					#          "vni": "101"
 | 
				
			||||||
 | 
					#        }
 | 
				
			||||||
 | 
					#      ],
 | 
				
			||||||
 | 
					#      "script": [contents of this file]
 | 
				
			||||||
 | 
					#      "script_arguments": {
 | 
				
			||||||
 | 
					#        "deb_mirror": "http://ftp.debian.org/debian",
 | 
				
			||||||
 | 
					#        "deb_release": "bullseye"
 | 
				
			||||||
 | 
					#      },
 | 
				
			||||||
 | 
					#      "system_architecture": "x86_64",
 | 
				
			||||||
 | 
					#      "system_details": {
 | 
				
			||||||
 | 
					#        "id": 78,
 | 
				
			||||||
 | 
					#        "migration_method": "live",
 | 
				
			||||||
 | 
					#        "name": "small",
 | 
				
			||||||
 | 
					#        "node_autostart": false,
 | 
				
			||||||
 | 
					#        "node_limit": null,
 | 
				
			||||||
 | 
					#        "node_selector": null,
 | 
				
			||||||
 | 
					#        "ova": null,
 | 
				
			||||||
 | 
					#        "serial": true,
 | 
				
			||||||
 | 
					#        "vcpu_count": 2,
 | 
				
			||||||
 | 
					#        "vnc": false,
 | 
				
			||||||
 | 
					#        "vnc_bind": null,
 | 
				
			||||||
 | 
					#        "vram_mb": 2048
 | 
				
			||||||
 | 
					#      },
 | 
				
			||||||
 | 
					#      "volumes": [
 | 
				
			||||||
 | 
					#        {
 | 
				
			||||||
 | 
					#          "disk_id": "sda",
 | 
				
			||||||
 | 
					#          "disk_size_gb": 4,
 | 
				
			||||||
 | 
					#          "filesystem": "ext4",
 | 
				
			||||||
 | 
					#          "filesystem_args": "-L=root",
 | 
				
			||||||
 | 
					#          "id": 9,
 | 
				
			||||||
 | 
					#          "mountpoint": "/",
 | 
				
			||||||
 | 
					#          "pool": "vms",
 | 
				
			||||||
 | 
					#          "source_volume": null,
 | 
				
			||||||
 | 
					#          "storage_template": 67
 | 
				
			||||||
 | 
					#        },
 | 
				
			||||||
 | 
					#        {
 | 
				
			||||||
 | 
					#          "disk_id": "sdb",
 | 
				
			||||||
 | 
					#          "disk_size_gb": 4,
 | 
				
			||||||
 | 
					#          "filesystem": "ext4",
 | 
				
			||||||
 | 
					#          "filesystem_args": "-L=var",
 | 
				
			||||||
 | 
					#          "id": 10,
 | 
				
			||||||
 | 
					#          "mountpoint": "/var",
 | 
				
			||||||
 | 
					#          "pool": "vms",
 | 
				
			||||||
 | 
					#          "source_volume": null,
 | 
				
			||||||
 | 
					#          "storage_template": 67
 | 
				
			||||||
 | 
					#        },
 | 
				
			||||||
 | 
					#        {
 | 
				
			||||||
 | 
					#          "disk_id": "sdc",
 | 
				
			||||||
 | 
					#          "disk_size_gb": 4,
 | 
				
			||||||
 | 
					#          "filesystem": "ext4",
 | 
				
			||||||
 | 
					#          "filesystem_args": "-L=log",
 | 
				
			||||||
 | 
					#          "id": 11,
 | 
				
			||||||
 | 
					#          "mountpoint": "/var/log",
 | 
				
			||||||
 | 
					#          "pool": "vms",
 | 
				
			||||||
 | 
					#          "source_volume": null,
 | 
				
			||||||
 | 
					#          "storage_template": 67
 | 
				
			||||||
 | 
					#        }
 | 
				
			||||||
 | 
					#      ]
 | 
				
			||||||
 | 
					#    }
 | 
				
			||||||
 | 
					#
 | 
				
			||||||
 | 
					# Any other information you may require must be obtained manually.
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					# WARNING:
 | 
				
			||||||
 | 
					#
 | 
				
			||||||
 | 
					# For safety reasons, the script runs in a modified chroot. It will have full access to
 | 
				
			||||||
 | 
					# the entire / (root partition) of the hypervisor, but read-only. In addition it has
 | 
				
			||||||
 | 
					# access to /dev, /sys, /run, and a fresh /tmp to write to; use /tmp/target (as
 | 
				
			||||||
 | 
					# convention) as the destination for any mounting of volumes and installation.
 | 
				
			||||||
 | 
					# Of course, in addition to this safety, it is VERY IMPORTANT to be aware that this
 | 
				
			||||||
 | 
					# script runs AS ROOT ON THE HYPERVISOR SYSTEM. You should never allow arbitrary,
 | 
				
			||||||
 | 
					# untrusted users the ability to add provisioning scripts even with this safeguard,
 | 
				
			||||||
 | 
					# since they could still do destructive things to /dev and the like!
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					# This import is always required here, as VMBuilder is used by the VMBuilderScript class
 | 
				
			||||||
 | 
					# and ProvisioningError is the primary exception that should be raised within the class.
 | 
				
			||||||
 | 
					from pvcapid.vmbuilder import VMBuilder, ProvisioningError
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					# The VMBuilderScript class must be named as such, and extend VMBuilder.
 | 
				
			||||||
 | 
					class VMBuilderScript(VMBuilder):
 | 
				
			||||||
 | 
					    def setup(self):
 | 
				
			||||||
 | 
					        """
 | 
				
			||||||
 | 
					        setup(): Perform special setup steps or validation before proceeding
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        This example uses the PVC built-in command runner to verify that rinse is
 | 
				
			||||||
 | 
					        installed and throws and error if not.
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        Note that, due to the aforementioned chroot, you *cannot* install or otherwise
 | 
				
			||||||
 | 
					        modify the hypervisor system here: any tooling, etc. must be pre-installed.
 | 
				
			||||||
 | 
					        """
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        # Run any imports first; as shown here, you can import anything from the PVC
 | 
				
			||||||
 | 
					        # namespace, as well as (of course) the main Python namespaces
 | 
				
			||||||
 | 
					        import daemon_lib.common as pvc_common
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        # Ensure we have debootstrap intalled on the provisioner system
 | 
				
			||||||
 | 
					        retcode, stdout, stderr = pvc_common.run_os_command(f"which rinse")
 | 
				
			||||||
 | 
					        if retcode:
 | 
				
			||||||
 | 
					            # Raise a ProvisioningError for any exception; the provisioner will handle
 | 
				
			||||||
 | 
					            # this gracefully and properly, avoiding dangling mounts, RBD maps, etc.
 | 
				
			||||||
 | 
					            raise ProvisioningError("Failed to find critical dependency: rinse")
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    def create(self):
 | 
				
			||||||
 | 
					        """
 | 
				
			||||||
 | 
					        create(): Create the VM libvirt schema definition
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        This step *must* return a fully-formed Libvirt XML document as a string or the
 | 
				
			||||||
 | 
					        provisioning task will fail.
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        This example leverages the built-in libvirt_schema objects provided by PVC; these
 | 
				
			||||||
 | 
					        can be used as-is, or replaced with your own schema(s) on a per-script basis.
 | 
				
			||||||
 | 
					        """
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        # Run any imports first
 | 
				
			||||||
 | 
					        import pvcapid.libvirt_schema as libvirt_schema
 | 
				
			||||||
 | 
					        import datetime
 | 
				
			||||||
 | 
					        import random
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        # Create the empty schema document that we will append to and return at the end
 | 
				
			||||||
 | 
					        schema = ""
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        # Prepare a description based on the VM profile
 | 
				
			||||||
 | 
					        description = (
 | 
				
			||||||
 | 
					            f"PVC provisioner @ {datetime.datetime.now()}, profile '{self.vm_profile}'"
 | 
				
			||||||
 | 
					        )
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        # Format the header
 | 
				
			||||||
 | 
					        schema += libvirt_schema.libvirt_header.format(
 | 
				
			||||||
 | 
					            vm_name=self.vm_name,
 | 
				
			||||||
 | 
					            vm_uuid=self.vm_uuid,
 | 
				
			||||||
 | 
					            vm_description=description,
 | 
				
			||||||
 | 
					            vm_memory=self.vm_data["system_details"]["vram_mb"],
 | 
				
			||||||
 | 
					            vm_vcpus=self.vm_data["system_details"]["vcpu_count"],
 | 
				
			||||||
 | 
					            vm_architecture=self.vm_data["system_architecture"],
 | 
				
			||||||
 | 
					        )
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        # Add the disk devices
 | 
				
			||||||
 | 
					        monitor_list = self.vm_data["ceph_monitor_list"]
 | 
				
			||||||
 | 
					        monitor_port = self.vm_data["ceph_monitor_port"]
 | 
				
			||||||
 | 
					        monitor_secret = self.vm_data["ceph_monitor_secret"]
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        for volume in self.vm_data["volumes"]:
 | 
				
			||||||
 | 
					            schema += libvirt_schema.devices_disk_header.format(
 | 
				
			||||||
 | 
					                ceph_storage_secret=monitor_secret,
 | 
				
			||||||
 | 
					                disk_pool=volume["pool"],
 | 
				
			||||||
 | 
					                vm_name=self.vm_name,
 | 
				
			||||||
 | 
					                disk_id=volume["disk_id"],
 | 
				
			||||||
 | 
					            )
 | 
				
			||||||
 | 
					            for monitor in monitor_list:
 | 
				
			||||||
 | 
					                schema += libvirt_schema.devices_disk_coordinator.format(
 | 
				
			||||||
 | 
					                    coordinator_name=monitor,
 | 
				
			||||||
 | 
					                    coordinator_ceph_mon_port=monitor_port,
 | 
				
			||||||
 | 
					                )
 | 
				
			||||||
 | 
					            schema += libvirt_schema.devices_disk_footer
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        # Add the special vhostmd device for hypervisor information inside the VM
 | 
				
			||||||
 | 
					        schema += libvirt_schema.devices_vhostmd
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        # Add the network devices
 | 
				
			||||||
 | 
					        network_id = 0
 | 
				
			||||||
 | 
					        for network in self.vm_data["networks"]:
 | 
				
			||||||
 | 
					            vm_id_hex = "{:x}".format(int(self.vm_id % 16))
 | 
				
			||||||
 | 
					            net_id_hex = "{:x}".format(int(network_id % 16))
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					            if self.vm_data.get("mac_template") is not None:
 | 
				
			||||||
 | 
					                mac_prefix = "52:54:01"
 | 
				
			||||||
 | 
					                macgen_template = self.vm_data["mac_template"]
 | 
				
			||||||
 | 
					                eth_macaddr = macgen_template.format(
 | 
				
			||||||
 | 
					                    prefix=mac_prefix, vmid=vm_id_hex, netid=net_id_hex
 | 
				
			||||||
 | 
					                )
 | 
				
			||||||
 | 
					            else:
 | 
				
			||||||
 | 
					                mac_prefix = "52:54:00"
 | 
				
			||||||
 | 
					                random_octet_A = "{:x}".format(random.randint(16, 238))
 | 
				
			||||||
 | 
					                random_octet_B = "{:x}".format(random.randint(16, 238))
 | 
				
			||||||
 | 
					                random_octet_C = "{:x}".format(random.randint(16, 238))
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					                macgen_template = "{prefix}:{octetA}:{octetB}:{octetC}"
 | 
				
			||||||
 | 
					                eth_macaddr = macgen_template.format(
 | 
				
			||||||
 | 
					                    prefix=mac_prefix,
 | 
				
			||||||
 | 
					                    octetA=random_octet_A,
 | 
				
			||||||
 | 
					                    octetB=random_octet_B,
 | 
				
			||||||
 | 
					                    octetC=random_octet_C,
 | 
				
			||||||
 | 
					                )
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					            schema += libvirt_schema.devices_net_interface.format(
 | 
				
			||||||
 | 
					                eth_macaddr=eth_macaddr,
 | 
				
			||||||
 | 
					                eth_bridge=network["eth_bridge"],
 | 
				
			||||||
 | 
					            )
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					            network_id += 1
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        # Add default devices
 | 
				
			||||||
 | 
					        schema += libvirt_schema.devices_default
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        # Add serial device
 | 
				
			||||||
 | 
					        if self.vm_data["system_details"]["serial"]:
 | 
				
			||||||
 | 
					            schema += libvirt_schema.devices_serial.format(vm_name=self.vm_name)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        # Add VNC device
 | 
				
			||||||
 | 
					        if self.vm_data["system_details"]["vnc"]:
 | 
				
			||||||
 | 
					            if self.vm_data["system_details"]["vnc_bind"]:
 | 
				
			||||||
 | 
					                vm_vnc_bind = self.vm_data["system_details"]["vnc_bind"]
 | 
				
			||||||
 | 
					            else:
 | 
				
			||||||
 | 
					                vm_vnc_bind = "127.0.0.1"
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					            vm_vncport = 5900
 | 
				
			||||||
 | 
					            vm_vnc_autoport = "yes"
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					            schema += libvirt_schema.devices_vnc.format(
 | 
				
			||||||
 | 
					                vm_vncport=vm_vncport,
 | 
				
			||||||
 | 
					                vm_vnc_autoport=vm_vnc_autoport,
 | 
				
			||||||
 | 
					                vm_vnc_bind=vm_vnc_bind,
 | 
				
			||||||
 | 
					            )
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        # Add SCSI controller
 | 
				
			||||||
 | 
					        schema += libvirt_schema.devices_scsi_controller
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        # Add footer
 | 
				
			||||||
 | 
					        schema += libvirt_schema.libvirt_footer
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        return schema
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    def prepare(self):
 | 
				
			||||||
 | 
					        """
 | 
				
			||||||
 | 
					        prepare(): Prepare any disks/volumes for the install() step
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        This function should use the various exposed PVC commands as indicated to create
 | 
				
			||||||
 | 
					        RBD block devices and map them to the host as required.
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        open_zk is exposed from pvcapid.vmbuilder to provide a context manager for opening
 | 
				
			||||||
 | 
					        connections to the PVC Zookeeper cluster; ensure you also import (and pass it)
 | 
				
			||||||
 | 
					        the config object from pvcapid.Daemon as well. This context manager then allows
 | 
				
			||||||
 | 
					        the use of various common daemon library functions, without going through the API.
 | 
				
			||||||
 | 
					        """
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        # Run any imports first
 | 
				
			||||||
 | 
					        import os
 | 
				
			||||||
 | 
					        from pvcapid.vmbuilder import open_zk
 | 
				
			||||||
 | 
					        from pvcapid.Daemon import config
 | 
				
			||||||
 | 
					        import daemon_lib.common as pvc_common
 | 
				
			||||||
 | 
					        import daemon_lib.ceph as pvc_ceph
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        # First loop: Create the disks, either by cloning (pvc_ceph.clone_volume), or by
 | 
				
			||||||
 | 
					        # new creation (pvc_ceph.add_volume), depending on the source_volume entry
 | 
				
			||||||
 | 
					        for volume in self.vm_data["volumes"]:
 | 
				
			||||||
 | 
					            if volume.get("source_volume") is not None:
 | 
				
			||||||
 | 
					                with open_zk(config) as zkhandler:
 | 
				
			||||||
 | 
					                    success, message = pvc_ceph.clone_volume(
 | 
				
			||||||
 | 
					                        zkhandler,
 | 
				
			||||||
 | 
					                        volume["pool"],
 | 
				
			||||||
 | 
					                        volume["source_volume"],
 | 
				
			||||||
 | 
					                        f"{self.vm_name}_{volume['disk_id']}",
 | 
				
			||||||
 | 
					                    )
 | 
				
			||||||
 | 
					                    print(message)
 | 
				
			||||||
 | 
					                    if not success:
 | 
				
			||||||
 | 
					                        raise ProvisioningError(
 | 
				
			||||||
 | 
					                            f"Failed to clone volume '{volume['source_volume']}' to '{volume['disk_id']}'."
 | 
				
			||||||
 | 
					                        )
 | 
				
			||||||
 | 
					            else:
 | 
				
			||||||
 | 
					                with open_zk(config) as zkhandler:
 | 
				
			||||||
 | 
					                    success, message = pvc_ceph.add_volume(
 | 
				
			||||||
 | 
					                        zkhandler,
 | 
				
			||||||
 | 
					                        volume["pool"],
 | 
				
			||||||
 | 
					                        f"{self.vm_name}_{volume['disk_id']}",
 | 
				
			||||||
 | 
					                        f"{volume['disk_size_gb']}G",
 | 
				
			||||||
 | 
					                    )
 | 
				
			||||||
 | 
					                    print(message)
 | 
				
			||||||
 | 
					                    if not success:
 | 
				
			||||||
 | 
					                        raise ProvisioningError(
 | 
				
			||||||
 | 
					                            f"Failed to create volume '{volume['disk_id']}'."
 | 
				
			||||||
 | 
					                        )
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        # Second loop: Map the disks to the local system
 | 
				
			||||||
 | 
					        for volume in self.vm_data["volumes"]:
 | 
				
			||||||
 | 
					            dst_volume_name = f"{self.vm_name}_{volume['disk_id']}"
 | 
				
			||||||
 | 
					            dst_volume = f"{volume['pool']}/{dst_volume_name}"
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					            with open_zk(config) as zkhandler:
 | 
				
			||||||
 | 
					                success, message = pvc_ceph.map_volume(
 | 
				
			||||||
 | 
					                    zkhandler,
 | 
				
			||||||
 | 
					                    volume["pool"],
 | 
				
			||||||
 | 
					                    dst_volume_name,
 | 
				
			||||||
 | 
					                )
 | 
				
			||||||
 | 
					                print(message)
 | 
				
			||||||
 | 
					                if not success:
 | 
				
			||||||
 | 
					                    raise ProvisioningError(f"Failed to map volume '{dst_volume}'.")
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        # Third loop: Create filesystems on the volumes
 | 
				
			||||||
 | 
					        for volume in self.vm_data["volumes"]:
 | 
				
			||||||
 | 
					            dst_volume_name = f"{self.vm_name}_{volume['disk_id']}"
 | 
				
			||||||
 | 
					            dst_volume = f"{volume['pool']}/{dst_volume_name}"
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					            if volume.get("source_volume") is not None:
 | 
				
			||||||
 | 
					                continue
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					            if volume.get("filesystem") is None:
 | 
				
			||||||
 | 
					                continue
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					            filesystem_args_list = list()
 | 
				
			||||||
 | 
					            for arg in volume["filesystem_args"].split():
 | 
				
			||||||
 | 
					                arg_entry, *arg_data = arg.split("=")
 | 
				
			||||||
 | 
					                arg_data = "=".join(arg_data)
 | 
				
			||||||
 | 
					                filesystem_args_list.append(arg_entry)
 | 
				
			||||||
 | 
					                filesystem_args_list.append(arg_data)
 | 
				
			||||||
 | 
					            filesystem_args = " ".join(filesystem_args_list)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					            if volume["filesystem"] == "swap":
 | 
				
			||||||
 | 
					                retcode, stdout, stderr = pvc_common.run_os_command(
 | 
				
			||||||
 | 
					                    f"mkswap -f /dev/rbd/{dst_volume}"
 | 
				
			||||||
 | 
					                )
 | 
				
			||||||
 | 
					                if retcode:
 | 
				
			||||||
 | 
					                    raise ProvisioningError(
 | 
				
			||||||
 | 
					                        f"Failed to create swap on '{dst_volume}': {stderr}"
 | 
				
			||||||
 | 
					                    )
 | 
				
			||||||
 | 
					            else:
 | 
				
			||||||
 | 
					                retcode, stdout, stderr = pvc_common.run_os_command(
 | 
				
			||||||
 | 
					                    f"mkfs.{volume['filesystem']} {filesystem_args} /dev/rbd/{dst_volume}"
 | 
				
			||||||
 | 
					                )
 | 
				
			||||||
 | 
					                if retcode:
 | 
				
			||||||
 | 
					                    raise ProvisioningError(
 | 
				
			||||||
 | 
					                        f"Faield to create {volume['filesystem']} file on '{dst_volume}': {stderr}"
 | 
				
			||||||
 | 
					                    )
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					            print(stdout)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        # Create a temporary directory to use during install
 | 
				
			||||||
 | 
					        temp_dir = "/tmp/target"
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        if not os.path.isdir(temp_dir):
 | 
				
			||||||
 | 
					            os.mkdir(temp_dir)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        # Fourth loop: Mount the volumes to a set of temporary directories
 | 
				
			||||||
 | 
					        for volume in self.vm_data["volumes"]:
 | 
				
			||||||
 | 
					            dst_volume_name = f"{self.vm_name}_{volume['disk_id']}"
 | 
				
			||||||
 | 
					            dst_volume = f"{volume['pool']}/{dst_volume_name}"
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					            if volume.get("source_volume") is not None:
 | 
				
			||||||
 | 
					                continue
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					            if volume.get("filesystem") is None:
 | 
				
			||||||
 | 
					                continue
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					            mapped_dst_volume = f"/dev/rbd/{dst_volume}"
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					            mount_path = f"{temp_dir}/{volume['mountpoint']}"
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					            if not os.path.isdir(mount_path):
 | 
				
			||||||
 | 
					                os.mkdir(mount_path)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					            # Mount filesystem
 | 
				
			||||||
 | 
					            retcode, stdout, stderr = pvc_common.run_os_command(
 | 
				
			||||||
 | 
					                f"mount {mapped_dst_volume} {mount_path}"
 | 
				
			||||||
 | 
					            )
 | 
				
			||||||
 | 
					            if retcode:
 | 
				
			||||||
 | 
					                raise ProvisioningError(
 | 
				
			||||||
 | 
					                    f"Failed to mount '{mapped_dst_volume}' on '{mount_path}': {stderr}"
 | 
				
			||||||
 | 
					                )
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    def install(self):
 | 
				
			||||||
 | 
					        """
 | 
				
			||||||
 | 
					        install(): Perform the installation
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        This example, unlike noop, performs a full rinse install and base config of a
 | 
				
			||||||
 | 
					        RHEL-like system, including installing GRUB for fully-virtualized boot (required
 | 
				
			||||||
 | 
					        by PVC) and cloud-init for later configuration with the PVC userdata
 | 
				
			||||||
 | 
					        functionality, leveraging a PVC managed network on the first NIC for DHCP.
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        Several arguments are also supported; these can be set either in the provisioner
 | 
				
			||||||
 | 
					        profile itself, or on the command line at runtime.
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        To show the options, this function does not use the previous PVC-exposed
 | 
				
			||||||
 | 
					        run_os_command function, but instead just uses os.system. The downside here is
 | 
				
			||||||
 | 
					        a lack of response and error handling, but the upside is simpler-to-read code.
 | 
				
			||||||
 | 
					        Use whichever you feel is appropriate for your situation.
 | 
				
			||||||
 | 
					        """
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        # Run any imports first
 | 
				
			||||||
 | 
					        import os
 | 
				
			||||||
 | 
					        from pvcapid.vmbuilder import chroot
 | 
				
			||||||
 | 
					        import daemon_lib.common as pvc_common
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        # The directory we mounted things on earlier during prepare(); this could very well
 | 
				
			||||||
 | 
					        # be exposed as a module-level variable if you so choose
 | 
				
			||||||
 | 
					        temporary_directory = "/tmp/target"
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        # A cache directory for rinse under tmp
 | 
				
			||||||
 | 
					        rinse_cache = "/tmp/rinse"
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        # Use these convenient aliases for later (avoiding lots of "self.vm_data" everywhere)
 | 
				
			||||||
 | 
					        vm_name = self.vm_name
 | 
				
			||||||
 | 
					        volumes = self.vm_data["volumes"]
 | 
				
			||||||
 | 
					        networks = self.vm_data["networks"]
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        # Parse these arguments out of self.vm_data["script_arguments"]
 | 
				
			||||||
 | 
					        if self.vm_data["script_arguments"].get("rinse_release") is not None:
 | 
				
			||||||
 | 
					            rinse_release = self.vm_data["script_arguments"].get("rinse_release")
 | 
				
			||||||
 | 
					        else:
 | 
				
			||||||
 | 
					            rinse_release = "rocky-9"
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        if self.vm_data["script_arguments"].get("rinse_mirror") is not None:
 | 
				
			||||||
 | 
					            rinse_mirror = self.vm_data["script_arguments"].get("rinse_mirror")
 | 
				
			||||||
 | 
					        else:
 | 
				
			||||||
 | 
					            # Use the defaults from /etc/rinse/rinse.conf
 | 
				
			||||||
 | 
					            rinse_mirror = None
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        if self.vm_data["script_arguments"].get("rinse_packages") is not None:
 | 
				
			||||||
 | 
					            deb_packages = (
 | 
				
			||||||
 | 
					                self.vm_data["script_arguments"].get("rinse_packages").split(",")
 | 
				
			||||||
 | 
					            )
 | 
				
			||||||
 | 
					        else:
 | 
				
			||||||
 | 
					            rinse_packages = [
 | 
				
			||||||
 | 
					                "kernel",
 | 
				
			||||||
 | 
					                "kernel-core",
 | 
				
			||||||
 | 
					                "linux-firmware",
 | 
				
			||||||
 | 
					                "grub2-pc",
 | 
				
			||||||
 | 
					                "grub2-pc-modules",
 | 
				
			||||||
 | 
					                "NetworkManager",
 | 
				
			||||||
 | 
					                "libndp",
 | 
				
			||||||
 | 
					            ]
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        if self.vm_data["script_arguments"].get("post_packages") is not None:
 | 
				
			||||||
 | 
					            deb_packages = (
 | 
				
			||||||
 | 
					                self.vm_data["script_arguments"].get("post_packages").split(",")
 | 
				
			||||||
 | 
					            )
 | 
				
			||||||
 | 
					        else:
 | 
				
			||||||
 | 
					            post_packages = ["cloud-init"]
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        # We need to know our root disk for later GRUB-ing
 | 
				
			||||||
 | 
					        root_disk = None
 | 
				
			||||||
 | 
					        for volume in volumes:
 | 
				
			||||||
 | 
					            if volume["mountpoint"] == "/":
 | 
				
			||||||
 | 
					                root_volume = volume
 | 
				
			||||||
 | 
					        if not root_volume:
 | 
				
			||||||
 | 
					            raise ProvisioningError("Failed to find root volume in volumes list")
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        if rinse_mirror is not None:
 | 
				
			||||||
 | 
					            mirror_arg = f"--mirror {rinse_mirror}"
 | 
				
			||||||
 | 
					        else:
 | 
				
			||||||
 | 
					            mirror_arg = ""
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        # Get the current architecture for rinse to use
 | 
				
			||||||
 | 
					        _, stdout, _ = pvc_common.run_os_command("dpkg --print-architecture")
 | 
				
			||||||
 | 
					        rinse_architecture = stdout.strip()
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        # Write a temporary file containing the additional packages
 | 
				
			||||||
 | 
					        with open(f"/tmp/addpkg", "w") as fh:
 | 
				
			||||||
 | 
					            for pkg in rinse_packages:
 | 
				
			||||||
 | 
					                fh.write(f"{pkg}\n")
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        # Perform a rinse installation
 | 
				
			||||||
 | 
					        print(
 | 
				
			||||||
 | 
					            f"Installing system with rinse: rinse --arch {rinse_architecture} --directory {temporary_directory} --distribution {rinse_release} --cache-dir {rinse_cache} --add-pkg-list /tmp/addpkg --verbose {mirror_arg}"
 | 
				
			||||||
 | 
					        )
 | 
				
			||||||
 | 
					        os.system(
 | 
				
			||||||
 | 
					            f"rinse --arch {rinse_architecture} --directory {temporary_directory} --distribution {rinse_release} --cache-dir {rinse_cache} --add-pkg-list /tmp/addpkg --verbose {mirror_arg}"
 | 
				
			||||||
 | 
					        )
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        # Bind mount the devfs, sysfs, and procfs so we can grub-install later
 | 
				
			||||||
 | 
					        os.system("mount --bind /dev {}/dev".format(temporary_directory))
 | 
				
			||||||
 | 
					        os.system("mount --bind /sys {}/sys".format(temporary_directory))
 | 
				
			||||||
 | 
					        os.system("mount --bind /proc {}/proc".format(temporary_directory))
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        # Create an fstab entry for each volume
 | 
				
			||||||
 | 
					        fstab_file = "{}/etc/fstab".format(temporary_directory)
 | 
				
			||||||
 | 
					        # The volume ID starts at zero and increments by one for each volume in the fixed-order
 | 
				
			||||||
 | 
					        # volume list. This lets us work around the insanity of Libvirt IDs not matching guest IDs,
 | 
				
			||||||
 | 
					        # while still letting us have some semblance of control here without enforcing things
 | 
				
			||||||
 | 
					        # like labels. It increments in the for loop below at the end of each iteration, and is
 | 
				
			||||||
 | 
					        # used to craft a /dev/disk/by-id/scsi-0QEMU_QEMU_HARDDISK_drive-scsi0-0-0-X device ID
 | 
				
			||||||
 | 
					        # which will always match the correct order from Libvirt (unlike sdX/vdX names).
 | 
				
			||||||
 | 
					        volume_id = 0
 | 
				
			||||||
 | 
					        for volume in volumes:
 | 
				
			||||||
 | 
					            # We assume SSD-based/-like storage (because Ceph behaves this way), and dislike atimes
 | 
				
			||||||
 | 
					            options = "defaults,discard,noatime,nodiratime"
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					            # The root, var, and log volumes have specific values
 | 
				
			||||||
 | 
					            if volume["mountpoint"] == "/":
 | 
				
			||||||
 | 
					                # This will be used later by GRUB's cmdline
 | 
				
			||||||
 | 
					                root_volume["scsi_id"] = volume_id
 | 
				
			||||||
 | 
					                dump = 0
 | 
				
			||||||
 | 
					                cpass = 1
 | 
				
			||||||
 | 
					            elif volume["mountpoint"] == "/var" or volume["mountpoint"] == "/var/log":
 | 
				
			||||||
 | 
					                dump = 0
 | 
				
			||||||
 | 
					                cpass = 2
 | 
				
			||||||
 | 
					            else:
 | 
				
			||||||
 | 
					                dump = 0
 | 
				
			||||||
 | 
					                cpass = 0
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					            # Append the fstab line
 | 
				
			||||||
 | 
					            with open(fstab_file, "a") as fh:
 | 
				
			||||||
 | 
					                # Using these /dev/disk/by-id/scsi-0QEMU_QEMU_HARDDISK entries guarantees
 | 
				
			||||||
 | 
					                # proper ordering; /dev/sdX (or similar) names are NOT guaranteed to be
 | 
				
			||||||
 | 
					                # in any order nor are they guaranteed to match the volume's sdX/vdX name
 | 
				
			||||||
 | 
					                # when inside the VM due to Linux's quirks.
 | 
				
			||||||
 | 
					                data = "/dev/disk/by-id/scsi-0QEMU_QEMU_HARDDISK_drive-scsi0-0-0-{volume} {mountpoint} {filesystem} {options} {dump} {cpass}\n".format(
 | 
				
			||||||
 | 
					                    volume=volume_id,
 | 
				
			||||||
 | 
					                    mountpoint=volume["mountpoint"],
 | 
				
			||||||
 | 
					                    filesystem=volume["filesystem"],
 | 
				
			||||||
 | 
					                    options=options,
 | 
				
			||||||
 | 
					                    dump=dump,
 | 
				
			||||||
 | 
					                    cpass=cpass,
 | 
				
			||||||
 | 
					                )
 | 
				
			||||||
 | 
					                fh.write(data)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					            # Increment the volume_id
 | 
				
			||||||
 | 
					            volume_id += 1
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        # Write the hostname; you could also take an FQDN argument for this as an example
 | 
				
			||||||
 | 
					        hostname_file = "{}/etc/hostname".format(temporary_directory)
 | 
				
			||||||
 | 
					        with open(hostname_file, "w") as fh:
 | 
				
			||||||
 | 
					            fh.write("{}".format(vm_name))
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        # Fix the cloud-init.target since it's broken by default
 | 
				
			||||||
 | 
					        cloudinit_target_file = "{}/etc/systemd/system/cloud-init.target".format(
 | 
				
			||||||
 | 
					            temporary_directory
 | 
				
			||||||
 | 
					        )
 | 
				
			||||||
 | 
					        with open(cloudinit_target_file, "w") as fh:
 | 
				
			||||||
 | 
					            # We lose our indent on these raw blocks to preserve the apperance of the files
 | 
				
			||||||
 | 
					            # inside the VM itself
 | 
				
			||||||
 | 
					            data = """[Install]
 | 
				
			||||||
 | 
					WantedBy=multi-user.target
 | 
				
			||||||
 | 
					[Unit]
 | 
				
			||||||
 | 
					Description=Cloud-init target
 | 
				
			||||||
 | 
					After=multi-user.target
 | 
				
			||||||
 | 
					"""
 | 
				
			||||||
 | 
					            fh.write(data)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        # Due to device ordering within the Libvirt XML configuration, the first Ethernet interface
 | 
				
			||||||
 | 
					        # will always be on PCI bus ID 2, hence the name "ens2".
 | 
				
			||||||
 | 
					        # Write a DHCP stanza for ens2
 | 
				
			||||||
 | 
					        if not os.path.exists(f"{temporary_directory}/etc/sysconfig/network-scripts"):
 | 
				
			||||||
 | 
					            os.mkdir(f"{temporary_directory}/etc/sysconfig/network-scripts")
 | 
				
			||||||
 | 
					        ens2_network_file = "{}/etc/sysconfig/network-scripts/ifcfg-ens2".format(
 | 
				
			||||||
 | 
					            temporary_directory
 | 
				
			||||||
 | 
					        )
 | 
				
			||||||
 | 
					        with open(ens2_network_file, "w") as fh:
 | 
				
			||||||
 | 
					            data = """# Written by the PVC provisioner
 | 
				
			||||||
 | 
					DEVICE=ens2
 | 
				
			||||||
 | 
					ONBOOT=yes
 | 
				
			||||||
 | 
					TYPE=Ethernet
 | 
				
			||||||
 | 
					BOOTPROTO=dhcp
 | 
				
			||||||
 | 
					USERCTL=no
 | 
				
			||||||
 | 
					"""
 | 
				
			||||||
 | 
					            fh.write(data)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        # Write the DHCP config for ens2
 | 
				
			||||||
 | 
					        dhclient_file = "{}/etc/dhcp/dhclient.conf".format(temporary_directory)
 | 
				
			||||||
 | 
					        with open(dhclient_file, "w") as fh:
 | 
				
			||||||
 | 
					            # We can use fstrings too, since PVC will always have Python 3.6+, though
 | 
				
			||||||
 | 
					            # using format() might be preferable for clarity in some situations
 | 
				
			||||||
 | 
					            data = f"""# DHCP client configuration
 | 
				
			||||||
 | 
					# Written by the PVC provisioner
 | 
				
			||||||
 | 
					option rfc3442-classless-static-routes code 121 = array of unsigned integer 8;
 | 
				
			||||||
 | 
					interface "ens2" {{
 | 
				
			||||||
 | 
					       send fqdn.fqdn = "{vm_name}";
 | 
				
			||||||
 | 
					       send host-name = "{vm_name}";
 | 
				
			||||||
 | 
					       request subnet-mask, broadcast-address, time-offset, routers,
 | 
				
			||||||
 | 
					               domain-name, domain-name-servers, domain-search, host-name,
 | 
				
			||||||
 | 
					               dhcp6.name-servers, dhcp6.domain-search, dhcp6.fqdn, dhcp6.sntp-servers,
 | 
				
			||||||
 | 
					               netbios-name-servers, netbios-scope, interface-mtu,
 | 
				
			||||||
 | 
					               rfc3442-classless-static-routes, ntp-servers;
 | 
				
			||||||
 | 
					}}
 | 
				
			||||||
 | 
					"""
 | 
				
			||||||
 | 
					            fh.write(data)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        # Write the GRUB configuration
 | 
				
			||||||
 | 
					        grubcfg_file = "{}/etc/default/grub".format(temporary_directory)
 | 
				
			||||||
 | 
					        with open(grubcfg_file, "w") as fh:
 | 
				
			||||||
 | 
					            data = """# Written by the PVC provisioner
 | 
				
			||||||
 | 
					GRUB_TIMEOUT=5
 | 
				
			||||||
 | 
					GRUB_DEFAULT=saved
 | 
				
			||||||
 | 
					GRUB_DISABLE_SUBMENU=true
 | 
				
			||||||
 | 
					GRUB_DISABLE_RECOVERY=true
 | 
				
			||||||
 | 
					GRUB_DISABLE_LINUX_UUID=false
 | 
				
			||||||
 | 
					GRUB_ENABLE_BLSCFG=false
 | 
				
			||||||
 | 
					GRUB_DISTRIBUTOR="PVC Virtual Machine"
 | 
				
			||||||
 | 
					GRUB_CMDLINE_LINUX="root=/dev/disk/by-id/scsi-0QEMU_QEMU_HARDDISK_drive-scsi0-0-0-{root_volume} console=tty0 console=ttyS0,115200n8 no_timer_check crashkernel=1G-4G:192M,4G-64G:256M,64G-:512M"
 | 
				
			||||||
 | 
					GRUB_TERMINAL=console
 | 
				
			||||||
 | 
					GRUB_SERIAL_COMMAND="serial --speed=115200 --unit=0 --word=8 --parity=no --stop=1"
 | 
				
			||||||
 | 
					""".format(
 | 
				
			||||||
 | 
					                root_volume=root_volume["scsi_id"]
 | 
				
			||||||
 | 
					            )
 | 
				
			||||||
 | 
					            fh.write(data)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        # Do some tasks inside the chroot using the provided context manager
 | 
				
			||||||
 | 
					        with chroot(temporary_directory):
 | 
				
			||||||
 | 
					            # Fix the broken kernel from rinse by setting a systemd machine ID and running the post scripts
 | 
				
			||||||
 | 
					            os.system("systemd-machine-id-setup")
 | 
				
			||||||
 | 
					            os.system(
 | 
				
			||||||
 | 
					                "rpm -q --scripts kernel-core | grep -A20  'posttrans scriptlet' | tail -n+2 | bash -x"
 | 
				
			||||||
 | 
					            )
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					            # Install any post packages
 | 
				
			||||||
 | 
					            os.system(f"dnf install -y {' '.join(post_packages)}")
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					            # Install and update GRUB config
 | 
				
			||||||
 | 
					            os.system(
 | 
				
			||||||
 | 
					                "grub2-install --force /dev/rbd/{}/{}_{}".format(
 | 
				
			||||||
 | 
					                    root_volume["pool"], vm_name, root_volume["disk_id"]
 | 
				
			||||||
 | 
					                )
 | 
				
			||||||
 | 
					            )
 | 
				
			||||||
 | 
					            os.system("grub2-mkconfig -o /boot/grub2/grub.cfg")
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					            # Set a really dumb root password so the VM can be debugged
 | 
				
			||||||
 | 
					            # EITHER CHANGE THIS YOURSELF, here or in Userdata, or run something after install
 | 
				
			||||||
 | 
					            # to change the root password: don't leave it like this on an Internet-facing machine!
 | 
				
			||||||
 | 
					            os.system("echo root:test123 | chpasswd")
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					            # Enable dbus-broker
 | 
				
			||||||
 | 
					            os.system("systemctl enable dbus-broker.service")
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					            # Enable NetworkManager
 | 
				
			||||||
 | 
					            os.system("systemctl enable NetworkManager.service")
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					            # Enable cloud-init target on (first) boot
 | 
				
			||||||
 | 
					            # Your user-data should handle this and disable it once done, or things get messy.
 | 
				
			||||||
 | 
					            # That cloud-init won't run without this hack seems like a bug... but even the official
 | 
				
			||||||
 | 
					            # Debian cloud images are affected, so who knows.
 | 
				
			||||||
 | 
					            os.system("systemctl enable cloud-init.target")
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					            # Set the timezone to UTC
 | 
				
			||||||
 | 
					            os.system("ln -sf ../usr/share/zoneinfo/UTC /etc/localtime")
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        # Unmount the bound devfs and sysfs
 | 
				
			||||||
 | 
					        os.system("umount {}/dev".format(temporary_directory))
 | 
				
			||||||
 | 
					        os.system("umount {}/sys".format(temporary_directory))
 | 
				
			||||||
 | 
					        os.system("umount {}/proc".format(temporary_directory))
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    def cleanup(self):
 | 
				
			||||||
 | 
					        """
 | 
				
			||||||
 | 
					        cleanup(): Perform any cleanup required due to prepare()/install()
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        It is important to now reverse *all* steps taken in those functions that might
 | 
				
			||||||
 | 
					        need cleanup before teardown of the upper chroot environment.
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        This function is also called if there is ANY exception raised in the prepare()
 | 
				
			||||||
 | 
					        or install() steps. While this doesn't mean you shouldn't or can't raise exceptions
 | 
				
			||||||
 | 
					        here, be warned that doing so might cause loops. Do this only if you really need to.
 | 
				
			||||||
 | 
					        """
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        # Run any imports first
 | 
				
			||||||
 | 
					        from pvcapid.vmbuilder import open_zk
 | 
				
			||||||
 | 
					        from pvcapid.Daemon import config
 | 
				
			||||||
 | 
					        import daemon_lib.common as pvc_common
 | 
				
			||||||
 | 
					        import daemon_lib.ceph as pvc_ceph
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        # Set the tempdir we used in the prepare() and install() steps
 | 
				
			||||||
 | 
					        temp_dir = "/tmp/target"
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        # Use this construct for reversing the list, as the normal reverse() messes with the list
 | 
				
			||||||
 | 
					        for volume in list(reversed(self.vm_data["volumes"])):
 | 
				
			||||||
 | 
					            dst_volume_name = f"{self.vm_name}_{volume['disk_id']}"
 | 
				
			||||||
 | 
					            dst_volume = f"{volume['pool']}/{dst_volume_name}"
 | 
				
			||||||
 | 
					            mapped_dst_volume = f"/dev/rbd/{dst_volume}"
 | 
				
			||||||
 | 
					            mount_path = f"{temp_dir}/{volume['mountpoint']}"
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					            if (
 | 
				
			||||||
 | 
					                volume.get("source_volume") is None
 | 
				
			||||||
 | 
					                and volume.get("filesystem") is not None
 | 
				
			||||||
 | 
					            ):
 | 
				
			||||||
 | 
					                # Unmount filesystem
 | 
				
			||||||
 | 
					                retcode, stdout, stderr = pvc_common.run_os_command(
 | 
				
			||||||
 | 
					                    f"umount {mount_path}"
 | 
				
			||||||
 | 
					                )
 | 
				
			||||||
 | 
					                if retcode:
 | 
				
			||||||
 | 
					                    raise ProvisioningError(
 | 
				
			||||||
 | 
					                        f"Failed to unmount '{mapped_dst_volume}' on '{mount_path}': {stderr}"
 | 
				
			||||||
 | 
					                    )
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					                # Unmap volume
 | 
				
			||||||
 | 
					                with open_zk(config) as zkhandler:
 | 
				
			||||||
 | 
					                    success, message = pvc_ceph.unmap_volume(
 | 
				
			||||||
 | 
					                        zkhandler,
 | 
				
			||||||
 | 
					                        volume["pool"],
 | 
				
			||||||
 | 
					                        dst_volume_name,
 | 
				
			||||||
 | 
					                    )
 | 
				
			||||||
 | 
					                if not success:
 | 
				
			||||||
 | 
					                    raise ProvisioningError(
 | 
				
			||||||
 | 
					                        f"Failed to unmap '{mapped_dst_volume}': {stderr}"
 | 
				
			||||||
 | 
					                    )
 | 
				
			||||||
		Reference in New Issue
	
	Block a user