Complete integration of OVA provisioner

Finishes a basic form of OVA provisioning within the existing create_vm
function. Future plans should include separating out the functions and
cleaning them up a bit more, but this is sufficient for basic operation.

Closes #71
This commit is contained in:
Joshua Boniface 2020-02-18 14:42:45 -05:00
parent 4c58addead
commit b322841edf
3 changed files with 139 additions and 136 deletions

View File

@ -53,6 +53,7 @@ libvirt_header = """<domain type='kvm'>
<on_reboot>restart</on_reboot> <on_reboot>restart</on_reboot>
<on_crash>restart</on_crash> <on_crash>restart</on_crash>
<devices> <devices>
<console type='pty'/>
""" """
# File footer, closing devices and domain elements # File footer, closing devices and domain elements
@ -75,7 +76,6 @@ devices_default = """ <emulator>/usr/bin/kvm</emulator>
devices_serial = """ <serial type='pty'> devices_serial = """ <serial type='pty'>
<log file='/var/log/libvirt/{vm_name}.log' append='on'/> <log file='/var/log/libvirt/{vm_name}.log' append='on'/>
</serial> </serial>
<console type='pty'/>
""" """
# VNC device # VNC device

View File

@ -1,46 +0,0 @@
#
# TEMP
#
def tempstuff():
# Verify that the cluster has enough space to store all OVA disk volumes
total_size_bytes = 0
for disk in disk_map:
# Normalize the dev size to MB
# The function always return XXXXB, so strip off the B and convert to an integer
dev_size_bytes = int(pvc_ceph.format_bytes_fromhuman(disk.get('capacity', 0))[:-1])
ova_size_bytes = int(pvc_ceph.format_bytes_fromhuman(ova_size)[:-1])
# Get the actual image size
total_size_bytes += dev_size_bytes
# Add on the OVA size to account for the VMDK
total_size_bytes += ova_size_bytes
zk_conn = pvc_common.startZKConnection(config['coordinators'])
pool_information = pvc_ceph.getPoolInformation(zk_conn, pool)
pvc_common.stopZKConnection(zk_conn)
pool_free_space_bytes = int(pool_information['stats']['free_bytes'])
if total_size_bytes >= pool_free_space_bytes:
output = {
'message': "ERROR: The cluster does not have enough free space ({}) to store the VM ({}).".format(
pvc_ceph.format_bytes_tohuman(pool_free_space_bytes),
pvc_ceph.format_bytes_tohuman(total_size_bytes)
)
}
retcode = 400
cleanup_ova_maps_and_volumes()
return output, retcode
# Convert from the temporary to destination format on the blockdevs
retcode, stdout, stderr = pvc_common.run_os_command(
'qemu-img convert -C -f {} -O raw {} {}'.format(img_type, temp_blockdev, dest_blockdev)
)
if retcode:
output = {
'message': "ERROR: Failed to convert image '{}' format from '{}' to 'raw': {}".format(disk.get('src'), img_type, stderr)
}
retcode = 400
cleanup_img_maps_and_volumes()
cleanup_ova_maps_and_volumes()
return output, retcode

View File

@ -918,12 +918,6 @@ def delete_profile(name):
close_database(conn, cur) close_database(conn, cur)
return retmsg, retcode return retmsg, retcode
#
# Cloned VM provisioning function - executed by the Celery worker
#
def clone_vm(self, vm_name, vm_profile, source_volumes):
pass
# #
# Main VM provisioning function - executed by the Celery worker # Main VM provisioning function - executed by the Celery worker
# #
@ -972,11 +966,17 @@ def create_vm(self, vm_name, vm_profile, define_vm=True, start_vm=True):
args = (vm_profile,) args = (vm_profile,)
db_cur.execute(query, args) db_cur.execute(query, args)
profile_data = db_cur.fetchone() profile_data = db_cur.fetchone()
if profile_data['arguments']: if profile_data.get('arguments'):
vm_data['script_arguments'] = profile_data['arguments'].split('|') vm_data['script_arguments'] = profile_data.get('arguments').split('|')
else: else:
vm_data['script_arguments'] = [] vm_data['script_arguments'] = []
if profile_data.get('profile_type') == 'ova':
is_ova_install = True
is_script_install = False # By definition
else:
is_ova_install = False
# Get the system details # Get the system details
query = 'SELECT * FROM system_template WHERE id = %s' query = 'SELECT * FROM system_template WHERE id = %s'
args = (profile_data['system_template'],) args = (profile_data['system_template'],)
@ -984,10 +984,14 @@ def create_vm(self, vm_name, vm_profile, define_vm=True, start_vm=True):
vm_data['system_details'] = db_cur.fetchone() vm_data['system_details'] = db_cur.fetchone()
# Get the MAC template # Get the MAC template
query = 'SELECT * FROM network_template WHERE id = %s' query = 'SELECT mac_template FROM network_template WHERE id = %s'
args = (profile_data['network_template'],) args = (profile_data['network_template'],)
db_cur.execute(query, args) db_cur.execute(query, args)
vm_data['mac_template'] = db_cur.fetchone()['mac_template'] db_row = db_cur.fetchone()
if db_row:
vm_data['mac_template'] = db_row.get('mac_template')
else:
vm_data['mac_template'] = None
# Get the networks # Get the networks
query = 'SELECT * FROM network WHERE network_template = %s' query = 'SELECT * FROM network WHERE network_template = %s'
@ -1003,17 +1007,27 @@ def create_vm(self, vm_name, vm_profile, define_vm=True, start_vm=True):
vm_data['volumes'] = db_cur.fetchall() vm_data['volumes'] = db_cur.fetchall()
# Get the script # Get the script
query = 'SELECT * FROM script WHERE id = %s' query = 'SELECT script FROM script WHERE id = %s'
args = (profile_data['script'],) args = (profile_data['script'],)
db_cur.execute(query, args) db_cur.execute(query, args)
vm_data['script'] = db_cur.fetchone()['script'] vm_data['script'] = db_cur.fetchone()
if not vm_data['script']: if vm_data['script'] and not is_ova_install:
# We do not have a script; skip it
is_script_install = False
else:
is_script_install = True is_script_install = True
else:
is_script_install = False
# Get the OVA details
if is_ova_install:
query = 'SELECT * FROM ova WHERE id = %s'
args = (profile_data['ova'],)
db_cur.execute(query, args)
vm_data['ova_details'] = db_cur.fetchone()
query = 'SELECT * FROM ova_volume WHERE ova = %s'
args = (profile_data['ova'],)
db_cur.execute(query, args)
vm_data['volumes'] = db_cur.fetchall()
close_database(db_conn, db_cur) close_database(db_conn, db_cur)
@ -1064,7 +1078,7 @@ def create_vm(self, vm_name, vm_profile, define_vm=True, start_vm=True):
# Verify that there is enough disk space free to provision all VM disks # Verify that there is enough disk space free to provision all VM disks
pools = dict() pools = dict()
for volume in vm_data['volumes']: for volume in vm_data['volumes']:
if volume['source_volume'] is not None: if volume.get('source_volume') is not None:
if not volume['pool'] in pools: if not volume['pool'] in pools:
volume_data, status = pvc_ceph.getVolumeInformation(zk_conn, volume['pool'], volume['source_volume']) volume_data, status = pvc_ceph.getVolumeInformation(zk_conn, volume['pool'], volume['source_volume'])
pools[volume['pool']] = volume_data['disk_size_gb'] pools[volume['pool']] = volume_data['disk_size_gb']
@ -1093,6 +1107,7 @@ def create_vm(self, vm_name, vm_profile, define_vm=True, start_vm=True):
print("There is enough space on cluster to store VM volumes") print("There is enough space on cluster to store VM volumes")
if not is_ova_install:
# Verify that every specified filesystem is valid # Verify that every specified filesystem is valid
used_filesystems = list() used_filesystems = list()
for volume in vm_data['volumes']: for volume in vm_data['volumes']:
@ -1176,9 +1191,8 @@ def create_vm(self, vm_name, vm_profile, define_vm=True, start_vm=True):
vm_id_hex = '{:x}'.format(int(vm_id % 16)) vm_id_hex = '{:x}'.format(int(vm_id % 16))
net_id_hex = '{:x}'.format(int(network_id % 16)) net_id_hex = '{:x}'.format(int(network_id % 16))
if vm_data['mac_template']: if vm_data.get('mac_template') is not None:
mac_prefix = '52:54:01' mac_prefix = '52:54:01'
mactemplate = "{prefix}:ff:f6:{vmid}{netid}"
macgen_template = vm_data['mac_template'] macgen_template = vm_data['mac_template']
eth_macaddr = macgen_template.format( eth_macaddr = macgen_template.format(
prefix=mac_prefix, prefix=mac_prefix,
@ -1286,7 +1300,7 @@ def create_vm(self, vm_name, vm_profile, define_vm=True, start_vm=True):
time.sleep(1) time.sleep(1)
for volume in vm_data['volumes']: for volume in vm_data['volumes']:
if volume['source_volume'] is not None: if volume.get('source_volume') is not None:
success, message = pvc_ceph.clone_volume(zk_conn, volume['pool'], "{}_{}".format(vm_name, volume['disk_id']), volume['source_volume']) success, message = pvc_ceph.clone_volume(zk_conn, volume['pool'], "{}_{}".format(vm_name, volume['disk_id']), volume['source_volume'])
print(message) print(message)
if not success: if not success:
@ -1306,14 +1320,48 @@ def create_vm(self, vm_name, vm_profile, define_vm=True, start_vm=True):
time.sleep(1) time.sleep(1)
for volume in vm_data['volumes']: for volume in vm_data['volumes']:
if volume['source_volume'] is not None: dst_volume_name = "{}_{}".format(vm_name, volume['disk_id'])
dst_volume = "{}/{}".format(volume['pool'], dst_volume_name)
if is_ova_install:
src_volume_name = volume['volume_name']
src_volume = "{}/{}".format(volume['pool'], src_volume_name)
# Map the target RBD device
retcode, retmsg = pvc_ceph.map_volume(zk_conn, volume['pool'], dst_volume_name)
if not retcode:
raise ProvisioningError('Failed to map destination volume "{}": {}'.format(dst_volume_name, retmsg))
# Map the source RBD device
retcode, retmsg = pvc_ceph.map_volume(zk_conn, volume['pool'], src_volume_name)
if not retcode:
raise ProvisioningError('Failed to map source volume "{}": {}'.format(src_volume_name, retmsg))
# Convert from source to target
retcode, stdout, stderr = pvc_common.run_os_command(
'qemu-img convert -C -f {} -O raw {} {}'.format(
volume['volume_format'],
"/dev/rbd/{}".format(src_volume),
"/dev/rbd/{}".format(dst_volume)
)
)
if retcode:
raise ProvisioningError('Failed to convert {} volume "{}" to raw volume "{}": {}'.format(volume['volume_format'], src_volume, dst_volume, stderr))
# Unmap the source RBD device (don't bother later)
retcode, retmsg = pvc_ceph.unmap_volume(zk_conn, volume['pool'], src_volume_name)
if not retcode:
raise ProvisioningError('Failed to unmap source volume "{}": {}'.format(src_volume_name, retmsg))
# Unmap the target RBD device (don't bother later)
retcode, retmsg = pvc_ceph.unmap_volume(zk_conn, volume['pool'], dst_volume_name)
if not retcode:
raise ProvisioningError('Failed to unmap destination volume "{}": {}'.format(dst_volume_name, retmsg))
print('Converted {} source volume {} to raw format on {}'.format(volume['volume_format'], src_volume, dst_volume))
else:
if volume.get('source_volume') is not None:
continue continue
if not volume['filesystem']: if volume.get('filesystem') is None:
continue continue
rbd_volume = "{}/{}_{}".format(volume['pool'], vm_name, volume['disk_id'])
filesystem_args_list = list() filesystem_args_list = list()
for arg in volume['filesystem_args'].split(): for arg in volume['filesystem_args'].split():
arg_entry, arg_data = arg.split('=') arg_entry, arg_data = arg.split('=')
@ -1322,21 +1370,21 @@ def create_vm(self, vm_name, vm_profile, define_vm=True, start_vm=True):
filesystem_args = ' '.join(filesystem_args_list) filesystem_args = ' '.join(filesystem_args_list)
# Map the RBD device # Map the RBD device
retcode, stdout, stderr = pvc_common.run_os_command("rbd map {}".format(rbd_volume)) retcode, retmsg = pvc_ceph.map_volume(zk_conn, volume['pool'], dst_volume_name)
if retcode: if not retcode:
raise ProvisioningError('Failed to map volume "{}": {}'.format(rbd_volume, stderr)) raise ProvisioningError('Failed to map volume "{}": {}'.format(dst_volume, retmsg))
# Create the filesystem # Create the filesystem
if volume['filesystem'] == 'swap': if volume['filesystem'] == 'swap':
retcode, stdout, stderr = pvc_common.run_os_command("mkswap -f /dev/rbd/{}".format(rbd_volume)) retcode, stdout, stderr = pvc_common.run_os_command("mkswap -f /dev/rbd/{}".format(dst_volume))
if retcode: if retcode:
raise ProvisioningError('Failed to create swap on "{}": {}'.format(rbd_volume, stderr)) raise ProvisioningError('Failed to create swap on "{}": {}'.format(dst_volume, stderr))
else: else:
retcode, stdout, stderr = pvc_common.run_os_command("mkfs.{} {} /dev/rbd/{}".format(volume['filesystem'], filesystem_args, rbd_volume)) retcode, stdout, stderr = pvc_common.run_os_command("mkfs.{} {} /dev/rbd/{}".format(volume['filesystem'], filesystem_args, dst_volume))
if retcode: if retcode:
raise ProvisioningError('Failed to create {} filesystem on "{}": {}'.format(volume['filesystem'], rbd_volume, stderr)) raise ProvisioningError('Failed to create {} filesystem on "{}": {}'.format(volume['filesystem'], dst_volume, stderr))
print("Created {} filesystem on {}:\n{}".format(volume['filesystem'], rbd_volume, stdout)) print("Created {} filesystem on {}:\n{}".format(volume['filesystem'], dst_volume, stdout))
if is_script_install: if is_script_install:
# Create temporary directory # Create temporary directory
@ -1352,7 +1400,7 @@ def create_vm(self, vm_name, vm_profile, define_vm=True, start_vm=True):
if not volume['mountpoint'] or volume['mountpoint'] == 'swap': if not volume['mountpoint'] or volume['mountpoint'] == 'swap':
continue continue
mapped_rbd_volume = "/dev/rbd/{}/{}_{}".format(volume['pool'], vm_name, volume['disk_id']) mapped_dst_volume = "/dev/rbd/{}/{}_{}".format(volume['pool'], vm_name, volume['disk_id'])
mount_path = "{}{}".format(temp_dir, volume['mountpoint']) mount_path = "{}{}".format(temp_dir, volume['mountpoint'])
# Ensure the mount path exists (within the filesystems) # Ensure the mount path exists (within the filesystems)
@ -1361,11 +1409,11 @@ def create_vm(self, vm_name, vm_profile, define_vm=True, start_vm=True):
raise ProvisioningError('Failed to create mountpoint "{}": {}'.format(mount_path, stderr)) raise ProvisioningError('Failed to create mountpoint "{}": {}'.format(mount_path, stderr))
# Mount filesystems to temporary directory # Mount filesystems to temporary directory
retcode, stdout, stderr = pvc_common.run_os_command("mount {} {}".format(mapped_rbd_volume, mount_path)) retcode, stdout, stderr = pvc_common.run_os_command("mount {} {}".format(mapped_dst_volume, mount_path))
if retcode: if retcode:
raise ProvisioningError('Failed to mount "{}" on "{}": {}'.format(mapped_rbd_volume, mount_path, stderr)) raise ProvisioningError('Failed to mount "{}" on "{}": {}'.format(mapped_dst_volume, mount_path, stderr))
print("Successfully mounted {} on {}".format(mapped_rbd_volume, mount_path)) print("Successfully mounted {} on {}".format(mapped_dst_volume, mount_path))
# Phase 8 - provisioning script execution # Phase 8 - provisioning script execution
# * Execute the provisioning script main function ("install") passing any custom arguments # * Execute the provisioning script main function ("install") passing any custom arguments
@ -1400,13 +1448,14 @@ def create_vm(self, vm_name, vm_profile, define_vm=True, start_vm=True):
self.update_state(state='RUNNING', meta={'current': 9, 'total': 10, 'status': 'Cleaning up local mounts and directories'}) self.update_state(state='RUNNING', meta={'current': 9, 'total': 10, 'status': 'Cleaning up local mounts and directories'})
time.sleep(1) time.sleep(1)
if not is_ova_install:
for volume in list(reversed(vm_data['volumes'])): for volume in list(reversed(vm_data['volumes'])):
if volume['source_volume'] is not None: if volume.get('source_volume') is not None:
continue continue
if is_script_install: if is_script_install:
# Unmount the volume # Unmount the volume
if volume['mountpoint'] and volume['mountpoint'] != 'swap': if volume.get('mountpoint') is not None and volume.get('mountpoint') != 'swap':
print("Cleaning up mount {}{}".format(temp_dir, volume['mountpoint'])) print("Cleaning up mount {}{}".format(temp_dir, volume['mountpoint']))
mount_path = "{}{}".format(temp_dir, volume['mountpoint']) mount_path = "{}{}".format(temp_dir, volume['mountpoint'])