Get to installation completion
This commit is contained in:
parent
cff6a49660
commit
e045eafbc1
|
@ -43,7 +43,6 @@ def install(**kwargs):
|
||||||
# by this point, so we can get right to running the debootstrap after setting
|
# by this point, so we can get right to running the debootstrap after setting
|
||||||
# some nicer variable names; you don't necessarily have to do this.
|
# some nicer variable names; you don't necessarily have to do this.
|
||||||
vm_name = kwargs['vm_name']
|
vm_name = kwargs['vm_name']
|
||||||
vm_id = kwargs['vm_id']
|
|
||||||
temporary_directory = kwargs['temporary_directory']
|
temporary_directory = kwargs['temporary_directory']
|
||||||
disks = kwargs['disks']
|
disks = kwargs['disks']
|
||||||
networks = kwargs['networks']
|
networks = kwargs['networks']
|
||||||
|
@ -69,7 +68,6 @@ def install(**kwargs):
|
||||||
root_disk = disk
|
root_disk = disk
|
||||||
if not root_disk:
|
if not root_disk:
|
||||||
return
|
return
|
||||||
print(root_disk)
|
|
||||||
|
|
||||||
# Ensure we have debootstrap intalled on the provisioner system; this is a
|
# Ensure we have debootstrap intalled on the provisioner system; this is a
|
||||||
# good idea to include if you plan to use anything that is not part of the
|
# good idea to include if you plan to use anything that is not part of the
|
||||||
|
@ -116,7 +114,7 @@ def install(**kwargs):
|
||||||
# Append the fstab line
|
# Append the fstab line
|
||||||
with open(fstab_file, 'a') as fh:
|
with open(fstab_file, 'a') as fh:
|
||||||
fh.write("/dev/{disk} {mountpoint} {filesystem} {options} {dump} {cpass}\n".format(
|
fh.write("/dev/{disk} {mountpoint} {filesystem} {options} {dump} {cpass}\n".format(
|
||||||
disk=disk['name'],
|
disk=disk['disk_id'],
|
||||||
mountpoint=disk['mountpoint'],
|
mountpoint=disk['mountpoint'],
|
||||||
filesystem=disk['filesystem'],
|
filesystem=disk['filesystem'],
|
||||||
options=options,
|
options=options,
|
||||||
|
@ -136,7 +134,7 @@ GRUB_CMDLINE_LINUX=""
|
||||||
GRUB_TERMINAL=console
|
GRUB_TERMINAL=console
|
||||||
GRUB_SERIAL_COMMAND="serial --speed=115200 --unit=0 --word=8 --parity=no --stop=1"
|
GRUB_SERIAL_COMMAND="serial --speed=115200 --unit=0 --word=8 --parity=no --stop=1"
|
||||||
GRUB_DISABLE_LINUX_UUID=false
|
GRUB_DISABLE_LINUX_UUID=false
|
||||||
""".format(root_disk=root_disk['name']))
|
""".format(root_disk=root_disk['disk_id']))
|
||||||
|
|
||||||
# Chroot and install GRUB so we can boot, then exit the chroot
|
# Chroot and install GRUB so we can boot, then exit the chroot
|
||||||
# EXITING THE CHROOT IS VERY IMPORTANT OR THE FOLLOWING STAGES OF THE PROVISIONER
|
# EXITING THE CHROOT IS VERY IMPORTANT OR THE FOLLOWING STAGES OF THE PROVISIONER
|
||||||
|
@ -146,7 +144,7 @@ GRUB_DISABLE_LINUX_UUID=false
|
||||||
fake_root = os.open("/", os.O_RDONLY)
|
fake_root = os.open("/", os.O_RDONLY)
|
||||||
os.fchdir(fake_root)
|
os.fchdir(fake_root)
|
||||||
os.system(
|
os.system(
|
||||||
"grub-install /dev/rbd/{}".format(root_disk['volume'])
|
"grub-install /dev/rbd/{}/{}_{}".format(root_disk['pool'], vm_name, root_disk['disk_id'])
|
||||||
)
|
)
|
||||||
os.system(
|
os.system(
|
||||||
"update-grub"
|
"update-grub"
|
||||||
|
|
|
@ -27,6 +27,8 @@ import psycopg2.extras
|
||||||
import os
|
import os
|
||||||
import re
|
import re
|
||||||
import time
|
import time
|
||||||
|
import shlex
|
||||||
|
import subprocess
|
||||||
|
|
||||||
import client_lib.common as pvc_common
|
import client_lib.common as pvc_common
|
||||||
import client_lib.node as pvc_node
|
import client_lib.node as pvc_node
|
||||||
|
@ -49,6 +51,12 @@ class ClusterError(Exception):
|
||||||
"""
|
"""
|
||||||
pass
|
pass
|
||||||
|
|
||||||
|
class ProvisioningError(Exception):
|
||||||
|
"""
|
||||||
|
An exception that results from a failure of a provisioning command.
|
||||||
|
"""
|
||||||
|
pass
|
||||||
|
|
||||||
#
|
#
|
||||||
# Common functions
|
# Common functions
|
||||||
#
|
#
|
||||||
|
@ -256,7 +264,7 @@ def create_template_storage(name):
|
||||||
close_database(conn, cur)
|
close_database(conn, cur)
|
||||||
return flask.jsonify(retmsg), retcode
|
return flask.jsonify(retmsg), retcode
|
||||||
|
|
||||||
def create_template_storage_element(name, pool, disk_id, disk_size_gb, mountpoint=None, filesystem=None):
|
def create_template_storage_element(name, pool, disk_id, disk_size_gb, filesystem=None, filesystem_args=[], mountpoint=None):
|
||||||
if not list_template_storage(name, is_fuzzy=False):
|
if not list_template_storage(name, is_fuzzy=False):
|
||||||
retmsg = { "message": "The storage template {} does not exist".format(name) }
|
retmsg = { "message": "The storage template {} does not exist".format(name) }
|
||||||
retcode = 400
|
retcode = 400
|
||||||
|
@ -283,8 +291,8 @@ def create_template_storage_element(name, pool, disk_id, disk_size_gb, mountpoin
|
||||||
args = (name,)
|
args = (name,)
|
||||||
cur.execute(query, args)
|
cur.execute(query, args)
|
||||||
template_id = cur.fetchone()['id']
|
template_id = cur.fetchone()['id']
|
||||||
query = "INSERT INTO storage (storage_template, pool, disk_id, disk_size_gb, mountpoint, filesystem) VALUES (%s, %s, %s, %s, %s, %s);"
|
query = "INSERT INTO storage (storage_template, pool, disk_id, disk_size_gb, mountpoint, filesystem, filesystem_args) VALUES (%s, %s, %s, %s, %s, %s, %s);"
|
||||||
args = (template_id, pool, disk_id, disk_size_gb, mountpoint, filesystem)
|
args = (template_id, pool, disk_id, disk_size_gb, mountpoint, filesystem, ' '.join(filesystem_args))
|
||||||
cur.execute(query, args)
|
cur.execute(query, args)
|
||||||
retmsg = { "name": name, "disk_id": disk_id }
|
retmsg = { "name": name, "disk_id": disk_id }
|
||||||
retcode = 200
|
retcode = 200
|
||||||
|
@ -495,7 +503,7 @@ def delete_script(name):
|
||||||
retmsg = { "name": name }
|
retmsg = { "name": name }
|
||||||
retcode = 200
|
retcode = 200
|
||||||
except psycopg2.IntegrityError as e:
|
except psycopg2.IntegrityError as e:
|
||||||
retmsg = { "message": "Failed to delete entry {}".format(name), "error": e }
|
retmsg = { "message": "Failed to delete entry {}".format(name), "error": str(e) }
|
||||||
retcode = 400
|
retcode = 400
|
||||||
close_database(conn, cur)
|
close_database(conn, cur)
|
||||||
return flask.jsonify(retmsg), retcode
|
return flask.jsonify(retmsg), retcode
|
||||||
|
@ -626,6 +634,30 @@ def delete_profile(name):
|
||||||
#
|
#
|
||||||
# VM provisioning helper functions
|
# VM provisioning helper functions
|
||||||
#
|
#
|
||||||
|
def run_os_command(command_string, background=False, environment=None, timeout=None):
|
||||||
|
command = shlex.split(command_string)
|
||||||
|
try:
|
||||||
|
command_output = subprocess.run(
|
||||||
|
command,
|
||||||
|
env=environment,
|
||||||
|
timeout=timeout,
|
||||||
|
stdout=subprocess.PIPE,
|
||||||
|
stderr=subprocess.PIPE,
|
||||||
|
)
|
||||||
|
retcode = command_output.returncode
|
||||||
|
except subprocess.TimeoutExpired:
|
||||||
|
retcode = 128
|
||||||
|
|
||||||
|
try:
|
||||||
|
stdout = command_output.stdout.decode('ascii')
|
||||||
|
except:
|
||||||
|
stdout = ''
|
||||||
|
try:
|
||||||
|
stderr = command_output.stderr.decode('ascii')
|
||||||
|
except:
|
||||||
|
stderr = ''
|
||||||
|
return retcode, stdout, stderr
|
||||||
|
|
||||||
|
|
||||||
#
|
#
|
||||||
# Main VM provisioning function - executed by the Celery worker
|
# Main VM provisioning function - executed by the Celery worker
|
||||||
|
@ -635,6 +667,8 @@ def create_vm(self, vm_name, vm_profile):
|
||||||
import time
|
import time
|
||||||
import importlib
|
import importlib
|
||||||
|
|
||||||
|
time.sleep(2)
|
||||||
|
|
||||||
print("Starting provisioning of VM '{}' with profile '{}'".format(vm_name, vm_profile))
|
print("Starting provisioning of VM '{}' with profile '{}'".format(vm_name, vm_profile))
|
||||||
|
|
||||||
# Phase 0 - connect to databases
|
# Phase 0 - connect to databases
|
||||||
|
@ -655,7 +689,7 @@ def create_vm(self, vm_name, vm_profile):
|
||||||
# * Get the details from these elements
|
# * Get the details from these elements
|
||||||
# * Assemble a VM configuration dictionary
|
# * Assemble a VM configuration dictionary
|
||||||
self.update_state(state='RUNNING', meta={'current': 1, 'total': 10, 'status': 'Collecting configuration'})
|
self.update_state(state='RUNNING', meta={'current': 1, 'total': 10, 'status': 'Collecting configuration'})
|
||||||
time.sleep(3)
|
time.sleep(1)
|
||||||
|
|
||||||
vm_data = dict()
|
vm_data = dict()
|
||||||
|
|
||||||
|
@ -685,7 +719,7 @@ def create_vm(self, vm_name, vm_profile):
|
||||||
vm_data['networks'] = db_cur.fetchall()
|
vm_data['networks'] = db_cur.fetchall()
|
||||||
|
|
||||||
# Get the storage volumes
|
# Get the storage volumes
|
||||||
query = 'SELECT pool, disk_id, disk_size_gb, mountpoint, filesystem FROM storage WHERE storage_template = %s'
|
query = 'SELECT pool, disk_id, disk_size_gb, mountpoint, filesystem, filesystem_args FROM storage WHERE storage_template = %s'
|
||||||
args = (profile_data['storage_template'],)
|
args = (profile_data['storage_template'],)
|
||||||
db_cur.execute(query, args)
|
db_cur.execute(query, args)
|
||||||
vm_data['volumes'] = db_cur.fetchall()
|
vm_data['volumes'] = db_cur.fetchall()
|
||||||
|
@ -698,15 +732,13 @@ def create_vm(self, vm_name, vm_profile):
|
||||||
|
|
||||||
close_database(db_conn, db_cur)
|
close_database(db_conn, db_cur)
|
||||||
|
|
||||||
print(json.dumps(vm_data))
|
|
||||||
|
|
||||||
# Phase 2 - verification
|
# Phase 2 - verification
|
||||||
# * Ensure that at least one node has enough free RAM to hold the VM (becomes main host)
|
# * Ensure that at least one node has enough free RAM to hold the VM (becomes main host)
|
||||||
# * Ensure that all networks are valid
|
# * Ensure that all networks are valid
|
||||||
# * Ensure that there is enough disk space in the Ceph cluster for the disks
|
# * Ensure that there is enough disk space in the Ceph cluster for the disks
|
||||||
# This is the "safe fail" step when an invalid configuration will be caught
|
# This is the "safe fail" step when an invalid configuration will be caught
|
||||||
self.update_state(state='RUNNING', meta={'current': 2, 'total': 10, 'status': 'Verifying configuration against cluster'})
|
self.update_state(state='RUNNING', meta={'current': 2, 'total': 10, 'status': 'Verifying configuration against cluster'})
|
||||||
time.sleep(3)
|
time.sleep(1)
|
||||||
|
|
||||||
# Verify that at least one host has enough free RAM to run the VM
|
# Verify that at least one host has enough free RAM to run the VM
|
||||||
_discard, nodes = pvc_node.get_list(zk_conn, None)
|
_discard, nodes = pvc_node.get_list(zk_conn, None)
|
||||||
|
@ -736,58 +768,169 @@ def create_vm(self, vm_name, vm_profile):
|
||||||
if not vni in cluster_networks:
|
if not vni in cluster_networks:
|
||||||
raise ClusterError("The network VNI {} is not present on the cluster".format(vni))
|
raise ClusterError("The network VNI {} is not present on the cluster".format(vni))
|
||||||
|
|
||||||
print("All configured networks {} for VM are valid".format(vm_data['networks']))
|
print("All configured networks for VM are valid")
|
||||||
|
|
||||||
# Verify that there is enough disk space free to provision all VM disks
|
# Verify that there is enough disk space free to provision all VM disks
|
||||||
pools = dict()
|
pools = dict()
|
||||||
for volume in vm_data['volumes']:
|
for volume in vm_data['volumes']:
|
||||||
if not volume['pool'] in pools:
|
if not volume['pool'] in pools:
|
||||||
pools[volume['pool']] = 0
|
pools[volume['pool']] = volume['disk_size_gb']
|
||||||
for volume in vm_data['volumes']:
|
else:
|
||||||
pools[volume['pool']] += volume['disk_size_gb']
|
pools[volume['pool']] += volume['disk_size_gb']
|
||||||
|
|
||||||
print(pools)
|
|
||||||
|
|
||||||
for pool in pools:
|
for pool in pools:
|
||||||
pool_free_space = pvc_ceph.getPoolInformation(zk_conn, pool)
|
pool_information = pvc_ceph.getPoolInformation(zk_conn, pool)
|
||||||
pool_vm_usage = pools[pool]
|
if not pool_information:
|
||||||
|
raise ClusterError("Pool {} is not present on the cluster".format(pool))
|
||||||
|
pool_free_space_gb = int(pool_information['stats']['free_bytes'] / 1024 / 1024 / 1024)
|
||||||
|
pool_vm_usage_gb = int(pools[pool])
|
||||||
|
|
||||||
print(pool_free_space)
|
if pool_vm_usage_gb >= pool_free_space_gb:
|
||||||
print(pool_vm_usage)
|
raise ClusterError("Pool {} has only {} GB free and VM requires {} GB".format(pool, pool_free_space_gb, pool_vm_usage_gb))
|
||||||
|
|
||||||
print(pvc_ceph.get_radosdf(zk_conn))
|
print("There is enough space on cluster to store VM volumes")
|
||||||
|
|
||||||
return
|
# Verify that every specified filesystem is valid
|
||||||
|
used_filesystems = list()
|
||||||
|
for volume in vm_data['volumes']:
|
||||||
|
if volume['filesystem'] and volume['filesystem'] not in used_filesystems:
|
||||||
|
used_filesystems.append(volume['filesystem'])
|
||||||
|
|
||||||
# Phase 3 - disk creation
|
for filesystem in used_filesystems:
|
||||||
|
retcode, stdout, stderr = run_os_command("which mkfs.{}".format(filesystem))
|
||||||
|
if retcode:
|
||||||
|
raise ProvisioningError("Failed to find binary for mkfs.{}: {}".format(filesystem, stderr))
|
||||||
|
|
||||||
|
print("All selected filesystems are valid")
|
||||||
|
|
||||||
|
# Phase 3 - provisioning script preparation
|
||||||
|
# * Import the provisioning script as a library with importlib
|
||||||
|
# * Ensure the required function(s) are present
|
||||||
|
self.update_state(state='RUNNING', meta={'current': 3, 'total': 10, 'status': 'Preparing provisioning script'})
|
||||||
|
time.sleep(1)
|
||||||
|
|
||||||
|
# Write the script out to a temporary file
|
||||||
|
retcode, stdout, stderr = run_os_command("mktemp")
|
||||||
|
if retcode:
|
||||||
|
raise ProvisioningError("Failed to create a temporary file: {}".format(stderr))
|
||||||
|
script_file = stdout.strip()
|
||||||
|
with open(script_file, 'w') as fh:
|
||||||
|
fh.write(vm_data['script'])
|
||||||
|
fh.write('\n')
|
||||||
|
|
||||||
|
# Import the script file
|
||||||
|
loader = importlib.machinery.SourceFileLoader('installer_script', script_file)
|
||||||
|
spec = importlib.util.spec_from_loader(loader.name, loader)
|
||||||
|
installer_script = importlib.util.module_from_spec(spec)
|
||||||
|
loader.exec_module(installer_script)
|
||||||
|
|
||||||
|
# Verify that the install() function is valid
|
||||||
|
if not "install" in dir(installer_script):
|
||||||
|
raise ProvisioningError("Specified script does not contain an install() function")
|
||||||
|
|
||||||
|
print("Provisioning script imported successfully")
|
||||||
|
|
||||||
|
# Phase 4 - disk creation
|
||||||
# * Create each Ceph storage volume for the disks
|
# * Create each Ceph storage volume for the disks
|
||||||
self.update_state(state='RUNNING', meta={'current': 3, 'total': 10, 'status': 'Creating storage volumes'})
|
self.update_state(state='RUNNING', meta={'current': 4, 'total': 10, 'status': 'Creating storage volumes'})
|
||||||
time.sleep(5)
|
time.sleep(1)
|
||||||
|
|
||||||
# Phase 4 - disk mapping
|
for volume in vm_data['volumes']:
|
||||||
|
success, message = pvc_ceph.add_volume(zk_conn, volume['pool'], "{}_{}".format(vm_name, volume['disk_id']), "{}G".format(volume['disk_size_gb']))
|
||||||
|
print(message)
|
||||||
|
if not success:
|
||||||
|
raise ClusterError("Failed to create volume {}".format(volume['disk_id']))
|
||||||
|
|
||||||
|
# Phase 5 - disk mapping
|
||||||
# * Map each volume to the local host in order
|
# * Map each volume to the local host in order
|
||||||
# * Format each volume with any specified filesystems
|
# * Format each volume with any specified filesystems
|
||||||
# * If any mountpoints are specified, create a temporary mount directory
|
# * If any mountpoints are specified, create a temporary mount directory
|
||||||
# * Mount any volumes to their respective mountpoints
|
# * Mount any volumes to their respective mountpoints
|
||||||
self.update_state(state='RUNNING', meta={'current': 4, 'total': 10, 'status': 'Mapping, formatting, and mounting storage volumes locally'})
|
self.update_state(state='RUNNING', meta={'current': 5, 'total': 10, 'status': 'Mapping, formatting, and mounting storage volumes locally'})
|
||||||
time.sleep(5)
|
time.sleep(1)
|
||||||
|
|
||||||
# Phase 5 - provisioning script preparation
|
for volume in vm_data['volumes']:
|
||||||
# * Import the provisioning script as a library with importlib
|
if not volume['filesystem']:
|
||||||
# * Ensure the required function(s) are present
|
continue
|
||||||
self.update_state(state='RUNNING', meta={'current': 5, 'total': 10, 'status': 'Preparing provisioning script'})
|
|
||||||
time.sleep(5)
|
rbd_volume = "{}/{}_{}".format(volume['pool'], vm_name, volume['disk_id'])
|
||||||
|
|
||||||
|
filesystem_args_list = list()
|
||||||
|
for arg in volume['filesystem_args'].split(' '):
|
||||||
|
arg_entry, arg_data = arg.split('=')
|
||||||
|
filesystem_args_list.append(arg_entry)
|
||||||
|
filesystem_args_list.append(arg_data)
|
||||||
|
filesystem_args = ' '.join(filesystem_args_list)
|
||||||
|
|
||||||
|
# Map the RBD device
|
||||||
|
retcode, stdout, stderr = run_os_command("rbd map {}".format(rbd_volume))
|
||||||
|
if retcode:
|
||||||
|
raise ProvisioningError("Failed to map volume {}: {}".format(rbd_volume, stderr))
|
||||||
|
|
||||||
|
# Create the filesystem
|
||||||
|
retcode, stdout, stderr = run_os_command("mkfs.{} {} /dev/rbd/{}".format(volume['filesystem'], filesystem_args, rbd_volume))
|
||||||
|
if retcode:
|
||||||
|
raise ProvisioningError("Failed to create {} filesystem on {}: {}".format(volume['filesystem'], rbd_volume, stderr))
|
||||||
|
|
||||||
|
print("Created {} filesystem on {}:\n{}".format(volume['filesystem'], rbd_volume, stdout))
|
||||||
|
|
||||||
|
# Create temporary directory
|
||||||
|
retcode, stdout, stderr = run_os_command("mktemp -d")
|
||||||
|
if retcode:
|
||||||
|
raise ProvisioningError("Failed to create a temporary directory: {}".format(stderr))
|
||||||
|
temp_dir = stdout.strip()
|
||||||
|
|
||||||
|
for volume in vm_data['volumes']:
|
||||||
|
if not volume['mountpoint']:
|
||||||
|
continue
|
||||||
|
|
||||||
|
mapped_rbd_volume = "/dev/rbd/{}/{}_{}".format(volume['pool'], vm_name, volume['disk_id'])
|
||||||
|
mount_path = "{}{}".format(temp_dir, volume['mountpoint'])
|
||||||
|
|
||||||
|
# Ensure the mount path exists (within the filesystems)
|
||||||
|
retcode, stdout, stderr = run_os_command("mkdir -p {}".format(mount_path))
|
||||||
|
if retcode:
|
||||||
|
raise ProvisioningError("Failed to create mountpoint {}: {}".format(mount_path, stderr))
|
||||||
|
|
||||||
|
# Mount filesystems to temporary directory
|
||||||
|
retcode, stdout, stderr = run_os_command("mount {} {}".format(mapped_rbd_volume, mount_path))
|
||||||
|
if retcode:
|
||||||
|
raise ProvisioningError("Failed to mount {} on {}: {}".format(mapped_rbd_volume, mount_path, stderr))
|
||||||
|
|
||||||
|
print("Successfully mounted {} on {}".format(mapped_rbd_volume, mount_path))
|
||||||
|
|
||||||
# Phase 6 - provisioning script execution
|
# Phase 6 - provisioning script execution
|
||||||
# * Execute the provisioning script main function ("install") passing any custom arguments
|
# * Execute the provisioning script main function ("install") passing any custom arguments
|
||||||
self.update_state(state='RUNNING', meta={'current': 6, 'total': 10, 'status': 'Executing provisioning script'})
|
self.update_state(state='RUNNING', meta={'current': 6, 'total': 10, 'status': 'Executing provisioning script'})
|
||||||
time.sleep(5)
|
time.sleep(1)
|
||||||
|
|
||||||
|
print("Running installer script")
|
||||||
|
|
||||||
|
# Parse the script arguments
|
||||||
|
script_arguments = dict()
|
||||||
|
for argument in vm_data['script_arguments']:
|
||||||
|
argument_name, argument_data = argument.split('=')
|
||||||
|
script_arguments[argument_name] = argument_data
|
||||||
|
|
||||||
|
# Run the script
|
||||||
|
installer_script.install(
|
||||||
|
vm_name=vm_name,
|
||||||
|
vm_id=re.findall(r'/(\d+)$/', vm_name),
|
||||||
|
temporary_directory=temp_dir,
|
||||||
|
disks=vm_data['volumes'],
|
||||||
|
networks=vm_data['networks'],
|
||||||
|
**script_arguments
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
return
|
||||||
|
|
||||||
|
|
||||||
# Phase 7 - install cleanup
|
# Phase 7 - install cleanup
|
||||||
# * Unmount any mounted volumes
|
# * Unmount any mounted volumes
|
||||||
# * Remove any temporary directories
|
# * Remove any temporary directories
|
||||||
self.update_state(state='RUNNING', meta={'current': 7, 'total': 10, 'status': 'Cleaning up local mounts and directories'})
|
self.update_state(state='RUNNING', meta={'current': 7, 'total': 10, 'status': 'Cleaning up local mounts and directories'})
|
||||||
time.sleep(5)
|
time.sleep(1)
|
||||||
|
|
||||||
# Phase 8 - configuration creation
|
# Phase 8 - configuration creation
|
||||||
# * Create the libvirt XML configuration
|
# * Create the libvirt XML configuration
|
||||||
|
|
|
@ -593,6 +593,10 @@ def api_template_storage_disk_root(template):
|
||||||
* type: Valid Linux filesystem
|
* type: Valid Linux filesystem
|
||||||
* optional: true
|
* optional: true
|
||||||
* requires: N/A
|
* requires: N/A
|
||||||
|
?filesystem_arg: Argument for the guest filesystem
|
||||||
|
* type: Valid mkfs.<filesystem> argument, multiple
|
||||||
|
* optional: true
|
||||||
|
* requires: N/A
|
||||||
?mountpoint: The Linux guest mountpoint for the disk
|
?mountpoint: The Linux guest mountpoint for the disk
|
||||||
* default: unmounted in guest
|
* default: unmounted in guest
|
||||||
* type: Valid Linux mountpoint (e.g. '/', '/var', etc.)
|
* type: Valid Linux mountpoint (e.g. '/', '/var', etc.)
|
||||||
|
@ -629,12 +633,17 @@ def api_template_storage_disk_root(template):
|
||||||
else:
|
else:
|
||||||
filesystem = None
|
filesystem = None
|
||||||
|
|
||||||
|
if 'filesystem_arg' in flask.request.values:
|
||||||
|
filesystem_args = flask.request.values.getlist('filesystem_arg')
|
||||||
|
else:
|
||||||
|
filesystem_args = None
|
||||||
|
|
||||||
if 'mountpoint' in flask.request.values:
|
if 'mountpoint' in flask.request.values:
|
||||||
mountpoint = flask.request.values['mountpoint']
|
mountpoint = flask.request.values['mountpoint']
|
||||||
else:
|
else:
|
||||||
mountpoint = None
|
mountpoint = None
|
||||||
|
|
||||||
return pvcprovisioner.create_template_storage_element(template, pool, disk_id, disk_size, filesystem, mountpoint)
|
return pvcprovisioner.create_template_storage_element(template, pool, disk_id, disk_size, filesystem, filesystem_args, mountpoint)
|
||||||
|
|
||||||
if flask.request.method == 'DELETE':
|
if flask.request.method == 'DELETE':
|
||||||
if 'disk_id' in flask.request.values:
|
if 'disk_id' in flask.request.values:
|
||||||
|
@ -697,12 +706,17 @@ def api_template_storage_disk_element(template, disk_id):
|
||||||
else:
|
else:
|
||||||
filesystem = None
|
filesystem = None
|
||||||
|
|
||||||
|
if 'filesystem_arg' in flask.request.values:
|
||||||
|
filesystem_args = flask.request.values.getlist('filesystem_arg')
|
||||||
|
else:
|
||||||
|
filesystem_args = None
|
||||||
|
|
||||||
if 'mountpoint' in flask.request.values:
|
if 'mountpoint' in flask.request.values:
|
||||||
mountpoint = flask.request.values['mountpoint']
|
mountpoint = flask.request.values['mountpoint']
|
||||||
else:
|
else:
|
||||||
mountpoint = None
|
mountpoint = None
|
||||||
|
|
||||||
return pvcprovisioner.create_template_storage_element(template, pool, disk_id, disk_size, mountpoint, filesystem)
|
return pvcprovisioner.create_template_storage_element(template, pool, disk_id, disk_size, filesystem, filesystem_args, mountpoint)
|
||||||
|
|
||||||
if flask.request.method == 'DELETE':
|
if flask.request.method == 'DELETE':
|
||||||
return pvcprovisioner.delete_template_storage_element(template, disk_id)
|
return pvcprovisioner.delete_template_storage_element(template, disk_id)
|
||||||
|
|
|
@ -4,7 +4,7 @@ create table system_template (id SERIAL PRIMARY KEY, name TEXT NOT NULL UNIQUE,
|
||||||
create table network_template (id SERIAL PRIMARY KEY, name TEXT NOT NULL UNIQUE, mac_template TEXT);
|
create table network_template (id SERIAL PRIMARY KEY, name TEXT NOT NULL UNIQUE, mac_template TEXT);
|
||||||
create table network (id SERIAL PRIMARY KEY, network_template INT REFERENCES network_template(id), vni INT NOT NULL);
|
create table network (id SERIAL PRIMARY KEY, network_template INT REFERENCES network_template(id), vni INT NOT NULL);
|
||||||
create table storage_template (id SERIAL PRIMARY KEY, name TEXT NOT NULL UNIQUE);
|
create table storage_template (id SERIAL PRIMARY KEY, name TEXT NOT NULL UNIQUE);
|
||||||
create table storage (id SERIAL PRIMARY KEY, storage_template INT REFERENCES storage_template(id), pool TEXT NOT NULL, disk_id TEXT NOT NULL, disk_size_gb INT NOT NULL, mountpoint TEXT, filesystem TEXT);
|
create table storage (id SERIAL PRIMARY KEY, storage_template INT REFERENCES storage_template(id), pool TEXT NOT NULL, disk_id TEXT NOT NULL, disk_size_gb INT NOT NULL, mountpoint TEXT, filesystem TEXT, filesystem_args TEXT);
|
||||||
create table script (id SERIAL PRIMARY KEY, name TEXT NOT NULL UNIQUE, script TEXT NOT NULL);
|
create table script (id SERIAL PRIMARY KEY, name TEXT NOT NULL UNIQUE, script TEXT NOT NULL);
|
||||||
create table profile (id SERIAL PRIMARY KEY, name TEXT NOT NULL UNIQUE, system_template INT REFERENCES system_template(id), network_template INT REFERENCES network_template(id), storage_template INT REFERENCES storage_template(id), script INT REFERENCES script(id), arguments text);
|
create table profile (id SERIAL PRIMARY KEY, name TEXT NOT NULL UNIQUE, system_template INT REFERENCES system_template(id), network_template INT REFERENCES network_template(id), storage_template INT REFERENCES storage_template(id), script INT REFERENCES script(id), arguments text);
|
||||||
grant all privileges on database pvcprov to pvcprov;
|
grant all privileges on database pvcprov to pvcprov;
|
||||||
|
|
Loading…
Reference in New Issue