2018-09-20 03:25:58 -04:00
#!/usr/bin/env python3
# vm.py - PVC client function library, VM fuctions
# Part of the Parallel Virtual Cluster (PVC) system
#
2021-03-25 17:01:55 -04:00
# Copyright (C) 2018-2021 Joshua M. Boniface <joshua@boniface.me>
2018-09-20 03:25:58 -04:00
#
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
2021-03-25 16:57:17 -04:00
# the Free Software Foundation, version 3.
2018-09-20 03:25:58 -04:00
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with this program. If not, see <https://www.gnu.org/licenses/>.
#
###############################################################################
import time
import re
import lxml . objectify
2021-05-23 16:41:42 -04:00
import lxml . etree
2019-04-11 19:06:06 -04:00
2021-07-13 19:04:56 -04:00
from distutils . util import strtobool
2021-06-28 12:27:43 -04:00
from uuid import UUID
2021-07-01 17:24:47 -04:00
from concurrent . futures import ThreadPoolExecutor
2021-06-28 12:27:43 -04:00
2020-02-08 18:48:59 -05:00
import daemon_lib . common as common
2018-09-20 03:25:58 -04:00
2020-02-08 18:48:59 -05:00
import daemon_lib . ceph as ceph
2021-06-21 22:21:54 -04:00
from daemon_lib . network import set_sriov_vf_vm , unset_sriov_vf_vm
2019-06-27 11:19:48 -04:00
2020-11-07 14:45:24 -05:00
2018-09-20 03:25:58 -04:00
#
# Cluster search functions
#
2021-05-29 21:17:19 -04:00
def getClusterDomainList ( zkhandler ) :
2018-09-20 03:25:58 -04:00
# Get a list of UUIDs by listing the children of /domains
2021-06-13 13:41:21 -04:00
uuid_list = zkhandler . children ( ' base.domain ' )
2018-09-20 03:25:58 -04:00
name_list = [ ]
# For each UUID, get the corresponding name from the data
for uuid in uuid_list :
2021-06-13 13:41:21 -04:00
name_list . append ( zkhandler . read ( ( ' domain ' , uuid ) ) )
2018-09-20 03:25:58 -04:00
return uuid_list , name_list
2020-11-07 14:45:24 -05:00
2021-05-29 21:17:19 -04:00
def searchClusterByUUID ( zkhandler , uuid ) :
2018-09-20 03:25:58 -04:00
try :
# Get the lists
2021-05-29 21:17:19 -04:00
uuid_list , name_list = getClusterDomainList ( zkhandler )
2018-09-20 03:25:58 -04:00
# We're looking for UUID, so find that element ID
index = uuid_list . index ( uuid )
# Get the name_list element at that index
name = name_list [ index ]
except ValueError :
# We didn't find anything
return None
return name
2020-11-07 14:45:24 -05:00
2021-05-29 21:17:19 -04:00
def searchClusterByName ( zkhandler , name ) :
2018-09-20 03:25:58 -04:00
try :
# Get the lists
2021-05-29 21:17:19 -04:00
uuid_list , name_list = getClusterDomainList ( zkhandler )
2018-09-20 03:25:58 -04:00
# We're looking for name, so find that element ID
index = name_list . index ( name )
# Get the uuid_list element at that index
uuid = uuid_list [ index ]
except ValueError :
# We didn't find anything
return None
return uuid
2020-11-07 14:45:24 -05:00
2021-05-29 21:17:19 -04:00
def getDomainUUID ( zkhandler , domain ) :
2019-06-27 11:19:48 -04:00
# Validate that VM exists in cluster
2018-09-20 03:25:58 -04:00
if common . validateUUID ( domain ) :
2021-05-29 21:17:19 -04:00
dom_name = searchClusterByUUID ( zkhandler , domain )
dom_uuid = searchClusterByName ( zkhandler , dom_name )
2018-09-20 03:25:58 -04:00
else :
2021-05-29 21:17:19 -04:00
dom_uuid = searchClusterByName ( zkhandler , domain )
dom_name = searchClusterByUUID ( zkhandler , dom_uuid )
2018-09-20 03:25:58 -04:00
return dom_uuid
2020-11-07 14:45:24 -05:00
2021-05-29 21:17:19 -04:00
def getDomainName ( zkhandler , domain ) :
2019-06-27 11:19:48 -04:00
# Validate that VM exists in cluster
2018-09-25 02:26:37 -04:00
if common . validateUUID ( domain ) :
2021-05-29 21:17:19 -04:00
dom_name = searchClusterByUUID ( zkhandler , domain )
dom_uuid = searchClusterByName ( zkhandler , dom_name )
2018-09-25 02:26:37 -04:00
else :
2021-05-29 21:17:19 -04:00
dom_uuid = searchClusterByName ( zkhandler , domain )
dom_name = searchClusterByUUID ( zkhandler , dom_uuid )
2018-09-25 02:26:37 -04:00
return dom_name
2020-11-07 14:45:24 -05:00
2021-05-29 21:17:19 -04:00
#
# Helper functions
#
def change_state ( zkhandler , dom_uuid , new_state ) :
2021-09-12 15:41:05 -04:00
lock = zkhandler . exclusivelock ( ( ' domain.state ' , dom_uuid ) )
with lock :
zkhandler . write ( [
( ( ' domain.state ' , dom_uuid ) , new_state )
] )
2021-05-29 21:17:19 -04:00
2021-09-12 15:41:05 -04:00
# Wait for 1/2 second to allow state to flow to all nodes
time . sleep ( 0.5 )
2021-05-29 21:17:19 -04:00
2018-09-20 03:25:58 -04:00
#
# Direct functions
#
2021-05-29 21:17:19 -04:00
def is_migrated ( zkhandler , domain ) :
2019-07-25 14:33:50 -04:00
# Validate that VM exists in cluster
2021-05-29 21:17:19 -04:00
dom_uuid = getDomainUUID ( zkhandler , domain )
2019-07-25 14:33:50 -04:00
if not dom_uuid :
return False , ' ERROR: Could not find VM " {} " in the cluster! ' . format ( domain )
2021-06-13 13:41:21 -04:00
last_node = zkhandler . read ( ( ' domain.last_node ' , dom_uuid ) )
2019-07-25 14:33:50 -04:00
if last_node :
return True
2019-07-25 15:45:45 -04:00
else :
2019-07-25 14:33:50 -04:00
return False
2020-11-07 14:45:24 -05:00
2021-05-29 21:17:19 -04:00
def flush_locks ( zkhandler , domain ) :
2019-08-07 13:38:49 -04:00
# Validate that VM exists in cluster
2021-05-29 21:17:19 -04:00
dom_uuid = getDomainUUID ( zkhandler , domain )
2019-08-07 13:38:49 -04:00
if not dom_uuid :
return False , ' ERROR: Could not find VM " {} " in the cluster! ' . format ( domain )
# Verify that the VM is in a stopped state; freeing locks is not safe otherwise
2021-06-13 13:41:21 -04:00
state = zkhandler . read ( ( ' domain.state ' , dom_uuid ) )
2019-08-07 13:38:49 -04:00
if state != ' stop ' :
return False , ' ERROR: VM " {} " is not in stopped state; flushing RBD locks on a running VM is dangerous. ' . format ( domain )
# Tell the cluster to create a new OSD for the host
flush_locks_string = ' flush_locks {} ' . format ( dom_uuid )
2021-05-29 21:17:19 -04:00
zkhandler . write ( [
2021-06-13 13:41:21 -04:00
( ' base.cmd.domain ' , flush_locks_string )
2021-05-29 21:17:19 -04:00
] )
2019-08-07 13:38:49 -04:00
# Wait 1/2 second for the cluster to get the message and start working
time . sleep ( 0.5 )
# Acquire a read lock, so we get the return exclusively
2021-06-13 13:41:21 -04:00
lock = zkhandler . readlock ( ' base.cmd.domain ' )
2019-08-07 13:38:49 -04:00
with lock :
try :
2021-06-13 13:41:21 -04:00
result = zkhandler . read ( ' base.cmd.domain ' ) . split ( ) [ 0 ]
2019-08-07 13:38:49 -04:00
if result == ' success-flush_locks ' :
message = ' Flushed locks on VM " {} " ' . format ( domain )
success = True
else :
message = ' ERROR: Failed to flush locks on VM " {} " ; check node logs for details. ' . format ( domain )
success = False
2020-11-06 19:24:10 -05:00
except Exception :
2019-08-07 13:38:49 -04:00
message = ' ERROR: Command ignored by node. '
success = False
# Acquire a write lock to ensure things go smoothly
2021-06-13 13:41:21 -04:00
lock = zkhandler . writelock ( ' base.cmd.domain ' )
2019-08-07 13:38:49 -04:00
with lock :
time . sleep ( 0.5 )
2021-05-29 21:17:19 -04:00
zkhandler . write ( [
2021-06-13 13:41:21 -04:00
( ' base.cmd.domain ' , ' ' )
2021-05-29 21:17:19 -04:00
] )
2019-08-07 13:38:49 -04:00
return success , message
2020-11-07 14:45:24 -05:00
2021-07-13 01:46:50 -04:00
def define_vm ( zkhandler , config_data , target_node , node_limit , node_selector , node_autostart , migration_method = None , profile = None , tags = [ ] , initial_state = ' stop ' ) :
2018-09-20 03:25:58 -04:00
# Parse the XML data
2019-07-09 10:22:23 -04:00
try :
parsed_xml = lxml . objectify . fromstring ( config_data )
2020-11-06 19:24:10 -05:00
except Exception :
2019-07-09 10:22:23 -04:00
return False , ' ERROR: Failed to parse XML data. '
2018-09-20 03:25:58 -04:00
dom_uuid = parsed_xml . uuid . text
dom_name = parsed_xml . name . text
2019-12-09 09:56:59 -05:00
# Ensure that the UUID and name are unique
2021-05-29 21:17:19 -04:00
if searchClusterByUUID ( zkhandler , dom_uuid ) or searchClusterByName ( zkhandler , dom_name ) :
2019-12-09 09:56:59 -05:00
return False , ' ERROR: Specified VM " {} " or UUID " {} " matches an existing VM on the cluster ' . format ( dom_name , dom_uuid )
2019-06-24 13:37:56 -04:00
if not target_node :
2021-05-29 21:17:19 -04:00
target_node = common . findTargetNode ( zkhandler , dom_uuid )
2019-06-24 13:25:24 -04:00
else :
# Verify node is valid
2021-05-29 21:17:19 -04:00
valid_node = common . verifyNode ( zkhandler , target_node )
2019-06-24 13:25:24 -04:00
if not valid_node :
2019-07-09 10:22:23 -04:00
return False , ' ERROR: Specified node " {} " is invalid. ' . format ( target_node )
2021-09-13 01:46:53 -04:00
# Validate the new RAM against the current active node
node_total_memory = int ( zkhandler . read ( ( ' node.memory.total ' , target_node ) ) )
if int ( parsed_xml . memory . text ) > = node_total_memory :
return False , ' ERROR: VM configuration specifies more memory ( {} MiB) than node " {} " has available ( {} MiB). ' . format ( parsed_xml . memory . text , target_node , node_total_memory )
# Validate the number of vCPUs against the current active node
node_total_cpus = int ( zkhandler . read ( ( ' node.data.static ' , target_node ) ) . split ( ) [ 0 ] )
if ( node_total_cpus - 2 ) < = int ( parsed_xml . vcpu . text ) :
return False , ' ERROR: VM configuration specifies more vCPUs ( {} ) than node " {} " has available ( {} minus 2). ' . format ( parsed_xml . vcpu . text , target_node , node_total_cpus )
2021-06-21 22:21:54 -04:00
# If a SR-IOV network device is being added, set its used state
dnetworks = common . getDomainNetworks ( parsed_xml , { } )
for network in dnetworks :
if network [ ' type ' ] in [ ' direct ' , ' hostdev ' ] :
dom_node = zkhandler . read ( ( ' domain.node ' , dom_uuid ) )
# Check if the network is already in use
is_used = zkhandler . read ( ( ' node.sriov.vf ' , dom_node , ' sriov_vf.used ' , network [ ' source ' ] ) )
if is_used == ' True ' :
used_by_name = searchClusterByUUID ( zkhandler , zkhandler . read ( ( ' node.sriov.vf ' , dom_node, ' sriov_vf . used_by ' , network[ ' source ' ])))
return False , ' ERROR: Attempted to use SR-IOV network " {} " which is already used by VM " {} " on node " {} " . ' . format ( network [ ' source ' ] , used_by_name , dom_node )
# We must update the "used" section
set_sriov_vf_vm ( zkhandler , dom_uuid , dom_node , network [ ' source ' ] , network [ ' mac ' ] , network [ ' type ' ] )
2019-07-09 10:22:23 -04:00
# Obtain the RBD disk list using the common functions
2020-06-07 00:40:21 -04:00
ddisks = common . getDomainDisks ( parsed_xml , { } )
2019-07-09 10:22:23 -04:00
rbd_list = [ ]
for disk in ddisks :
if disk [ ' type ' ] == ' rbd ' :
rbd_list . append ( disk [ ' name ' ] )
2018-09-20 03:25:58 -04:00
2019-12-19 12:03:46 -05:00
# Join the limit
2019-12-19 13:22:38 -05:00
if isinstance ( node_limit , list ) and node_limit :
2019-12-19 12:03:46 -05:00
formatted_node_limit = ' , ' . join ( node_limit )
else :
2019-12-19 13:22:38 -05:00
formatted_node_limit = ' '
2019-12-19 12:03:46 -05:00
# Join the RBD list
2019-12-19 13:22:38 -05:00
if isinstance ( rbd_list , list ) and rbd_list :
2019-12-19 12:03:46 -05:00
formatted_rbd_list = ' , ' . join ( rbd_list )
else :
2019-12-19 13:22:38 -05:00
formatted_rbd_list = ' '
2019-12-19 12:03:46 -05:00
2018-09-20 03:25:58 -04:00
# Add the new domain to Zookeeper
2021-06-21 22:21:54 -04:00
zkhandler . write ( [
2021-06-13 13:41:21 -04:00
( ( ' domain ' , dom_uuid ) , dom_name ) ,
( ( ' domain.xml ' , dom_uuid ) , config_data ) ,
( ( ' domain.state ' , dom_uuid ) , initial_state ) ,
( ( ' domain.profile ' , dom_uuid ) , profile ) ,
( ( ' domain.stats ' , dom_uuid ) , ' ' ) ,
( ( ' domain.node ' , dom_uuid ) , target_node ) ,
( ( ' domain.last_node ' , dom_uuid ) , ' ' ) ,
( ( ' domain.failed_reason ' , dom_uuid ) , ' ' ) ,
( ( ' domain.storage.volumes ' , dom_uuid ) , formatted_rbd_list ) ,
( ( ' domain.console.log ' , dom_uuid ) , ' ' ) ,
( ( ' domain.console.vnc ' , dom_uuid ) , ' ' ) ,
( ( ' domain.meta.autostart ' , dom_uuid ) , node_autostart ) ,
( ( ' domain.meta.migrate_method ' , dom_uuid ) , migration_method ) ,
( ( ' domain.meta.node_limit ' , dom_uuid ) , formatted_node_limit ) ,
( ( ' domain.meta.node_selector ' , dom_uuid ) , node_selector ) ,
2021-07-13 19:04:56 -04:00
( ( ' domain.meta.tags ' , dom_uuid ) , ' ' ) ,
2021-06-13 15:49:30 -04:00
( ( ' domain.migrate.sync_lock ' , dom_uuid ) , ' ' ) ,
2021-05-29 21:17:19 -04:00
] )
2018-09-20 03:25:58 -04:00
2021-07-13 19:04:56 -04:00
for tag in tags :
tag_name = tag [ ' name ' ]
zkhandler . write ( [
( ( ' domain.meta.tags ' , dom_uuid , ' tag.name ' , tag_name ) , tag [ ' name ' ] ) ,
( ( ' domain.meta.tags ' , dom_uuid , ' tag.type ' , tag_name ) , tag [ ' type ' ] ) ,
( ( ' domain.meta.tags ' , dom_uuid , ' tag.protected ' , tag_name ) , tag [ ' protected ' ] ) ,
] )
2021-06-21 22:21:54 -04:00
return True , ' Added new VM with Name " {} " and UUID " {} " to database. ' . format ( dom_name , dom_uuid )
2018-09-20 03:25:58 -04:00
2020-11-07 14:45:24 -05:00
2021-09-12 15:41:05 -04:00
def attach_vm_device ( zkhandler , domain , device_spec_xml ) :
# Validate that VM exists in cluster
dom_uuid = getDomainUUID ( zkhandler , domain )
if not dom_uuid :
return False , ' ERROR: Could not find VM " {} " in the cluster! ' . format ( domain )
# Verify that the VM is in a stopped state; freeing locks is not safe otherwise
state = zkhandler . read ( ( ' domain.state ' , dom_uuid ) )
if state != ' start ' :
return False , ' ERROR: VM " {} " is not in started state; live-add unneccessary. ' . format ( domain )
# Tell the cluster to attach the device
attach_device_string = ' attach_device {} {} ' . format ( dom_uuid , device_spec_xml )
zkhandler . write ( [
( ' base.cmd.domain ' , attach_device_string )
] )
# Wait 1/2 second for the cluster to get the message and start working
time . sleep ( 0.5 )
# Acquire a read lock, so we get the return exclusively
lock = zkhandler . readlock ( ' base.cmd.domain ' )
with lock :
try :
result = zkhandler . read ( ' base.cmd.domain ' ) . split ( ) [ 0 ]
if result == ' success-attach_device ' :
message = ' Attached device on VM " {} " ' . format ( domain )
success = True
else :
message = ' ERROR: Failed to attach device on VM " {} " ; check node logs for details. ' . format ( domain )
success = False
except Exception :
message = ' ERROR: Command ignored by node. '
success = False
# Acquire a write lock to ensure things go smoothly
lock = zkhandler . writelock ( ' base.cmd.domain ' )
with lock :
time . sleep ( 0.5 )
zkhandler . write ( [
( ' base.cmd.domain ' , ' ' )
] )
return success , message
def detach_vm_device ( zkhandler , domain , device_spec_xml ) :
# Validate that VM exists in cluster
dom_uuid = getDomainUUID ( zkhandler , domain )
if not dom_uuid :
return False , ' ERROR: Could not find VM " {} " in the cluster! ' . format ( domain )
# Verify that the VM is in a stopped state; freeing locks is not safe otherwise
state = zkhandler . read ( ( ' domain.state ' , dom_uuid ) )
if state != ' start ' :
return False , ' ERROR: VM " {} " is not in started state; live-add unneccessary. ' . format ( domain )
# Tell the cluster to detach the device
detach_device_string = ' detach_device {} {} ' . format ( dom_uuid , device_spec_xml )
zkhandler . write ( [
( ' base.cmd.domain ' , detach_device_string )
] )
# Wait 1/2 second for the cluster to get the message and start working
time . sleep ( 0.5 )
# Acquire a read lock, so we get the return exclusively
lock = zkhandler . readlock ( ' base.cmd.domain ' )
with lock :
try :
result = zkhandler . read ( ' base.cmd.domain ' ) . split ( ) [ 0 ]
if result == ' success-detach_device ' :
message = ' Attached device on VM " {} " ' . format ( domain )
success = True
else :
message = ' ERROR: Failed to detach device on VM " {} " ; check node logs for details. ' . format ( domain )
success = False
except Exception :
message = ' ERROR: Command ignored by node. '
success = False
# Acquire a write lock to ensure things go smoothly
lock = zkhandler . writelock ( ' base.cmd.domain ' )
with lock :
time . sleep ( 0.5 )
zkhandler . write ( [
( ' base.cmd.domain ' , ' ' )
] )
return success , message
2021-05-29 21:17:19 -04:00
def modify_vm_metadata ( zkhandler , domain , node_limit , node_selector , node_autostart , provisioner_profile , migration_method ) :
dom_uuid = getDomainUUID ( zkhandler , domain )
2019-10-12 01:17:39 -04:00
if not dom_uuid :
return False , ' ERROR: Could not find VM " {} " in the cluster! ' . format ( domain )
2021-06-13 13:41:21 -04:00
update_list = list ( )
2019-10-12 01:17:39 -04:00
if node_limit is not None :
2021-06-13 13:41:21 -04:00
update_list . append ( ( ( ' domain.meta.node_limit ' , dom_uuid ) , node_limit ) )
2019-10-12 01:17:39 -04:00
if node_selector is not None :
2021-06-13 13:41:21 -04:00
update_list . append ( ( ( ' domain.meta.node_selector ' , dom_uuid ) , node_selector ) )
2019-10-12 01:17:39 -04:00
if node_autostart is not None :
2021-06-13 13:41:21 -04:00
update_list . append ( ( ( ' domain.meta.autostart ' , dom_uuid ) , node_autostart ) )
2019-10-12 01:17:39 -04:00
2020-01-30 11:45:46 -05:00
if provisioner_profile is not None :
2021-06-13 13:41:21 -04:00
update_list . append ( ( ( ' domain.profile ' , dom_uuid ) , provisioner_profile ) )
2020-01-30 11:45:46 -05:00
2020-10-29 11:31:32 -04:00
if migration_method is not None :
2021-06-13 13:41:21 -04:00
update_list . append ( ( ( ' domain.meta.migrate_method ' , dom_uuid ) , migration_method ) )
if len ( update_list ) < 1 :
return False , ' ERROR: No updates to apply. '
zkhandler . write ( update_list )
2020-10-29 11:31:32 -04:00
2019-10-12 01:17:39 -04:00
return True , ' Successfully modified PVC metadata of VM " {} " . ' . format ( domain )
2020-11-07 14:45:24 -05:00
2021-07-13 19:04:56 -04:00
def modify_vm_tag ( zkhandler , domain , action , tag , protected = False ) :
2021-07-13 01:46:50 -04:00
dom_uuid = getDomainUUID ( zkhandler , domain )
if not dom_uuid :
return False , ' ERROR: Could not find VM " {} " in the cluster! ' . format ( domain )
2021-07-13 19:04:56 -04:00
if action == ' add ' :
2021-07-13 01:46:50 -04:00
zkhandler . write ( [
2021-07-13 19:04:56 -04:00
( ( ' domain.meta.tags ' , dom_uuid , ' tag.name ' , tag ) , tag ) ,
( ( ' domain.meta.tags ' , dom_uuid , ' tag.type ' , tag ) , ' user ' ) ,
( ( ' domain.meta.tags ' , dom_uuid , ' tag.protected ' , tag ) , protected ) ,
2021-07-13 01:46:50 -04:00
] )
2021-07-13 19:04:56 -04:00
return True , ' Successfully added tag " {} " to VM " {} " . ' . format ( tag , domain )
elif action == ' remove ' :
if not zkhandler . exists ( ( ' domain.meta.tags ' , dom_uuid , ' tag ' , tag ) ) :
return False , ' The tag " {} " does not exist. ' . format ( tag )
if zkhandler . read ( ( ' domain.meta.tags ' , dom_uuid , ' tag.type ' , tag ) ) != ' user ' :
return False , ' The tag " {} " is not a user tag and cannot be removed. ' . format ( tag )
if bool ( strtobool ( zkhandler . read ( ( ' domain.meta.tags ' , dom_uuid , ' tag.protected ' , tag ) ) ) ) :
return False , ' The tag " {} " is protected and cannot be removed. ' . format ( tag )
zkhandler . delete ( [
( ( ' domain.meta.tags ' , dom_uuid , ' tag ' , tag ) )
2021-07-13 01:46:50 -04:00
] )
2021-07-13 19:04:56 -04:00
return True , ' Successfully removed tag " {} " from VM " {} " . ' . format ( tag , domain )
2021-07-13 01:46:50 -04:00
else :
return False , ' Specified tag action is not available. '
2021-05-29 21:17:19 -04:00
def modify_vm ( zkhandler , domain , restart , new_vm_config ) :
dom_uuid = getDomainUUID ( zkhandler , domain )
2019-04-11 19:06:06 -04:00
if not dom_uuid :
2018-09-20 03:25:58 -04:00
return False , ' ERROR: Could not find VM " {} " in the cluster! ' . format ( domain )
2021-05-29 21:17:19 -04:00
dom_name = getDomainName ( zkhandler , domain )
2018-09-20 03:25:58 -04:00
2020-05-12 11:07:58 -04:00
# Parse and valiate the XML
try :
parsed_xml = lxml . objectify . fromstring ( new_vm_config )
2020-11-06 19:24:10 -05:00
except Exception :
2021-06-21 22:21:54 -04:00
return False , ' ERROR: Failed to parse new XML data. '
2021-06-22 02:21:55 -04:00
# Get our old network list for comparison purposes
old_vm_config = zkhandler . read ( ( ' domain.xml ' , dom_uuid ) )
old_parsed_xml = lxml . objectify . fromstring ( old_vm_config )
old_dnetworks = common . getDomainNetworks ( old_parsed_xml , { } )
2021-09-13 01:46:53 -04:00
# Validate the new RAM against the current active node
node_name = zkhandler . read ( ( ' domain.node ' , dom_uuid ) )
node_total_memory = int ( zkhandler . read ( ( ' node.memory.total ' , node_name ) ) )
if int ( parsed_xml . memory . text ) > = node_total_memory :
return False , ' ERROR: Updated VM configuration specifies more memory ( {} MiB) than node " {} " has available ( {} MiB). ' . format ( parsed_xml . memory . text , node_name , node_total_memory )
# Validate the number of vCPUs against the current active node
node_total_cpus = int ( zkhandler . read ( ( ' node.data.static ' , node_name ) ) . split ( ) [ 0 ] )
if ( node_total_cpus - 2 ) < = int ( parsed_xml . vcpu . text ) :
return False , ' ERROR: Updated VM configuration specifies more vCPUs ( {} ) than node " {} " has available ( {} minus 2). ' . format ( parsed_xml . vcpu . text , node_name , node_total_cpus )
2021-06-21 22:21:54 -04:00
# If a SR-IOV network device is being added, set its used state
dnetworks = common . getDomainNetworks ( parsed_xml , { } )
for network in dnetworks :
2021-06-22 02:21:55 -04:00
# Ignore networks that are already there
2021-06-22 02:31:14 -04:00
if network [ ' source ' ] in [ net [ ' source ' ] for net in old_dnetworks ] :
2021-06-22 02:21:55 -04:00
continue
2021-06-21 22:21:54 -04:00
if network [ ' type ' ] in [ ' direct ' , ' hostdev ' ] :
dom_node = zkhandler . read ( ( ' domain.node ' , dom_uuid ) )
# Check if the network is already in use
is_used = zkhandler . read ( ( ' node.sriov.vf ' , dom_node , ' sriov_vf.used ' , network [ ' source ' ] ) )
if is_used == ' True ' :
used_by_name = searchClusterByUUID ( zkhandler , zkhandler . read ( ( ' node.sriov.vf ' , dom_node , ' sriov_vf.used_by ' , network [ ' source ' ] ) ) )
return False , ' ERROR: Attempted to use SR-IOV network " {} " which is already used by VM " {} " on node " {} " . ' . format ( network [ ' source ' ] , used_by_name , dom_node )
# We must update the "used" section
set_sriov_vf_vm ( zkhandler , dom_uuid , dom_node , network [ ' source ' ] , network [ ' mac ' ] , network [ ' type ' ] )
# If a SR-IOV network device is being removed, unset its used state
for network in old_dnetworks :
if network [ ' type ' ] in [ ' direct ' , ' hostdev ' ] :
if network [ ' mac ' ] not in [ n [ ' mac ' ] for n in dnetworks ] :
dom_node = zkhandler . read ( ( ' domain.node ' , dom_uuid ) )
# We must update the "used" section
unset_sriov_vf_vm ( zkhandler , dom_node , network [ ' source ' ] )
2020-05-12 11:07:58 -04:00
# Obtain the RBD disk list using the common functions
2020-06-07 00:40:21 -04:00
ddisks = common . getDomainDisks ( parsed_xml , { } )
2020-05-12 11:07:58 -04:00
rbd_list = [ ]
for disk in ddisks :
if disk [ ' type ' ] == ' rbd ' :
rbd_list . append ( disk [ ' name ' ] )
# Join the RBD list
if isinstance ( rbd_list , list ) and rbd_list :
formatted_rbd_list = ' , ' . join ( rbd_list )
else :
formatted_rbd_list = ' '
2018-09-20 03:25:58 -04:00
# Add the modified config to Zookeeper
2021-05-29 21:17:19 -04:00
zkhandler . write ( [
2021-06-13 13:41:21 -04:00
( ( ' domain ' , dom_uuid ) , dom_name ) ,
( ( ' domain.storage.volumes ' , dom_uuid ) , formatted_rbd_list ) ,
( ( ' domain.xml ' , dom_uuid ) , new_vm_config ) ,
2021-05-29 21:17:19 -04:00
] )
2018-09-20 03:25:58 -04:00
2020-01-30 11:15:07 -05:00
if restart :
2021-05-29 21:17:19 -04:00
change_state ( zkhandler , dom_uuid , ' restart ' )
2020-01-30 11:15:07 -05:00
2020-11-07 18:01:26 -05:00
return True , ' Successfully modified configuration of VM " {} " . ' . format ( domain )
2018-09-20 03:25:58 -04:00
2020-11-07 14:45:24 -05:00
2021-05-29 21:17:19 -04:00
def dump_vm ( zkhandler , domain ) :
dom_uuid = getDomainUUID ( zkhandler , domain )
2019-04-11 19:06:06 -04:00
if not dom_uuid :
2019-03-12 21:09:54 -04:00
return False , ' ERROR: Could not find VM " {} " in the cluster! ' . format ( domain )
# Gram the domain XML and dump it to stdout
2021-06-13 13:41:21 -04:00
vm_xml = zkhandler . read ( ( ' domain.xml ' , dom_uuid ) )
2019-03-12 21:09:54 -04:00
2019-05-20 22:15:28 -04:00
return True , vm_xml
2019-03-12 21:09:54 -04:00
2020-11-07 14:45:24 -05:00
2021-05-29 21:17:19 -04:00
def rename_vm ( zkhandler , domain , new_domain ) :
dom_uuid = getDomainUUID ( zkhandler , domain )
2021-05-23 16:41:42 -04:00
if not dom_uuid :
return False , ' ERROR: Could not find VM " {} " in the cluster! ' . format ( domain )
# Verify that the VM is in a stopped state; renaming is not supported otherwise
2021-06-13 13:41:21 -04:00
state = zkhandler . read ( ( ' domain.state ' , dom_uuid ) )
2021-05-23 16:41:42 -04:00
if state != ' stop ' :
return False , ' ERROR: VM " {} " is not in stopped state; VMs cannot be renamed while running. ' . format ( domain )
# Parse and valiate the XML
2021-05-29 21:17:19 -04:00
vm_config = common . getDomainXML ( zkhandler , dom_uuid )
2021-05-23 16:41:42 -04:00
# Obtain the RBD disk list using the common functions
ddisks = common . getDomainDisks ( vm_config , { } )
pool_list = [ ]
rbd_list = [ ]
for disk in ddisks :
if disk [ ' type ' ] == ' rbd ' :
pool_list . append ( disk [ ' name ' ] . split ( ' / ' ) [ 0 ] )
rbd_list . append ( disk [ ' name ' ] . split ( ' / ' ) [ 1 ] )
# Rename each volume in turn
for idx , rbd in enumerate ( rbd_list ) :
rbd_new = re . sub ( r " {} " . format ( domain ) , new_domain , rbd )
# Skip renaming if nothing changed
if rbd_new == rbd :
continue
2021-05-29 21:17:19 -04:00
ceph . rename_volume ( zkhandler , pool_list [ idx ] , rbd , rbd_new )
2021-05-23 16:41:42 -04:00
# Replace the name in the config
vm_config_new = lxml . etree . tostring ( vm_config , encoding = ' ascii ' , method = ' xml ' ) . decode ( ) . replace ( domain , new_domain )
# Get VM information
2021-05-29 21:17:19 -04:00
_b , dom_info = get_info ( zkhandler , dom_uuid )
2021-05-23 16:41:42 -04:00
# Undefine the old VM
2021-05-29 21:17:19 -04:00
undefine_vm ( zkhandler , dom_uuid )
2021-05-23 16:41:42 -04:00
# Define the new VM
2021-07-13 19:04:56 -04:00
define_vm ( zkhandler , vm_config_new , dom_info [ ' node ' ] , dom_info [ ' node_limit ' ] , dom_info [ ' node_selector ' ] , dom_info [ ' node_autostart ' ] , migration_method = dom_info [ ' migration_method ' ] , profile = dom_info [ ' profile ' ] , tags = dom_info [ ' tags ' ] , initial_state = ' stop ' )
2021-05-23 16:41:42 -04:00
# If the VM is migrated, store that
if dom_info [ ' migrated ' ] != ' no ' :
2021-05-29 21:17:19 -04:00
zkhandler . write ( [
2021-06-13 13:41:21 -04:00
( ( ' domain.last_node ' , dom_uuid ) , dom_info [ ' last_node ' ] )
2021-05-29 21:17:19 -04:00
] )
2021-05-23 16:41:42 -04:00
return True , ' Successfully renamed VM " {} " to " {} " . ' . format ( domain , new_domain )
2021-05-29 21:17:19 -04:00
def undefine_vm ( zkhandler , domain ) :
2019-06-27 11:19:48 -04:00
# Validate that VM exists in cluster
2021-05-29 21:17:19 -04:00
dom_uuid = getDomainUUID ( zkhandler , domain )
2019-04-11 19:06:06 -04:00
if not dom_uuid :
2018-09-20 03:25:58 -04:00
return False , ' ERROR: Could not find VM " {} " in the cluster! ' . format ( domain )
# Shut down the VM
2021-06-13 13:41:21 -04:00
current_vm_state = zkhandler . read ( ( ' domain.state ' , dom_uuid ) )
2019-06-27 11:19:48 -04:00
if current_vm_state != ' stop ' :
2021-05-29 21:17:19 -04:00
change_state ( zkhandler , dom_uuid , ' stop ' )
2019-06-27 11:19:48 -04:00
# Gracefully terminate the class instances
2021-05-29 21:17:19 -04:00
change_state ( zkhandler , dom_uuid , ' delete ' )
2018-09-20 03:25:58 -04:00
# Delete the configurations
2021-06-13 13:41:21 -04:00
zkhandler . delete ( [
( ' domain ' , dom_uuid )
] )
2019-06-27 11:19:48 -04:00
return True , ' Undefined VM " {} " from the cluster. ' . format ( domain )
2020-11-07 14:45:24 -05:00
2021-05-29 21:17:19 -04:00
def remove_vm ( zkhandler , domain ) :
2019-06-27 11:19:48 -04:00
# Validate that VM exists in cluster
2021-05-29 21:17:19 -04:00
dom_uuid = getDomainUUID ( zkhandler , domain )
2019-06-27 11:19:48 -04:00
if not dom_uuid :
return False , ' ERROR: Could not find VM " {} " in the cluster! ' . format ( domain )
2021-05-29 21:17:19 -04:00
disk_list = common . getDomainDiskList ( zkhandler , dom_uuid )
2019-06-27 11:19:48 -04:00
# Shut down the VM
2021-06-13 13:41:21 -04:00
current_vm_state = zkhandler . read ( ( ' domain.state ' , dom_uuid ) )
2019-06-27 11:19:48 -04:00
if current_vm_state != ' stop ' :
2021-05-29 21:17:19 -04:00
change_state ( zkhandler , dom_uuid , ' stop ' )
2019-06-27 11:19:48 -04:00
2021-05-29 21:17:19 -04:00
# Wait for 1 second to allow state to flow to all nodes
time . sleep ( 1 )
2019-06-27 11:19:48 -04:00
# Remove disks
for disk in disk_list :
# vmpool/vmname_volume
2021-07-09 15:48:57 -04:00
try :
disk_pool , disk_name = disk . split ( ' / ' )
except ValueError :
continue
2021-07-09 15:39:06 -04:00
retcode , message = ceph . remove_volume ( zkhandler , disk_pool , disk_name )
if not retcode :
2021-07-09 15:48:57 -04:00
if re . match ( ' ^ERROR: No volume with name ' , message ) :
continue
else :
return False , message
2021-07-09 15:39:06 -04:00
# Gracefully terminate the class instances
change_state ( zkhandler , dom_uuid , ' delete ' )
# Wait for 1/2 second to allow state to flow to all nodes
time . sleep ( 0.5 )
# Delete the VM configuration from Zookeeper
zkhandler . delete ( [
( ' domain ' , dom_uuid )
] )
2018-09-20 03:25:58 -04:00
2021-07-09 15:39:06 -04:00
return True , ' Removed VM " {} " and its disks from the cluster. ' . format ( domain )
2018-09-20 03:25:58 -04:00
2020-11-07 14:45:24 -05:00
2021-05-29 21:17:19 -04:00
def start_vm ( zkhandler , domain ) :
2019-06-27 11:19:48 -04:00
# Validate that VM exists in cluster
2021-05-29 21:17:19 -04:00
dom_uuid = getDomainUUID ( zkhandler , domain )
2019-04-11 19:06:06 -04:00
if not dom_uuid :
2018-09-20 03:25:58 -04:00
return False , ' ERROR: Could not find VM " {} " in the cluster! ' . format ( domain )
# Set the VM to start
2021-05-29 21:17:19 -04:00
change_state ( zkhandler , dom_uuid , ' start ' )
2018-09-20 03:25:58 -04:00
2019-07-05 16:38:54 -04:00
return True , ' Starting VM " {} " . ' . format ( domain )
2018-09-20 03:25:58 -04:00
2020-11-07 14:45:24 -05:00
2021-05-29 21:17:19 -04:00
def restart_vm ( zkhandler , domain , wait = False ) :
2019-06-27 11:19:48 -04:00
# Validate that VM exists in cluster
2021-05-29 21:17:19 -04:00
dom_uuid = getDomainUUID ( zkhandler , domain )
2019-04-11 19:06:06 -04:00
if not dom_uuid :
2018-09-20 03:25:58 -04:00
return False , ' ERROR: Could not find VM " {} " in the cluster! ' . format ( domain )
# Get state and verify we're OK to proceed
2021-06-13 13:41:21 -04:00
current_state = zkhandler . read ( ( ' domain.state ' , dom_uuid ) )
2018-09-20 03:25:58 -04:00
if current_state != ' start ' :
2019-07-05 16:38:54 -04:00
return False , ' ERROR: VM " {} " is not in " start " state! ' . format ( domain )
2018-09-20 03:25:58 -04:00
2020-02-19 09:45:31 -05:00
retmsg = ' Restarting VM " {} " . ' . format ( domain )
# Set the VM to restart
2021-05-29 21:17:19 -04:00
change_state ( zkhandler , dom_uuid , ' restart ' )
2018-09-20 03:25:58 -04:00
2020-02-19 09:45:31 -05:00
if wait :
2021-06-13 13:41:21 -04:00
while zkhandler . read ( ( ' domain.state ' , dom_uuid ) ) == ' restart ' :
time . sleep ( 0.5 )
2020-02-19 09:45:31 -05:00
retmsg = ' Restarted VM " {} " ' . format ( domain )
2018-09-20 03:25:58 -04:00
2020-02-19 09:45:31 -05:00
return True , retmsg
2020-11-07 14:45:24 -05:00
2021-05-29 21:17:19 -04:00
def shutdown_vm ( zkhandler , domain , wait = False ) :
2019-06-27 11:19:48 -04:00
# Validate that VM exists in cluster
2021-05-29 21:17:19 -04:00
dom_uuid = getDomainUUID ( zkhandler , domain )
2019-04-11 19:06:06 -04:00
if not dom_uuid :
2018-09-20 03:25:58 -04:00
return False , ' ERROR: Could not find VM " {} " in the cluster! ' . format ( domain )
# Get state and verify we're OK to proceed
2021-06-13 13:41:21 -04:00
current_state = zkhandler . read ( ( ' domain.state ' , dom_uuid ) )
2018-09-20 03:25:58 -04:00
if current_state != ' start ' :
2019-07-05 16:38:54 -04:00
return False , ' ERROR: VM " {} " is not in " start " state! ' . format ( domain )
2018-09-20 03:25:58 -04:00
2020-02-19 09:45:31 -05:00
retmsg = ' Shutting down VM " {} " ' . format ( domain )
2018-09-20 03:25:58 -04:00
# Set the VM to shutdown
2021-05-29 21:17:19 -04:00
change_state ( zkhandler , dom_uuid , ' shutdown ' )
2020-02-19 09:45:31 -05:00
if wait :
2021-06-13 13:41:21 -04:00
while zkhandler . read ( ( ' domain.state ' , dom_uuid ) ) == ' shutdown ' :
time . sleep ( 0.5 )
2020-02-19 10:04:58 -05:00
retmsg = ' Shut down VM " {} " ' . format ( domain )
2018-09-20 03:25:58 -04:00
2020-02-19 09:45:31 -05:00
return True , retmsg
2018-09-20 03:25:58 -04:00
2020-11-07 14:45:24 -05:00
2021-05-29 21:17:19 -04:00
def stop_vm ( zkhandler , domain ) :
2019-06-27 11:19:48 -04:00
# Validate that VM exists in cluster
2021-05-29 21:17:19 -04:00
dom_uuid = getDomainUUID ( zkhandler , domain )
2019-04-11 19:06:06 -04:00
if not dom_uuid :
2018-09-20 03:25:58 -04:00
return False , ' ERROR: Could not find VM " {} " in the cluster! ' . format ( domain )
2021-05-29 21:17:19 -04:00
# Set the VM to stop
change_state ( zkhandler , dom_uuid , ' stop ' )
2018-09-20 03:25:58 -04:00
2019-07-05 16:38:54 -04:00
return True , ' Forcibly stopping VM " {} " . ' . format ( domain )
2018-09-20 03:25:58 -04:00
2020-11-07 14:45:24 -05:00
2021-05-29 21:17:19 -04:00
def disable_vm ( zkhandler , domain ) :
2019-10-23 23:37:42 -04:00
# Validate that VM exists in cluster
2021-05-29 21:17:19 -04:00
dom_uuid = getDomainUUID ( zkhandler , domain )
2019-10-23 23:37:42 -04:00
if not dom_uuid :
return False , ' ERROR: Could not find VM " {} " in the cluster! ' . format ( domain )
# Get state and verify we're OK to proceed
2021-06-13 13:41:21 -04:00
current_state = zkhandler . read ( ( ' domain.state ' , dom_uuid ) )
2019-10-23 23:37:42 -04:00
if current_state != ' stop ' :
return False , ' ERROR: VM " {} " must be stopped before disabling! ' . format ( domain )
2021-05-29 21:17:19 -04:00
# Set the VM to disable
change_state ( zkhandler , dom_uuid , ' disable ' )
2019-10-23 23:37:42 -04:00
2019-10-23 23:53:25 -04:00
return True , ' Marked VM " {} " as disable. ' . format ( domain )
2019-10-23 23:37:42 -04:00
2020-11-07 14:45:24 -05:00
2021-06-21 23:18:34 -04:00
def update_vm_sriov_nics ( zkhandler , dom_uuid , source_node , target_node ) :
# Update all the SR-IOV device states on both nodes, used during migrations but called by the node-side
vm_config = zkhandler . read ( ( ' domain.xml ' , dom_uuid ) )
parsed_xml = lxml . objectify . fromstring ( vm_config )
dnetworks = common . getDomainNetworks ( parsed_xml , { } )
retcode = True
retmsg = ' '
for network in dnetworks :
if network [ ' type ' ] in [ ' direct ' , ' hostdev ' ] :
# Check if the network is already in use
is_used = zkhandler . read ( ( ' node.sriov.vf ' , target_node , ' sriov_vf.used ' , network [ ' source ' ] ) )
if is_used == ' True ' :
used_by_name = searchClusterByUUID ( zkhandler , zkhandler . read ( ( ' node.sriov.vf ' , target_node , ' sriov_vf.used_by ' , network [ ' source ' ] ) ) )
if retcode :
2021-06-21 23:46:47 -04:00
retcode_this = False
2021-06-21 23:18:34 -04:00
retmsg = ' Attempting to use SR-IOV network " {} " which is already used by VM " {} " ' . format ( network [ ' source ' ] , used_by_name )
2021-06-21 23:46:47 -04:00
else :
retcode_this = True
2021-06-21 23:18:34 -04:00
# We must update the "used" section
2021-06-21 23:46:47 -04:00
if retcode_this :
2021-06-21 23:18:34 -04:00
# This conditional ensure that if we failed the is_used check, we don't try to overwrite the information of a VF that belongs to another VM
set_sriov_vf_vm ( zkhandler , dom_uuid , target_node , network [ ' source ' ] , network [ ' mac ' ] , network [ ' type ' ] )
# ... but we still want to free the old node in an case
unset_sriov_vf_vm ( zkhandler , source_node , network [ ' source ' ] )
2021-06-21 23:46:47 -04:00
if not retcode_this :
retcode = retcode_this
2021-06-21 23:18:34 -04:00
return retcode , retmsg
2021-05-29 21:17:19 -04:00
def move_vm ( zkhandler , domain , target_node , wait = False , force_live = False ) :
2019-06-27 11:19:48 -04:00
# Validate that VM exists in cluster
2021-05-29 21:17:19 -04:00
dom_uuid = getDomainUUID ( zkhandler , domain )
2019-04-11 19:06:06 -04:00
if not dom_uuid :
2018-09-20 03:25:58 -04:00
return False , ' ERROR: Could not find VM " {} " in the cluster! ' . format ( domain )
2020-02-19 09:45:31 -05:00
# Get state and verify we're OK to proceed
2021-06-13 13:41:21 -04:00
current_state = zkhandler . read ( ( ' domain.state ' , dom_uuid ) )
2020-02-19 09:45:31 -05:00
if current_state != ' start ' :
# If the current state isn't start, preserve it; we're not doing live migration
target_state = current_state
else :
2020-06-06 11:49:21 -04:00
if force_live :
target_state = ' migrate-live '
else :
target_state = ' migrate '
2020-02-19 09:45:31 -05:00
2021-06-13 13:41:21 -04:00
current_node = zkhandler . read ( ( ' domain.node ' , dom_uuid ) )
2018-09-20 03:25:58 -04:00
2019-06-24 13:37:56 -04:00
if not target_node :
2021-05-29 21:17:19 -04:00
target_node = common . findTargetNode ( zkhandler , dom_uuid )
2018-09-20 03:25:58 -04:00
else :
2019-06-24 13:25:24 -04:00
# Verify node is valid
2021-05-29 21:17:19 -04:00
valid_node = common . verifyNode ( zkhandler , target_node )
2019-06-24 13:25:24 -04:00
if not valid_node :
2019-10-12 01:50:15 -04:00
return False , ' ERROR: Specified node " {} " is invalid. ' . format ( target_node )
2019-06-24 13:25:24 -04:00
2019-10-12 01:45:44 -04:00
# Check if node is within the limit
2021-06-13 13:41:21 -04:00
node_limit = zkhandler . read ( ( ' domain.meta.node_limit ' , dom_uuid ) )
2020-01-05 13:42:23 -05:00
if node_limit and target_node not in node_limit . split ( ' , ' ) :
2019-10-12 01:50:15 -04:00
return False , ' ERROR: Specified node " {} " is not in the allowed list of nodes for VM " {} " . ' . format ( target_node , domain )
2019-10-12 01:45:44 -04:00
2019-06-24 13:25:24 -04:00
# Verify if node is current node
2018-10-14 02:01:35 -04:00
if target_node == current_node :
2021-06-13 13:41:21 -04:00
last_node = zkhandler . read ( ( ' domain.last_node ' , dom_uuid ) )
2020-06-06 11:21:58 -04:00
if last_node :
2021-05-29 21:17:19 -04:00
zkhandler . write ( [
2021-06-13 13:41:21 -04:00
( ( ' domain.last_node ' , dom_uuid ) , ' ' )
2021-05-29 21:17:19 -04:00
] )
2020-06-06 11:21:58 -04:00
return True , ' Making temporary migration permanent for VM " {} " . ' . format ( domain )
2019-07-05 16:38:54 -04:00
return False , ' ERROR: VM " {} " is already running on node " {} " . ' . format ( domain , current_node )
2018-09-20 03:25:58 -04:00
2019-10-17 12:16:21 -04:00
if not target_node :
return False , ' ERROR: Could not find a valid migration target for VM " {} " . ' . format ( domain )
2020-02-19 09:45:31 -05:00
retmsg = ' Permanently migrating VM " {} " to node " {} " . ' . format ( domain , target_node )
2018-09-20 03:25:58 -04:00
2021-06-13 13:41:21 -04:00
lock = zkhandler . exclusivelock ( ( ' domain.state ' , dom_uuid ) )
2021-05-29 21:17:19 -04:00
with lock :
zkhandler . write ( [
2021-06-13 13:41:21 -04:00
( ( ' domain.state ' , dom_uuid ) , target_state ) ,
( ( ' domain.node ' , dom_uuid ) , target_node ) ,
( ( ' domain.last_node ' , dom_uuid ) , ' ' )
2021-05-29 21:17:19 -04:00
] )
# Wait for 1/2 second for migration to start
time . sleep ( 0.5 )
2018-09-20 03:25:58 -04:00
2021-06-22 00:00:50 -04:00
# Update any SR-IOV NICs
update_vm_sriov_nics ( zkhandler , dom_uuid , current_node , target_node )
2020-02-19 09:45:31 -05:00
if wait :
2021-06-13 13:41:21 -04:00
while zkhandler . read ( ( ' domain.state ' , dom_uuid ) ) == target_state :
time . sleep ( 0.5 )
2020-02-19 09:45:31 -05:00
retmsg = ' Permanently migrated VM " {} " to node " {} " ' . format ( domain , target_node )
return True , retmsg
2020-11-07 14:45:24 -05:00
2021-05-29 21:17:19 -04:00
def migrate_vm ( zkhandler , domain , target_node , force_migrate , wait = False , force_live = False ) :
2019-06-27 11:19:48 -04:00
# Validate that VM exists in cluster
2021-05-29 21:17:19 -04:00
dom_uuid = getDomainUUID ( zkhandler , domain )
2019-04-11 19:06:06 -04:00
if not dom_uuid :
2018-09-20 03:25:58 -04:00
return False , ' ERROR: Could not find VM " {} " in the cluster! ' . format ( domain )
# Get state and verify we're OK to proceed
2021-06-13 13:41:21 -04:00
current_state = zkhandler . read ( ( ' domain.state ' , dom_uuid ) )
2018-09-20 03:25:58 -04:00
if current_state != ' start ' :
2020-02-19 09:45:31 -05:00
# If the current state isn't start, preserve it; we're not doing live migration
target_state = current_state
2018-09-20 03:25:58 -04:00
else :
2020-06-06 11:49:21 -04:00
if force_live :
target_state = ' migrate-live '
else :
target_state = ' migrate '
2018-09-20 03:25:58 -04:00
2021-06-13 13:41:21 -04:00
current_node = zkhandler . read ( ( ' domain.node ' , dom_uuid ) )
last_node = zkhandler . read ( ( ' domain.last_node ' , dom_uuid ) )
2018-09-20 03:25:58 -04:00
2019-06-24 13:37:56 -04:00
if last_node and not force_migrate :
2020-02-19 10:18:41 -05:00
return False , ' ERROR: VM " {} " has been previously migrated. ' . format ( domain )
2018-09-20 03:25:58 -04:00
2019-06-24 13:37:56 -04:00
if not target_node :
2021-05-29 21:17:19 -04:00
target_node = common . findTargetNode ( zkhandler , dom_uuid )
2018-09-20 03:25:58 -04:00
else :
2019-06-24 13:25:24 -04:00
# Verify node is valid
2021-05-29 21:17:19 -04:00
valid_node = common . verifyNode ( zkhandler , target_node )
2019-06-24 13:25:24 -04:00
if not valid_node :
2019-10-12 01:50:15 -04:00
return False , ' ERROR: Specified node " {} " is invalid. ' . format ( target_node )
2019-06-24 13:25:24 -04:00
2019-10-12 01:45:44 -04:00
# Check if node is within the limit
2021-06-13 13:41:21 -04:00
node_limit = zkhandler . read ( ( ' domain.meta.node_limit ' , dom_uuid ) )
2020-01-05 13:42:23 -05:00
if node_limit and target_node not in node_limit . split ( ' , ' ) :
2019-10-12 01:50:15 -04:00
return False , ' ERROR: Specified node " {} " is not in the allowed list of nodes for VM " {} " . ' . format ( target_node , domain )
2019-10-12 01:45:44 -04:00
2019-06-24 13:25:24 -04:00
# Verify if node is current node
2018-10-14 02:01:35 -04:00
if target_node == current_node :
2019-07-05 16:38:54 -04:00
return False , ' ERROR: VM " {} " is already running on node " {} " . ' . format ( domain , current_node )
2018-09-20 03:25:58 -04:00
2019-10-17 12:16:21 -04:00
if not target_node :
return False , ' ERROR: Could not find a valid migration target for VM " {} " . ' . format ( domain )
2019-07-07 15:10:48 -04:00
# Don't overwrite an existing last_node when using force_migrate
2021-06-22 00:00:50 -04:00
real_current_node = current_node # Used for the SR-IOV update
2019-07-07 15:10:48 -04:00
if last_node and force_migrate :
current_node = last_node
2020-02-19 09:45:31 -05:00
retmsg = ' Migrating VM " {} " to node " {} " . ' . format ( domain , target_node )
2021-06-13 13:41:21 -04:00
lock = zkhandler . exclusivelock ( ( ' domain.state ' , dom_uuid ) )
2021-05-29 21:17:19 -04:00
with lock :
zkhandler . write ( [
2021-06-13 13:41:21 -04:00
( ( ' domain.state ' , dom_uuid ) , target_state ) ,
( ( ' domain.node ' , dom_uuid ) , target_node ) ,
( ( ' domain.last_node ' , dom_uuid ) , current_node )
2021-05-29 21:17:19 -04:00
] )
# Wait for 1/2 second for migration to start
time . sleep ( 0.5 )
2018-09-20 03:25:58 -04:00
2021-06-22 00:00:50 -04:00
# Update any SR-IOV NICs
update_vm_sriov_nics ( zkhandler , dom_uuid , real_current_node , target_node )
2020-02-19 09:45:31 -05:00
if wait :
2021-06-13 13:41:21 -04:00
while zkhandler . read ( ( ' domain.state ' , dom_uuid ) ) == target_state :
time . sleep ( 0.5 )
2020-02-19 09:45:31 -05:00
retmsg = ' Migrated VM " {} " to node " {} " ' . format ( domain , target_node )
return True , retmsg
2018-09-20 03:25:58 -04:00
2020-11-07 14:45:24 -05:00
2021-05-29 21:17:19 -04:00
def unmigrate_vm ( zkhandler , domain , wait = False , force_live = False ) :
2019-06-27 11:19:48 -04:00
# Validate that VM exists in cluster
2021-05-29 21:17:19 -04:00
dom_uuid = getDomainUUID ( zkhandler , domain )
2019-04-11 19:06:06 -04:00
if not dom_uuid :
2018-09-20 03:25:58 -04:00
return False , ' ERROR: Could not find VM " {} " in the cluster! ' . format ( domain )
# Get state and verify we're OK to proceed
2021-06-13 13:41:21 -04:00
current_state = zkhandler . read ( ( ' domain.state ' , dom_uuid ) )
2018-09-20 03:25:58 -04:00
if current_state != ' start ' :
2019-03-20 10:19:01 -04:00
# If the current state isn't start, preserve it; we're not doing live migration
target_state = current_state
2018-09-20 03:25:58 -04:00
else :
2020-06-06 11:49:21 -04:00
if force_live :
target_state = ' migrate-live '
else :
target_state = ' migrate '
2018-09-20 03:25:58 -04:00
2021-06-22 00:00:50 -04:00
current_node = zkhandler . read ( ( ' domain.node ' , dom_uuid ) )
2021-06-13 13:41:21 -04:00
target_node = zkhandler . read ( ( ' domain.last_node ' , dom_uuid ) )
2018-09-20 03:25:58 -04:00
2018-10-14 02:01:35 -04:00
if target_node == ' ' :
2019-07-05 16:38:54 -04:00
return False , ' ERROR: VM " {} " has not been previously migrated. ' . format ( domain )
2018-09-20 03:25:58 -04:00
2020-02-19 09:45:31 -05:00
retmsg = ' Unmigrating VM " {} " back to node " {} " . ' . format ( domain , target_node )
2021-06-13 13:41:21 -04:00
lock = zkhandler . exclusivelock ( ( ' domain.state ' , dom_uuid ) )
2021-05-29 21:17:19 -04:00
with lock :
zkhandler . write ( [
2021-06-13 13:41:21 -04:00
( ( ' domain.state ' , dom_uuid ) , target_state ) ,
( ( ' domain.node ' , dom_uuid ) , target_node ) ,
( ( ' domain.last_node ' , dom_uuid ) , ' ' )
2021-05-29 21:17:19 -04:00
] )
# Wait for 1/2 second for migration to start
time . sleep ( 0.5 )
2018-09-20 03:25:58 -04:00
2021-06-22 00:00:50 -04:00
# Update any SR-IOV NICs
update_vm_sriov_nics ( zkhandler , dom_uuid , current_node , target_node )
2020-02-19 09:45:31 -05:00
if wait :
2021-06-13 13:41:21 -04:00
while zkhandler . read ( ( ' domain.state ' , dom_uuid ) ) == target_state :
time . sleep ( 0.5 )
2020-02-19 09:45:31 -05:00
retmsg = ' Unmigrated VM " {} " back to node " {} " ' . format ( domain , target_node )
return True , retmsg
2018-09-20 03:25:58 -04:00
2020-11-07 14:45:24 -05:00
2021-05-29 21:17:19 -04:00
def get_console_log ( zkhandler , domain , lines = 1000 ) :
2019-06-27 11:19:48 -04:00
# Validate that VM exists in cluster
2021-05-29 21:17:19 -04:00
dom_uuid = getDomainUUID ( zkhandler , domain )
2019-04-11 19:06:06 -04:00
if not dom_uuid :
return False , ' ERROR: Could not find VM " {} " in the cluster! ' . format ( domain )
# Get the data from ZK
2021-07-09 16:47:09 -04:00
console_log = zkhandler . read ( ( ' domain.console.log ' , dom_uuid ) )
2019-04-11 19:06:06 -04:00
2021-07-09 13:08:00 -04:00
if console_log is None :
return True , ' '
2019-04-11 19:06:06 -04:00
# Shrink the log buffer to length lines
shrunk_log = console_log . split ( ' \n ' ) [ - lines : ]
loglines = ' \n ' . join ( shrunk_log )
2019-12-25 19:10:12 -05:00
return True , loglines
2019-04-11 19:06:06 -04:00
2020-11-07 14:45:24 -05:00
2021-05-29 21:17:19 -04:00
def get_info ( zkhandler , domain ) :
2019-06-27 11:19:48 -04:00
# Validate that VM exists in cluster
2021-05-29 21:17:19 -04:00
dom_uuid = getDomainUUID ( zkhandler , domain )
2019-05-20 22:15:28 -04:00
if not dom_uuid :
return False , ' ERROR: No VM named " {} " is present in the cluster. ' . format ( domain )
# Gather information from XML config and print it
2021-05-29 21:17:19 -04:00
domain_information = common . getInformationFromXML ( zkhandler , dom_uuid )
2019-06-24 13:37:56 -04:00
if not domain_information :
2019-05-20 22:15:28 -04:00
return False , ' ERROR: Could not get information about VM " {} " . ' . format ( domain )
return True , domain_information
2020-11-07 14:45:24 -05:00
2021-10-07 11:13:30 -04:00
def get_list ( zkhandler , node , state , tag , limit , is_fuzzy = True , negate = False ) :
2019-06-24 13:37:56 -04:00
if node :
2018-09-20 03:25:58 -04:00
# Verify node is valid
2021-05-29 21:17:19 -04:00
if not common . verifyNode ( zkhandler , node ) :
2019-07-08 22:37:26 -04:00
return False , ' Specified node " {} " is invalid. ' . format ( node )
2018-09-20 03:25:58 -04:00
2019-06-24 13:37:56 -04:00
if state :
2020-11-07 13:02:54 -05:00
valid_states = [ ' start ' , ' restart ' , ' shutdown ' , ' stop ' , ' disable ' , ' fail ' , ' migrate ' , ' unmigrate ' , ' provision ' ]
2020-11-06 20:37:52 -05:00
if state not in valid_states :
2019-03-20 11:31:54 -04:00
return False , ' VM state " {} " is not valid. ' . format ( state )
2021-06-13 13:41:21 -04:00
full_vm_list = zkhandler . children ( ' base.domain ' )
2018-09-20 03:25:58 -04:00
2019-03-12 23:52:59 -04:00
# Set our limit to a sensible regex
2021-06-28 12:27:43 -04:00
if limit :
# Check if the limit is a UUID
is_limit_uuid = False
2019-03-12 23:52:59 -04:00
try :
2021-06-28 12:27:43 -04:00
uuid_obj = UUID ( limit , version = 4 )
limit = str ( uuid_obj )
is_limit_uuid = True
except ValueError :
pass
if is_fuzzy and not is_limit_uuid :
try :
# Implcitly assume fuzzy limits
if not re . match ( r ' \ ^.* ' , limit ) :
limit = ' .* ' + limit
if not re . match ( r ' .* \ $ ' , limit ) :
limit = limit + ' .* '
except Exception as e :
return False , ' Regex Error: {} ' . format ( e )
2019-03-12 23:52:59 -04:00
2021-06-23 19:14:26 -04:00
get_vm_info = dict ( )
2018-09-25 02:20:32 -04:00
for vm in full_vm_list :
2021-06-13 13:41:21 -04:00
name = zkhandler . read ( ( ' domain ' , vm ) )
2021-06-23 19:14:26 -04:00
is_limit_match = False
2021-07-14 00:51:48 -04:00
is_tag_match = False
2021-06-23 19:14:26 -04:00
is_node_match = False
is_state_match = False
# Check on limit
2019-06-24 13:37:56 -04:00
if limit :
2021-06-28 12:27:43 -04:00
# Try to match the limit against the UUID (if applicable) and name
2018-09-20 03:25:58 -04:00
try :
2021-06-28 12:27:43 -04:00
if is_limit_uuid and re . match ( limit , vm ) :
is_limit_match = True
2019-06-24 13:37:56 -04:00
if re . match ( limit , name ) :
2021-06-23 19:14:26 -04:00
is_limit_match = True
2018-09-20 03:25:58 -04:00
except Exception as e :
2018-09-25 02:20:32 -04:00
return False , ' Regex Error: {} ' . format ( e )
2018-09-20 03:25:58 -04:00
else :
2021-06-23 19:14:26 -04:00
is_limit_match = True
2021-07-14 00:51:48 -04:00
if tag :
vm_tags = zkhandler . children ( ( ' domain.meta.tags ' , vm ) )
2021-10-07 11:13:30 -04:00
if negate and tag not in vm_tags :
is_tag_match = True
if not negate and tag in vm_tags :
2021-07-14 00:51:48 -04:00
is_tag_match = True
else :
is_tag_match = True
2021-06-23 19:14:26 -04:00
# Check on node
if node :
vm_node = zkhandler . read ( ( ' domain.node ' , vm ) )
2021-10-07 11:13:30 -04:00
if negate and vm_node != node :
is_node_match = True
if not negate and vm_node == node :
2021-06-23 19:14:26 -04:00
is_node_match = True
else :
is_node_match = True
# Check on state
if state :
vm_state = zkhandler . read ( ( ' domain.state ' , vm ) )
2021-10-07 11:13:30 -04:00
if negate and vm_state != state :
is_state_match = True
if not negate and vm_state == state :
2021-06-23 19:14:26 -04:00
is_state_match = True
else :
is_state_match = True
2021-07-14 00:51:48 -04:00
get_vm_info [ vm ] = True if is_limit_match and is_tag_match and is_node_match and is_state_match else False
2021-07-01 17:24:47 -04:00
# Obtain our VM data in a thread pool
# This helps parallelize the numerous Zookeeper calls a bit, within the bounds of the GIL, and
# should help prevent this task from becoming absurdly slow with very large numbers of VMs.
# The max_workers is capped at 32 to avoid creating an absurd number of threads especially if
# the list gets called multiple times simultaneously by the API, but still provides a noticeable
# speedup.
vm_execute_list = [ vm for vm in full_vm_list if get_vm_info [ vm ] ]
vm_data_list = list ( )
with ThreadPoolExecutor ( max_workers = 32 , thread_name_prefix = ' vm_list ' ) as executor :
futures = [ ]
for vm_uuid in vm_execute_list :
futures . append ( executor . submit ( common . getInformationFromXML , zkhandler , vm_uuid ) )
for future in futures :
2021-07-09 13:05:37 -04:00
try :
vm_data_list . append ( future . result ( ) )
except Exception :
pass
2021-07-01 17:24:47 -04:00
return True , vm_data_list