Implement VM metadata and use it

Implements the storing of three VM metadata attributes:
1. Node limits - allows specifying a list of hosts on which the VM must
run. This limit influences the migration behaviour of VMs.
2. Per-VM node selectors - allows each VM to have its migration
autoselection method specified, to automatically allow different methods
per VM based on the administrator's preferences.
3. VM autorestart - allows a VM to be automatically restarted from a
stopped state, presumably due to a failure to find a target node (either
due to limits or otherwise) during a flush/fence recovery, on the next
node unflush/ready state of its home hypervisor. Useful mostly in
conjunction with limits to ensure that VMs which were shut down due to
there being no valid migration targets are started back up when their
node becomes ready again.

Includes the full client interaction with these metadata options,
including printing, as well as defining a new function to modify this
metadata. For the CLI it is set/modified either on `vm define` or via the
`vm meta` command. For the API it is set/modified either on a POST to
the `/vm` endpoint (during VM definition) or on POST to the `/vm/<vm>`
endpoint. For the API this replaces the previous reserved word for VM
creation from scratch as this will no longer be implemented in-daemon
(see #22).

Closes #52
This commit is contained in:
Joshua Boniface 2019-10-12 01:17:39 -04:00
parent 94501f400f
commit 5995353597
10 changed files with 233 additions and 51 deletions

View File

@ -272,13 +272,6 @@ def vm_list(node=None, state=None, limit=None, is_fuzzy=True):
pvc_common.stopZKConnection(zk_conn) pvc_common.stopZKConnection(zk_conn)
return flask.jsonify(retdata), retcode return flask.jsonify(retdata), retcode
# TODO: #22
#def vm_add():
# """
# Add a VM named NAME to the PVC cluster.
# """
# return '', 200
def vm_define(name, xml, node, selector): def vm_define(name, xml, node, selector):
""" """
Define a VM from Libvirt XML in the PVC cluster. Define a VM from Libvirt XML in the PVC cluster.
@ -296,6 +289,23 @@ def vm_define(name, xml, node, selector):
} }
return flask.jsonify(output), retcode return flask.jsonify(output), retcode
def vm_meta(vm, limit, selector, autostart):
"""
Update metadata of a VM.
"""
zk_conn = pvc_common.startZKConnection(config['coordinators'])
retflag, retdata = pvc_vm.modify_vm_metadata(zk_conn, vm. limit, selector, autostart)
if retflag:
retcode = 200
else:
retcode = 400
pvc_common.stopZKConnection(zk_conn)
output = {
'message': retdata.replace('\"', '\'')
}
return flask.jsonify(output), retcode
def vm_modify(name, restart, xml): def vm_modify(name, restart, xml):
""" """
Modify a VM Libvirt XML in the PVC cluster. Modify a VM Libvirt XML in the PVC cluster.

View File

@ -235,11 +235,23 @@ def api_vm_root():
else: else:
node = None node = None
# Get target selector # Set target limit metadata
if 'limit' in flask.request.values:
limit = flask.request.values['limit']
else:
limit = None
# Set target selector metadata
if 'selector' in flask.request.values: if 'selector' in flask.request.values:
selector = flask.request.values['selector'] selector = flask.request.values['selector']
else: else:
selector = None selector = 'mem'
# Set target autostart metadata
if 'autostart' in flask.request.values:
autostart = True
else:
autostart = False
return pvcapi.vm_define(vm, libvirt_xml, node, selector) return pvcapi.vm_define(vm, libvirt_xml, node, selector)
@ -251,8 +263,27 @@ def api_vm_element(vm):
return pvcapi.vm_list(None, None, vm, is_fuzzy=False) return pvcapi.vm_list(None, None, vm, is_fuzzy=False)
if flask.request.method == 'POST': if flask.request.method == 'POST':
# TODO: #22 # Set target limit metadata
flask.abort(501) if 'limit' in flask.request.values:
limit = flask.request.values['limit']
else:
limit = None
# Set target selector metadata
if 'selector' in flask.request.values:
selector = flask.request.values['selector']
else:
selector = None
# Set target autostart metadata
if 'no-autostart' in flask.request.values:
autostart = False
elif 'autostart' in flask.request.values:
autostart = True
else:
autostart = None
return pvcapi.vm_meta(vm, limit, selector, autostart)
if flask.request.method == 'PUT': if flask.request.method == 'PUT':
libvirt_xml = flask.request.data libvirt_xml = flask.request.data

View File

@ -217,17 +217,25 @@ def cli_vm():
@click.command(name='define', short_help='Define a new virtual machine from a Libvirt XML file.') @click.command(name='define', short_help='Define a new virtual machine from a Libvirt XML file.')
@click.option( @click.option(
'-t', '--target', 'target_node', '-t', '--target', 'target_node',
help='Home node for this domain; autodetect if unspecified.' help='Home node for this domain; autoselect if unspecified.'
) )
@click.option( @click.option(
'-s', '--selector', 'selector', default='mem', show_default=True, '-l', '--limit', 'node_limit', default=None, show_default=False,
help='Comma-separated list of nodes to limit VM operation to; saved with VM.'
)
@click.option(
'-s', '--selector', 'node_selector', default='mem', show_default=True,
type=click.Choice(['mem','load','vcpus','vms']), type=click.Choice(['mem','load','vcpus','vms']),
help='Method to determine optimal target node during autodetect.' help='Method to determine optimal target node during autoselect; saved with VM.'
)
@click.option(
'-a'/'-A', '--autostart'/'--no-autostart', 'node_autostart', is_flag=True, default=False,
help='Start VM automatically on next unflush/ready state of home node; unset by daemon once used.'
) )
@click.argument( @click.argument(
'config', type=click.File() 'config', type=click.File()
) )
def vm_define(config, target_node, selector): def vm_define(config, target_node, node_limit, node_selector, node_autostart):
""" """
Define a new virtual machine from Libvirt XML configuration file CONFIG. Define a new virtual machine from Libvirt XML configuration file CONFIG.
""" """
@ -237,7 +245,39 @@ def vm_define(config, target_node, selector):
config.close() config.close()
zk_conn = pvc_common.startZKConnection(zk_host) zk_conn = pvc_common.startZKConnection(zk_host)
retcode, retmsg = pvc_vm.define_vm(zk_conn, config_data, target_node, selector) retcode, retmsg = pvc_vm.define_vm(zk_conn, config_data, target_node, node_limit, node_selector, node_autostart)
cleanup(retcode, retmsg, zk_conn)
###############################################################################
# pvc vm meta
###############################################################################
@click.command(name='meta', short_help='Modify PVC metadata of an existing VM.')
@click.option(
'-l', '--limit', 'node_limit', default=None, show_default=False,
help='Comma-separated list of nodes to limit VM operation to.'
)
@click.option(
'-s', '--selector', 'node_selector', default=None, show_default=False,
type=click.Choice(['mem','load','vcpus','vms']),
help='Method to determine optimal target node during autoselect; saved with VM.'
)
@click.option(
'-a'/'-A', '--autostart'/'--no-autostart', 'node_autostart', is_flag=True, default=None,
help='Start VM automatically on next unflush/ready state of home node; unset by daemon once used.'
)
@click.argument(
'domain'
)
def vm_meta(domain, node_limit, node_selector, node_autostart):
"""
Modify the PVC metadata of existing virtual machine DOMAIN. At least one option to update must be specified. DOMAIN may be a UUID or name.
"""
if node_limit is None and node_selector is None and node_autostart is None:
cleanup(False, 'At least one metadata option must be specified to update.')
zk_conn = pvc_common.startZKConnection(zk_host)
retcode, retmsg = pvc_vm.modify_vm_metadata(zk_conn, domain, node_limit, node_selector, node_autostart)
cleanup(retcode, retmsg, zk_conn) cleanup(retcode, retmsg, zk_conn)
############################################################################### ###############################################################################
@ -1774,6 +1814,7 @@ cli_node.add_command(node_info)
cli_node.add_command(node_list) cli_node.add_command(node_list)
cli_vm.add_command(vm_define) cli_vm.add_command(vm_define)
cli_vm.add_command(vm_meta)
cli_vm.add_command(vm_modify) cli_vm.add_command(vm_modify)
cli_vm.add_command(vm_undefine) cli_vm.add_command(vm_undefine)
cli_vm.add_command(vm_remove) cli_vm.add_command(vm_remove)

View File

@ -164,6 +164,10 @@ def getInformationFromXML(zk_conn, uuid):
domain_lastnode = zkhandler.readdata(zk_conn, '/domains/{}/lastnode'.format(uuid)) domain_lastnode = zkhandler.readdata(zk_conn, '/domains/{}/lastnode'.format(uuid))
domain_failedreason = zkhandler.readdata(zk_conn, '/domains/{}/failedreason'.format(uuid)) domain_failedreason = zkhandler.readdata(zk_conn, '/domains/{}/failedreason'.format(uuid))
domain_node_limit = zkhandler.readdata(zk_conn, '/domains/{}/node_limit'.format(uuid))
domain_node_selector = zkhandler.readdata(zk_conn, '/domains/{}/node_selector'.format(uuid))
domain_node_autostart = zkhandler.readdata(zk_conn, '/domains/{}/node_autostart'.format(uuid))
parsed_xml = getDomainXML(zk_conn, uuid) parsed_xml = getDomainXML(zk_conn, uuid)
domain_uuid, domain_name, domain_description, domain_memory, domain_vcpu, domain_vcputopo = getDomainMainDetails(parsed_xml) domain_uuid, domain_name, domain_description, domain_memory, domain_vcpu, domain_vcputopo = getDomainMainDetails(parsed_xml)
@ -188,6 +192,9 @@ def getInformationFromXML(zk_conn, uuid):
'last_node': domain_lastnode, 'last_node': domain_lastnode,
'migrated': domain_migrated, 'migrated': domain_migrated,
'failed_reason': domain_failedreason, 'failed_reason': domain_failedreason,
'node_limit': domain_node_limit,
'node_selector': domain_node_selector,
'node_autostart': domain_node_autostart,
'description': domain_description, 'description': domain_description,
'memory': domain_memory, 'memory': domain_memory,
'vcpu': domain_vcpu, 'vcpu': domain_vcpu,

View File

@ -160,7 +160,7 @@ def flush_locks(zk_conn, domain):
return success, message return success, message
def define_vm(zk_conn, config_data, target_node, selector): def define_vm(zk_conn, config_data, target_node, node_limit, node_selector, node_autostart):
# Parse the XML data # Parse the XML data
try: try:
parsed_xml = lxml.objectify.fromstring(config_data) parsed_xml = lxml.objectify.fromstring(config_data)
@ -190,6 +190,9 @@ def define_vm(zk_conn, config_data, target_node, selector):
'/domains/{}/state'.format(dom_uuid): 'stop', '/domains/{}/state'.format(dom_uuid): 'stop',
'/domains/{}/node'.format(dom_uuid): target_node, '/domains/{}/node'.format(dom_uuid): target_node,
'/domains/{}/lastnode'.format(dom_uuid): '', '/domains/{}/lastnode'.format(dom_uuid): '',
'/domains/{}/node_limit'.format(dom_uuid): node_limit,
'/domains/{}/node_selector'.format(dom_uuid): node_selector,
'/domains/{}/node_autostart'.format(dom_uuid): node_autostart,
'/domains/{}/failedreason'.format(dom_uuid): '', '/domains/{}/failedreason'.format(dom_uuid): '',
'/domains/{}/consolelog'.format(dom_uuid): '', '/domains/{}/consolelog'.format(dom_uuid): '',
'/domains/{}/rbdlist'.format(dom_uuid): ','.join(rbd_list), '/domains/{}/rbdlist'.format(dom_uuid): ','.join(rbd_list),
@ -198,6 +201,28 @@ def define_vm(zk_conn, config_data, target_node, selector):
return True, 'Added new VM with Name "{}" and UUID "{}" to database.'.format(dom_name, dom_uuid) return True, 'Added new VM with Name "{}" and UUID "{}" to database.'.format(dom_name, dom_uuid)
def modify_vm_metadata(zk_conn, domain, node_limit, node_selector, node_autostart):
dom_uuid = getDomainUUID(zk_conn, domain)
if not dom_uuid:
return False, 'ERROR: Could not find VM "{}" in the cluster!'.format(domain)
if node_limit is not None:
zkhandler.writedata(zk_conn, {
'/domains/{}/node_limit'.format(dom_uuid): node_limit
})
if node_selector is not None:
zkhandler.writedata(zk_conn, {
'/domains/{}/node_selector'.format(dom_uuid): node_selector
})
if node_autostart is not None:
zkhandler.writedata(zk_conn, {
'/domains/{}/node_autostart'.format(dom_uuid): node_autostart
})
return True, 'Successfully modified PVC metadata of VM "{}".'.format(domain)
def modify_vm(zk_conn, domain, restart, new_vm_config): def modify_vm(zk_conn, domain, restart, new_vm_config):
dom_uuid = getDomainUUID(zk_conn, domain) dom_uuid = getDomainUUID(zk_conn, domain)
if not dom_uuid: if not dom_uuid:
@ -681,6 +706,11 @@ def format_info(zk_conn, domain_information, long_output):
ainformation.append('') ainformation.append('')
ainformation.append('{}Failure reason:{} {}'.format(ansiprint.purple(), ansiprint.end(), domain_information['failed_reason'])) ainformation.append('{}Failure reason:{} {}'.format(ansiprint.purple(), ansiprint.end(), domain_information['failed_reason']))
ainformation.append('')
ainformation.append('{}Node limit:{} {}'.format(ansiprint.purple(), ansiprint.end(), domain_information['node_limit']))
ainformation.append('{}Migration selector:{} {}'.format(ansiprint.purple(), ansiprint.end(), domain_information['node_selector']))
ainformation.append('{}Autostart:{} {}'.format(ansiprint.purple(), ansiprint.end(), domain_information['node_autostart']))
# Network list # Network list
net_list = [] net_list = []
for net in domain_information['networks']: for net in domain_information['networks']:

View File

@ -256,20 +256,26 @@ Return a JSON document containing information about all cluster VMs. If `limit`
###### `POST` ###### `POST`
* Mandatory values: `xml` * Mandatory values: `xml`
* Optional values: `node`, `selector` * Optional values: `node`, `limit`, `selector`, `autostart`
Define a new VM with Libvirt XML configuration `xml` (either single-line or human-readable multi-line). Define a new VM with Libvirt XML configuration `xml` (either single-line or human-readable multi-line).
If `node` is specified and is valid, the VM will be assigned to `node` instead of automatically determining the target node. If `node` is specified and not valid, auto-selection occurs instead. If `node` is specified and is valid, the VM will be assigned to `node` instead of automatically determining the target node. If `node` is specified and not valid, auto-selection occurs instead.
If `selector` is specified and no specific and valid `node` is specified, the automatic node determination will use `selector` to determine the optimal node instead of the default for the cluster. If `limit` is speficied, the node will not be allowed to run on nodes not specified in the limit.
The `limit` value must be a comma-separated list of nodes; invalid nodes are ignored.
If `selector` is specified and no specific and valid `node` is specified, the automatic node determination will use `selector` to determine the optimal node instead of the default for the cluster. This value is stored as PVC metadata for this VM and is used in subsequent migrate (including node flush) and fence recovery operations.
Valid `selector` values are: `mem`: the node with the least allocated VM memory; `vcpus`: the node with the least allocated VM vCPUs; `load`: the node with the least current load average; `vms`: the node with the least number of provisioned VMs. Valid `selector` values are: `mem`: the node with the least allocated VM memory; `vcpus`: the node with the least allocated VM vCPUs; `load`: the node with the least current load average; `vms`: the node with the least number of provisioned VMs.
If `autostart` is specified, the VM will be set to autostart on the next node unflush/ready operation of the home node. This metadata value is reset to False by the node daemon on each successful use.
**NOTE:** The `POST` operation assumes that the VM resources (i.e. disks, operating system, etc.) are already created. This is equivalent to the `pvc vm define` command in the PVC CLI client. *[todo v0.6]* Creating a new VM using the provisioner uses the `POST /api/vm/<vm>` endpoint instead. **NOTE:** The `POST` operation assumes that the VM resources (i.e. disks, operating system, etc.) are already created. This is equivalent to the `pvc vm define` command in the PVC CLI client. *[todo v0.6]* Creating a new VM using the provisioner uses the `POST /api/vm/<vm>` endpoint instead.
#### `/api/v1/vm/<vm>` #### `/api/v1/vm/<vm>`
* Methods: `GET`, *[todo v0.6]* `POST`, `PUT`, `DELETE` * Methods: `GET`, `POST`, `PUT`, `DELETE`
###### `GET` ###### `GET`
* Mandatory values: N/A * Mandatory values: N/A
@ -277,11 +283,23 @@ Valid `selector` values are: `mem`: the node with the least allocated VM memory;
Return a JSON document containing information about `<vm>`. The output is identical to `GET /api/v1/vm?limit=<vm>` without fuzzy regex matching. Return a JSON document containing information about `<vm>`. The output is identical to `GET /api/v1/vm?limit=<vm>` without fuzzy regex matching.
###### *[todo v0.6]* `POST` ###### `POST`
* Mandatory values: `vm_template` * Mandatory values: At least one of optional values must be specified
* Optional values: `description` * Optional values: `limit`, `selector`, `autostart`/`no-autostart`
Create a new virtual machine `<vm>` with the specified VM template `vm_template` and optional text `description`. Update the PVC metadata of `<vm>` with the specified values.
If `limit` is speficied, the node will not be allowed to run on nodes not specified in the limit.
The `limit` value must be a comma-separated list of nodes; invalid nodes are ignored.
If `selector` is specified and no specific and valid `node` is specified, the automatic node determination will use `selector` to determine the optimal node instead of the default for the cluster. This value is stored as PVC metadata for this VM and is used in subsequent migrate (including node flush) and fence recovery operations.
Valid `selector` values are: `mem`: the node with the least allocated VM memory; `vcpus`: the node with the least allocated VM vCPUs; `load`: the node with the least current load average; `vms`: the node with the least number of provisioned VMs.
If `autostart` is specified, the VM will be set to autostart on the next node unflush/ready operation of the home node. This metadata value is reset to False by the node daemon on each successful use.
If `no-autostart` is specified, an existing autostart will be disabled if applicable.
###### `PUT` ###### `PUT`
* Mandatory values: `xml` * Mandatory values: `xml`

View File

@ -382,7 +382,7 @@ class NodeInstance(object):
self.logger.out('Selecting target to migrate VM "{}"'.format(dom_uuid), state='i') self.logger.out('Selecting target to migrate VM "{}"'.format(dom_uuid), state='i')
target_node = common.findTargetHypervisor(self.zk_conn, 'mem', dom_uuid) target_node = common.findTargetHypervisor(self.zk_conn, self.config, dom_uuid)
# Don't replace the previous node if the VM is already migrated # Don't replace the previous node if the VM is already migrated
if zkhandler.readdata(self.zk_conn, '/domains/{}/lastnode'.format(dom_uuid)): if zkhandler.readdata(self.zk_conn, '/domains/{}/lastnode'.format(dom_uuid)):
@ -390,9 +390,10 @@ class NodeInstance(object):
else: else:
current_node = zkhandler.readdata(self.zk_conn, '/domains/{}/node'.format(dom_uuid)) current_node = zkhandler.readdata(self.zk_conn, '/domains/{}/node'.format(dom_uuid))
if target_node == None: if target_node is None:
self.logger.out('Failed to find migration target for VM "{}"; shutting down'.format(dom_uuid), state='e') self.logger.out('Failed to find migration target for VM "{}"; shutting down and setting autostart flag'.format(dom_uuid), state='e')
zkhandler.writedata(self.zk_conn, { '/domains/{}/state'.format(dom_uuid): 'shutdown' }) zkhandler.writedata(self.zk_conn, { '/domains/{}/state'.format(dom_uuid): 'shutdown' })
zkhandler.writedata(self.zk_conn, { '/domains/{}/node_autostart'.format(dom_uuid): 'True' })
# Wait for the VM to shut down # Wait for the VM to shut down
while zkhandler.readdata(self.zk_conn, '/domains/{}/state'.format(dom_uuid)) != 'stop': while zkhandler.readdata(self.zk_conn, '/domains/{}/state'.format(dom_uuid)) != 'stop':
@ -427,6 +428,19 @@ class NodeInstance(object):
self.flush_stopper = False self.flush_stopper = False
return return
# Handle autostarts
autostart = zkhandler.readdata(self.zk_conn, '/domains/{}/node_autostart'.format(dom_uuid))
node = zkhandler.readdata(self.zk_conn, '/domains/{}/node'.format(dom_uuid))
if autostart == 'True' and node == self.name:
self.logger.out('Starting autostart VM "{}"'.format(dom_uuid), state='i')
zkhandler.writedata(self.zk_conn, {
'/domains/{}/state'.format(dom_uuid): 'start',
'/domains/{}/node'.format(dom_uuid): self.name,
'/domains/{}/lastnode'.format(dom_uuid): '',
'/domains/{}/node_autostart'.format(dom_uuid): 'False'
})
continue
try: try:
last_node = zkhandler.readdata(self.zk_conn, '/domains/{}/lastnode'.format(dom_uuid)) last_node = zkhandler.readdata(self.zk_conn, '/domains/{}/lastnode'.format(dom_uuid))
except: except:

View File

@ -105,6 +105,10 @@ class VMInstance(object):
self.domname = zkhandler.readdata(zk_conn, '/domains/{}'.format(domuuid)) self.domname = zkhandler.readdata(zk_conn, '/domains/{}'.format(domuuid))
self.state = zkhandler.readdata(self.zk_conn, '/domains/{}/state'.format(self.domuuid)) self.state = zkhandler.readdata(self.zk_conn, '/domains/{}/state'.format(self.domuuid))
self.node = zkhandler.readdata(self.zk_conn, '/domains/{}/node'.format(self.domuuid)) self.node = zkhandler.readdata(self.zk_conn, '/domains/{}/node'.format(self.domuuid))
try:
self.pinpolicy = zkhandler.readdata(self.zk_conn, '/domains/{}/pinpolicy'.format(self.domuuid))
except:
self.pinpolicy = "None"
# These will all be set later # These will all be set later
self.instart = False self.instart = False

View File

@ -139,24 +139,44 @@ def removeIPAddress(ipaddr, cidrnetmask, dev):
# #
# Find a migration target # Find a migration target
# #
def findTargetHypervisor(zk_conn, search_field, dom_uuid): def findTargetHypervisor(zk_conn, config, dom_uuid):
# Determine VM node limits; set config value if read fails
try:
node_limit = zkhandler.readdata(zk_conn, '/domains/{}/node_limit'.format(node)).split(',')
except:
node_limit = None
zkhandler.writedata(zk_conn, { '/domains/{}/node_limit'.format(node): 'None' })
# Determine VM search field or use default; set config value if read fails
try:
search_field = zkhandler.readdata(zk_conn, '/domains/{}/node_selector'.format(node)).split(',')
except:
search_field = config.migration_target_selector
zkhandler.writedata(zk_conn, { '/domains/{}/node_selector'.format(node): config.migration_target_selector })
# Execute the search
if search_field == 'mem': if search_field == 'mem':
return findTargetHypervisorMem(zk_conn, dom_uuid) return findTargetHypervisorMem(zk_conn, node_limit, dom_uuid)
if search_field == 'load': if search_field == 'load':
return findTargetHypervisorLoad(zk_conn, dom_uuid) return findTargetHypervisorLoad(zk_conn, node_limit, dom_uuid)
if search_field == 'vcpus': if search_field == 'vcpus':
return findTargetHypervisorVCPUs(zk_conn, dom_uuid) return findTargetHypervisorVCPUs(zk_conn, node_limit, dom_uuid)
if search_field == 'vms': if search_field == 'vms':
return findTargetHypervisorVMs(zk_conn, dom_uuid) return findTargetHypervisorVMs(zk_conn, node_limit, dom_uuid)
# Nothing was found
return None return None
# Get the list of valid target nodes # Get the list of valid target nodes
def getHypervisors(zk_conn, dom_uuid): def getHypervisors(zk_conn, node_limit, dom_uuid):
valid_node_list = [] valid_node_list = []
full_node_list = zkhandler.listchildren(zk_conn, '/nodes') full_node_list = zkhandler.listchildren(zk_conn, '/nodes')
current_node = zkhandler.readdata(zk_conn, '/domains/{}/node'.format(dom_uuid)) current_node = zkhandler.readdata(zk_conn, '/domains/{}/node'.format(dom_uuid))
for node in full_node_list: for node in full_node_list:
if node_limit and node not in node_limit:
continue
daemon_state = zkhandler.readdata(zk_conn, '/nodes/{}/daemonstate'.format(node)) daemon_state = zkhandler.readdata(zk_conn, '/nodes/{}/daemonstate'.format(node))
domain_state = zkhandler.readdata(zk_conn, '/nodes/{}/domainstate'.format(node)) domain_state = zkhandler.readdata(zk_conn, '/nodes/{}/domainstate'.format(node))
@ -171,11 +191,11 @@ def getHypervisors(zk_conn, dom_uuid):
return valid_node_list return valid_node_list
# via free memory (relative to allocated memory) # via free memory (relative to allocated memory)
def findTargetHypervisorMem(zk_conn, dom_uuid): def findTargetHypervisorMem(zk_conn, node_limit, dom_uuid):
most_allocfree = 0 most_allocfree = 0
target_node = None target_node = None
node_list = getHypervisors(zk_conn, dom_uuid) node_list = getHypervisors(zk_conn, node_limit, dom_uuid)
for node in node_list: for node in node_list:
memalloc = int(zkhandler.readdata(zk_conn, '/nodes/{}/memalloc'.format(node))) memalloc = int(zkhandler.readdata(zk_conn, '/nodes/{}/memalloc'.format(node)))
memused = int(zkhandler.readdata(zk_conn, '/nodes/{}/memused'.format(node))) memused = int(zkhandler.readdata(zk_conn, '/nodes/{}/memused'.format(node)))
@ -190,11 +210,11 @@ def findTargetHypervisorMem(zk_conn, dom_uuid):
return target_node return target_node
# via load average # via load average
def findTargetHypervisorLoad(zk_conn, dom_uuid): def findTargetHypervisorLoad(zk_conn, node_limit, dom_uuid):
least_load = 9999 least_load = 9999
target_node = None target_node = None
node_list = getHypervisors(zk_conn, dom_uuid) node_list = getHypervisors(zk_conn, node_limit, dom_uuid)
for node in node_list: for node in node_list:
load = int(zkhandler.readdata(zk_conn, '/nodes/{}/load'.format(node))) load = int(zkhandler.readdata(zk_conn, '/nodes/{}/load'.format(node)))
@ -205,11 +225,11 @@ def findTargetHypervisorLoad(zk_conn, dom_uuid):
return target_node return target_node
# via total vCPUs # via total vCPUs
def findTargetHypervisorVCPUs(zk_conn, dom_uuid): def findTargetHypervisorVCPUs(zk_conn, node_limit, dom_uuid):
least_vcpus = 9999 least_vcpus = 9999
target_node = None target_node = None
node_list = getHypervisors(zk_conn, dom_uuid) node_list = getHypervisors(zk_conn, node_limit, dom_uuid)
for node in node_list: for node in node_list:
vcpus = int(zkhandler.readdata(zk_conn, '/nodes/{}/vcpualloc'.format(node))) vcpus = int(zkhandler.readdata(zk_conn, '/nodes/{}/vcpualloc'.format(node)))
@ -220,11 +240,11 @@ def findTargetHypervisorVCPUs(zk_conn, dom_uuid):
return target_node return target_node
# via total VMs # via total VMs
def findTargetHypervisorVMs(zk_conn, dom_uuid): def findTargetHypervisorVMs(zk_conn, node_limit, dom_uuid):
least_vms = 9999 least_vms = 9999
target_node = None target_node = None
node_list = getHypervisors(zk_conn, dom_uuid) node_list = getHypervisors(zk_conn, node_limit, dom_uuid)
for node in node_list: for node in node_list:
vms = int(zkhandler.readdata(zk_conn, '/nodes/{}/domainscount'.format(node))) vms = int(zkhandler.readdata(zk_conn, '/nodes/{}/domainscount'.format(node)))

View File

@ -67,27 +67,34 @@ def fenceNode(node_name, zk_conn, config, logger):
# If the fence succeeded and successful_fence is migrate # If the fence succeeded and successful_fence is migrate
if fence_status == True and config['successful_fence'] == 'migrate': if fence_status == True and config['successful_fence'] == 'migrate':
migrateFromFencedNode(zk_conn, node_name, logger) migrateFromFencedNode(zk_conn, node_name, config, logger)
# If the fence failed and failed_fence is migrate # If the fence failed and failed_fence is migrate
if fence_status == False and config['failed_fence'] == 'migrate' and config['suicide_intervals'] != '0': if fence_status == False and config['failed_fence'] == 'migrate' and config['suicide_intervals'] != '0':
migrateFromFencedNode(zk_conn, node_name, logger) migrateFromFencedNode(zk_conn, node_name, config, logger)
# Migrate hosts away from a fenced node # Migrate hosts away from a fenced node
def migrateFromFencedNode(zk_conn, node_name, logger): def migrateFromFencedNode(zk_conn, node_name, config, logger):
logger.out('Migrating VMs from dead node "{}" to new hosts'.format(node_name), state='i') logger.out('Migrating VMs from dead node "{}" to new hosts'.format(node_name), state='i')
dead_node_running_domains = zkhandler.readdata(zk_conn, '/nodes/{}/runningdomains'.format(node_name)).split() dead_node_running_domains = zkhandler.readdata(zk_conn, '/nodes/{}/runningdomains'.format(node_name)).split()
for dom_uuid in dead_node_running_domains: for dom_uuid in dead_node_running_domains:
VMInstance.flush_locks(zk_conn, logger, dom_uuid) VMInstance.flush_locks(zk_conn, logger, dom_uuid)
target_node = common.findTargetHypervisor(zk_conn, 'mem', dom_uuid) target_node = common.findTargetHypervisor(zk_conn, config, dom_uuid)
logger.out('Migrating VM "{}" to node "{}"'.format(dom_uuid, target_node), state='i') if target_node is not None:
zkhandler.writedata(zk_conn, { logger.out('Migrating VM "{}" to node "{}"'.format(dom_uuid, target_node), state='i')
'/domains/{}/state'.format(dom_uuid): 'start', zkhandler.writedata(zk_conn, {
'/domains/{}/node'.format(dom_uuid): target_node, '/domains/{}/state'.format(dom_uuid): 'start',
'/domains/{}/lastnode'.format(dom_uuid): node_name '/domains/{}/node'.format(dom_uuid): target_node,
}) '/domains/{}/lastnode'.format(dom_uuid): node_name
})
else:
logger.out('No target node found for VM "{}"; VM will autostart on next unflush/ready of current node'.format(dom_uuid), state='i')
zkhandler.writedata(zk_conn, {
'/domains/{}/state'.format(dom_uuid): 'stopped',
'/domains/{}/node_autostart'.format(dom_uuid): 'True'
}
# Set node in flushed state for easy remigrating when it comes back # Set node in flushed state for easy remigrating when it comes back
zkhandler.writedata(zk_conn, { '/nodes/{}/domainstate'.format(node_name): 'flushed' }) zkhandler.writedata(zk_conn, { '/nodes/{}/domainstate'.format(node_name): 'flushed' })