#!/usr/bin/env python3 # NodeInstance.py - Class implementing a PVC node and run by pvcvd # Part of the Parallel Virtual Cluster (PVC) system # # Copyright (C) 2018 Joshua M. Boniface # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see . # ############################################################################### import os import sys import psutil import socket import time import libvirt import kazoo.client import threading import subprocess import daemon_lib.ansiiprint as ansiiprint import daemon_lib.zkhandler as zkhandler class NodeInstance(): # Initialization function def __init__(self, this_node, name, t_node, s_domain, zk_conn, config): # Passed-in variables on creation self.zk_conn = zk_conn self.config = config self.this_node = this_node self.name = name self.daemon_state = 'stop' self.domain_state = 'ready' self.t_node = t_node self.active_node_list = [] self.flushed_node_list = [] self.inactive_node_list = [] self.s_domain = s_domain self.domain_list = [] self.ipmi_hostname = self.config['ipmi_hostname'] self.domains_count = 0 self.memused = 0 self.memfree = 0 self.memalloc = 0 self.vcpualloc = 0 self.inflush = False # Zookeeper handlers for changed states @self.zk_conn.DataWatch('/nodes/{}/daemonstate'.format(self.name)) def watch_hypervisor_daemonstate(data, stat, event=""): if event and event.type == 'DELETED': # The key has been deleted after existing before; terminate this watcher # because this class instance is about to be reaped in Daemon.py return False try: self.daemon_state = data.decode('ascii') except AttributeError: self.daemon_state = 'stop' @self.zk_conn.DataWatch('/nodes/{}/domainstate'.format(self.name)) def watch_hypervisor_domainstate(data, stat, event=""): if event and event.type == 'DELETED': # The key has been deleted after existing before; terminate this watcher # because this class instance is about to be reaped in Daemon.py return False try: self.domain_state = data.decode('ascii') except AttributeError: self.domain_state = 'unknown' # toggle state management of this node if self.name == self.this_node: if self.domain_state == 'flush' and self.inflush == False: # Do flushing in a thread so it doesn't block the migrates out flush_thread = threading.Thread(target=self.flush, args=(), kwargs={}) flush_thread.start() if self.domain_state == 'unflush' and self.inflush == False: self.unflush() @self.zk_conn.DataWatch('/nodes/{}/memfree'.format(self.name)) def watch_hypervisor_memfree(data, stat, event=""): if event and event.type == 'DELETED': # The key has been deleted after existing before; terminate this watcher # because this class instance is about to be reaped in Daemon.py return False try: self.memfree = data.decode('ascii') except AttributeError: self.memfree = 0 @self.zk_conn.DataWatch('/nodes/{}/memused'.format(self.name)) def watch_hypervisor_memused(data, stat, event=""): if event and event.type == 'DELETED': # The key has been deleted after existing before; terminate this watcher # because this class instance is about to be reaped in Daemon.py return False try: self.memused = data.decode('ascii') except AttributeError: self.memused = 0 @self.zk_conn.DataWatch('/nodes/{}/memalloc'.format(self.name)) def watch_hypervisor_memalloc(data, stat, event=""): if event and event.type == 'DELETED': # The key has been deleted after existing before; terminate this watcher # because this class instance is about to be reaped in Daemon.py return False try: self.memalloc = data.decode('ascii') except AttributeError: self.memalloc = 0 @self.zk_conn.DataWatch('/nodes/{}/vcpualloc'.format(self.name)) def watch_hypervisor_vcpualloc(data, stat, event=""): if event and event.type == 'DELETED': # The key has been deleted after existing before; terminate this watcher # because this class instance is about to be reaped in Daemon.py return False try: self.vcpualloc = data.decode('ascii') except AttributeError: self.vcpualloc = 0 @self.zk_conn.DataWatch('/nodes/{}/runningdomains'.format(self.name)) def watch_hypervisor_runningdomains(data, stat, event=""): if event and event.type == 'DELETED': # The key has been deleted after existing before; terminate this watcher # because this class instance is about to be reaped in Daemon.py return False try: self.domain_list = data.decode('ascii').split() except AttributeError: self.domain_list = [] @self.zk_conn.DataWatch('/nodes/{}/domainscount'.format(self.name)) def watch_hypervisor_domainscount(data, stat, event=""): if event and event.type == 'DELETED': # The key has been deleted after existing before; terminate this watcher # because this class instance is about to be reaped in Daemon.py return False try: self.domains_count = data.decode('ascii') except AttributeError: self.domains_count = 0 # Get value functions def getfreemem(self): return self.memfree def getallocmem(self): return self.memalloc def getallocvcpu(self): return self.vcpualloc def getcpuload(self): return self.cpuload def getname(self): return self.name def getdaemonstate(self): return self.daemon_state def getdomainstate(self): return self.domain_state def getdomainlist(self): return self.domain_list # Update value functions def updatenodelist(self, t_node): self.t_node = t_node def updatedomainlist(self, s_domain): self.s_domain = s_domain # Flush all VMs on the host def flush(self): self.inflush = True ansiiprint.echo('Flushing node "{}" of running VMs'.format(self.name), '', 'i') ansiiprint.echo('Domain list: {}'.format(', '.join(self.domain_list)), '', 'c') fixed_domain_list = self.domain_list.copy() for dom_uuid in fixed_domain_list: ansiiprint.echo('Selecting target to migrate VM "{}"'.format(dom_uuid), '', 'i') current_hypervisor = zkhandler.readdata(self.zk_conn, '/domains/{}/hypervisor'.format(dom_uuid)) target_hypervisor = findTargetHypervisor(self.zk_conn, 'mem', dom_uuid) if target_hypervisor == None: ansiiprint.echo('Failed to find migration target for VM "{}"; shutting down'.format(dom_uuid), '', 'e') zkhandler.writedata(self.zk_conn, { '/domains/{}/state'.format(dom_uuid): 'shutdown' }) else: ansiiprint.echo('Migrating VM "{}" to hypervisor "{}"'.format(dom_uuid, target_hypervisor), '', 'i') zkhandler.writedata(self.zk_conn, { '/domains/{}/state'.format(dom_uuid): 'migrate', '/domains/{}/hypervisor'.format(dom_uuid): target_hypervisor, '/domains/{}/lasthypervisor'.format(dom_uuid): current_hypervisor }) # Wait for the VM to migrate so the next VM's free RAM count is accurate (they migrate in serial anyways) while True: time.sleep(1) vm_current_state = zkhandler.readdata(self.zk_conn, '/domains/{}/state'.format(dom_uuid)) if vm_current_state == "start": break zkhandler.writedata(self.zk_conn, { '/nodes/{}/runningdomains'.format(self.name): '' }) zkhandler.writedata(self.zk_conn, { '/nodes/{}/domainstate'.format(self.name): 'flushed' }) self.inflush = False def unflush(self): self.inflush = True ansiiprint.echo('Restoring node {} to active service.'.format(self.name), '', 'i') zkhandler.writedata(self.zk_conn, { '/nodes/{}/domainstate'.format(self.name): 'ready' }) fixed_domain_list = self.s_domain.copy() for dom_uuid in fixed_domain_list: try: last_hypervisor = zkhandler.readdata(self.zk_conn, '/domains/{}/lasthypervisor'.format(dom_uuid)) except: continue if last_hypervisor != self.name: continue ansiiprint.echo('Setting unmigration for VM "{}"'.format(dom_uuid), '', 'i') zkhandler.writedata(self.zk_conn, { '/domains/{}/state'.format(dom_uuid): 'migrate', '/domains/{}/hypervisor'.format(dom_uuid): self.name, '/domains/{}/lasthypervisor'.format(dom_uuid): '' }) self.inflush = False def update_zookeeper(self): # Connect to libvirt libvirt_name = "qemu:///system" lv_conn = libvirt.open(libvirt_name) if lv_conn == None: ansiiprint.echo('Failed to open connection to "{}"'.format(libvirt_name), '', 'e') return # Get past state and update if needed past_state = zkhandler.readdata(self.zk_conn, '/nodes/{}/daemonstate'.format(self.name)) if past_state != 'run': self.daemon_state = 'run' zkhandler.writedata(self.zk_conn, { '/nodes/{}/daemonstate'.format(self.name): 'run' }) else: self.daemon_state = 'run' # Toggle state management of dead VMs to restart them memalloc = 0 vcpualloc = 0 for domain, instance in self.s_domain.items(): if instance.inshutdown == False and domain in self.domain_list: # Add the allocated memory to our memalloc value memalloc += instance.getmemory() vcpualloc += instance.getvcpus() if instance.getstate() == 'start' and instance.gethypervisor() == self.name: if instance.getdom() != None: try: if instance.getdom().state()[0] != libvirt.VIR_DOMAIN_RUNNING: raise except Exception as e: # Toggle a state "change" zkhandler.writedata(self.zk_conn, { '/domains/{}/state'.format(domain): instance.getstate() }) # Ensure that any running VMs are readded to the domain_list running_domains = lv_conn.listAllDomains(libvirt.VIR_CONNECT_LIST_DOMAINS_ACTIVE) for domain in running_domains: domain_uuid = domain.UUIDString() if domain_uuid not in self.domain_list: self.domain_list.append(domain_uuid) # Set our information in zookeeper self.name = lv_conn.getHostname() self.memused = int(psutil.virtual_memory().used / 1024 / 1024) self.memfree = int(psutil.virtual_memory().free / 1024 / 1024) self.memalloc = memalloc self.vcpualloc = vcpualloc self.cpuload = os.getloadavg()[0] self.domains_count = len(lv_conn.listDomainsID()) keepalive_time = int(time.time()) try: zkhandler.writedata(self.zk_conn, { '/nodes/{}/memused'.format(self.name): str(self.memused), '/nodes/{}/memfree'.format(self.name): str(self.memfree), '/nodes/{}/memalloc'.format(self.name): str(self.memalloc), '/nodes/{}/vcpualloc'.format(self.name): str(self.vcpualloc), '/nodes/{}/cpuload'.format(self.name): str(self.cpuload), '/nodes/{}/runningdomains'.format(self.name): ' '.join(self.domain_list), '/nodes/{}/domainscount'.format(self.name): str(self.domains_count), '/nodes/{}/keepalive'.format(self.name): str(keepalive_time) }) except: ansiiprint.echo('Failed to set keepalive data', '', 'e') return # Close the Libvirt connection lv_conn.close() # Display node information to the terminal ansiiprint.echo('{}{} keepalive{}'.format(ansiiprint.purple(), self.name, ansiiprint.end()), '', 't') ansiiprint.echo('{0}Active domains:{1} {2} {0}Allocated memory [MiB]:{1} {6} {0}Free memory [MiB]:{1} {3} {0}Used memory [MiB]:{1} {4} {0}Load:{1} {5}'.format(ansiiprint.bold(), ansiiprint.end(), self.domains_count, self.memfree, self.memused, self.cpuload, self.memalloc), '', 'c') # Update our local node lists for node_name in self.t_node: try: node_daemon_state = zkhandler.readdata(self.zk_conn, '/nodes/{}/daemonstate'.format(node_name)) node_domain_state = zkhandler.readdata(self.zk_conn, '/nodes/{}/domainstate'.format(node_name)) node_keepalive = int(zkhandler.readdata(self.zk_conn, '/nodes/{}/keepalive'.format(node_name))) except: node_daemon_state = 'unknown' node_domain_state = 'unknown' node_keepalive = 0 # Handle deadtime and fencng if needed # (A node is considered dead when its keepalive timer is >6*keepalive_interval seconds # out-of-date while in 'start' state) node_deadtime = int(time.time()) - ( int(self.config['keepalive_interval']) * int(self.config['fence_intervals']) ) if node_keepalive < node_deadtime and node_daemon_state == 'run': ansiiprint.echo('Node {} seems dead - starting monitor for fencing'.format(node_name), '', 'w') zkhandler.writedata(self.zk_conn, { '/nodes/{}/daemonstate'.format(node_name): 'dead' }) fence_thread = threading.Thread(target=fenceNode, args=(node_name, self.zk_conn, self.config), kwargs={}) fence_thread.start() # Update the arrays if node_daemon_state == 'run' and node_domain_state != 'flushed' and node_name not in self.active_node_list: self.active_node_list.append(node_name) try: self.flushed_node_list.remove(node_name) except ValueError: pass try: self.inactive_node_list.remove(node_name) except ValueError: pass if node_daemon_state != 'run' and node_domain_state != 'flushed' and node_name not in self.inactive_node_list: self.inactive_node_list.append(node_name) try: self.active_node_list.remove(node_name) except ValueError: pass try: self.flushed_node_list.remove(node_name) except ValueError: pass if node_domain_state == 'flushed' and node_name not in self.flushed_node_list: self.flushed_node_list.append(node_name) try: self.active_node_list.remove(node_name) except ValueError: pass try: self.inactive_node_list.remove(node_name) except ValueError: pass # Display cluster information to the terminal ansiiprint.echo('{}Cluster status{}'.format(ansiiprint.purple(), ansiiprint.end()), '', 't') ansiiprint.echo('{}Active nodes:{} {}'.format(ansiiprint.bold(), ansiiprint.end(), ' '.join(self.active_node_list)), '', 'c') ansiiprint.echo('{}Inactive nodes:{} {}'.format(ansiiprint.bold(), ansiiprint.end(), ' '.join(self.inactive_node_list)), '', 'c') ansiiprint.echo('{}Flushed nodes:{} {}'.format(ansiiprint.bold(), ansiiprint.end(), ' '.join(self.flushed_node_list)), '', 'c') # # Find a migration target # def findTargetHypervisor(zk_conn, search_field, dom_uuid): if search_field == 'mem': return findTargetHypervisorMem(zk_conn, dom_uuid) if search_field == 'load': return findTargetHypervisorLoad(zk_conn, dom_uuid) if search_field == 'vcpus': return findTargetHypervisorVCPUs(zk_conn, dom_uuid) if search_field == 'vms': return findTargetHypervisorVMs(zk_conn, dom_uuid) return None # Get the list of valid target hypervisors def getHypervisors(zk_conn, dom_uuid): valid_hypervisor_list = [] full_hypervisor_list = zkhandler.listchildren(zk_conn, '/nodes') current_hypervisor = zkhandler.readdata(zk_conn, '/domains/{}/hypervisor'.format(dom_uuid)) for hypervisor in full_hypervisor_list: daemon_state = zkhandler.readdata(zk_conn, '/nodes/{}/daemonstate'.format(hypervisor)) domain_state = zkhandler.readdata(zk_conn, '/nodes/{}/domainstate'.format(hypervisor)) if hypervisor == current_hypervisor: continue if daemon_state != 'run' or domain_state != 'ready': continue valid_hypervisor_list.append(hypervisor) return valid_hypervisor_list # via free memory (relative to allocated memory) def findTargetHypervisorMem(zk_conn, dom_uuid): most_allocfree = 0 target_hypervisor = None hypervisor_list = getHypervisors(zk_conn, dom_uuid) for hypervisor in hypervisor_list: memalloc = int(zkhandler.readdata(zk_conn, '/nodes/{}/memalloc'.format(hypervisor))) memused = int(zkhandler.readdata(zk_conn, '/nodes/{}/memused'.format(hypervisor))) memfree = int(zkhandler.readdata(zk_conn, '/nodes/{}/memfree'.format(hypervisor))) memtotal = memused + memfree allocfree = memtotal - memalloc if allocfree > most_allocfree: most_allocfree = allocfree target_hypervisor = hypervisor return target_hypervisor # via load average def findTargetHypervisorLoad(zk_conn, dom_uuid): least_load = 9999 target_hypervisor = None hypervisor_list = getHypervisors(zk_conn, dom_uuid) for hypervisor in hypervisor_list: load = int(zkhandler.readdata(zk_conn, '/nodes/{}/load'.format(hypervisor))) if load < least_load: least_load = load target_hypevisor = hypervisor return target_hypervisor # via total vCPUs def findTargetHypervisorVCPUs(zk_conn, dom_uuid): least_vcpus = 9999 target_hypervisor = None hypervisor_list = getHypervisors(zk_conn, dom_uuid) for hypervisor in hypervisor_list: vcpus = int(zkhandler.readdata(zk_conn, '/nodes/{}/vcpualloc'.format(hypervisor))) if vcpus < least_vcpus: least_vcpus = vcpus target_hypervisor = hypervisor return target_hypervisor # via total VMs def findTargetHypervisorVMs(zk_conn, dom_uuid): least_vms = 9999 target_hypervisor = None hypervisor_list = getHypervisors(zk_conn, dom_uuid) for hypervisor in hypervisor_list: vms = int(zkhandler.readdata(zk_conn, '/nodes/{}/domainscount'.format(hypervisor))) if vms < least_vms: least_vms = vms target_hypervisor = hypervisor return target_hypervisor # # Fence thread entry function # def fenceNode(node_name, zk_conn, config): failcount = 0 # We allow exactly 3 saving throws for the host to come back online while failcount < 3: # Wait 5 seconds time.sleep(5) # Get the state node_daemon_state = zkhandler.readdata(zk_conn, '/nodes/{}/daemonstate'.format(node_name)) # Is it still 'dead' if node_daemon_state == 'dead': failcount += 1 ansiiprint.echo('Node "{}" failed {} saving throws'.format(node_name, failcount), '', 'w') # It changed back to something else so it must be alive else: ansiiprint.echo('Node "{}" passed a saving throw; canceling fence'.format(node_name), '', 'o') return ansiiprint.echo('Fencing node "{}" via IPMI reboot signal'.format(node_name), '', 'e') # Get IPMI information ipmi_hostname = zkhandler.readdata(zk_conn, '/nodes/{}/ipmihostname'.format(node_name)) ipmi_username = zkhandler.readdata(zk_conn, '/nodes/{}/ipmiusername'.format(node_name)) ipmi_password = zkhandler.readdata(zk_conn, '/nodes/{}/ipmipassword'.format(node_name)) # Shoot it in the head fence_status = rebootViaIPMI(ipmi_hostname, ipmi_username, ipmi_password) # Hold to ensure the fence takes effect time.sleep(3) # If the fence succeeded and successful_fence is migrate if fence_status == True and config['successful_fence'] == 'migrate': migrateFromFencedHost(zk_conn, node_name) # If the fence failed and failed_fence is migrate if fence_status == False and config['failed_fence'] == 'migrate' and config['suicide_intervals'] != '0': migrateFromFencedHost(zk_conn, node_name) # Migrate hosts away from a fenced node def migrateFromFencedHost(zk_conn, node_name): ansiiprint.echo('Moving VMs from dead hypervisor "{}" to new hosts'.format(node_name), '', 'i') dead_node_running_domains = zkhandler.readdata(zk_conn, '/nodes/{}/runningdomains'.format(node_name)).split() for dom_uuid in dead_node_running_domains: target_hypervisor = findTargetHypervisor(zk_conn, 'mem', dom_uuid) ansiiprint.echo('Moving VM "{}" to hypervisor "{}"'.format(dom_uuid, target_hypervisor), '', 'i') zkhandler.writedata(zk_conn, { '/domains/{}/state'.format(dom_uuid): 'start', '/domains/{}/hypervisor'.format(dom_uuid): target_hypervisor, '/domains/{}/lasthypervisor'.format(dom_uuid): node_name }) # Set node in flushed state for easy remigrating when it comes back zkhandler.writedata(zk_conn, { '/nodes/{}/domainstate'.format(node_name): 'flushed' }) # # Perform an IPMI fence # def rebootViaIPMI(ipmi_hostname, ipmi_user, ipmi_password): ipmi_command = ['/usr/bin/ipmitool', '-I', 'lanplus', '-H', ipmi_hostname, '-U', ipmi_user, '-P', ipmi_password, 'chassis', 'power', 'reset'] ipmi_command_output = subprocess.run(ipmi_command, stdout=subprocess.PIPE, stderr=subprocess.PIPE) if ipmi_command_output.returncode == 0: ansiiprint.echo('Successfully rebooted dead node', '', 'o') return True else: ansiiprint.echo('Failed to reboot dead node', '', 'e') return False