Uses same method as state or node filtering, rather than altering how the main LIMIT field works.
		
			
				
	
	
		
			340 lines
		
	
	
		
			12 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
			
		
		
	
	
			340 lines
		
	
	
		
			12 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
| #!/usr/bin/env python3
 | |
| 
 | |
| # cluster.py - PVC client function library, cluster management
 | |
| # Part of the Parallel Virtual Cluster (PVC) system
 | |
| #
 | |
| #    Copyright (C) 2018-2021 Joshua M. Boniface <joshua@boniface.me>
 | |
| #
 | |
| #    This program is free software: you can redistribute it and/or modify
 | |
| #    it under the terms of the GNU General Public License as published by
 | |
| #    the Free Software Foundation, version 3.
 | |
| #
 | |
| #    This program is distributed in the hope that it will be useful,
 | |
| #    but WITHOUT ANY WARRANTY; without even the implied warranty of
 | |
| #    MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
 | |
| #    GNU General Public License for more details.
 | |
| #
 | |
| #    You should have received a copy of the GNU General Public License
 | |
| #    along with this program.  If not, see <https://www.gnu.org/licenses/>.
 | |
| #
 | |
| ###############################################################################
 | |
| 
 | |
| import re
 | |
| 
 | |
| import daemon_lib.common as common
 | |
| import daemon_lib.vm as pvc_vm
 | |
| import daemon_lib.node as pvc_node
 | |
| import daemon_lib.network as pvc_network
 | |
| import daemon_lib.ceph as pvc_ceph
 | |
| 
 | |
| 
 | |
| def set_maintenance(zkhandler, maint_state):
 | |
|     current_maint_state = zkhandler.read('base.config.maintenance')
 | |
|     if maint_state == current_maint_state:
 | |
|         if maint_state == 'true':
 | |
|             return True, 'Cluster is already in maintenance mode'
 | |
|         else:
 | |
|             return True, 'Cluster is already in normal mode'
 | |
| 
 | |
|     if maint_state == 'true':
 | |
|         zkhandler.write([
 | |
|             ('base.config.maintenance', 'true')
 | |
|         ])
 | |
|         return True, 'Successfully set cluster in maintenance mode'
 | |
|     else:
 | |
|         zkhandler.write([
 | |
|             ('base.config.maintenance', 'false')
 | |
|         ])
 | |
|         return True, 'Successfully set cluster in normal mode'
 | |
| 
 | |
| 
 | |
| def getClusterInformation(zkhandler):
 | |
|     # Get cluster maintenance state
 | |
|     maint_state = zkhandler.read('base.config.maintenance')
 | |
| 
 | |
|     # List of messages to display to the clients
 | |
|     cluster_health_msg = []
 | |
|     storage_health_msg = []
 | |
| 
 | |
|     # Get node information object list
 | |
|     retcode, node_list = pvc_node.get_list(zkhandler, None)
 | |
| 
 | |
|     # Get vm information object list
 | |
|     retcode, vm_list = pvc_vm.get_list(zkhandler, None, None, None, None)
 | |
| 
 | |
|     # Get network information object list
 | |
|     retcode, network_list = pvc_network.get_list(zkhandler, None, None)
 | |
| 
 | |
|     # Get storage information object list
 | |
|     retcode, ceph_osd_list = pvc_ceph.get_list_osd(zkhandler, None)
 | |
|     retcode, ceph_pool_list = pvc_ceph.get_list_pool(zkhandler, None)
 | |
|     retcode, ceph_volume_list = pvc_ceph.get_list_volume(zkhandler, None, None)
 | |
|     retcode, ceph_snapshot_list = pvc_ceph.get_list_snapshot(zkhandler, None, None, None)
 | |
| 
 | |
|     # Determine, for each subsection, the total count
 | |
|     node_count = len(node_list)
 | |
|     vm_count = len(vm_list)
 | |
|     network_count = len(network_list)
 | |
|     ceph_osd_count = len(ceph_osd_list)
 | |
|     ceph_pool_count = len(ceph_pool_list)
 | |
|     ceph_volume_count = len(ceph_volume_list)
 | |
|     ceph_snapshot_count = len(ceph_snapshot_list)
 | |
| 
 | |
|     # Determinations for general cluster health
 | |
|     cluster_healthy_status = True
 | |
|     # Check for (n-1) overprovisioning
 | |
|     #   Assume X nodes. If the total VM memory allocation (counting only running VMss) is greater than
 | |
|     #   the total memory of the (n-1) smallest nodes, trigger this warning.
 | |
|     n_minus_1_total = 0
 | |
|     alloc_total = 0
 | |
| 
 | |
|     node_largest_index = None
 | |
|     node_largest_count = 0
 | |
|     for index, node in enumerate(node_list):
 | |
|         node_mem_total = node['memory']['total']
 | |
|         node_mem_alloc = node['memory']['allocated']
 | |
|         alloc_total += node_mem_alloc
 | |
| 
 | |
|         # Determine if this node is the largest seen so far
 | |
|         if node_mem_total > node_largest_count:
 | |
|             node_largest_index = index
 | |
|             node_largest_count = node_mem_total
 | |
|     n_minus_1_node_list = list()
 | |
|     for index, node in enumerate(node_list):
 | |
|         if index == node_largest_index:
 | |
|             continue
 | |
|         n_minus_1_node_list.append(node)
 | |
|     for index, node in enumerate(n_minus_1_node_list):
 | |
|         n_minus_1_total += node['memory']['total']
 | |
|     if alloc_total > n_minus_1_total:
 | |
|         cluster_healthy_status = False
 | |
|         cluster_health_msg.append("Total VM memory ({}) is overprovisioned (max {}) for (n-1) failure scenarios".format(alloc_total, n_minus_1_total))
 | |
| 
 | |
|     # Determinations for node health
 | |
|     node_healthy_status = list(range(0, node_count))
 | |
|     node_report_status = list(range(0, node_count))
 | |
|     for index, node in enumerate(node_list):
 | |
|         daemon_state = node['daemon_state']
 | |
|         domain_state = node['domain_state']
 | |
|         if daemon_state != 'run' and domain_state != 'ready':
 | |
|             node_healthy_status[index] = False
 | |
|             cluster_health_msg.append("Node '{}' in {},{} state".format(node['name'], daemon_state, domain_state))
 | |
|         else:
 | |
|             node_healthy_status[index] = True
 | |
|         node_report_status[index] = daemon_state + ',' + domain_state
 | |
| 
 | |
|     # Determinations for VM health
 | |
|     vm_healthy_status = list(range(0, vm_count))
 | |
|     vm_report_status = list(range(0, vm_count))
 | |
|     for index, vm in enumerate(vm_list):
 | |
|         vm_state = vm['state']
 | |
|         if vm_state not in ['start', 'disable', 'migrate', 'unmigrate', 'provision']:
 | |
|             vm_healthy_status[index] = False
 | |
|             cluster_health_msg.append("VM '{}' in {} state".format(vm['name'], vm_state))
 | |
|         else:
 | |
|             vm_healthy_status[index] = True
 | |
|         vm_report_status[index] = vm_state
 | |
| 
 | |
|     # Determinations for OSD health
 | |
|     ceph_osd_healthy_status = list(range(0, ceph_osd_count))
 | |
|     ceph_osd_report_status = list(range(0, ceph_osd_count))
 | |
|     for index, ceph_osd in enumerate(ceph_osd_list):
 | |
|         try:
 | |
|             ceph_osd_up = ceph_osd['stats']['up']
 | |
|         except KeyError:
 | |
|             ceph_osd_up = 0
 | |
| 
 | |
|         try:
 | |
|             ceph_osd_in = ceph_osd['stats']['in']
 | |
|         except KeyError:
 | |
|             ceph_osd_in = 0
 | |
| 
 | |
|         up_texts = {1: 'up', 0: 'down'}
 | |
|         in_texts = {1: 'in', 0: 'out'}
 | |
| 
 | |
|         if not ceph_osd_up or not ceph_osd_in:
 | |
|             ceph_osd_healthy_status[index] = False
 | |
|             cluster_health_msg.append('OSD {} in {},{} state'.format(ceph_osd['id'], up_texts[ceph_osd_up], in_texts[ceph_osd_in]))
 | |
|         else:
 | |
|             ceph_osd_healthy_status[index] = True
 | |
|         ceph_osd_report_status[index] = up_texts[ceph_osd_up] + ',' + in_texts[ceph_osd_in]
 | |
| 
 | |
|     # Find out the overall cluster health; if any element of a healthy_status is false, it's unhealthy
 | |
|     if maint_state == 'true':
 | |
|         cluster_health = 'Maintenance'
 | |
|     elif cluster_healthy_status is False or False in node_healthy_status or False in vm_healthy_status or False in ceph_osd_healthy_status:
 | |
|         cluster_health = 'Degraded'
 | |
|     else:
 | |
|         cluster_health = 'Optimal'
 | |
| 
 | |
|     # Find out our storage health from Ceph
 | |
|     ceph_status = zkhandler.read('base.storage').split('\n')
 | |
|     ceph_health = ceph_status[2].split()[-1]
 | |
| 
 | |
|     # Parse the status output to get the health indicators
 | |
|     line_record = False
 | |
|     for index, line in enumerate(ceph_status):
 | |
|         if re.search('services:', line):
 | |
|             line_record = False
 | |
|         if line_record and len(line.strip()) > 0:
 | |
|             storage_health_msg.append(line.strip())
 | |
|         if re.search('health:', line):
 | |
|             line_record = True
 | |
| 
 | |
|     if maint_state == 'true':
 | |
|         storage_health = 'Maintenance'
 | |
|     elif ceph_health != 'HEALTH_OK':
 | |
|         storage_health = 'Degraded'
 | |
|     else:
 | |
|         storage_health = 'Optimal'
 | |
| 
 | |
|     # State lists
 | |
|     node_state_combinations = [
 | |
|         'run,ready', 'run,flush', 'run,flushed', 'run,unflush',
 | |
|         'init,ready', 'init,flush', 'init,flushed', 'init,unflush',
 | |
|         'stop,ready', 'stop,flush', 'stop,flushed', 'stop,unflush',
 | |
|         'dead,ready', 'dead,flush', 'dead,flushed', 'dead,unflush'
 | |
|     ]
 | |
|     vm_state_combinations = [
 | |
|         'start', 'restart', 'shutdown', 'stop', 'disable', 'fail', 'migrate', 'unmigrate', 'provision'
 | |
|     ]
 | |
|     ceph_osd_state_combinations = [
 | |
|         'up,in', 'up,out', 'down,in', 'down,out'
 | |
|     ]
 | |
| 
 | |
|     # Format the Node states
 | |
|     formatted_node_states = {'total': node_count}
 | |
|     for state in node_state_combinations:
 | |
|         state_count = 0
 | |
|         for node_state in node_report_status:
 | |
|             if node_state == state:
 | |
|                 state_count += 1
 | |
|         if state_count > 0:
 | |
|             formatted_node_states[state] = state_count
 | |
| 
 | |
|     # Format the VM states
 | |
|     formatted_vm_states = {'total': vm_count}
 | |
|     for state in vm_state_combinations:
 | |
|         state_count = 0
 | |
|         for vm_state in vm_report_status:
 | |
|             if vm_state == state:
 | |
|                 state_count += 1
 | |
|         if state_count > 0:
 | |
|             formatted_vm_states[state] = state_count
 | |
| 
 | |
|     # Format the OSD states
 | |
|     formatted_osd_states = {'total': ceph_osd_count}
 | |
|     for state in ceph_osd_state_combinations:
 | |
|         state_count = 0
 | |
|         for ceph_osd_state in ceph_osd_report_status:
 | |
|             if ceph_osd_state == state:
 | |
|                 state_count += 1
 | |
|         if state_count > 0:
 | |
|             formatted_osd_states[state] = state_count
 | |
| 
 | |
|     # Format the status data
 | |
|     cluster_information = {
 | |
|         'health': cluster_health,
 | |
|         'health_msg': cluster_health_msg,
 | |
|         'storage_health': storage_health,
 | |
|         'storage_health_msg': storage_health_msg,
 | |
|         'primary_node': common.getPrimaryNode(zkhandler),
 | |
|         'upstream_ip': zkhandler.read('base.config.upstream_ip'),
 | |
|         'nodes': formatted_node_states,
 | |
|         'vms': formatted_vm_states,
 | |
|         'networks': network_count,
 | |
|         'osds': formatted_osd_states,
 | |
|         'pools': ceph_pool_count,
 | |
|         'volumes': ceph_volume_count,
 | |
|         'snapshots': ceph_snapshot_count
 | |
|     }
 | |
| 
 | |
|     return cluster_information
 | |
| 
 | |
| 
 | |
| def get_info(zkhandler):
 | |
|     # This is a thin wrapper function for naming purposes
 | |
|     cluster_information = getClusterInformation(zkhandler)
 | |
|     if cluster_information:
 | |
|         return True, cluster_information
 | |
|     else:
 | |
|         return False, 'ERROR: Failed to obtain cluster information!'
 | |
| 
 | |
| 
 | |
| def cluster_initialize(zkhandler, overwrite=False):
 | |
|     # Abort if we've initialized the cluster before
 | |
|     if zkhandler.exists('base.config.primary_node') and not overwrite:
 | |
|         return False, 'ERROR: Cluster contains data and overwrite not set.'
 | |
| 
 | |
|     if overwrite:
 | |
|         # Delete the existing keys
 | |
|         for key in zkhandler.schema.keys('base'):
 | |
|             if key == 'root':
 | |
|                 # Don't delete the root key
 | |
|                 continue
 | |
| 
 | |
|             status = zkhandler.delete('base.{}'.format(key), recursive=True)
 | |
|             if not status:
 | |
|                 return False, 'ERROR: Failed to delete data in cluster; running nodes perhaps?'
 | |
| 
 | |
|     # Create the root keys
 | |
|     zkhandler.schema.apply(zkhandler)
 | |
| 
 | |
|     return True, 'Successfully initialized cluster'
 | |
| 
 | |
| 
 | |
| def cluster_backup(zkhandler):
 | |
|     # Dictionary of values to come
 | |
|     cluster_data = dict()
 | |
| 
 | |
|     def get_data(path):
 | |
|         data = zkhandler.read(path)
 | |
|         children = zkhandler.children(path)
 | |
| 
 | |
|         cluster_data[path] = data
 | |
| 
 | |
|         if children:
 | |
|             if path == '/':
 | |
|                 child_prefix = '/'
 | |
|             else:
 | |
|                 child_prefix = path + '/'
 | |
| 
 | |
|             for child in children:
 | |
|                 if child_prefix + child == '/zookeeper':
 | |
|                     # We must skip the built-in /zookeeper tree
 | |
|                     continue
 | |
|                 if child_prefix + child == '/patroni':
 | |
|                     # We must skip the /patroni tree
 | |
|                     continue
 | |
| 
 | |
|                 get_data(child_prefix + child)
 | |
| 
 | |
|     try:
 | |
|         get_data('/')
 | |
|     except Exception as e:
 | |
|         return False, 'ERROR: Failed to obtain backup: {}'.format(e)
 | |
| 
 | |
|     return True, cluster_data
 | |
| 
 | |
| 
 | |
| def cluster_restore(zkhandler, cluster_data):
 | |
|     # Build a key+value list
 | |
|     kv = []
 | |
|     schema_version = None
 | |
|     for key in cluster_data:
 | |
|         if key == zkhandler.schema.path('base.schema.version'):
 | |
|             schema_version = cluster_data[key]
 | |
|         data = cluster_data[key]
 | |
|         kv.append((key, data))
 | |
| 
 | |
|     if int(schema_version) != int(zkhandler.schema.version):
 | |
|         return False, 'ERROR: Schema version of backup ({}) does not match cluster schema version ({}).'.format(schema_version, zkhandler.schema.version)
 | |
| 
 | |
|     # Close the Zookeeper connection
 | |
|     result = zkhandler.write(kv)
 | |
| 
 | |
|     if result:
 | |
|         return True, 'Restore completed successfully.'
 | |
|     else:
 | |
|         return False, 'Restore failed.'
 |