pvc/client-cli/pvc_init.py

663 lines
24 KiB
Python
Raw Normal View History

2019-03-10 20:40:03 -04:00
#!/usr/bin/env python3
# pvcd.py - PVC client command-line interface
# Part of the Parallel Virtual Cluster (PVC) system
#
# Copyright (C) 2018 Joshua M. Boniface <joshua@boniface.me>
#
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
# the Free Software Foundation, either version 3 of the License, or
# (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with this program. If not, see <https://www.gnu.org/licenses/>.
#
###############################################################################
import locale
import socket
import click
import tempfile
import sys
import os
import tempfile
import subprocess
import difflib
import re
import yaml
import colorama
import netifaces
import ipaddress
import urllib.request
import tarfile
from dialog import Dialog
import client_lib.common as pvc_common
import client_lib.node as pvc_node
# Repository configurations
#deb_mirror = "ftp.debian.org"
deb_mirror = "deb1.i.bonilan.net:3142"
deb_release = "buster"
deb_arch = "amd64"
deb_packages = "mdadm,lvm2,parted,gdisk,debootstrap,grub-pc,linux-image-amd64"
# Scripts
cluster_floating_ip = "10.10.1.254"
bootstrap_script = """#!/bin/bash
# Check in and get our nodename, pvcd.conf, and install script
output="$( curl {}:10080/node_checkin )"
# Export node_id
node_id="$( jq -r '.node_id' <<<"${output}" )"
export node_id
# Export pvcd.conf
pvcd_conf="$( jq -r '.pvcd_conf' <<<"${output}" )"
export pvcd_conf
# Execute install script
jq -r '.install_script' <<<"${output}" | bash
"""
install_script = """#!/bin/bash
#
"""
# Run a oneshot command, optionally without blocking
def run_os_command(command_string, environment=None):
command = command_string.split()
try:
command_output = subprocess.run(
command,
env=environment,
stdout=subprocess.PIPE,
stderr=subprocess.PIPE,
)
except FileNotFoundError:
return 1, "", ""
retcode = command_output.returncode
try:
stdout = command_output.stdout.decode('ascii')
except:
stdout = ''
try:
stderr = command_output.stderr.decode('ascii')
except:
stderr = ''
return retcode, stdout, stderr
# Start the initalization of a new cluster
def orun():
locale.setlocale(locale.LC_ALL, '')
# You may want to use 'autowidgetsize=True' here (requires pythondialog >= 3.1)
d = Dialog(dialog="dialog", autowidgetsize=True)
de = Dialog(dialog="dialog", autowidgetsize=True)
# Dialog.set_background_title() requires pythondialog 2.13 or later
d.set_background_title("PVC Cluster Initialization")
# Initial message
d.msgbox("""Welcome to the PVC cluster initalization tool. This tool
will ask you several questions about the cluster, and then
perform the required tasks to bootstrap the cluster.
PLEASE READ ALL SCREENS CAREFULLY.
Before proceeding, ensure that:
(a) This system is connected, wired if possible, to a switch.
(b) This system has a second network connection with Internet
connectivity and is able to download files.
(c) The initial nodes are powered off, connected to the
mentioned switch, and are configured to boot from PXE.
(d) All non-system disks are disconnected from all nodes.
Storage disks will be added after bootstrapping.
Once these prerequisites are complete, press Enter to proceed.
""")
#
# Phase 0 - get our local interface
#
interfaces = netifaces.interfaces()
interface_list = list()
for idx, val in enumerate(interfaces):
interface_list.append(("{}".format(idx), "{}".format(val)))
code, index = d.menu("""Select a network interface to use for cluster bootstrapping.""",
choices=interface_list)
if code == d.CANCEL:
print("Aborted.")
exit(0)
interface = interfaces[int(index)]
#
# Phase 1 - coordinator list
#
code, coordinator_count = d.menu("Select the number of initial (coordinator) nodes:",
choices=[("1", "Testing and very small non-redundant deployments"),
("3", "Standard (3-20) hypervisor deployments"),
("5", "Large (21-99) hypervisor deployments")])
coordinator_count = int(coordinator_count)
if code == d.CANCEL:
print("Aborted.")
exit(0)
#
# Phase 2 - Get the networks
#
d.msgbox("""The next screens will ask for the cluster networks in CIDR
format as well as a floating IP in each. The networks are:
(a) Cluster: Used by the nodes to communicate VXLANs and
pass virtualization (migration) traffic between each
other. Each node will be assigned an address in this
network equal to its node ID (e.g. node1 at .1, etc.).
Each node with IPMI support will be assigned an IPMI
address in this network equal to its node ID plus 120
(e.g. node1-lom at .121, etc.). IPs 241-254 will be
reserved for cluster management; the floating IP should
be in this range.
(b) Storage: Used by the nodes to pass storage traffic
between each other, both for Ceph OSDs and for RBD
access. Each node will be assigned an address in this
network equal to its node ID. IPs 241-254 will be
reserved for cluster management; the floating IP should
be in this range.
(c) Upstream: Used by the nodes to communicate upstream
outside of the cluster. This network has several
functions depending on the configuration of the
virtual networks; relevant questions will be asked
later in the configuration.
* The first two networks are dedicated to the cluster. They
should be RFC1918 private networks and be sized sufficiently
for the future growth of the cluster; a /24 is recommended
for most situations and will support up to 99 nodes.
* The third network, as mentioned, has several potential
functions depending on the final network configuration of the
cluster. It should already exist, and nodes may or may not
have individual addresses in this network. Further questions
about this network will be asked later during setup.
* All networks should have a DNS domain which will be asked
during this stage. For the first two networks, the domain
may be private and unresolvable outside the network if
desired; the third should be a valid but will generally
be unused in the administration of the cluster. The FQDNs
of each node will contain the Cluster domain.
""")
# Get the primary cluster network
valid_network = False
message = "Enter the new cluster's primary network in CIDR format."
while not valid_network:
code, network = d.inputbox(message)
if code == d.CANCEL:
print("Aborted.")
exit(0)
try:
cluster_network = ipaddress.ip_network(network)
valid_network = True
except ValueError:
message = "Error - network {} is not valid.\n\nEnter the new cluster's primary network in CIDR format.".format(network)
continue
valid_address = False
message = "Enter the CIDR floating IP address for the cluster's primary network."
while not valid_address:
code, address = d.inputbox(message)
if code == d.CANCEL:
print("Aborted.")
exit (0)
try:
cluster_floating_ip = ipaddress.ip_address(address)
if not cluster_floating_ip in list(cluster_network.hosts()):
message = "Error - address {} is not in network {}.\n\nEnter the CIDR floating IP address for the cluster's primary network.".format(cluster_floating_ip, cluster_network)
continue
valid_address = True
except ValueError:
message = "Error - address {} is not valid.\n\nEnter the CIDR floating IP address for the cluster's primary network.".format(cluster_floating_ip, cluster_network)
continue
code, cluster_domain = d.inputbox("""Enter the new cluster's primary DNS domain.""")
if code == d.CANCEL:
print("Aborted.")
exit(0)
# Get the storage network
valid_network = False
message = "Enter the new cluster's storage network in CIDR format."
while not valid_network:
code, network = d.inputbox(message)
if code == d.CANCEL:
print("Aborted.")
exit(0)
try:
storage_network = ipaddress.ip_network(network)
valid_network = True
except ValueError:
message = "Error - network {} is not valid.\n\nEnter the new cluster's storage network in CIDR format.".format(network)
continue
valid_address = False
message = "Enter the CIDR floating IP address for the cluster's storage network."
while not valid_address:
code, address = d.inputbox(message)
if code == d.CANCEL:
print("Aborted.")
exit (0)
try:
storage_floating_ip = ipaddress.ip_address(address)
if not storage_floating_ip in list(storage_network.hosts()):
message = "Error - address {} is not in network {}.\n\nEnter the CIDR floating IP address for the cluster's storage network.".format(storage_floating_ip, storage_network)
continue
valid_address = True
except ValueError:
message = "Error - address {} is not valid.\n\nEnter the CIDR floating IP address for the cluster's storage network.".format(storage_floating_ip, storage_network)
continue
code, storage_domain = d.inputbox("""Enter the new cluster's storage DNS domain.""")
if code == d.CANCEL:
print("Aborted.")
exit(0)
# Get the upstream network
valid_network = False
message = "Enter the new cluster's upstream network in CIDR format."
while not valid_network:
code, network = d.inputbox(message)
if code == d.CANCEL:
print("Aborted.")
exit(0)
try:
upstream_network = ipaddress.ip_network(network)
valid_network = True
except ValueError:
message = "Error - network {} is not valid.\n\nEnter the new cluster's upstream network in CIDR format.".format(network)
continue
valid_address = False
message = "Enter the CIDR floating IP address for the cluster's upstream network."
while not valid_address:
code, address = d.inputbox(message)
if code == d.CANCEL:
print("Aborted.")
exit (0)
try:
upstream_floating_ip = ipaddress.ip_address(address)
if not upstream_floating_ip in list(upstream_network.hosts()):
message = "Error - address {} is not in network {}.\n\nEnter the CIDR floating IP address for the cluster's upstream network.".format(upstream_floating_ip, upstream_network)
continue
valid_address = True
except ValueError:
message = "Error - address {} is not valid.\n\nEnter the CIDR floating IP address for the cluster's upstream network.".format(upstream_floating_ip, upstream_network)
continue
code, upstream_domain = d.inputbox("""Enter the new cluster's upstream DNS domain.""")
if code == d.CANCEL:
print("Aborted.")
exit(0)
#
# Phase 3 - Upstream settings
#
d.msgbox("""The next screens will present several questions regarding
the upstream and guest network configuration for the new
cluster, in an attempt to determine some default values
for the initial template files. Most of these options can
be overridden later by the client configuration tool or by
manual modification of the node configuration files, but
will shape the initial VM configuration and node config
file.
""")
if d.yesno("""Should the PVC cluster manage client IP addressing?""") == d.OK:
enable_routing = True
else:
enable_routing = False
if d.yesno("""Should the PVC cluster provide NAT functionality?""") == d.OK:
enable_nat = True
else:
enable_nat = False
if d.yesno("""Should the PVC cluster manage client DNS records?""") == d.OK:
enable_dns = True
else:
enable_dns = False
#
# Phase 4 - Configure templates
#
d.msgbox("""The next screens will present templates for several
configuration files in your $EDITOR, based on the options
selected above. These templates will be distributed to the
cluster nodes during bootstrapping.
Various values are indicated for '<replacement>' by you,
as 'TEMPLATE' values to be filled in from other information,
gained during these dialogs, or as default values.
Once you are finished editing the files, write and quit the
editor.
For more information on any particular field, see the PVC
documentation.
""")
# Generate the node interfaces file template
interfaces_configuration = """#
# pvc node network interfaces file
#
# Writing this template requires knowledge of the default
# persistent network names of the target server class.
#
# Configure any required bonding here, however do not
# configure any vLANs or VXLANs as those are managed
# by the PVC daemon itself.
#
# Make note of the interfaces specified for each type,
# as these will be required in the daemon config as
# well.
#
# Note that the Cluster and Storage networks *may* use
# the same underlying network device; in which case,
# only define one here and specify the same device
# for both networks in the daemon config.
auto lo
iface lo inet loopback
# Upstream physical interface
auto <upstream_dev_interface>
iface <upstream_dev_interface> inet manual
# Cluster physical interface
auto <cluster_dev_interface>
iface <cluster_dev_interface> inet manual
# Storage physical interface
auto <storage_dev_interface>
iface <storage_dev_interface> inet manual
"""
with tempfile.NamedTemporaryFile(suffix=".tmp") as tf:
EDITOR = os.environ.get('EDITOR', 'vi')
tf.write(interfaces_configuration.encode("utf-8"))
tf.flush()
subprocess.call([EDITOR, tf.name])
tf.seek(0)
interfaces_configuration = tf.read().decode("utf-8")
# Generate the configuration file template
coordinator_list = list()
for i in range(0,coordinator_count):
coordinator_list.append("node{}".format(i + 1))
dnssql_password = "Sup3rS3cr37SQL"
ipmi_password = "Sup3rS3cr37IPMI"
pvcd_configuration = {
"pvc": {
"node": "NODENAME",
"cluster": {
"coordinators": coordinator_list,
"networks": {
"upstream": {
"domain": upstream_domain,
"network": str(upstream_network),
"floating_ip": str(upstream_floating_ip)
},
"cluster": {
"domain": cluster_domain,
"network": str(cluster_network),
"floating_ip": str(cluster_floating_ip)
},
"storage": {
"domain": storage_domain,
"network": str(storage_network),
"floating_ip": str(storage_floating_ip)
},
}
},
"coordinator": {
"dns": {
"database": {
"host": "localhost",
"port": "3306",
"name": "pvcdns",
"user": "pvcdns",
"pass": dnssql_password
}
}
},
"system": {
"fencing": {
"intervals": {
"keepalive_interval": "5",
"fence_intervals": "6",
"suicide_intervals": "0"
},
"actions": {
"successful_fence": "migrate",
"failed_fence": "None"
},
"ipmi": {
"address": "by-id",
"user": "pvcipmi",
"pass": ipmi_password
}
},
"migration": {
"target_selector": "mem"
},
"configuration":{
"directories": {
"dynamic_directory": "/run/pvc",
"log_directory": "/var/log/pvc"
},
"logging": {
"file_logging": "True",
"stdout_logging": "True"
},
"networking": {
"upstream": {
"device": "<upstream_interface_dev>",
"address": "None"
},
"cluster": {
"device": "<cluster_interface_dev>",
"address": "by-id"
},
"storage": {
"device": "<storage_interface_dev>",
"address": "by-id"
}
}
}
}
}
}
pvcd_configuration_header = """#
# pvcd node configuration file
#
# For full details on the available options, consult the PVC documentation.
#
# The main pertanent replacements are:
# <upstream_interface_dev>: the upstream device name from the interface template
# <cluster_interface_dev>: the cluster device name from the interface template
# <storage_interface_dev>: the storage device name from the interface template
"""
with tempfile.NamedTemporaryFile(suffix=".tmp") as tf:
EDITOR = os.environ.get('EDITOR', 'vi')
pvcd_configuration_string = pvcd_configuration_header + yaml.dump(pvcd_configuration, default_style='"', default_flow_style=False)
tf.write(pvcd_configuration_string.encode("utf-8"))
tf.flush()
subprocess.call([EDITOR, tf.name])
tf.seek(0)
pvcd_configuration = yaml.load(tf.read().decode("utf-8"))
# We now have all the details to begin
# - interface
# - coordinator_count
# - cluster_network
# - cluster_floating_ip
# - cluster_domain
# - storage_network
# - storage_floating_ip
# - storage_domain
# - upstream_network
# - upstream_floating_ip
# - upstream_domain
# - enable_routing
# - enable_nat
# - enable_dns
# - interfaces_configuration [template]
# - coordinator_list
# - dnssql_password
# - ipmi_password
# - pvcd_configuration [ template]
d.msgbox("""Information gathering complete. The PVC bootstrap
utility will now prepare the local system:
(a) Generate the node bootstrap image(s).
(b) Start up dnsmasq listening on the interface.
""")
def run():
# Begin preparing the local system - install required packages
required_packages = [
'dnsmasq',
'debootstrap',
'debconf-utils',
'squashfs-tools',
'live-boot',
'ansible'
]
apt_command = "sudo apt install -y " + ' '.join(required_packages)
retcode, stdout, stderr = run_os_command(apt_command)
print(stdout)
if retcode:
print("ERROR: Package installation failed. Aborting setup.")
print(stderr)
exit(1)
#
# Generate a TFTP image for the installer
#
# Create our temporary working directory
print("Create temporary directory...")
tempdir = tempfile.mkdtemp()
print(" > " + tempdir)
# Download the netboot files
print("Download PXE boot files...")
download_path = "http://{mirror}/debian/dists/{release}/main/installer-{arch}/current/images/netboot/netboot.tar.gz".format(
mirror=deb_mirror,
release=deb_release,
arch=deb_arch
)
bootarchive_file, headers = urllib.request.urlretrieve (download_path, tempdir + "/netboot.tar.gz")
print(" > " + bootarchive_file)
# Extract the netboot files
print("Extract PXE boot files...")
with tarfile.open(bootarchive_file) as tar:
tar.extractall(tempdir + "/bootfiles")
# Prepare a bare system with debootstrap
print("Prepare installer debootstrap install...")
debootstrap_command = "sudo -u root debootstrap --include={instpkg} {release} {tempdir}/rootfs http://{mirror}/debian".format(
instpkg=deb_packages,
release=deb_release,
tempdir=tempdir,
mirror=deb_mirror
)
retcode, stdout, stderr = run_os_command(debootstrap_command)
if retcode:
print("ERROR: Debootstrap failed. Aborting setup.")
print(stdout)
exit(1)
# Prepare some useful configuration tweaks
print("Tweaking installed image for boot...")
sedtty_command = """sudo -u root sed -i
's|/sbin/agetty --noclear|/sbin/agetty --noclear --autologin root|g'
{}/rootfs/etc/systemd/system/getty@tty1.service""".format(tempdir)
retcode, stdout, stderr = run_os_command(sedtty_command)
# "Fix" permissions so we can write
retcode, stdout, stderr = run_os_command("sudo chmod 777 {}/rootfs/root".format(tempdir))
retcode, stdout, stderr = run_os_command("sudo chmod 666 {}/rootfs/root/.bashrc".format(tempdir))
# Write the install script to root's bashrc
with open("{}/rootfs/root/.bashrc".format(tempdir), "w") as bashrcf:
bashrcf.write(bootstrap_script)
# Restore permissions
retcode, stdout, stderr = run_os_command("sudo chmod 600 {}/rootfs/root/.bashrc".format(tempdir))
retcode, stdout, stderr = run_os_command("sudo chmod 700 {}/rootfs/root".format(tempdir))
# Create the squashfs
print("Create the squashfs...")
squashfs_command = "sudo nice mksquashfs {tempdir}/rootfs {tempdir}/bootfiles/installer.squashfs".format(
tempdir=tempdir
)
retcode, stdout, stderr = run_os_command(squashfs_command)
if retcode:
print("ERROR: SquashFS creation failed. Aborting setup.")
print(stderr)
exit(1)
#
# Prepare the DHCP and TFTP dnsmasq daemon
#
#
# Prepare the HTTP listenener for the first node
#
#
# Initialize the Zookeeper cluster
#
def init_zookeeper(zk_host):
click.echo('Initializing a new cluster with Zookeeper address "{}".'.format(zk_host))
# Open a Zookeeper connection
zk_conn = pvc_common.startZKConnection(zk_host)
# Destroy the existing data
try:
zk_conn.delete('/networks', recursive=True)
zk_conn.delete('/domains', recursive=True)
zk_conn.delete('/nodes', recursive=True)
zk_conn.delete('/primary_node', recursive=True)
zk_conn.delete('/ceph', recursive=True)
except:
pass
# Create the root keys
transaction = zk_conn.transaction()
transaction.create('/nodes', ''.encode('ascii'))
transaction.create('/primary_node', 'none'.encode('ascii'))
transaction.create('/domains', ''.encode('ascii'))
transaction.create('/networks', ''.encode('ascii'))
transaction.create('/ceph', ''.encode('ascii'))
transaction.create('/ceph/osds', ''.encode('ascii'))
transaction.create('/ceph/pools', ''.encode('ascii'))
2019-06-19 00:23:14 -04:00
transaction.create('/ceph/volumes', ''.encode('ascii'))
transaction.create('/ceph/snapshots', ''.encode('ascii'))
transaction.create('/locks', ''.encode('ascii'))
transaction.create('/locks/flush_lock', 'False'.encode('ascii'))
2019-03-10 20:40:03 -04:00
transaction.commit()
# Close the Zookeeper connection
pvc_common.stopZKConnection(zk_conn)
click.echo('Successfully initialized new cluster. Any running PVC daemons will need to be restarted.')