From de8c79755b0c98b06756e54cfc4c4403c297419b Mon Sep 17 00:00:00 2001 From: "Joshua M. Boniface" Date: Thu, 19 Dec 2019 23:44:10 -0500 Subject: [PATCH] Add documentation about the provisioner API --- docs/manuals/api.md | 143 +++++++++++++++++++++++++++++++++++++++++++- 1 file changed, 140 insertions(+), 3 deletions(-) diff --git a/docs/manuals/api.md b/docs/manuals/api.md index 4e4b1fdb..b3bf02c5 100644 --- a/docs/manuals/api.md +++ b/docs/manuals/api.md @@ -1,6 +1,6 @@ # PVC HTTP API manual -The PVC HTTP API client is built with Flask, a Python framework for creating API interfaces, and run directly with the PyWSGI framework. It interfaces directly with the Zookeeper cluster to send and receive information about the cluster. It supports authentication configured statically via tokens in the configuration file as well as SSL. +The PVC HTTP API client is built with Flask, a Python framework for creating API interfaces, and run directly with the PyWSGI framework. It interfaces directly with the Zookeeper cluster to send and receive information about the cluster. It supports authentication configured statically via tokens in the configuration file as well as SSL. It also includes the provisioner client, an optional section that can be used to create VMs automatically using a set of templates and standardized scripts. The [`pvc-ansible`](https://github.com/parallelvirtualcluster/pvc-ansible) framework will install and configure the API by default, and enable the node daemon option for an instance of the API to follow the primary node, thus ensuring the API is listening on the upstream floating IP at all times. @@ -30,7 +30,128 @@ The PVC API consistently accepts HTTP POST commands of HTML form documents. Howe The PCI API consistently returns JSON bodies as its responses, with the one exception of the `vm dump` endpoint which returns an XML body. For all POST endpoints, unless otherwise specified below, this is a `message` value containing a human-readable message about the success or failure of the command. The HTTP return code is always 200 for a success or 510 for a failure. For all GET endpoints except the mentioned `vm dump`, this is a JSON body containing the requested data. -## Configuration +## Provisioner + +The provisioner subsection (`/api/v1/provisioner`) is used to create new virtual machines on a PVC cluster. By creating templates and scripts, then grouping these into profiles, VMs can be created based on dynamic, declarative configurations via direct installation or templating. Administrators can use this facility to automate the creation VMs running most *NIX instances that can be installed in a parent host, or by using templates as a base for new VMs. It can also create VMs based on existing templates or ISO images to facilitate installing alternate operating systems such as Microsoft Windows. + +### Templates + +Templates are used to configure the four components that define a VM configuration. Templates can be created and managed via the API, then grouped into profiles. + +#### System Templates + +System templates define the basic configuration of a VM. This includes the number of vCPUs and amount vRAM, as well as console access (either VNC or serial) and several pieces of PVC metadata. + +Generally, a system template is usable across multiple VM profiles, so there will generally be a small number of system templates defining several standard resource profiles that can then be reused. + +Some elements of the system template are mandatory, but most are optional. + +###### Example: Creating a system template + +* Note: vRAM sizes are always specified in MB. + +``` +curl -X POST http://localhost:7370/api/v1/provisioner/template/system?name=2cpu-1gb-serial\&vcpus=2\&vram=1024\&serial=true\&vnc=false\&node_limit='pvchv1,pvchv2'\&node_selector=mem\&start_with_node=false +curl -X GET http://localhost:7370/api/v1/provisioner/template/system/2cpu-1gb-serial +``` + +#### Network Templates + +Network templates define the network configuration of a VM. These are tied into the PVC networking facility, and are quite simple. A MAC template is assigned to each template, which defines how MAC addresses are generated (either randomly, or via a simple templating system for static MAC addresses). + +With a network template, various "nets" can be configured. A "net" defines a PVC virtual network VNI, which must be valid on the PVC cluster. The first net is assigned to the first Ethernet device (usually eth0 or ens2 in Linux), with each subsequent network being added as an additional interface in order. + +###### Example: Creating a network template with two networks + +``` +curl -X POST http://localhost:7370/api/v1/provisioner/template/network?name=net200+net300 +curl -X POST http://localhost:7370/api/v1/provisioner/template/network/net200+net300/net?vni=200 +curl -X POST http://localhost:7370/api/v1/provisioner/template/network/net200+net300/net/300 +curl -X GET http://localhost:7370/api/v1/provisioner/template/net200+net300 + +#### Storage Templates + +Storage templates define the Ceph RBD disks, as well as optional filesystems and mountpoints for Linux-based guests, of a VM. The template itself consists only of a name; disk or image entries are configured as additional elements similar to network templates. + +Each disk in a storage template is identified by a sequential ID, usually "sda"/"vda", "sdb"/"vdb", etc., a size, and a Ceph RBD pool within the PVC cluster. These alone are all that are required, and will create raw, unformatted images of the specified size, on the specified pool, and attached to the VM at the ID value. In addition to these basics, filesystems (with argument support) and mountpoints can also be specified. Filesystems specified here will be used to format the volume during the provisioning process, and mountpoints will mount the volume at the specified mountpoint during provisioning, so that a guest operating system can be installed on them during the process with a provisioning script. + +In addition to disks, storage templates can also contain image entries. Like disk entries, they are identified by a sequential ID, as well as a source Ceph RBD pool and volume name. The specified volume may belong to a (shutdown) VM or be a dedicated template uploaded to the Ceph cluster. + +###### Example: Creating a storage template with three mounted disks + +* Note: You can also include the template name during creation. +* Note: Disk sizes are always specified in GB. +* Note: Filesystem arguments are passed as-is to the `mkfs` command and must use an `--opt=val` format to prevent splitting. + +``` +curl -X POST http://localhost:7370/api/v1/provisioner/template/storage/ext4-root-var-log +curl -X POST http://localhost:7370/api/v1/provisioner/template/storage/ext4-root-var-log/disk?disk_id=sda\&disk_size=4\&filesystem=ext4\&mountpoint=/\&pool=vms\&filesystem_arg='-L=root' +curl -X POST http://localhost:7370/api/v1/provisioner/template/storage/ext4-root-var-log/disk/sdb?disk_size=4\&filesystem=ext4\&mountpoint=/var\&pool=vms\&filesystem_arg='-L=var' +curl -X POST http://localhost:7370/api/v1/provisioner/template/storage/ext4-root-var-log/disk/sdc -d "disk_size=4\&filesystem=ext4\&mountpoint=/var/log\&pool=vms\&filesystem_arg='-L=log'\&filesystem_arg='-m=1'" +curl -X GET http://localhost:7370/api/v1/provisioner/template/storage/ext4-root-var-log +``` + +#### Userdata Templates + +Userdata templates contain cloud-init metadata that can be provided to VMs on their first boot. It is accessible via an EC2-compatible API running on the PVC cluster to VMs. A userdata template contains the full text of the userdata, including optional multi-part sections if desired. + +A default userdata template called "empty" is created by default, and this can be used for any profile which does not require cloud-init userdata, since a template must always be specified. + +Examples of userdata templates can be found in `/usr/share/pvc/provisioner/examples` when the API is installed. + +###### Example: Creating a userdata template from the `userdata.yaml` example file + +* Note: For the block text commands (userdata and scripts), using the HTTP POST body for the data is always better than a URL argument. + +``` +curl -X POST http://localhost:7370/api/v1/provisioner/template/userdata?name=example-userdata -d "data=$( cat /usr/share/pvc/provisioner/examples/userdata.yaml )" +curl -X GET http://localhost:7370/api/v1/provisioner/template/userdata?name=example-userdata +``` + +### Scripts + +Scripts automate the installation of VMs with Python. To make use of a script, at least one disk volume must be both formatted with a Linux-compatible filesyste, and have a mountpoint (very likely `/`) configured. The specified disk is then mounted in a temporary directory on the active coordinator, and the script run against it. This script can then do any task required to set up and configure the VM, such as installing a Debian or Ubuntu system with debootstrap, obtaining a chroot and configuring GRUB, or almost any other task that the administrator may wish. All scripts are written in Python 3, which is then integrated into the provisioner's worker during VM creation and executed at the appropriate point. + +Each script must contain a function called `install()` which accepts `**kwargs` and no other arguments. A number of default arguments are provided, including `vm_name`, the `temporary_directory`, and dictionaries of the `disks` and `networks`. Additional arguments can be specified in VM profiles to facilitate advanced configurations specific to particular VM types. + +Examples of scripts can be found in `/usr/share/pvc/provisioner/examples` when the API is installed. + +###### Example: Creating a script from the `debootstrap_script.py` example file + +* Note: For the block text commands (userdata and scripts), using the HTTP POST body for the data is always better than a URL argument. + +``` +curl -X POST http://localhost:7370/api/v1/provisioner/script/debootstrap-example -d "data=$( cat /usr/share/pvc/provisioner/examples/userdata.yaml )" +curl -X GET http://localhost:7370/api/v1/provisioner/script/debootstrap-example +``` + +### Profiles + +Profiles group together the four template types and scripts, as well as optional script arguments, into a named profile which can be assigned to VMs on creation. When creating a VM, templates and scripts themselves are not explicitly specified; rather a profile is specified which then maps to these other values. This allows maximum flexibility, allowing a VM profile to combine the various templates and scripts in an arbitrary way. One potential usecase is to create a profile for a particular VM role, for instance a webserver, which will have a specific system, disk, network, and userdata configuration; multiple VMs can then be created with this profile to ensure they all contain the same resources and configuration. + +###### Example: Creating a profile with the previously-created templates and some script arguments + +* Note: Script arguments are specified as `name=value` pairs after the `arg=` argument. + +``` +curl -X POST http://localhost:7370/api/v1/provisioner/profile/test-profile?system_template=2cpu-1gb-serial\&network_template=net200+net300\&disk_template=ext4-root-var-log\&userdata_template=example-userdata\&script=debootstrap-example\&arg=deb_release=buster\&arg=deb_mirror=http://deb.debian.org/debian\&arg=deb_packages=linux-image-amd64,grub-pc,cloud-init,python3-cffi-backend,wget +curl -X GET http://localhost:7370/api/v1/provisioner/profile/test-profile +``` + +### Creating VMs + +VMs are created by specifying a name and a profile value. The provisioner API will then collect the details of the profile, and trigger the Celery worker (`pvc-provisioner-worker.service`) to begin creating the VM. The administrator can, at any point, obtain the status of the process via the Task ID, which is returned in the JSON body of the creation command. Once completed, by default, the resulting VM will be defined and started on the cluster, ready to use. If the VM uses cloud-init, it will then hit the Metadata API on startup to obtain the details of the VM as well as the userdata specified in the profile. + +Additional options can also be specified at install time. Automatic definition of the VM and automatic startup of the VM can both be disabled via options to the creation command. The former is most useful when creating disk images from an installed set of VM disks, and the latter provides flexibility for the administrator to edit or review the final VM before starting it for the first time. + +###### Example: Creating a VM and viewing its status + +``` +curl -X POST http://localhost:7370/api/v1/provisioner/create?name=test1\&profile=test-profile +curl -X GET http://localhost:7370/api/v1/provisioner/status/ +``` + +## API Daemon Configuration The API is configured using a YAML configuration file which is passed in to the API process by the environment variable `PVC_CONFIG_FILE`. When running with the default package and SystemD unit, this file is located at `/etc/pvc/pvc-api.yaml`. @@ -135,7 +256,7 @@ The path to the SSL certificate file for the API to use. The path to the SSL private key file for the API to use. -## API endpoint documentation +## Primary Client API endpoint documentation ### General endpoints @@ -714,3 +835,19 @@ Return a JSON document containing information about Ceph RBD volume snapshot `` of Ceph RBD volume `` on Ceph RBD pool ``. + +## Provisioner API endpoint documentation + +### Node endpoints + +These endpoints manage PVC node state and operation. + +#### `/api/v1/node` + * Methods: `GET` + +###### `GET` + * Mandatory values: N/A + * Optional values: `limit` + +Return a JSON document containing information about all cluster nodes. If `limit` is specified, return a JSON document containing information about cluster nodes with names matching `limit` as fuzzy regex. +