Joshua Boniface
58789f1db4
This helps parallelize the numerous Zookeeper calls a little bit, at least within the bounds of the GIL, to improve performance when getting a large list of VMs. The max_workers value is capped at 32 to avoid causing too many threads during concurrent executions, but still provides a noticeable speedup (on the order of 0.2-0.4 seconds with 75 VMs, scaling up further as counts grow). |
||
---|---|---|
.github/workflows | ||
.hooks | ||
api-daemon | ||
client-cli | ||
daemon-common | ||
debian | ||
docs | ||
node-daemon | ||
.file-header | ||
.gitignore | ||
.gitlab-ci.yml | ||
.version | ||
LICENSE | ||
README.md | ||
build-and-deploy.sh | ||
build-deb.sh | ||
build-unstable-deb.sh | ||
bump-version | ||
gen-api-doc | ||
gen-api-migrations | ||
gen-zk-migrations | ||
lint | ||
mkdocs.yml | ||
test-cluster.sh |
README.md
What is PVC?
PVC is a virtual machine-based hyperconverged infrastructure (HCI) virtualization cluster solution that is fully Free Software, scalable, redundant, self-healing, self-managing, and designed for administrator simplicity. It is an alternative to other HCI solutions such as Harvester, Nutanix, and VMWare, as well as to other common virtualization stacks such as ProxMox and OpenStack.
PVC is a complete HCI solution, built from well-known and well-trusted Free Software tools, to assist an administrator in creating and managing a cluster of servers to run virtual machines, as well as self-managing several important aspects including storage failover, node failure and recovery, virtual machine failure and recovery, and network plumbing. It is designed to act consistently, reliably, and unobtrusively, letting the administrator concentrate on more important things.
PVC is highly scalable. From a minimum (production) node count of 3, up to 12 or more, and supporting many dozens of VMs, PVC scales along with your workload and requirements. Deploy a cluster once and grow it as your needs expand.
As a consequence of its features, PVC makes administrating very high-uptime VMs extremely easy, featuring VM live migration, built-in always-enabled shared storage with transparent multi-node replication, and consistent network plumbing throughout the cluster. Nodes can also be seamlessly removed from or added to service, with zero VM downtime, to facilitate maintenance, upgrades, or other work.
PVC also features an optional, fully customizable VM provisioning framework, designed to automate and simplify VM deployments using custom provisioning profiles, scripts, and CloudInit userdata API support.
Installation of PVC is accomplished by two main components: a Node installer ISO which creates on-demand installer ISOs, and an Ansible role framework to configure, bootstrap, and administrate the nodes. Once up, the cluster is managed via an HTTP REST API, accessible via a Python Click CLI client or WebUI.
Just give it physical servers, and it will run your VMs without you having to think about it, all in just an hour or two of setup time.
What is it based on?
The core node and API daemons, as well as the CLI API client, are written in Python 3 and are fully Free Software (GNU GPL v3). In addition to these, PVC makes use of the following software tools to provide a holistic hyperconverged infrastructure solution:
- Debian GNU/Linux as the base OS.
- Linux KVM, QEMU, and Libvirt for VM management.
- Linux
ip
, FRRouting, NFTables, DNSMasq, and PowerDNS for network management. - Ceph for storage management.
- Apache Zookeeper for the primary cluster state database.
- Patroni PostgreSQL manager for the secondary relation databases (DNS aggregation, Provisioner configuration).
Getting Started
To get started with PVC, please see the About page for general information about the project, and the Getting Started page for details on configuring your first cluster.
Changelog
v0.9.21
- [API Daemon] Ensures VMs stop before removing them
- [Node Daemon] Fixes a bug with VM shutdowns not timing out
- [Documentation] Adds information about georedundancy caveats
- [All] Adds support for SR-IOV NICs (hostdev and macvtap) and surrounding documentation
- [Node Daemon] Fixes a bug where shutdown aborted migrations unexpectedly
- [Node Daemon] Fixes a bug where the migration method was not updated realtime
- [Node Daemon] Adjusts the Patroni commands to remove reference to Zookeeper path
- [CLI Client] Adjusts several help messages and fixes some typos
- [CLI Client] Converts the CLI client to a proper Python module
- [API Daemon] Improves VM list performance
- [API Daemon] Adjusts VM list matching critera (only matches against the UUID if it's a full UUID)
- [API Worker] Fixes incompatibility between Deb 10 and 11 in launching Celery worker
- [API Daemon] Corrects several bugs with initialization command
- [Documentation] Adds a shiny new logo and revamps introduction text
v0.9.20
- [Daemons] Implemented a Zookeeper schema handler and version 0 schema
- [Daemons] Completes major refactoring of codebase to make use of the schema handler
- [Daemons] Adds support for dynamic chema changges and "hot reloading" of pvcnoded processes
- [Daemons] Adds a functional testing script for verifying operation against a test cluster
- [Daemons, CLI] Fixes several minor bugs found by the above script
- [Daemons, CLI] Add support for Debian 11 "Bullseye"
v0.9.19
- [CLI] Corrects some flawed conditionals
- [API] Disables SQLAlchemy modification tracking functionality (not used by us)
- [Daemons] Implements new zkhandler module for improved reliability and reusability
- [Daemons] Refactors some code to use new zkhandler module
- [API, CLI] Adds support for "none" migration selector (uses cluster default instead)
- [Daemons] Moves some configuration keys to new /config tree
- [Node Daemon] Increases initial lock timeout for VM migrations to avoid out-of-sync potential
- [Provisioner] Support storing and using textual cluster network labels ("upstream", "storage", "cluster") in templates
- [API] Avoid duplicating existing node states
v0.9.18
- Adds VM rename functionality to API and CLI client
v0.9.17
- [CLI] Fixes bugs in log follow output
v0.9.16
- Improves some CLI help messages
- Skips empty local cluster in CLI
- Adjusts how confirmations happen during VM modify restarts
- Fixes bug around corrupted VM log files
- Fixes bug around subprocess pipe exceptions
v0.9.15
- [CLI] Adds additional verification (--yes) to several VM management commands
- [CLI] Adds a method to override --yes/confirmation requirements via envvar (PVC_UNSAFE)
- [CLI] Adds description fields to PVC clusters in CLI
v0.9.14
- Fixes bugs around cloned volume provisioning
- Fixes some minor visual bugs
- Minor license update (from GPL3+ to GPL3)
- Adds qemu-guest-agent support to provisioner-created VMs by default
v0.9.13
- Adds nicer startup messages for daemons
- Adds additional API field for stored_bytes to pool stats
- Fixes sorting issues with snapshot lists
- Fixes missing increment/decrement of snapshot_count on volumes
- Fixes bad calls in pool element API endpoints
- Fixes inconsistent bytes_tohuman behaviour in daemons
- Adds validation and maximum volume size on creation (must be smaller than the pool free space)
v0.9.12
- Fixes a bug in the pvcnoded service unit file causing a Zookeeper startup race condition
v0.9.11
- Documentation updates
- Adds VNC information to VM info
- Goes back to external Ceph commands for disk usage
v0.9.10
- Moves OSD stats uploading to primary, eliminating reporting failures while hosts are down
- Documentation updates
- Significantly improves RBD locking behaviour in several situations, eliminating cold-cluster start issues and failed VM boot-ups after crashes
- Fixes some timeout delays with fencing
- Fixes bug in validating YAML provisioner userdata
v0.9.9
- Adds documentation updates
- Removes single-element list stripping and fixes surrounding bugs
- Adds additional fields to some API endpoints for ease of parsing by clients
- Fixes bugs with network configuration
v0.9.8
- Adds support for cluster backup/restore
- Moves location of
init
command in CLI to make room for the above - Cleans up some invalid help messages from the API
v0.9.7
- Fixes bug with provisioner system template modifications
v0.9.6
- Fixes bug with migrations
v0.9.5
- Fixes bug with line count in log follow
- Fixes bug with disk stat output being None
- Adds short pretty health output
- Documentation updates
v0.9.4
- Fixes major bug in OVA parser
v0.9.3
- Fixes bugs with image & OVA upload parsing
v0.9.2
- Major linting of the codebase with flake8; adds linting tools
- Implements CLI-based modification of VM vCPUs, memory, networks, and disks without directly editing XML
- Fixes bug where
pvc vm log -f
would show all 1000 lines before starting - Fixes bug in default provisioner libvirt schema (
drive
->driver
typo)
v0.9.1
- Added per-VM migration method feature
- Fixed bug with provisioner system template listing
v0.9.0
Numerous small improvements and bugfixes. This release is suitable for general use and is pre-release-quality software.
This release introduces an updated version scheme; all future stable releases until 1.0.0 is ready will be made under this 0.9.z naming. This does not represent semantic versioning and all changes (feature, improvement, or bugfix) will be considered for inclusion in this release train.
v0.8
Numerous improvements and bugfixes. This release is suitable for general use and is pre-release-quality software.
v0.7
Numerous improvements and bugfixes, revamped documentation. This release is suitable for general use and is beta-quality software.
v0.6
Numerous improvements and bugfixes, full implementation of the provisioner, full implementation of the API CLI client (versus direct CLI client). This release is suitable for general use and is beta-quality software.
v0.5
First public release; fully implements the VM, network, and storage managers, the HTTP API, and the pvc-ansible framework for deploying and bootstrapping a cluster. This release is suitable for general use, though it is still alpha-quality software and should be expected to change significantly until 1.0 is released.
v0.4
Full implementation of virtual management and virtual networking functionality. Partial implementation of storage functionality.
v0.3
Basic implementation of virtual management functionality.