Implement new provisioner setup

This commit is contained in:
2022-10-05 16:03:05 -04:00
parent f1df1cfe93
commit 4df70cf086
8 changed files with 2363 additions and 1166 deletions

View File

@ -29,6 +29,7 @@ from pvcapid.Daemon import config, strtobool, API_VERSION
import pvcapid.helper as api_helper
import pvcapid.provisioner as api_provisioner
import pvcapid.vmbuilder as api_vmbuilder
import pvcapid.benchmark as api_benchmark
import pvcapid.ova as api_ova
@ -144,7 +145,7 @@ def Authenticator(function):
def create_vm(
self, vm_name, profile_name, define_vm=True, start_vm=True, script_run_args=[]
):
return api_provisioner.create_vm(
return api_vmbuilder.create_vm(
self,
vm_name,
profile_name,

View File

@ -19,23 +19,12 @@
#
###############################################################################
import json
import psycopg2
import psycopg2.extras
import re
from pvcapid.Daemon import config, strtobool
from daemon_lib.zkhandler import ZKHandler
import daemon_lib.common as pvc_common
import daemon_lib.node as pvc_node
import daemon_lib.vm as pvc_vm
import daemon_lib.network as pvc_network
import daemon_lib.ceph as pvc_ceph
import pvcapid.libvirt_schema as libvirt_schema
from pvcapid.ova import list_ova
@ -1229,866 +1218,3 @@ def delete_profile(name):
retcode = 400
close_database(conn, cur)
return retmsg, retcode
#
# Main VM provisioning function - executed by the Celery worker
#
def create_vm(
self, vm_name, vm_profile, define_vm=True, start_vm=True, script_run_args=[]
):
# Runtime imports
import time
import importlib
import uuid
import datetime
import random
temp_dir = None
time.sleep(2)
print(
"Starting provisioning of VM '{}' with profile '{}'".format(vm_name, vm_profile)
)
# Phase 0 - connect to databases
try:
db_conn, db_cur = open_database(config)
except Exception:
raise ClusterError("Failed to connect to Postgres")
try:
zkhandler = ZKHandler(config)
zkhandler.connect()
except Exception:
raise ClusterError("Failed to connect to Zookeeper")
# Phase 1 - setup
# * Get the profile elements
# * Get the details from these elements
# * Assemble a VM configuration dictionary
self.update_state(
state="RUNNING",
meta={"current": 1, "total": 10, "status": "Collecting configuration"},
)
time.sleep(1)
vm_id = re.findall(r"/(\d+)$/", vm_name)
if not vm_id:
vm_id = 0
else:
vm_id = vm_id[0]
vm_data = dict()
# Get the profile information
query = "SELECT * FROM profile WHERE name = %s"
args = (vm_profile,)
db_cur.execute(query, args)
profile_data = db_cur.fetchone()
if profile_data.get("arguments"):
vm_data["script_arguments"] = profile_data.get("arguments").split("|")
else:
vm_data["script_arguments"] = []
if profile_data.get("profile_type") == "ova":
is_ova_install = True
is_script_install = False # By definition
else:
is_ova_install = False
# Get the system details
query = "SELECT * FROM system_template WHERE id = %s"
args = (profile_data["system_template"],)
db_cur.execute(query, args)
vm_data["system_details"] = db_cur.fetchone()
# Get the MAC template
query = "SELECT mac_template FROM network_template WHERE id = %s"
args = (profile_data["network_template"],)
db_cur.execute(query, args)
db_row = db_cur.fetchone()
if db_row:
vm_data["mac_template"] = db_row.get("mac_template")
else:
vm_data["mac_template"] = None
# Get the networks
query = "SELECT * FROM network WHERE network_template = %s"
args = (profile_data["network_template"],)
db_cur.execute(query, args)
vm_data["networks"] = db_cur.fetchall()
# Get the storage volumes
# ORDER BY ensures disks are always in the sdX/vdX order, regardless of add order
query = "SELECT * FROM storage WHERE storage_template = %s ORDER BY disk_id"
args = (profile_data["storage_template"],)
db_cur.execute(query, args)
vm_data["volumes"] = db_cur.fetchall()
# Get the script
query = "SELECT script FROM script WHERE id = %s"
args = (profile_data["script"],)
db_cur.execute(query, args)
db_row = db_cur.fetchone()
if db_row:
vm_data["script"] = db_row.get("script")
else:
vm_data["script"] = None
if vm_data["script"] and not is_ova_install:
is_script_install = True
else:
is_script_install = False
# Get the OVA details
if is_ova_install:
query = "SELECT * FROM ova WHERE id = %s"
args = (profile_data["ova"],)
db_cur.execute(query, args)
vm_data["ova_details"] = db_cur.fetchone()
query = "SELECT * FROM ova_volume WHERE ova = %s"
args = (profile_data["ova"],)
db_cur.execute(query, args)
vm_data["volumes"] = db_cur.fetchall()
close_database(db_conn, db_cur)
print(
"VM configuration data:\n{}".format(
json.dumps(vm_data, sort_keys=True, indent=2)
)
)
# Phase 2 - verification
# * Ensure that at least one node has enough free RAM to hold the VM (becomes main host)
# * Ensure that all networks are valid
# * Ensure that there is enough disk space in the Ceph cluster for the disks
# This is the "safe fail" step when an invalid configuration will be caught
self.update_state(
state="RUNNING",
meta={
"current": 2,
"total": 10,
"status": "Verifying configuration against cluster",
},
)
time.sleep(1)
# Verify that a VM with this name does not already exist
if pvc_vm.searchClusterByName(zkhandler, vm_name):
raise ClusterError(
"A VM with the name '{}' already exists in the cluster.".format(vm_name)
)
# Verify that at least one host has enough free RAM to run the VM
_discard, nodes = pvc_node.get_list(zkhandler, None)
target_node = None
last_free = 0
for node in nodes:
# Skip the node if it is not ready to run VMs
if node["daemon_state"] != "run" or node["domain_state"] != "ready":
continue
# Skip the node if its free memory is less than the new VM's size, plus a 512MB buffer
if node["memory"]["free"] < (vm_data["system_details"]["vram_mb"] + 512):
continue
# If this node has the most free, use it
if node["memory"]["free"] > last_free:
last_free = node["memory"]["free"]
target_node = node["name"]
# Raise if no node was found
if not target_node:
raise ClusterError(
"No ready cluster node contains at least {}+512 MB of free RAM.".format(
vm_data["system_details"]["vram_mb"]
)
)
print(
'Selecting target node "{}" with "{}" MB free RAM'.format(
target_node, last_free
)
)
# Verify that all configured networks are present on the cluster
cluster_networks, _discard = pvc_network.getClusterNetworkList(zkhandler)
for network in vm_data["networks"]:
vni = str(network["vni"])
if vni not in cluster_networks and vni not in [
"upstream",
"cluster",
"storage",
]:
raise ClusterError(
'The network VNI "{}" is not present on the cluster.'.format(vni)
)
print("All configured networks for VM are valid")
# Verify that there is enough disk space free to provision all VM disks
pools = dict()
for volume in vm_data["volumes"]:
if volume.get("source_volume") is not None:
volume_data = pvc_ceph.getVolumeInformation(
zkhandler, volume["pool"], volume["source_volume"]
)
if not volume_data:
raise ClusterError(
"The source volume {}/{} could not be found.".format(
volume["pool"], volume["source_volume"]
)
)
if not volume["pool"] in pools:
pools[volume["pool"]] = int(
pvc_ceph.format_bytes_fromhuman(volume_data["stats"]["size"])
/ 1024
/ 1024
/ 1024
)
else:
pools[volume["pool"]] += int(
pvc_ceph.format_bytes_fromhuman(volume_data["stats"]["size"])
/ 1024
/ 1024
/ 1024
)
else:
if not volume["pool"] in pools:
pools[volume["pool"]] = volume["disk_size_gb"]
else:
pools[volume["pool"]] += volume["disk_size_gb"]
for pool in pools:
try:
pool_information = pvc_ceph.getPoolInformation(zkhandler, pool)
if not pool_information:
raise
except Exception:
raise ClusterError('Pool "{}" is not present on the cluster.'.format(pool))
pool_free_space_gb = int(
pool_information["stats"]["free_bytes"] / 1024 / 1024 / 1024
)
pool_vm_usage_gb = int(pools[pool])
if pool_vm_usage_gb >= pool_free_space_gb:
raise ClusterError(
'Pool "{}" has only {} GB free and VM requires {} GB.'.format(
pool, pool_free_space_gb, pool_vm_usage_gb
)
)
print("There is enough space on cluster to store VM volumes")
if not is_ova_install:
# Verify that every specified filesystem is valid
used_filesystems = list()
for volume in vm_data["volumes"]:
if volume["source_volume"] is not None:
continue
if volume["filesystem"] and volume["filesystem"] not in used_filesystems:
used_filesystems.append(volume["filesystem"])
for filesystem in used_filesystems:
if filesystem == "swap":
retcode, stdout, stderr = pvc_common.run_os_command("which mkswap")
if retcode:
raise ProvisioningError(
"Failed to find binary for mkswap: {}".format(stderr)
)
else:
retcode, stdout, stderr = pvc_common.run_os_command(
"which mkfs.{}".format(filesystem)
)
if retcode:
raise ProvisioningError(
"Failed to find binary for mkfs.{}: {}".format(
filesystem, stderr
)
)
print("All selected filesystems are valid")
# Phase 3 - provisioning script preparation
# * Import the provisioning script as a library with importlib
# * Ensure the required function(s) are present
self.update_state(
state="RUNNING",
meta={"current": 3, "total": 10, "status": "Preparing provisioning script"},
)
time.sleep(1)
if is_script_install:
# Write the script out to a temporary file
retcode, stdout, stderr = pvc_common.run_os_command("mktemp")
if retcode:
raise ProvisioningError(
"Failed to create a temporary file: {}".format(stderr)
)
script_file = stdout.strip()
with open(script_file, "w") as fh:
fh.write(vm_data["script"])
fh.write("\n")
# Import the script file
loader = importlib.machinery.SourceFileLoader("installer_script", script_file)
spec = importlib.util.spec_from_loader(loader.name, loader)
installer_script = importlib.util.module_from_spec(spec)
loader.exec_module(installer_script)
# Verify that the install() function is valid
if "install" not in dir(installer_script):
raise ProvisioningError(
"Specified script does not contain an install() function."
)
print("Provisioning script imported successfully")
# Phase 4 - configuration creation
# * Create the libvirt XML configuration
self.update_state(
state="RUNNING",
meta={
"current": 4,
"total": 10,
"status": "Preparing Libvirt XML configuration",
},
)
time.sleep(1)
print("Creating Libvirt configuration")
# Get information about VM
vm_uuid = uuid.uuid4()
vm_description = "PVC provisioner @ {}, profile '{}'".format(
datetime.datetime.now(), vm_profile
)
retcode, stdout, stderr = pvc_common.run_os_command("uname -m")
system_architecture = stdout.strip()
# Begin assembling libvirt schema
vm_schema = ""
vm_schema += libvirt_schema.libvirt_header.format(
vm_name=vm_name,
vm_uuid=vm_uuid,
vm_description=vm_description,
vm_memory=vm_data["system_details"]["vram_mb"],
vm_vcpus=vm_data["system_details"]["vcpu_count"],
vm_architecture=system_architecture,
)
# Add disk devices
monitor_list = list()
coordinator_names = config["storage_hosts"]
for coordinator in coordinator_names:
monitor_list.append("{}.{}".format(coordinator, config["storage_domain"]))
ceph_storage_secret = config["ceph_storage_secret_uuid"]
for volume in vm_data["volumes"]:
vm_schema += libvirt_schema.devices_disk_header.format(
ceph_storage_secret=ceph_storage_secret,
disk_pool=volume["pool"],
vm_name=vm_name,
disk_id=volume["disk_id"],
)
for monitor in monitor_list:
vm_schema += libvirt_schema.devices_disk_coordinator.format(
coordinator_name=monitor,
coordinator_ceph_mon_port=config["ceph_monitor_port"],
)
vm_schema += libvirt_schema.devices_disk_footer
vm_schema += libvirt_schema.devices_vhostmd
# Add network devices
network_id = 0
for network in vm_data["networks"]:
vni = network["vni"]
if vni in ["upstream", "cluster", "storage"]:
eth_bridge = "br{}".format(vni)
else:
eth_bridge = "vmbr{}".format(vni)
vm_id_hex = "{:x}".format(int(vm_id % 16))
net_id_hex = "{:x}".format(int(network_id % 16))
if vm_data.get("mac_template") is not None:
mac_prefix = "52:54:01"
macgen_template = vm_data["mac_template"]
eth_macaddr = macgen_template.format(
prefix=mac_prefix,
vmid=vm_id_hex,
netid=net_id_hex,
)
else:
mac_prefix = "52:54:00"
random_octet_A = "{:x}".format(random.randint(16, 238))
random_octet_B = "{:x}".format(random.randint(16, 238))
random_octet_C = "{:x}".format(random.randint(16, 238))
macgen_template = "{prefix}:{octetA}:{octetB}:{octetC}"
eth_macaddr = macgen_template.format(
prefix=mac_prefix,
octetA=random_octet_A,
octetB=random_octet_B,
octetC=random_octet_C,
)
vm_schema += libvirt_schema.devices_net_interface.format(
eth_macaddr=eth_macaddr, eth_bridge=eth_bridge
)
network_id += 1
# Add default devices
vm_schema += libvirt_schema.devices_default
# Add serial device
if vm_data["system_details"]["serial"]:
vm_schema += libvirt_schema.devices_serial.format(vm_name=vm_name)
# Add VNC device
if vm_data["system_details"]["vnc"]:
if vm_data["system_details"]["vnc_bind"]:
vm_vnc_bind = vm_data["system_details"]["vnc_bind"]
else:
vm_vnc_bind = "127.0.0.1"
vm_vncport = 5900
vm_vnc_autoport = "yes"
vm_schema += libvirt_schema.devices_vnc.format(
vm_vncport=vm_vncport,
vm_vnc_autoport=vm_vnc_autoport,
vm_vnc_bind=vm_vnc_bind,
)
# Add SCSI controller
vm_schema += libvirt_schema.devices_scsi_controller
# Add footer
vm_schema += libvirt_schema.libvirt_footer
print("Final VM schema:\n{}\n".format(vm_schema))
# All the following steps may require cleanup later on, so catch them here and do cleanup in a Finally block
try:
# Phase 5 - definition
# * Create the VM in the PVC cluster
self.update_state(
state="RUNNING",
meta={"current": 5, "total": 10, "status": "Defining VM on the cluster"},
)
time.sleep(1)
if define_vm:
print("Defining VM on cluster")
node_limit = vm_data["system_details"]["node_limit"]
if node_limit:
node_limit = node_limit.split(",")
node_selector = vm_data["system_details"]["node_selector"]
node_autostart = vm_data["system_details"]["node_autostart"]
migration_method = vm_data["system_details"]["migration_method"]
retcode, retmsg = pvc_vm.define_vm(
zkhandler,
vm_schema.strip(),
target_node,
node_limit,
node_selector,
node_autostart,
migration_method,
vm_profile,
initial_state="provision",
)
print(retmsg)
else:
print("Skipping VM definition")
# Phase 6 - disk creation
# * Create each Ceph storage volume for the disks
self.update_state(
state="RUNNING",
meta={"current": 6, "total": 10, "status": "Creating storage volumes"},
)
time.sleep(1)
for volume in vm_data["volumes"]:
if volume.get("source_volume") is not None:
success, message = pvc_ceph.clone_volume(
zkhandler,
volume["pool"],
volume["source_volume"],
"{}_{}".format(vm_name, volume["disk_id"]),
)
print(message)
if not success:
raise ProvisioningError(
'Failed to clone volume "{}" to "{}".'.format(
volume["source_volume"], volume["disk_id"]
)
)
else:
success, message = pvc_ceph.add_volume(
zkhandler,
volume["pool"],
"{}_{}".format(vm_name, volume["disk_id"]),
"{}G".format(volume["disk_size_gb"]),
)
print(message)
if not success:
raise ProvisioningError(
'Failed to create volume "{}".'.format(volume["disk_id"])
)
# Phase 7 - disk mapping
# * Map each volume to the local host in order
# * Format each volume with any specified filesystems
# * If any mountpoints are specified, create a temporary mount directory
# * Mount any volumes to their respective mountpoints
self.update_state(
state="RUNNING",
meta={
"current": 7,
"total": 10,
"status": "Mapping, formatting, and mounting storage volumes locally",
},
)
time.sleep(1)
for volume in vm_data["volumes"]:
dst_volume_name = "{}_{}".format(vm_name, volume["disk_id"])
dst_volume = "{}/{}".format(volume["pool"], dst_volume_name)
if is_ova_install:
src_volume_name = volume["volume_name"]
src_volume = "{}/{}".format(volume["pool"], src_volume_name)
print(
"Converting {} source volume {} to raw format on {}".format(
volume["volume_format"], src_volume, dst_volume
)
)
# Map the target RBD device
retcode, retmsg = pvc_ceph.map_volume(
zkhandler, volume["pool"], dst_volume_name
)
if not retcode:
raise ProvisioningError(
'Failed to map destination volume "{}": {}'.format(
dst_volume_name, retmsg
)
)
# Map the source RBD device
retcode, retmsg = pvc_ceph.map_volume(
zkhandler, volume["pool"], src_volume_name
)
if not retcode:
raise ProvisioningError(
'Failed to map source volume "{}": {}'.format(
src_volume_name, retmsg
)
)
# Convert from source to target
retcode, stdout, stderr = pvc_common.run_os_command(
"qemu-img convert -C -f {} -O raw {} {}".format(
volume["volume_format"],
"/dev/rbd/{}".format(src_volume),
"/dev/rbd/{}".format(dst_volume),
)
)
if retcode:
raise ProvisioningError(
'Failed to convert {} volume "{}" to raw volume "{}": {}'.format(
volume["volume_format"], src_volume, dst_volume, stderr
)
)
# Unmap the source RBD device (don't bother later)
retcode, retmsg = pvc_ceph.unmap_volume(
zkhandler, volume["pool"], src_volume_name
)
if not retcode:
raise ProvisioningError(
'Failed to unmap source volume "{}": {}'.format(
src_volume_name, retmsg
)
)
# Unmap the target RBD device (don't bother later)
retcode, retmsg = pvc_ceph.unmap_volume(
zkhandler, volume["pool"], dst_volume_name
)
if not retcode:
raise ProvisioningError(
'Failed to unmap destination volume "{}": {}'.format(
dst_volume_name, retmsg
)
)
else:
if volume.get("source_volume") is not None:
continue
if volume.get("filesystem") is None:
continue
filesystem_args_list = list()
for arg in volume["filesystem_args"].split():
arg_entry, *arg_data = arg.split("=")
arg_data = "=".join(arg_data)
filesystem_args_list.append(arg_entry)
filesystem_args_list.append(arg_data)
filesystem_args = " ".join(filesystem_args_list)
print(
"Creating {} filesystem on {}".format(
volume["filesystem"], dst_volume
)
)
print("Args: {}".format(filesystem_args))
# Map the RBD device
retcode, retmsg = pvc_ceph.map_volume(
zkhandler, volume["pool"], dst_volume_name
)
if not retcode:
raise ProvisioningError(
'Failed to map volume "{}": {}'.format(dst_volume, retmsg)
)
# Create the filesystem
if volume["filesystem"] == "swap":
retcode, stdout, stderr = pvc_common.run_os_command(
"mkswap -f /dev/rbd/{}".format(dst_volume)
)
if retcode:
raise ProvisioningError(
'Failed to create swap on "{}": {}'.format(
dst_volume, stderr
)
)
else:
retcode, stdout, stderr = pvc_common.run_os_command(
"mkfs.{} {} /dev/rbd/{}".format(
volume["filesystem"], filesystem_args, dst_volume
)
)
if retcode:
raise ProvisioningError(
'Failed to create {} filesystem on "{}": {}'.format(
volume["filesystem"], dst_volume, stderr
)
)
print(stdout)
if is_script_install:
# Create temporary directory
retcode, stdout, stderr = pvc_common.run_os_command("mktemp -d")
if retcode:
raise ProvisioningError(
"Failed to create a temporary directory: {}".format(stderr)
)
temp_dir = stdout.strip()
for volume in vm_data["volumes"]:
if volume["source_volume"] is not None:
continue
if not volume["mountpoint"] or volume["mountpoint"] == "swap":
continue
mapped_dst_volume = "/dev/rbd/{}/{}_{}".format(
volume["pool"], vm_name, volume["disk_id"]
)
mount_path = "{}{}".format(temp_dir, volume["mountpoint"])
# Ensure the mount path exists (within the filesystems)
retcode, stdout, stderr = pvc_common.run_os_command(
"mkdir -p {}".format(mount_path)
)
if retcode:
raise ProvisioningError(
'Failed to create mountpoint "{}": {}'.format(
mount_path, stderr
)
)
# Mount filesystems to temporary directory
retcode, stdout, stderr = pvc_common.run_os_command(
"mount {} {}".format(mapped_dst_volume, mount_path)
)
if retcode:
raise ProvisioningError(
'Failed to mount "{}" on "{}": {}'.format(
mapped_dst_volume, mount_path, stderr
)
)
print(
"Successfully mounted {} on {}".format(
mapped_dst_volume, mount_path
)
)
# Phase 8 - provisioning script execution
# * Execute the provisioning script main function ("install") passing any custom arguments
self.update_state(
state="RUNNING",
meta={"current": 8, "total": 10, "status": "Executing provisioning script"},
)
time.sleep(1)
if is_script_install:
print("Running installer script")
# Parse the script arguments
script_arguments = dict()
for argument in vm_data["script_arguments"]:
argument_name, argument_data = argument.split("=")
script_arguments[argument_name] = argument_data
# Parse the runtime arguments
if script_run_args is not None:
for argument in script_run_args:
argument_name, argument_data = argument.split("=")
script_arguments[argument_name] = argument_data
print("Script arguments: {}".format(script_arguments))
# Run the script
try:
installer_script.install(
vm_name=vm_name,
vm_id=vm_id,
temporary_directory=temp_dir,
disks=vm_data["volumes"],
networks=vm_data["networks"],
**script_arguments
)
except Exception as e:
raise ProvisioningError("Failed to run install script: {}".format(e))
except Exception as e:
start_vm = False
raise e
# Always perform the cleanup steps
finally:
# Phase 9 - install cleanup
# * Unmount any mounted volumes
# * Remove any temporary directories
self.update_state(
state="RUNNING",
meta={
"current": 9,
"total": 10,
"status": "Cleaning up local mounts and directories",
},
)
time.sleep(1)
if not is_ova_install:
for volume in list(reversed(vm_data["volumes"])):
if volume.get("source_volume") is not None:
continue
if is_script_install:
# Unmount the volume
if (
volume.get("mountpoint") is not None
and volume.get("mountpoint") != "swap"
):
print(
"Cleaning up mount {}{}".format(
temp_dir, volume["mountpoint"]
)
)
mount_path = "{}{}".format(temp_dir, volume["mountpoint"])
# Make sure any bind mounts or submounts are unmounted first
if volume["mountpoint"] == "/":
retcode, stdout, stderr = pvc_common.run_os_command(
"umount {}/**/**".format(mount_path)
)
retcode, stdout, stderr = pvc_common.run_os_command(
"umount {}/**".format(mount_path)
)
retcode, stdout, stderr = pvc_common.run_os_command(
"umount {}".format(mount_path)
)
if retcode:
print(
'Failed to unmount "{}": {}'.format(mount_path, stderr)
)
# Unmap the RBD device
if volume["filesystem"]:
print(
"Cleaning up RBD mapping /dev/rbd/{}/{}_{}".format(
volume["pool"], vm_name, volume["disk_id"]
)
)
rbd_volume = "/dev/rbd/{}/{}_{}".format(
volume["pool"], vm_name, volume["disk_id"]
)
retcode, stdout, stderr = pvc_common.run_os_command(
"rbd unmap {}".format(rbd_volume)
)
if retcode:
print(
'Failed to unmap volume "{}": {}'.format(rbd_volume, stderr)
)
print("Cleaning up temporary directories and files")
if is_script_install:
# Remove temporary mount directory (don't fail if not removed)
retcode, stdout, stderr = pvc_common.run_os_command(
"rmdir {}".format(temp_dir)
)
if retcode:
print(
'Failed to delete temporary directory "{}": {}'.format(
temp_dir, stderr
)
)
# Remote temporary script (don't fail if not removed)
retcode, stdout, stderr = pvc_common.run_os_command(
"rm -f {}".format(script_file)
)
if retcode:
print(
'Failed to delete temporary script file "{}": {}'.format(
script_file, stderr
)
)
# Phase 10 - startup
# * Start the VM in the PVC cluster
if start_vm:
self.update_state(
state="RUNNING",
meta={"current": 10, "total": 10, "status": "Starting VM"},
)
time.sleep(1)
retcode, retmsg = pvc_vm.start_vm(zkhandler, vm_name)
print(retmsg)
zkhandler.disconnect()
del zkhandler
return {
"status": 'VM "{}" with profile "{}" has been provisioned successfully'.format(
vm_name, vm_profile
),
"current": 10,
"total": 10,
}

730
api-daemon/pvcapid/vmbuilder.py Executable file
View File

@ -0,0 +1,730 @@
#!/usr/bin/env python3
# vmbuilder.py - pvc api vm builder (provisioner) functions
# part of the parallel virtual cluster (pvc) system
#
# copyright (c) 2018-2022 joshua m. boniface <joshua@boniface.me>
#
# this program is free software: you can redistribute it and/or modify
# it under the terms of the gnu general public license as published by
# the free software foundation, version 3.
#
# this program is distributed in the hope that it will be useful,
# but without any warranty; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with this program. If not, see <https://www.gnu.org/licenses/>.
#
###############################################################################
import json
import psycopg2
import psycopg2.extras
import re
import os
# import sys
import time
import importlib.util
import uuid
from contextlib import contextmanager
from pvcapid.Daemon import config
from daemon_lib.zkhandler import ZKHandler
import daemon_lib.common as pvc_common
import daemon_lib.node as pvc_node
import daemon_lib.vm as pvc_vm
import daemon_lib.network as pvc_network
import daemon_lib.ceph as pvc_ceph
#
# Exceptions (used by Celery tasks)
#
class ValidationError(Exception):
"""
An exception that results from some value being un- or mis-defined.
"""
pass
class ClusterError(Exception):
"""
An exception that results from the PVC cluster being out of alignment with the action.
"""
pass
class ProvisioningError(Exception):
"""
An exception that results from a failure of a provisioning command.
"""
pass
#
# VMBuilder class - subclassed by install scripts
#
class VMBuilder(object):
def __init__(
self,
vm_name,
vm_id,
vm_profile,
vm_data,
):
self.vm_name = vm_name
self.vm_id = vm_id
self.vm_uuid = uuid.uuid4()
self.vm_profile = vm_profile
self.vm_data = vm_data
#
# Primary class functions; implemented by the individual scripts
#
def setup(self):
"""
setup(): Perform special setup steps before proceeding
OPTIONAL
"""
pass
def create(self):
"""
create(): Create the VM libvirt schema definition which is defined afterwards
"""
pass
def prepare(self):
"""
prepare(): Prepare any disks/volumes for the install step
"""
pass
def install(self):
"""
install(): Perform the installation
"""
pass
def cleanup(self):
"""
cleanup(): Perform any cleanup required after the prepare() step or on failure of the install() step
"""
pass
#
# Helper functions (as context managers)
#
@contextmanager
def chroot(destination):
"""
Change root directory to a given destination
"""
try:
real_root = os.open("/", os.O_RDONLY)
os.chroot(destination)
fake_root = os.open("/", os.O_RDONLY)
os.fchdir(fake_root)
yield
except Exception:
pass
finally:
os.fchdir(real_root)
os.chroot(".")
os.fchdir(fake_root)
os.close(fake_root)
os.close(real_root)
del fake_root
del real_root
@contextmanager
def open_db(config):
try:
conn = psycopg2.connect(
host=config["database_host"],
port=config["database_port"],
dbname=config["database_name"],
user=config["database_user"],
password=config["database_password"],
)
cur = conn.cursor(cursor_factory=psycopg2.extras.RealDictCursor)
yield cur
except Exception:
raise ClusterError("Failed to connect to Postgres")
finally:
conn.commit()
cur.close()
conn.close()
del conn
@contextmanager
def open_zk(config):
try:
zkhandler = ZKHandler(config)
zkhandler.connect()
yield zkhandler
except Exception:
raise ClusterError("Failed to connect to Zookeeper")
finally:
zkhandler.disconnect()
del zkhandler
#
# Main VM provisioning function - executed by the Celery worker
#
def create_vm(
self, vm_name, vm_profile, define_vm=True, start_vm=True, script_run_args=[]
):
print(f"Starting provisioning of VM '{vm_name}' with profile '{vm_profile}'")
# Phase 1 - setup
# * Get the profile elements
# * Get the details from these elements
# * Assemble a VM configuration dictionary
self.update_state(
state="RUNNING",
meta={"current": 1, "total": 10, "status": "Collecting configuration"},
)
time.sleep(1)
vm_id = re.findall(r"/(\d+)$/", vm_name)
if not vm_id:
vm_id = 0
else:
vm_id = vm_id[0]
vm_data = dict()
with open_db(config) as db_cur:
# Get the profile information
query = "SELECT * FROM profile WHERE name = %s"
args = (vm_profile,)
db_cur.execute(query, args)
profile_data = db_cur.fetchone()
if profile_data.get("arguments"):
vm_data["script_arguments"] = profile_data.get("arguments").split("|")
else:
vm_data["script_arguments"] = []
if profile_data.get("profile_type") == "ova":
is_ova_install = True
else:
is_ova_install = False
# Get the system details
query = "SELECT * FROM system_template WHERE id = %s"
args = (profile_data["system_template"],)
db_cur.execute(query, args)
vm_data["system_details"] = db_cur.fetchone()
# Get the MAC template
query = "SELECT mac_template FROM network_template WHERE id = %s"
args = (profile_data["network_template"],)
db_cur.execute(query, args)
db_row = db_cur.fetchone()
if db_row:
vm_data["mac_template"] = db_row.get("mac_template")
else:
vm_data["mac_template"] = None
# Get the networks
query = "SELECT * FROM network WHERE network_template = %s"
args = (profile_data["network_template"],)
db_cur.execute(query, args)
_vm_networks = db_cur.fetchall()
vm_networks = list()
# Set the eth_bridge for each network
for network in _vm_networks:
vni = network["vni"]
if vni in ["upstream", "cluster", "storage"]:
eth_bridge = "br{}".format(vni)
else:
eth_bridge = "vmbr{}".format(vni)
network["eth_bridge"] = eth_bridge
vm_networks.append(network)
vm_data["networks"] = vm_networks
# Get the storage volumes
# ORDER BY ensures disks are always in the sdX/vdX order, regardless of add order
query = "SELECT * FROM storage WHERE storage_template = %s ORDER BY disk_id"
args = (profile_data["storage_template"],)
db_cur.execute(query, args)
vm_data["volumes"] = db_cur.fetchall()
# Get the script
query = "SELECT script FROM script WHERE id = %s"
args = (profile_data["script"],)
db_cur.execute(query, args)
db_row = db_cur.fetchone()
if db_row:
vm_data["script"] = db_row.get("script")
else:
vm_data["script"] = None
query = "SELECT * FROM ova WHERE id = %s"
args = (profile_data["ova"],)
db_cur.execute(query, args)
vm_data["ova_details"] = db_cur.fetchone()
query = "SELECT * FROM ova_volume WHERE ova = %s"
args = (profile_data["ova"],)
db_cur.execute(query, args)
vm_data["volumes"] = db_cur.fetchall()
retcode, stdout, stderr = pvc_common.run_os_command("uname -m")
vm_data["system_architecture"] = stdout.strip()
monitor_list = list()
coordinator_names = config["storage_hosts"]
for coordinator in coordinator_names:
monitor_list.append("{}.{}".format(coordinator, config["storage_domain"]))
vm_data["ceph_monitor_list"] = monitor_list
vm_data["ceph_monitor_port"] = config["ceph_monitor_port"]
vm_data["ceph_monitor_secret"] = config["ceph_storage_secret_uuid"]
# Parse the script arguments
script_arguments = dict()
for argument in vm_data["script_arguments"]:
argument_name, argument_data = argument.split("=")
script_arguments[argument_name] = argument_data
# Parse the runtime arguments
if script_run_args is not None:
for argument in script_run_args:
argument_name, argument_data = argument.split("=")
script_arguments[argument_name] = argument_data
print("Script arguments: {}".format(script_arguments))
vm_data["script_arguments"] = script_arguments
print(
"VM configuration data:\n{}".format(
json.dumps(vm_data, sort_keys=True, indent=2)
)
)
# Phase 2 - verification
# * Ensure that at least one node has enough free RAM to hold the VM (becomes main host)
# * Ensure that all networks are valid
# * Ensure that there is enough disk space in the Ceph cluster for the disks
# This is the "safe fail" step when an invalid configuration will be caught
self.update_state(
state="RUNNING",
meta={
"current": 2,
"total": 10,
"status": "Verifying configuration against cluster",
},
)
time.sleep(1)
with open_zk(config) as zkhandler:
# Verify that a VM with this name does not already exist
if pvc_vm.searchClusterByName(zkhandler, vm_name):
raise ClusterError(
"A VM with the name '{}' already exists in the cluster.".format(vm_name)
)
# Verify that at least one host has enough free RAM to run the VM
_discard, nodes = pvc_node.get_list(zkhandler, None)
target_node = None
last_free = 0
for node in nodes:
# Skip the node if it is not ready to run VMs
if node["daemon_state"] != "run" or node["domain_state"] != "ready":
continue
# Skip the node if its free memory is less than the new VM's size, plus a 512MB buffer
if node["memory"]["free"] < (vm_data["system_details"]["vram_mb"] + 512):
continue
# If this node has the most free, use it
if node["memory"]["free"] > last_free:
last_free = node["memory"]["free"]
target_node = node["name"]
# Raise if no node was found
if not target_node:
raise ClusterError(
"No ready cluster node contains at least {}+512 MB of free RAM.".format(
vm_data["system_details"]["vram_mb"]
)
)
print(
'Selecting target node "{}" with "{}" MB free RAM'.format(
target_node, last_free
)
)
# Verify that all configured networks are present on the cluster
cluster_networks, _discard = pvc_network.getClusterNetworkList(zkhandler)
for network in vm_data["networks"]:
vni = str(network["vni"])
if vni not in cluster_networks and vni not in [
"upstream",
"cluster",
"storage",
]:
raise ClusterError(
'The network VNI "{}" is not present on the cluster.'.format(vni)
)
print("All configured networks for VM are valid")
# Verify that there is enough disk space free to provision all VM disks
pools = dict()
for volume in vm_data["volumes"]:
if volume.get("source_volume") is not None:
volume_data = pvc_ceph.getVolumeInformation(
zkhandler, volume["pool"], volume["source_volume"]
)
if not volume_data:
raise ClusterError(
"The source volume {}/{} could not be found.".format(
volume["pool"], volume["source_volume"]
)
)
if not volume["pool"] in pools:
pools[volume["pool"]] = int(
pvc_ceph.format_bytes_fromhuman(volume_data["stats"]["size"])
/ 1024
/ 1024
/ 1024
)
else:
pools[volume["pool"]] += int(
pvc_ceph.format_bytes_fromhuman(volume_data["stats"]["size"])
/ 1024
/ 1024
/ 1024
)
else:
if not volume["pool"] in pools:
pools[volume["pool"]] = volume["disk_size_gb"]
else:
pools[volume["pool"]] += volume["disk_size_gb"]
for pool in pools:
try:
pool_information = pvc_ceph.getPoolInformation(zkhandler, pool)
if not pool_information:
raise
except Exception:
raise ClusterError(
'Pool "{}" is not present on the cluster.'.format(pool)
)
pool_free_space_gb = int(
pool_information["stats"]["free_bytes"] / 1024 / 1024 / 1024
)
pool_vm_usage_gb = int(pools[pool])
if pool_vm_usage_gb >= pool_free_space_gb:
raise ClusterError(
'Pool "{}" has only {} GB free and VM requires {} GB.'.format(
pool, pool_free_space_gb, pool_vm_usage_gb
)
)
print("There is enough space on cluster to store VM volumes")
if not is_ova_install:
# Verify that every specified filesystem is valid
used_filesystems = list()
for volume in vm_data["volumes"]:
if volume["source_volume"] is not None:
continue
if volume["filesystem"] and volume["filesystem"] not in used_filesystems:
used_filesystems.append(volume["filesystem"])
for filesystem in used_filesystems:
if filesystem == "swap":
retcode, stdout, stderr = pvc_common.run_os_command("which mkswap")
if retcode:
raise ProvisioningError(
"Failed to find binary for mkswap: {}".format(stderr)
)
else:
retcode, stdout, stderr = pvc_common.run_os_command(
"which mkfs.{}".format(filesystem)
)
if retcode:
raise ProvisioningError(
"Failed to find binary for mkfs.{}: {}".format(
filesystem, stderr
)
)
print("All selected filesystems are valid")
# Phase 3 - provisioning script preparation
# * Import the provisioning script as a library with importlib
# * Ensure the required function(s) are present
self.update_state(
state="RUNNING",
meta={"current": 3, "total": 10, "status": "Preparing provisioning script"},
)
time.sleep(1)
# Write the script out to a temporary file
retcode, stdout, stderr = pvc_common.run_os_command("mktemp")
if retcode:
raise ProvisioningError("Failed to create a temporary file: {}".format(stderr))
script_file = stdout.strip()
with open(script_file, "w") as fh:
fh.write(vm_data["script"])
fh.write("\n")
# Import the script file
loader = importlib.machinery.SourceFileLoader("installer_script", script_file)
spec = importlib.util.spec_from_loader(loader.name, loader)
installer_script = importlib.util.module_from_spec(spec)
spec.loader.exec_module(installer_script)
# Set up the VMBuilderScript object
vm_builder = installer_script.VMBuilderScript(
vm_name=vm_name,
vm_id=vm_id,
vm_profile=vm_profile,
vm_data=vm_data,
)
print("Provisioning script imported successfully")
# Create temporary directory for external chroot
retcode, stdout, stderr = pvc_common.run_os_command("mktemp -d")
if retcode:
raise ProvisioningError(f"Failed to create a temporary directory: {stderr}")
temp_dir = stdout.strip()
# Bind mount / to the chroot location /
retcode, stdout, stderr = pvc_common.run_os_command(
f"mount --bind --options ro / {temp_dir}"
)
if retcode:
raise ProvisioningError(
f"Failed to mount rootfs onto {temp_dir} for chroot: {stderr}"
)
# Mount tmpfs to the chroot location /tmp
retcode, stdout, stderr = pvc_common.run_os_command(
f"mount --type tmpfs tmpfs {temp_dir}/tmp"
)
if retcode:
raise ProvisioningError(
f"Failed to mount tmpfs onto {temp_dir}/tmp for chroot: {stderr}"
)
# Bind mount /dev to the chroot location /dev
retcode, stdout, stderr = pvc_common.run_os_command(
f"mount --bind --options ro /dev {temp_dir}/dev"
)
if retcode:
raise ProvisioningError(
f"Failed to mount devfs onto {temp_dir}/dev for chroot: {stderr}"
)
# Bind mount /run to the chroot location /run
retcode, stdout, stderr = pvc_common.run_os_command(
f"mount --bind --options rw /run {temp_dir}/run"
)
if retcode:
raise ProvisioningError(
f"Failed to mount runfs onto {temp_dir}/run for chroot: {stderr}"
)
# Bind mount /sys to the chroot location /sys
retcode, stdout, stderr = pvc_common.run_os_command(
f"mount --bind --options rw /sys {temp_dir}/sys"
)
if retcode:
raise ProvisioningError(
f"Failed to mount sysfs onto {temp_dir}/sys for chroot: {stderr}"
)
print("Chroot environment prepared successfully")
# Phase 4 - script: setup()
# * Run pre-setup steps
self.update_state(
state="RUNNING",
meta={
"current": 4,
"total": 10,
"status": "Running script setup() step",
},
)
time.sleep(1)
print("Running script setup() step")
with chroot(temp_dir):
vm_builder.setup()
# Phase 5 - script: create()
# * Prepare the libvirt XML defintion for the VM
self.update_state(
state="RUNNING",
meta={
"current": 5,
"total": 10,
"status": "Running script create() step",
},
)
time.sleep(1)
if define_vm:
print("Running script create() step")
with chroot(temp_dir):
vm_schema = vm_builder.create()
print("Generated VM schema:\n{}\n".format(vm_schema))
print("Defining VM on cluster")
node_limit = vm_data["system_details"]["node_limit"]
if node_limit:
node_limit = node_limit.split(",")
node_selector = vm_data["system_details"]["node_selector"]
node_autostart = vm_data["system_details"]["node_autostart"]
migration_method = vm_data["system_details"]["migration_method"]
with open_zk(config) as zkhandler:
retcode, retmsg = pvc_vm.define_vm(
zkhandler,
vm_schema.strip(),
target_node,
node_limit,
node_selector,
node_autostart,
migration_method,
vm_profile,
initial_state="provision",
)
print(retmsg)
else:
print("Skipping VM definition due to define_vm=False")
# Phase 6 - script: prepare()
# * Run preparation steps (e.g. disk creation and mapping, filesystem creation, etc.)
self.update_state(
state="RUNNING",
meta={
"current": 6,
"total": 10,
"status": "Running script prepare() step",
},
)
time.sleep(1)
print("Running script prepare() step")
with chroot(temp_dir):
vm_builder.prepare()
# Phase 7 - script: install()
# * Run installation with arguments
self.update_state(
state="RUNNING",
meta={
"current": 7,
"total": 10,
"status": "Running script install() step",
},
)
time.sleep(1)
print("Running script install() step")
with chroot(temp_dir):
vm_builder.install()
# Phase 8 - script: cleanup()
# * Run cleanup steps
self.update_state(
state="RUNNING",
meta={
"current": 8,
"total": 10,
"status": "Running script cleanup() step",
},
)
time.sleep(1)
print("Running script cleanup() step")
with chroot(temp_dir):
vm_builder.cleanup()
# Phase 9 - general cleanup
# * Clean up the chroot from earlier
self.update_state(
state="RUNNING",
meta={
"current": 9,
"total": 10,
"status": "Running upper cleanup steps",
},
)
time.sleep(1)
print("Running upper cleanup steps")
# Remote temporary script (don't fail if not removed)
if not os.remove(script_file):
print(f"Failed to delete temporary script file '{script_file}'.")
# Unmount bind-mounted devfs on the chroot
retcode, stdout, stderr = pvc_common.run_os_command(f"umount {temp_dir}/dev")
if retcode:
raise ProvisioningError(f"Failed to unmount devfs from chroot: {stderr}")
# Unmount bind-mounted runfs on the chroot
retcode, stdout, stderr = pvc_common.run_os_command(f"umount {temp_dir}/run")
if retcode:
raise ProvisioningError(f"Failed to unmount runfs from chroot: {stderr}")
# Unmount bind-mounted sysfs on the chroot
retcode, stdout, stderr = pvc_common.run_os_command(f"umount {temp_dir}/sys")
if retcode:
raise ProvisioningError(f"Failed to unmount sysfs from chroot: {stderr}")
# Unmount bind-mounted tmpfs on the chroot
retcode, stdout, stderr = pvc_common.run_os_command(f"umount {temp_dir}/tmp")
if retcode:
raise ProvisioningError(f"Failed to unmount tmpfs from chroot: {stderr}")
# Unmount bind-mounted rootfs on the chroot
retcode, stdout, stderr = pvc_common.run_os_command(f"umount {temp_dir}")
if retcode:
raise ProvisioningError(f"Failed to unmount rootfs from chroot: {stderr}")
# Remove the temp_dir
if not os.rmdir(temp_dir):
print(f"Failed to delete temporary chroot directory '{temp_dir}'.")
# Phase 10 - startup
# * Start the VM in the PVC cluster
self.update_state(
state="RUNNING",
meta={
"current": 10,
"total": 10,
"status": "Starting VM",
},
)
time.sleep(1)
if start_vm:
print("Starting VM")
with open_zk(config) as zkhandler:
success, message = pvc_vm.start_vm(zkhandler, vm_name)
print(message)