Reformat code with Black code formatter
Unify the code style along PEP and Black principles using the tool.
This commit is contained in:
@ -29,28 +29,24 @@ import daemon_lib.ceph as pvc_ceph
|
||||
|
||||
|
||||
def set_maintenance(zkhandler, maint_state):
|
||||
current_maint_state = zkhandler.read('base.config.maintenance')
|
||||
current_maint_state = zkhandler.read("base.config.maintenance")
|
||||
if maint_state == current_maint_state:
|
||||
if maint_state == 'true':
|
||||
return True, 'Cluster is already in maintenance mode'
|
||||
if maint_state == "true":
|
||||
return True, "Cluster is already in maintenance mode"
|
||||
else:
|
||||
return True, 'Cluster is already in normal mode'
|
||||
return True, "Cluster is already in normal mode"
|
||||
|
||||
if maint_state == 'true':
|
||||
zkhandler.write([
|
||||
('base.config.maintenance', 'true')
|
||||
])
|
||||
return True, 'Successfully set cluster in maintenance mode'
|
||||
if maint_state == "true":
|
||||
zkhandler.write([("base.config.maintenance", "true")])
|
||||
return True, "Successfully set cluster in maintenance mode"
|
||||
else:
|
||||
zkhandler.write([
|
||||
('base.config.maintenance', 'false')
|
||||
])
|
||||
return True, 'Successfully set cluster in normal mode'
|
||||
zkhandler.write([("base.config.maintenance", "false")])
|
||||
return True, "Successfully set cluster in normal mode"
|
||||
|
||||
|
||||
def getClusterInformation(zkhandler):
|
||||
# Get cluster maintenance state
|
||||
maint_state = zkhandler.read('base.config.maintenance')
|
||||
maint_state = zkhandler.read("base.config.maintenance")
|
||||
|
||||
# List of messages to display to the clients
|
||||
cluster_health_msg = []
|
||||
@ -69,7 +65,9 @@ def getClusterInformation(zkhandler):
|
||||
retcode, ceph_osd_list = pvc_ceph.get_list_osd(zkhandler, None)
|
||||
retcode, ceph_pool_list = pvc_ceph.get_list_pool(zkhandler, None)
|
||||
retcode, ceph_volume_list = pvc_ceph.get_list_volume(zkhandler, None, None)
|
||||
retcode, ceph_snapshot_list = pvc_ceph.get_list_snapshot(zkhandler, None, None, None)
|
||||
retcode, ceph_snapshot_list = pvc_ceph.get_list_snapshot(
|
||||
zkhandler, None, None, None
|
||||
)
|
||||
|
||||
# Determine, for each subsection, the total count
|
||||
node_count = len(node_list)
|
||||
@ -91,8 +89,8 @@ def getClusterInformation(zkhandler):
|
||||
node_largest_index = None
|
||||
node_largest_count = 0
|
||||
for index, node in enumerate(node_list):
|
||||
node_mem_total = node['memory']['total']
|
||||
node_mem_alloc = node['memory']['allocated']
|
||||
node_mem_total = node["memory"]["total"]
|
||||
node_mem_alloc = node["memory"]["allocated"]
|
||||
alloc_total += node_mem_alloc
|
||||
|
||||
# Determine if this node is the largest seen so far
|
||||
@ -105,32 +103,42 @@ def getClusterInformation(zkhandler):
|
||||
continue
|
||||
n_minus_1_node_list.append(node)
|
||||
for index, node in enumerate(n_minus_1_node_list):
|
||||
n_minus_1_total += node['memory']['total']
|
||||
n_minus_1_total += node["memory"]["total"]
|
||||
if alloc_total > n_minus_1_total:
|
||||
cluster_healthy_status = False
|
||||
cluster_health_msg.append("Total VM memory ({}) is overprovisioned (max {}) for (n-1) failure scenarios".format(alloc_total, n_minus_1_total))
|
||||
cluster_health_msg.append(
|
||||
"Total VM memory ({}) is overprovisioned (max {}) for (n-1) failure scenarios".format(
|
||||
alloc_total, n_minus_1_total
|
||||
)
|
||||
)
|
||||
|
||||
# Determinations for node health
|
||||
node_healthy_status = list(range(0, node_count))
|
||||
node_report_status = list(range(0, node_count))
|
||||
for index, node in enumerate(node_list):
|
||||
daemon_state = node['daemon_state']
|
||||
domain_state = node['domain_state']
|
||||
if daemon_state != 'run' and domain_state != 'ready':
|
||||
daemon_state = node["daemon_state"]
|
||||
domain_state = node["domain_state"]
|
||||
if daemon_state != "run" and domain_state != "ready":
|
||||
node_healthy_status[index] = False
|
||||
cluster_health_msg.append("Node '{}' in {},{} state".format(node['name'], daemon_state, domain_state))
|
||||
cluster_health_msg.append(
|
||||
"Node '{}' in {},{} state".format(
|
||||
node["name"], daemon_state, domain_state
|
||||
)
|
||||
)
|
||||
else:
|
||||
node_healthy_status[index] = True
|
||||
node_report_status[index] = daemon_state + ',' + domain_state
|
||||
node_report_status[index] = daemon_state + "," + domain_state
|
||||
|
||||
# Determinations for VM health
|
||||
vm_healthy_status = list(range(0, vm_count))
|
||||
vm_report_status = list(range(0, vm_count))
|
||||
for index, vm in enumerate(vm_list):
|
||||
vm_state = vm['state']
|
||||
if vm_state not in ['start', 'disable', 'migrate', 'unmigrate', 'provision']:
|
||||
vm_state = vm["state"]
|
||||
if vm_state not in ["start", "disable", "migrate", "unmigrate", "provision"]:
|
||||
vm_healthy_status[index] = False
|
||||
cluster_health_msg.append("VM '{}' in {} state".format(vm['name'], vm_state))
|
||||
cluster_health_msg.append(
|
||||
"VM '{}' in {} state".format(vm["name"], vm_state)
|
||||
)
|
||||
else:
|
||||
vm_healthy_status[index] = True
|
||||
vm_report_status[index] = vm_state
|
||||
@ -140,70 +148,99 @@ def getClusterInformation(zkhandler):
|
||||
ceph_osd_report_status = list(range(0, ceph_osd_count))
|
||||
for index, ceph_osd in enumerate(ceph_osd_list):
|
||||
try:
|
||||
ceph_osd_up = ceph_osd['stats']['up']
|
||||
ceph_osd_up = ceph_osd["stats"]["up"]
|
||||
except KeyError:
|
||||
ceph_osd_up = 0
|
||||
|
||||
try:
|
||||
ceph_osd_in = ceph_osd['stats']['in']
|
||||
ceph_osd_in = ceph_osd["stats"]["in"]
|
||||
except KeyError:
|
||||
ceph_osd_in = 0
|
||||
|
||||
up_texts = {1: 'up', 0: 'down'}
|
||||
in_texts = {1: 'in', 0: 'out'}
|
||||
up_texts = {1: "up", 0: "down"}
|
||||
in_texts = {1: "in", 0: "out"}
|
||||
|
||||
if not ceph_osd_up or not ceph_osd_in:
|
||||
ceph_osd_healthy_status[index] = False
|
||||
cluster_health_msg.append('OSD {} in {},{} state'.format(ceph_osd['id'], up_texts[ceph_osd_up], in_texts[ceph_osd_in]))
|
||||
cluster_health_msg.append(
|
||||
"OSD {} in {},{} state".format(
|
||||
ceph_osd["id"], up_texts[ceph_osd_up], in_texts[ceph_osd_in]
|
||||
)
|
||||
)
|
||||
else:
|
||||
ceph_osd_healthy_status[index] = True
|
||||
ceph_osd_report_status[index] = up_texts[ceph_osd_up] + ',' + in_texts[ceph_osd_in]
|
||||
ceph_osd_report_status[index] = (
|
||||
up_texts[ceph_osd_up] + "," + in_texts[ceph_osd_in]
|
||||
)
|
||||
|
||||
# Find out the overall cluster health; if any element of a healthy_status is false, it's unhealthy
|
||||
if maint_state == 'true':
|
||||
cluster_health = 'Maintenance'
|
||||
elif cluster_healthy_status is False or False in node_healthy_status or False in vm_healthy_status or False in ceph_osd_healthy_status:
|
||||
cluster_health = 'Degraded'
|
||||
if maint_state == "true":
|
||||
cluster_health = "Maintenance"
|
||||
elif (
|
||||
cluster_healthy_status is False
|
||||
or False in node_healthy_status
|
||||
or False in vm_healthy_status
|
||||
or False in ceph_osd_healthy_status
|
||||
):
|
||||
cluster_health = "Degraded"
|
||||
else:
|
||||
cluster_health = 'Optimal'
|
||||
cluster_health = "Optimal"
|
||||
|
||||
# Find out our storage health from Ceph
|
||||
ceph_status = zkhandler.read('base.storage').split('\n')
|
||||
ceph_status = zkhandler.read("base.storage").split("\n")
|
||||
ceph_health = ceph_status[2].split()[-1]
|
||||
|
||||
# Parse the status output to get the health indicators
|
||||
line_record = False
|
||||
for index, line in enumerate(ceph_status):
|
||||
if re.search('services:', line):
|
||||
if re.search("services:", line):
|
||||
line_record = False
|
||||
if line_record and len(line.strip()) > 0:
|
||||
storage_health_msg.append(line.strip())
|
||||
if re.search('health:', line):
|
||||
if re.search("health:", line):
|
||||
line_record = True
|
||||
|
||||
if maint_state == 'true':
|
||||
storage_health = 'Maintenance'
|
||||
elif ceph_health != 'HEALTH_OK':
|
||||
storage_health = 'Degraded'
|
||||
if maint_state == "true":
|
||||
storage_health = "Maintenance"
|
||||
elif ceph_health != "HEALTH_OK":
|
||||
storage_health = "Degraded"
|
||||
else:
|
||||
storage_health = 'Optimal'
|
||||
storage_health = "Optimal"
|
||||
|
||||
# State lists
|
||||
node_state_combinations = [
|
||||
'run,ready', 'run,flush', 'run,flushed', 'run,unflush',
|
||||
'init,ready', 'init,flush', 'init,flushed', 'init,unflush',
|
||||
'stop,ready', 'stop,flush', 'stop,flushed', 'stop,unflush',
|
||||
'dead,ready', 'dead,flush', 'dead,flushed', 'dead,unflush'
|
||||
"run,ready",
|
||||
"run,flush",
|
||||
"run,flushed",
|
||||
"run,unflush",
|
||||
"init,ready",
|
||||
"init,flush",
|
||||
"init,flushed",
|
||||
"init,unflush",
|
||||
"stop,ready",
|
||||
"stop,flush",
|
||||
"stop,flushed",
|
||||
"stop,unflush",
|
||||
"dead,ready",
|
||||
"dead,flush",
|
||||
"dead,flushed",
|
||||
"dead,unflush",
|
||||
]
|
||||
vm_state_combinations = [
|
||||
'start', 'restart', 'shutdown', 'stop', 'disable', 'fail', 'migrate', 'unmigrate', 'provision'
|
||||
]
|
||||
ceph_osd_state_combinations = [
|
||||
'up,in', 'up,out', 'down,in', 'down,out'
|
||||
"start",
|
||||
"restart",
|
||||
"shutdown",
|
||||
"stop",
|
||||
"disable",
|
||||
"fail",
|
||||
"migrate",
|
||||
"unmigrate",
|
||||
"provision",
|
||||
]
|
||||
ceph_osd_state_combinations = ["up,in", "up,out", "down,in", "down,out"]
|
||||
|
||||
# Format the Node states
|
||||
formatted_node_states = {'total': node_count}
|
||||
formatted_node_states = {"total": node_count}
|
||||
for state in node_state_combinations:
|
||||
state_count = 0
|
||||
for node_state in node_report_status:
|
||||
@ -213,7 +250,7 @@ def getClusterInformation(zkhandler):
|
||||
formatted_node_states[state] = state_count
|
||||
|
||||
# Format the VM states
|
||||
formatted_vm_states = {'total': vm_count}
|
||||
formatted_vm_states = {"total": vm_count}
|
||||
for state in vm_state_combinations:
|
||||
state_count = 0
|
||||
for vm_state in vm_report_status:
|
||||
@ -223,7 +260,7 @@ def getClusterInformation(zkhandler):
|
||||
formatted_vm_states[state] = state_count
|
||||
|
||||
# Format the OSD states
|
||||
formatted_osd_states = {'total': ceph_osd_count}
|
||||
formatted_osd_states = {"total": ceph_osd_count}
|
||||
for state in ceph_osd_state_combinations:
|
||||
state_count = 0
|
||||
for ceph_osd_state in ceph_osd_report_status:
|
||||
@ -234,19 +271,19 @@ def getClusterInformation(zkhandler):
|
||||
|
||||
# Format the status data
|
||||
cluster_information = {
|
||||
'health': cluster_health,
|
||||
'health_msg': cluster_health_msg,
|
||||
'storage_health': storage_health,
|
||||
'storage_health_msg': storage_health_msg,
|
||||
'primary_node': common.getPrimaryNode(zkhandler),
|
||||
'upstream_ip': zkhandler.read('base.config.upstream_ip'),
|
||||
'nodes': formatted_node_states,
|
||||
'vms': formatted_vm_states,
|
||||
'networks': network_count,
|
||||
'osds': formatted_osd_states,
|
||||
'pools': ceph_pool_count,
|
||||
'volumes': ceph_volume_count,
|
||||
'snapshots': ceph_snapshot_count
|
||||
"health": cluster_health,
|
||||
"health_msg": cluster_health_msg,
|
||||
"storage_health": storage_health,
|
||||
"storage_health_msg": storage_health_msg,
|
||||
"primary_node": common.getPrimaryNode(zkhandler),
|
||||
"upstream_ip": zkhandler.read("base.config.upstream_ip"),
|
||||
"nodes": formatted_node_states,
|
||||
"vms": formatted_vm_states,
|
||||
"networks": network_count,
|
||||
"osds": formatted_osd_states,
|
||||
"pools": ceph_pool_count,
|
||||
"volumes": ceph_volume_count,
|
||||
"snapshots": ceph_snapshot_count,
|
||||
}
|
||||
|
||||
return cluster_information
|
||||
@ -258,29 +295,32 @@ def get_info(zkhandler):
|
||||
if cluster_information:
|
||||
return True, cluster_information
|
||||
else:
|
||||
return False, 'ERROR: Failed to obtain cluster information!'
|
||||
return False, "ERROR: Failed to obtain cluster information!"
|
||||
|
||||
|
||||
def cluster_initialize(zkhandler, overwrite=False):
|
||||
# Abort if we've initialized the cluster before
|
||||
if zkhandler.exists('base.config.primary_node') and not overwrite:
|
||||
return False, 'ERROR: Cluster contains data and overwrite not set.'
|
||||
if zkhandler.exists("base.config.primary_node") and not overwrite:
|
||||
return False, "ERROR: Cluster contains data and overwrite not set."
|
||||
|
||||
if overwrite:
|
||||
# Delete the existing keys
|
||||
for key in zkhandler.schema.keys('base'):
|
||||
if key == 'root':
|
||||
for key in zkhandler.schema.keys("base"):
|
||||
if key == "root":
|
||||
# Don't delete the root key
|
||||
continue
|
||||
|
||||
status = zkhandler.delete('base.{}'.format(key), recursive=True)
|
||||
status = zkhandler.delete("base.{}".format(key), recursive=True)
|
||||
if not status:
|
||||
return False, 'ERROR: Failed to delete data in cluster; running nodes perhaps?'
|
||||
return (
|
||||
False,
|
||||
"ERROR: Failed to delete data in cluster; running nodes perhaps?",
|
||||
)
|
||||
|
||||
# Create the root keys
|
||||
zkhandler.schema.apply(zkhandler)
|
||||
|
||||
return True, 'Successfully initialized cluster'
|
||||
return True, "Successfully initialized cluster"
|
||||
|
||||
|
||||
def cluster_backup(zkhandler):
|
||||
@ -294,25 +334,25 @@ def cluster_backup(zkhandler):
|
||||
cluster_data[path] = data
|
||||
|
||||
if children:
|
||||
if path == '/':
|
||||
child_prefix = '/'
|
||||
if path == "/":
|
||||
child_prefix = "/"
|
||||
else:
|
||||
child_prefix = path + '/'
|
||||
child_prefix = path + "/"
|
||||
|
||||
for child in children:
|
||||
if child_prefix + child == '/zookeeper':
|
||||
if child_prefix + child == "/zookeeper":
|
||||
# We must skip the built-in /zookeeper tree
|
||||
continue
|
||||
if child_prefix + child == '/patroni':
|
||||
if child_prefix + child == "/patroni":
|
||||
# We must skip the /patroni tree
|
||||
continue
|
||||
|
||||
get_data(child_prefix + child)
|
||||
|
||||
try:
|
||||
get_data('/')
|
||||
get_data("/")
|
||||
except Exception as e:
|
||||
return False, 'ERROR: Failed to obtain backup: {}'.format(e)
|
||||
return False, "ERROR: Failed to obtain backup: {}".format(e)
|
||||
|
||||
return True, cluster_data
|
||||
|
||||
@ -322,18 +362,23 @@ def cluster_restore(zkhandler, cluster_data):
|
||||
kv = []
|
||||
schema_version = None
|
||||
for key in cluster_data:
|
||||
if key == zkhandler.schema.path('base.schema.version'):
|
||||
if key == zkhandler.schema.path("base.schema.version"):
|
||||
schema_version = cluster_data[key]
|
||||
data = cluster_data[key]
|
||||
kv.append((key, data))
|
||||
|
||||
if int(schema_version) != int(zkhandler.schema.version):
|
||||
return False, 'ERROR: Schema version of backup ({}) does not match cluster schema version ({}).'.format(schema_version, zkhandler.schema.version)
|
||||
return (
|
||||
False,
|
||||
"ERROR: Schema version of backup ({}) does not match cluster schema version ({}).".format(
|
||||
schema_version, zkhandler.schema.version
|
||||
),
|
||||
)
|
||||
|
||||
# Close the Zookeeper connection
|
||||
result = zkhandler.write(kv)
|
||||
|
||||
if result:
|
||||
return True, 'Restore completed successfully.'
|
||||
return True, "Restore completed successfully."
|
||||
else:
|
||||
return False, 'Restore failed.'
|
||||
return False, "Restore failed."
|
||||
|
Reference in New Issue
Block a user