X-Git-Url: https://code.grnet.gr/git/ganeti-local/blobdiff_plain/9cdb95781e44c99134f20a331f744c995ff5a309..43ffd83959093e1cd56c45bf7eac1c86d77b8ee7:/tools/cfgupgrade diff --git a/tools/cfgupgrade b/tools/cfgupgrade index d3984bf..554877f 100755 --- a/tools/cfgupgrade +++ b/tools/cfgupgrade @@ -1,7 +1,7 @@ #!/usr/bin/python # -# Copyright (C) 2007, 2008, 2009 Google Inc. +# Copyright (C) 2007, 2008, 2009, 2010, 2011, 2012, 2013 Google Inc. # # This program is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by @@ -31,9 +31,9 @@ import os import os.path import sys import optparse -import tempfile import logging -import errno +import time +from cStringIO import StringIO from ganeti import constants from ganeti import serializer @@ -41,36 +41,22 @@ from ganeti import utils from ganeti import cli from ganeti import bootstrap from ganeti import config +from ganeti import netutils +from ganeti import pathutils options = None args = None -# Unique object to identify calls without default value -NoDefault = object() - -# Dictionary with instance old keys, and new hypervisor keys -INST_HV_CHG = { - 'hvm_pae': constants.HV_PAE, - 'vnc_bind_address': constants.HV_VNC_BIND_ADDRESS, - 'initrd_path': constants.HV_INITRD_PATH, - 'hvm_nic_type': constants.HV_NIC_TYPE, - 'kernel_path': constants.HV_KERNEL_PATH, - 'hvm_acpi': constants.HV_ACPI, - 'hvm_cdrom_image_path': constants.HV_CDROM_IMAGE_PATH, - 'hvm_boot_order': constants.HV_BOOT_ORDER, - 'hvm_disk_type': constants.HV_DISK_TYPE, - } - -# Instance beparams changes -INST_BE_CHG = { - 'vcpus': constants.BE_VCPUS, - 'memory': constants.BE_MEMORY, - 'auto_balance': constants.BE_AUTO_BALANCE, - } - -# Field names -F_SERIAL = 'serial_no' + +#: Target major version we will upgrade to +TARGET_MAJOR = 2 +#: Target minor version we will upgrade to +TARGET_MINOR = 9 +#: Target major version for downgrade +DOWNGRADE_MAJOR = 2 +#: Target minor version for downgrade +DOWNGRADE_MINOR = 8 class Error(Exception): @@ -78,57 +64,6 @@ class Error(Exception): pass -def SsconfName(key): - """Returns the file name of an (old) ssconf key. - - """ - return "%s/ssconf_%s" % (options.data_dir, key) - - -def ReadFile(file_name, default=NoDefault): - """Reads a file. - - """ - logging.debug("Reading %s", file_name) - try: - fh = open(file_name, 'r') - except IOError, err: - if default is not NoDefault and err.errno == errno.ENOENT: - return default - raise - - try: - return fh.read() - finally: - fh.close() - - -def WriteFile(file_name, data): - """Writes a configuration file. - - """ - logging.debug("Writing %s", file_name) - utils.WriteFile(file_name=file_name, data=data, mode=0600, - dry_run=options.dry_run, backup=True) - - -def GenerateSecret(all_secrets): - """Generate an unique DRBD secret. - - This is a copy from ConfigWriter. - - """ - retries = 64 - while retries > 0: - secret = utils.GenerateSecret() - if secret not in all_secrets: - break - retries -= 1 - else: - raise Error("Can't generate unique DRBD secret") - return secret - - def SetupLogging(): """Configures the logging module. @@ -142,279 +77,501 @@ def SetupLogging(): elif options.verbose: stderr_handler.setLevel(logging.INFO) else: - stderr_handler.setLevel(logging.CRITICAL) + stderr_handler.setLevel(logging.WARNING) root_logger = logging.getLogger("") root_logger.setLevel(logging.NOTSET) root_logger.addHandler(stderr_handler) -def Cluster12To20(cluster): - """Upgrades the cluster object from 1.2 to 2.0. +def CheckHostname(path): + """Ensures hostname matches ssconf value. - """ - logging.info("Upgrading the cluster object") - # Upgrade the configuration version - if 'config_version' in cluster: - del cluster['config_version'] - - # Add old ssconf keys back to config - logging.info(" - importing ssconf keys") - for key in ('master_node', 'master_ip', 'master_netdev', 'cluster_name'): - if key not in cluster: - cluster[key] = ReadFile(SsconfName(key)).strip() - - if 'default_hypervisor' not in cluster: - old_hyp = ReadFile(SsconfName('hypervisor')).strip() - if old_hyp == "xen-3.0": - hyp = "xen-pvm" - elif old_hyp == "xen-hvm-3.1": - hyp = "xen-hvm" - elif old_hyp == "fake": - hyp = "fake" - else: - raise Error("Unknown old hypervisor name '%s'" % old_hyp) + @param path: Path to ssconf file - logging.info("Setting the default and enabled hypervisor") - cluster['default_hypervisor'] = hyp - cluster['enabled_hypervisors'] = [hyp] + """ + ssconf_master_node = utils.ReadOneLineFile(path) + hostname = netutils.GetHostname().name + + if ssconf_master_node == hostname: + return True + + logging.warning("Warning: ssconf says master node is '%s', but this" + " machine's name is '%s'; this tool must be run on" + " the master node", ssconf_master_node, hostname) + return False + + +def _FillIPolicySpecs(default_ipolicy, ipolicy): + if "minmax" in ipolicy: + for (key, spec) in ipolicy["minmax"][0].items(): + for (par, val) in default_ipolicy["minmax"][0][key].items(): + if par not in spec: + spec[par] = val + + +def UpgradeIPolicy(ipolicy, default_ipolicy, isgroup): + minmax_keys = ["min", "max"] + if any((k in ipolicy) for k in minmax_keys): + minmax = {} + for key in minmax_keys: + if key in ipolicy: + if ipolicy[key]: + minmax[key] = ipolicy[key] + del ipolicy[key] + if minmax: + ipolicy["minmax"] = [minmax] + if isgroup and "std" in ipolicy: + del ipolicy["std"] + _FillIPolicySpecs(default_ipolicy, ipolicy) + + +def UpgradeNetworks(config_data): + networks = config_data.get("networks", None) + if not networks: + config_data["networks"] = {} + + +def UpgradeCluster(config_data): + cluster = config_data.get("cluster", None) + if cluster is None: + raise Error("Cannot find cluster") + ipolicy = cluster.setdefault("ipolicy", None) + if ipolicy: + UpgradeIPolicy(ipolicy, constants.IPOLICY_DEFAULTS, False) + + +def UpgradeGroups(config_data): + cl_ipolicy = config_data["cluster"].get("ipolicy") + for group in config_data["nodegroups"].values(): + networks = group.get("networks", None) + if not networks: + group["networks"] = {} + ipolicy = group.get("ipolicy", None) + if ipolicy: + if cl_ipolicy is None: + raise Error("A group defines an instance policy but there is no" + " instance policy at cluster level") + UpgradeIPolicy(ipolicy, cl_ipolicy, True) + + +def GetExclusiveStorageValue(config_data): + """Return a conservative value of the exclusive_storage flag. + + Return C{True} if the cluster or at least a nodegroup have the flag set. - # hv/be params - if 'hvparams' not in cluster: - logging.info(" - adding hvparams") - cluster['hvparams'] = constants.HVC_DEFAULTS - if 'beparams' not in cluster: - logging.info(" - adding beparams") - cluster['beparams'] = {constants.PP_DEFAULT: constants.BEC_DEFAULTS} + """ + ret = False + cluster = config_data["cluster"] + ndparams = cluster.get("ndparams") + if ndparams is not None and ndparams.get("exclusive_storage"): + ret = True + for group in config_data["nodegroups"].values(): + ndparams = group.get("ndparams") + if ndparams is not None and ndparams.get("exclusive_storage"): + ret = True + return ret + + +def UpgradeInstances(config_data): + network2uuid = dict((n["name"], n["uuid"]) + for n in config_data["networks"].values()) + if "instances" not in config_data: + raise Error("Can't find the 'instances' key in the configuration!") + + missing_spindles = False + for instance, iobj in config_data["instances"].items(): + for nic in iobj["nics"]: + name = nic.get("network", None) + if name: + uuid = network2uuid.get(name, None) + if uuid: + print("NIC with network name %s found." + " Substituting with uuid %s." % (name, uuid)) + nic["network"] = uuid + + if "disks" not in iobj: + raise Error("Instance '%s' doesn't have a disks entry?!" % instance) + disks = iobj["disks"] + for idx, dobj in enumerate(disks): + expected = "disk/%s" % idx + current = dobj.get("iv_name", "") + if current != expected: + logging.warning("Updating iv_name for instance %s/disk %s" + " from '%s' to '%s'", + instance, idx, current, expected) + dobj["iv_name"] = expected + if not "spindles" in dobj: + missing_spindles = True + + if GetExclusiveStorageValue(config_data) and missing_spindles: + # We cannot be sure that the instances that are missing spindles have + # exclusive storage enabled (the check would be more complicated), so we + # give a noncommittal message + logging.warning("Some instance disks could be needing to update the" + " spindles parameter; you can check by running" + " 'gnt-cluster verify', and fix any problem with" + " 'gnt-cluster repair-disk-sizes'") + + +def UpgradeRapiUsers(): + if (os.path.isfile(options.RAPI_USERS_FILE_PRE24) and + not os.path.islink(options.RAPI_USERS_FILE_PRE24)): + if os.path.exists(options.RAPI_USERS_FILE): + raise Error("Found pre-2.4 RAPI users file at %s, but another file" + " already exists at %s" % + (options.RAPI_USERS_FILE_PRE24, options.RAPI_USERS_FILE)) + logging.info("Found pre-2.4 RAPI users file at %s, renaming to %s", + options.RAPI_USERS_FILE_PRE24, options.RAPI_USERS_FILE) + if not options.dry_run: + utils.RenameFile(options.RAPI_USERS_FILE_PRE24, options.RAPI_USERS_FILE, + mkdir=True, mkdir_mode=0750) + + # Create a symlink for RAPI users file + if (not (os.path.islink(options.RAPI_USERS_FILE_PRE24) or + os.path.isfile(options.RAPI_USERS_FILE_PRE24)) and + os.path.isfile(options.RAPI_USERS_FILE)): + logging.info("Creating symlink from %s to %s", + options.RAPI_USERS_FILE_PRE24, options.RAPI_USERS_FILE) + if not options.dry_run: + os.symlink(options.RAPI_USERS_FILE, options.RAPI_USERS_FILE_PRE24) - # file storage - if 'file_storage_dir' not in cluster: - cluster['file_storage_dir'] = constants.DEFAULT_FILE_STORAGE_DIR - # candidate pool size - if 'candidate_pool_size' not in cluster: - cluster['candidate_pool_size'] = constants.MASTER_POOL_SIZE_DEFAULT +def UpgradeWatcher(): + # Remove old watcher state file if it exists + if os.path.exists(options.WATCHER_STATEFILE): + logging.info("Removing watcher state file %s", options.WATCHER_STATEFILE) + if not options.dry_run: + utils.RemoveFile(options.WATCHER_STATEFILE) -def Node12To20(node): - """Upgrades a node from 1.2 to 2.0. +def UpgradeFileStoragePaths(config_data): + # Write file storage paths + if not os.path.exists(options.FILE_STORAGE_PATHS_FILE): + cluster = config_data["cluster"] + file_storage_dir = cluster.get("file_storage_dir") + shared_file_storage_dir = cluster.get("shared_file_storage_dir") + del cluster + + logging.info("Ganeti 2.7 and later only allow whitelisted directories" + " for file storage; writing existing configuration values" + " into '%s'", + options.FILE_STORAGE_PATHS_FILE) + + if file_storage_dir: + logging.info("File storage directory: %s", file_storage_dir) + if shared_file_storage_dir: + logging.info("Shared file storage directory: %s", + shared_file_storage_dir) + + buf = StringIO() + buf.write("# List automatically generated from configuration by\n") + buf.write("# cfgupgrade at %s\n" % time.asctime()) + if file_storage_dir: + buf.write("%s\n" % file_storage_dir) + if shared_file_storage_dir: + buf.write("%s\n" % shared_file_storage_dir) + utils.WriteFile(file_name=options.FILE_STORAGE_PATHS_FILE, + data=buf.getvalue(), + mode=0600, + dry_run=options.dry_run, + backup=True) + + +def GetNewNodeIndex(nodes_by_old_key, old_key, new_key_field): + if old_key not in nodes_by_old_key: + logging.warning("Can't find node '%s' in configuration, assuming that it's" + " already up-to-date", old_key) + return old_key + return nodes_by_old_key[old_key][new_key_field] + + +def ChangeNodeIndices(config_data, old_key_field, new_key_field): + def ChangeDiskNodeIndices(disk): + if disk["dev_type"] in constants.LDS_DRBD: + for i in range(0, 2): + disk["logical_id"][i] = GetNewNodeIndex(nodes_by_old_key, + disk["logical_id"][i], + new_key_field) + if "children" in disk: + for child in disk["children"]: + ChangeDiskNodeIndices(child) + + nodes_by_old_key = {} + nodes_by_new_key = {} + for (_, node) in config_data["nodes"].items(): + nodes_by_old_key[node[old_key_field]] = node + nodes_by_new_key[node[new_key_field]] = node + + config_data["nodes"] = nodes_by_new_key + + cluster = config_data["cluster"] + cluster["master_node"] = GetNewNodeIndex(nodes_by_old_key, + cluster["master_node"], + new_key_field) + + for inst in config_data["instances"].values(): + inst["primary_node"] = GetNewNodeIndex(nodes_by_old_key, + inst["primary_node"], + new_key_field) + for disk in inst["disks"]: + ChangeDiskNodeIndices(disk) + + +def ChangeInstanceIndices(config_data, old_key_field, new_key_field): + insts_by_old_key = {} + insts_by_new_key = {} + for (_, inst) in config_data["instances"].items(): + insts_by_old_key[inst[old_key_field]] = inst + insts_by_new_key[inst[new_key_field]] = inst + + config_data["instances"] = insts_by_new_key + + +def UpgradeNodeIndices(config_data): + ChangeNodeIndices(config_data, "name", "uuid") + + +def UpgradeInstanceIndices(config_data): + ChangeInstanceIndices(config_data, "name", "uuid") + + +def UpgradeAll(config_data): + config_data["version"] = constants.BuildVersion(TARGET_MAJOR, + TARGET_MINOR, 0) + UpgradeRapiUsers() + UpgradeWatcher() + UpgradeFileStoragePaths(config_data) + UpgradeNetworks(config_data) + UpgradeCluster(config_data) + UpgradeGroups(config_data) + UpgradeInstances(config_data) + UpgradeNodeIndices(config_data) + UpgradeInstanceIndices(config_data) + + +def DowngradeDisks(disks, owner): + for disk in disks: + # Remove spindles to downgrade to 2.8 + if "spindles" in disk: + logging.warning("Removing spindles (value=%s) from disk %s (%s) of" + " instance %s", + disk["spindles"], disk["iv_name"], disk["uuid"], owner) + del disk["spindles"] + + +def DowngradeInstances(config_data): + if "instances" not in config_data: + raise Error("Cannot find the 'instances' key in the configuration!") + for (iname, iobj) in config_data["instances"].items(): + if "disks" not in iobj: + raise Error("Cannot find 'disks' key for instance %s" % iname) + DowngradeDisks(iobj["disks"], iname) + - """ - logging.info("Upgrading node %s" % node['name']) - if F_SERIAL not in node: - node[F_SERIAL] = 1 - if 'master_candidate' not in node: - node['master_candidate'] = True - for key in 'offline', 'drained': - if key not in node: - node[key] = False +def DowngradeNodeIndices(config_data): + ChangeNodeIndices(config_data, "uuid", "name") -def Instance12To20(drbd_minors, secrets, hypervisor, instance): - """Upgrades an instance from 1.2 to 2.0. +def DowngradeInstanceIndices(config_data): + ChangeInstanceIndices(config_data, "uuid", "name") - """ - if F_SERIAL not in instance: - instance[F_SERIAL] = 1 - - if 'hypervisor' not in instance: - instance['hypervisor'] = hypervisor - - # hvparams changes - if 'hvparams' not in instance: - instance['hvparams'] = hvp = {} - for old, new in INST_HV_CHG.items(): - if old in instance: - if (instance[old] is not None and - instance[old] != constants.VALUE_DEFAULT and # no longer valid in 2.0 - new in constants.HVC_DEFAULTS[hypervisor]): - hvp[new] = instance[old] - del instance[old] - - # beparams changes - if 'beparams' not in instance: - instance['beparams'] = bep = {} - for old, new in INST_BE_CHG.items(): - if old in instance: - if instance[old] is not None: - bep[new] = instance[old] - del instance[old] - - # disk changes - for disk in instance['disks']: - Disk12To20(drbd_minors, secrets, disk) - - # other instance changes - if 'status' in instance: - instance['admin_up'] = instance['status'] == 'up' - del instance['status'] - - -def Disk12To20(drbd_minors, secrets, disk): - """Upgrades a disk from 1.2 to 2.0. - """ - if 'mode' not in disk: - disk['mode'] = constants.DISK_RDWR - if disk['dev_type'] == constants.LD_DRBD8: - old_lid = disk['logical_id'] - for node in old_lid[:2]: - if node not in drbd_minors: - raise Error("Can't find node '%s' while upgrading disk" % node) - drbd_minors[node] += 1 - minor = drbd_minors[node] - old_lid.append(minor) - old_lid.append(GenerateSecret(secrets)) - del disk['physical_id'] - if disk['children']: - for child in disk['children']: - Disk12To20(drbd_minors, secrets, child) +def DowngradeAll(config_data): + # Any code specific to a particular version should be labeled that way, so + # it can be removed when updating to the next version. + config_data["version"] = constants.BuildVersion(DOWNGRADE_MAJOR, + DOWNGRADE_MINOR, 0) + DowngradeInstances(config_data) + DowngradeNodeIndices(config_data) + DowngradeInstanceIndices(config_data) def main(): """Main program. """ - global options, args - - program = os.path.basename(sys.argv[0]) + global options, args # pylint: disable=W0603 # Option parsing parser = optparse.OptionParser(usage="%prog [--debug|--verbose] [--force]") - parser.add_option('--dry-run', dest='dry_run', + parser.add_option("--dry-run", dest="dry_run", action="store_true", help="Try to do the conversion, but don't write" " output file") parser.add_option(cli.FORCE_OPT) parser.add_option(cli.DEBUG_OPT) parser.add_option(cli.VERBOSE_OPT) - parser.add_option('--path', help="Convert configuration in this" - " directory instead of '%s'" % constants.DATA_DIR, - default=constants.DATA_DIR, dest="data_dir") + parser.add_option("--ignore-hostname", dest="ignore_hostname", + action="store_true", default=False, + help="Don't abort if hostname doesn't match") + parser.add_option("--path", help="Convert configuration in this" + " directory instead of '%s'" % pathutils.DATA_DIR, + default=pathutils.DATA_DIR, dest="data_dir") + parser.add_option("--confdir", + help=("Use this directory instead of '%s'" % + pathutils.CONF_DIR), + default=pathutils.CONF_DIR, dest="conf_dir") + parser.add_option("--no-verify", + help="Do not verify configuration after upgrade", + action="store_true", dest="no_verify", default=False) + parser.add_option("--downgrade", + help="Downgrade to the previous stable version", + action="store_true", dest="downgrade", default=False) (options, args) = parser.parse_args() # We need to keep filenames locally because they might be renamed between # versions. + options.data_dir = os.path.abspath(options.data_dir) options.CONFIG_DATA_PATH = options.data_dir + "/config.data" options.SERVER_PEM_PATH = options.data_dir + "/server.pem" options.KNOWN_HOSTS_PATH = options.data_dir + "/known_hosts" options.RAPI_CERT_FILE = options.data_dir + "/rapi.pem" + options.SPICE_CERT_FILE = options.data_dir + "/spice.pem" + options.SPICE_CACERT_FILE = options.data_dir + "/spice-ca.pem" + options.RAPI_USERS_FILE = options.data_dir + "/rapi/users" + options.RAPI_USERS_FILE_PRE24 = options.data_dir + "/rapi_users" + options.CONFD_HMAC_KEY = options.data_dir + "/hmac.key" + options.CDS_FILE = options.data_dir + "/cluster-domain-secret" + options.SSCONF_MASTER_NODE = options.data_dir + "/ssconf_master_node" + options.WATCHER_STATEFILE = options.data_dir + "/watcher.data" + options.FILE_STORAGE_PATHS_FILE = options.conf_dir + "/file-storage-paths" SetupLogging() # Option checking if args: raise Error("No arguments expected") + if options.downgrade and not options.no_verify: + options.no_verify = True + + # Check master name + if not (CheckHostname(options.SSCONF_MASTER_NODE) or options.ignore_hostname): + logging.error("Aborting due to hostname mismatch") + sys.exit(constants.EXIT_FAILURE) if not options.force: - usertext = ("%s MUST be run on the master node. Is this the master" - " node and are ALL instances down?" % program) + if options.downgrade: + usertext = ("The configuration is going to be DOWNGRADED to version %s.%s" + " Some configuration data might be removed if they don't fit" + " in the old format. Please make sure you have read the" + " upgrade notes (available in the UPGRADE file and included" + " in other documentation formats) to understand what they" + " are. Continue with *DOWNGRADING* the configuration?" % + (DOWNGRADE_MAJOR, DOWNGRADE_MINOR)) + else: + usertext = ("Please make sure you have read the upgrade notes for" + " Ganeti %s (available in the UPGRADE file and included" + " in other documentation formats). Continue with upgrading" + " configuration?" % constants.RELEASE_VERSION) if not cli.AskUser(usertext): - sys.exit(1) + sys.exit(constants.EXIT_FAILURE) # Check whether it's a Ganeti configuration directory if not (os.path.isfile(options.CONFIG_DATA_PATH) and - os.path.isfile(options.SERVER_PEM_PATH) or + os.path.isfile(options.SERVER_PEM_PATH) and os.path.isfile(options.KNOWN_HOSTS_PATH)): - raise Error(("%s does not seem to be a known Ganeti configuration" + raise Error(("%s does not seem to be a Ganeti configuration" " directory") % options.data_dir) - config_version = ReadFile(SsconfName('config_version'), "1.2").strip() - logging.info("Found configuration version %s", config_version) - - config_data = serializer.LoadJson(ReadFile(options.CONFIG_DATA_PATH)) + if not os.path.isdir(options.conf_dir): + raise Error("Not a directory: %s" % options.conf_dir) - # Ganeti 1.2? - if config_version == "1.2": - logging.info("Found a Ganeti 1.2 configuration") + config_data = serializer.LoadJson(utils.ReadFile(options.CONFIG_DATA_PATH)) - cluster = config_data["cluster"] - - old_config_version = cluster.get("config_version", None) - logging.info("Found old configuration version %s", old_config_version) - if old_config_version not in (3, ): - raise Error("Unsupported configuration version: %s" % - old_config_version) - if 'version' not in config_data: - config_data['version'] = constants.BuildVersion(2, 0, 0) - if F_SERIAL not in config_data: - config_data[F_SERIAL] = 1 - - # Make sure no instance uses remote_raid1 anymore - remote_raid1_instances = [] - for instance in config_data["instances"].values(): - if instance["disk_template"] == "remote_raid1": - remote_raid1_instances.append(instance["name"]) - if remote_raid1_instances: - for name in remote_raid1_instances: - logging.error("Instance %s still using remote_raid1 disk template") - raise Error("Unable to convert configuration as long as there are" - " instances using remote_raid1 disk template") - - # Build content of new known_hosts file - cluster_name = ReadFile(SsconfName('cluster_name')).rstrip() - cluster_key = cluster['rsahostkeypub'] - known_hosts = "%s ssh-rsa %s\n" % (cluster_name, cluster_key) - - Cluster12To20(cluster) - - # Add node attributes - logging.info("Upgrading nodes") - # stable-sort the names to have repeatable runs - for node_name in utils.NiceSort(config_data['nodes'].keys()): - Node12To20(config_data['nodes'][node_name]) - - # Instance changes - logging.info("Upgrading instances") - drbd_minors = dict.fromkeys(config_data['nodes'], 0) - secrets = set() - # stable-sort the names to have repeatable runs - for instance_name in utils.NiceSort(config_data['instances'].keys()): - Instance12To20(drbd_minors, secrets, cluster['default_hypervisor'], - config_data['instances'][instance_name]) + try: + config_version = config_data["version"] + except KeyError: + raise Error("Unable to determine configuration version") + + (config_major, config_minor, config_revision) = \ + constants.SplitVersion(config_version) + + logging.info("Found configuration version %s (%d.%d.%d)", + config_version, config_major, config_minor, config_revision) + + if "config_version" in config_data["cluster"]: + raise Error("Inconsistent configuration: found config_version in" + " configuration file") + + # Downgrade to the previous stable version + if options.downgrade: + if not ((config_major == TARGET_MAJOR and config_minor == TARGET_MINOR) or + (config_major == DOWNGRADE_MAJOR and + config_minor == DOWNGRADE_MINOR)): + raise Error("Downgrade supported only from the latest version (%s.%s)," + " found %s (%s.%s.%s) instead" % + (TARGET_MAJOR, TARGET_MINOR, config_version, config_major, + config_minor, config_revision)) + DowngradeAll(config_data) + + # Upgrade from 2.{0..7} to 2.9 + elif config_major == 2 and config_minor in range(0, 10): + if config_revision != 0: + logging.warning("Config revision is %s, not 0", config_revision) + UpgradeAll(config_data) + + elif config_major == TARGET_MAJOR and config_minor == TARGET_MINOR: + logging.info("No changes necessary") else: - logging.info("Found a Ganeti 2.0 configuration") - - if "config_version" in config_data["cluster"]: - raise Error("Inconsistent configuration: found config_data in" - " configuration file") - - known_hosts = None + raise Error("Configuration version %d.%d.%d not supported by this tool" % + (config_major, config_minor, config_revision)) try: - logging.info("Writing configuration file") - WriteFile(options.CONFIG_DATA_PATH, serializer.DumpJson(config_data)) - - if known_hosts is not None: - logging.info("Writing SSH known_hosts file (%s)", known_hosts.strip()) - WriteFile(options.KNOWN_HOSTS_PATH, known_hosts) + logging.info("Writing configuration file to %s", options.CONFIG_DATA_PATH) + utils.WriteFile(file_name=options.CONFIG_DATA_PATH, + data=serializer.DumpJson(config_data), + mode=0600, + dry_run=options.dry_run, + backup=True) if not options.dry_run: - if not os.path.exists(options.RAPI_CERT_FILE): - bootstrap._GenerateSelfSignedSslCert(options.RAPI_CERT_FILE) - - except: - logging.critical("Writing configuration failed. It is proably in an" + bootstrap.GenerateClusterCrypto( + False, False, False, False, False, + nodecert_file=options.SERVER_PEM_PATH, + rapicert_file=options.RAPI_CERT_FILE, + spicecert_file=options.SPICE_CERT_FILE, + spicecacert_file=options.SPICE_CACERT_FILE, + hmackey_file=options.CONFD_HMAC_KEY, + cds_file=options.CDS_FILE) + + except Exception: + logging.critical("Writing configuration failed. It is probably in an" " inconsistent state and needs manual intervention.") raise # test loading the config file - if not options.dry_run: + all_ok = True + if not (options.dry_run or options.no_verify): logging.info("Testing the new config file...") cfg = config.ConfigWriter(cfg_file=options.CONFIG_DATA_PATH, + accept_foreign=options.ignore_hostname, offline=True) # if we reached this, it's all fine vrfy = cfg.VerifyConfig() if vrfy: logging.error("Errors after conversion:") for item in vrfy: - logging.error(" - %s" % item) + logging.error(" - %s", item) + all_ok = False + else: + logging.info("File loaded successfully after upgrading") del cfg - logging.info("File loaded successfully") + + if options.downgrade: + action = "downgraded" + out_ver = "%s.%s" % (DOWNGRADE_MAJOR, DOWNGRADE_MINOR) + else: + action = "upgraded" + out_ver = constants.RELEASE_VERSION + if all_ok: + cli.ToStderr("Configuration successfully %s to version %s.", + action, out_ver) + else: + cli.ToStderr("Configuration %s to version %s, but there are errors." + "\nPlease review the file.", action, out_ver) if __name__ == "__main__": main() - -# vim: set foldmethod=marker :