from ganeti import utils
from ganeti import cli
from ganeti import bootstrap
+from ganeti import config
-# We need to keep filenames locally because they might be renamed between
-# versions.
-CONFIG_DATA_PATH = constants.DATA_DIR + "/config.data"
-SERVER_PEM_PATH = constants.DATA_DIR + "/server.pem"
-KNOWN_HOSTS_PATH = constants.DATA_DIR + "/known_hosts"
-SSCONF_CLUSTER_NAME_PATH = constants.DATA_DIR + "/ssconf_cluster_name"
-SSCONF_CONFIG_VERSION_PATH = constants.DATA_DIR + "/ssconf_config_version"
-
options = None
args = None
# Unique object to identify calls without default value
NoDefault = object()
+# Dictionary with instance old keys, and new hypervisor keys
+INST_HV_CHG = {
+ 'hvm_pae': constants.HV_PAE,
+ 'vnc_bind_address': constants.HV_VNC_BIND_ADDRESS,
+ 'initrd_path': constants.HV_INITRD_PATH,
+ 'hvm_nic_type': constants.HV_NIC_TYPE,
+ 'kernel_path': constants.HV_KERNEL_PATH,
+ 'hvm_acpi': constants.HV_ACPI,
+ 'hvm_cdrom_image_path': constants.HV_CDROM_IMAGE_PATH,
+ 'hvm_boot_order': constants.HV_BOOT_ORDER,
+ 'hvm_disk_type': constants.HV_DISK_TYPE,
+ }
+
+# Instance beparams changes
+INST_BE_CHG = {
+ 'vcpus': constants.BE_VCPUS,
+ 'memory': constants.BE_MEMORY,
+ 'auto_balance': constants.BE_AUTO_BALANCE,
+ }
class Error(Exception):
"""Generic exception"""
"""Returns the file name of an (old) ssconf key.
"""
- return "%s/ssconf_%s" % (constants.DATA_DIR, key)
+ return "%s/ssconf_%s" % (options.data_dir, key)
def ReadFile(file_name, default=NoDefault):
dry_run=options.dry_run, backup=True)
+def GenerateSecret(all_secrets):
+ """Generate an unique DRBD secret.
+
+ This is a copy from ConfigWriter.
+
+ """
+ retries = 64
+ while retries > 0:
+ secret = utils.GenerateSecret()
+ if secret not in all_secrets:
+ break
+ retries -= 1
+ else:
+ raise Error("Can't generate unique DRBD secret")
+ return secret
+
+
def SetupLogging():
"""Configures the logging module.
if 'file_storage_dir' not in cluster:
cluster['file_storage_dir'] = constants.DEFAULT_FILE_STORAGE_DIR
+ # candidate pool size
+ if 'candidate_pool_size' not in cluster:
+ cluster['candidate_pool_size'] = constants.MASTER_POOL_SIZE_DEFAULT
+
def Node12To20(node):
"""Upgrades a node from 1.2 to 2.0.
node[key] = False
+def Instance12To20(drbd_minors, secrets, hypervisor, instance):
+ """Upgrades an instance from 1.2 to 2.0.
+
+ """
+ if 'hypervisor' not in instance:
+ instance['hypervisor'] = hypervisor
+
+ # hvparams changes
+ if 'hvparams' not in instance:
+ instance['hvparams'] = hvp = {}
+ for old, new in INST_HV_CHG.items():
+ if old in instance:
+ if (instance[old] is not None and
+ instance[old] != constants.VALUE_DEFAULT and # no longer valid in 2.0
+ new in constants.HVC_DEFAULTS[hypervisor]):
+ hvp[new] = instance[old]
+ del instance[old]
+
+ # beparams changes
+ if 'beparams' not in instance:
+ instance['beparams'] = bep = {}
+ for old, new in INST_BE_CHG.items():
+ if old in instance:
+ if instance[old] is not None:
+ bep[new] = instance[old]
+ del instance[old]
+
+ # disk changes
+ for disk in instance['disks']:
+ Disk12To20(drbd_minors, secrets, disk)
+
+ # other instance changes
+ if 'status' in instance:
+ instance['admin_up'] = instance['status'] == 'up'
+ del instance['status']
+
+
+def Disk12To20(drbd_minors, secrets, disk):
+ """Upgrades a disk from 1.2 to 2.0.
+
+ """
+ if 'mode' not in disk:
+ disk['mode'] = constants.DISK_RDWR
+ if disk['dev_type'] == constants.LD_DRBD8:
+ old_lid = disk['logical_id']
+ for node in old_lid[:2]:
+ if node not in drbd_minors:
+ raise Error("Can't find node '%s' while upgrading disk" % node)
+ drbd_minors[node] += 1
+ minor = drbd_minors[node]
+ old_lid.append(minor)
+ old_lid.append(GenerateSecret(secrets))
+ del disk['physical_id']
+ if disk['children']:
+ for child in disk['children']:
+ Disk12To20(drbd_minors, secrets, child)
+
+
def main():
"""Main program.
parser.add_option('-v', '--verbose', dest='verbose',
action="store_true",
help="Verbose output")
+ parser.add_option('--path', help="Convert configuration in this"
+ " directory instead of '%s'" % constants.DATA_DIR,
+ default=constants.DATA_DIR, dest="data_dir")
(options, args) = parser.parse_args()
+ # We need to keep filenames locally because they might be renamed between
+ # versions.
+ options.CONFIG_DATA_PATH = options.data_dir + "/config.data"
+ options.SERVER_PEM_PATH = options.data_dir + "/server.pem"
+ options.KNOWN_HOSTS_PATH = options.data_dir + "/known_hosts"
+ options.RAPI_CERT_FILE = options.data_dir + "/rapi.pem"
+
SetupLogging()
# Option checking
sys.exit(1)
# Check whether it's a Ganeti configuration directory
- if not (os.path.isfile(CONFIG_DATA_PATH) and
- os.path.isfile(SERVER_PEM_PATH) or
- os.path.isfile(KNOWN_HOSTS_PATH)):
+ if not (os.path.isfile(options.CONFIG_DATA_PATH) and
+ os.path.isfile(options.SERVER_PEM_PATH) or
+ os.path.isfile(options.KNOWN_HOSTS_PATH)):
raise Error(("%s does not seem to be a known Ganeti configuration"
- " directory") % constants.DATA_DIR)
+ " directory") % options.data_dir)
- config_version = ReadFile(SSCONF_CONFIG_VERSION_PATH, "1.2").strip()
+ config_version = ReadFile(SsconfName('config_version'), "1.2").strip()
logging.info("Found configuration version %s", config_version)
- config_data = serializer.LoadJson(ReadFile(CONFIG_DATA_PATH))
+ config_data = serializer.LoadJson(ReadFile(options.CONFIG_DATA_PATH))
# Ganeti 1.2?
if config_version == "1.2":
# Make sure no instance uses remote_raid1 anymore
remote_raid1_instances = []
- for instance in config_data["instances"]:
+ for instance in config_data["instances"].values():
if instance["disk_template"] == "remote_raid1":
remote_raid1_instances.append(instance["name"])
if remote_raid1_instances:
" instances using remote_raid1 disk template")
# Build content of new known_hosts file
- cluster_name = ReadFile(SSCONF_CLUSTER_NAME_PATH).rstrip()
+ cluster_name = ReadFile(SsconfName('cluster_name')).rstrip()
cluster_key = cluster['rsahostkeypub']
known_hosts = "%s ssh-rsa %s\n" % (cluster_name, cluster_key)
for node_name in utils.NiceSort(config_data['nodes'].keys()):
Node12To20(config_data['nodes'][node_name])
- # instance changes
- # TODO: add instance upgrade
- for instance in config_data['instances'].values():
- pass
+ # Instance changes
+ logging.info("Upgrading instances")
+ drbd_minors = dict.fromkeys(config_data['nodes'], 0)
+ secrets = set()
+ # stable-sort the names to have repeatable runs
+ for instance_name in utils.NiceSort(config_data['instances'].keys()):
+ Instance12To20(drbd_minors, secrets, cluster['default_hypervisor'],
+ config_data['instances'][instance_name])
else:
logging.info("Found a Ganeti 2.0 configuration")
try:
logging.info("Writing configuration file")
- WriteFile(CONFIG_DATA_PATH, serializer.DumpJson(config_data))
+ WriteFile(options.CONFIG_DATA_PATH, serializer.DumpJson(config_data))
if known_hosts is not None:
logging.info("Writing SSH known_hosts file (%s)", known_hosts.strip())
- WriteFile(KNOWN_HOSTS_PATH, known_hosts)
+ WriteFile(options.KNOWN_HOSTS_PATH, known_hosts)
if not options.dry_run:
- if not os.path.exists(constants.RAPI_CERT_FILE):
- bootstrap._GenerateSelfSignedSslCert(constants.RAPI_CERT_FILE)
+ if not os.path.exists(options.RAPI_CERT_FILE):
+ bootstrap._GenerateSelfSignedSslCert(options.RAPI_CERT_FILE)
except:
logging.critical("Writing configuration failed. It is proably in an"
" inconsistent state and needs manual intervention.")
raise
+ # test loading the config file
+ if not options.dry_run:
+ logging.info("Testing the new config file...")
+ cfg = config.ConfigWriter(cfg_file=options.CONFIG_DATA_PATH,
+ offline=True)
+ # if we reached this, it's all fine
+ vrfy = cfg.VerifyConfig()
+ if vrfy:
+ logging.error("Errors after conversion:")
+ for item in vrfy:
+ logging.error(" - %s" % item)
+ del cfg
+ logging.info("File loaded successfully")
+
if __name__ == "__main__":
main()