X-Git-Url: https://code.grnet.gr/git/ganeti-local/blobdiff_plain/6f287cf3c893e170a92957c6cf640248532d79dd..d0d1f2c459be0e90183f474b8fff708fd4cf3a99:/lib/cli.py diff --git a/lib/cli.py b/lib/cli.py index f81f1aa..8e8402c 100644 --- a/lib/cli.py +++ b/lib/cli.py @@ -1,7 +1,7 @@ # # -# Copyright (C) 2006, 2007, 2008, 2009, 2010, 2011, 2012 Google Inc. +# Copyright (C) 2006, 2007, 2008, 2009, 2010, 2011, 2012, 2013 Google Inc. # # This program is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by @@ -44,6 +44,7 @@ from ganeti import compat from ganeti import netutils from ganeti import qlang from ganeti import objects +from ganeti import pathutils from optparse import (OptionParser, TitledHelpFormatter, Option, OptionValueError) @@ -53,6 +54,7 @@ __all__ = [ # Command line options "ABSOLUTE_OPT", "ADD_UIDS_OPT", + "ADD_RESERVED_IPS_OPT", "ALLOCATABLE_OPT", "ALLOC_POLICY_OPT", "ALL_OPT", @@ -79,16 +81,21 @@ __all__ = [ "DST_NODE_OPT", "EARLY_RELEASE_OPT", "ENABLED_HV_OPT", + "ENABLED_DISK_TEMPLATES_OPT", "ERROR_CODES_OPT", + "FAILURE_ONLY_OPT", "FIELDS_OPT", "FILESTORE_DIR_OPT", "FILESTORE_DRIVER_OPT", "FORCE_FILTER_OPT", "FORCE_OPT", "FORCE_VARIANT_OPT", + "GATEWAY_OPT", + "GATEWAY6_OPT", "GLOBAL_FILEDIR_OPT", "HID_OS_OPT", "GLOBAL_SHARED_FILEDIR_OPT", + "HOTPLUG_OPT", "HVLIST_OPT", "HVOPTS_OPT", "HYPERVISOR_OPT", @@ -102,6 +109,7 @@ __all__ = [ "IGNORE_REMOVE_FAILURES_OPT", "IGNORE_SECONDARIES_OPT", "IGNORE_SIZE_OPT", + "INCLUDEDEFAULTS_OPT", "INTERVAL_OPT", "MAC_PREFIX_OPT", "MAINTAIN_NODE_HEALTH_OPT", @@ -109,14 +117,19 @@ __all__ = [ "MASTER_NETMASK_OPT", "MC_OPT", "MIGRATION_MODE_OPT", + "MODIFY_ETCHOSTS_OPT", "NET_OPT", + "NETWORK_OPT", + "NETWORK6_OPT", "NEW_CLUSTER_CERT_OPT", "NEW_CLUSTER_DOMAIN_SECRET_OPT", "NEW_CONFD_HMAC_KEY_OPT", "NEW_RAPI_CERT_OPT", + "NEW_PRIMARY_OPT", "NEW_SECONDARY_OPT", "NEW_SPICE_CERT_OPT", "NIC_PARAMS_OPT", + "NOCONFLICTSCHECK_OPT", "NODE_FORCE_JOIN_OPT", "NODE_LIST_OPT", "NODE_PLACEMENT_OPT", @@ -157,8 +170,10 @@ __all__ = [ "PRIORITY_OPT", "RAPI_CERT_OPT", "READD_OPT", + "REASON_OPT", "REBOOT_TYPE_OPT", "REMOVE_INSTANCE_OPT", + "REMOVE_RESERVED_IPS_OPT", "REMOVE_UIDS_OPT", "RESERVED_LVS_OPT", "RUNTIME_MEM_OPT", @@ -168,6 +183,7 @@ __all__ = [ "SELECT_OS_OPT", "SEP_OPT", "SHOWCMD_OPT", + "SHOW_MACHINE_OPT", "SHUTDOWN_TIMEOUT_OPT", "SINGLE_NODE_OPT", "SPECS_CPU_COUNT_OPT", @@ -175,6 +191,8 @@ __all__ = [ "SPECS_DISK_SIZE_OPT", "SPECS_MEM_SIZE_OPT", "SPECS_NIC_COUNT_OPT", + "SPLIT_ISPECS_OPTS", + "IPOLICY_STD_SPECS_OPT", "IPOLICY_DISK_TEMPLATES", "IPOLICY_VCPU_RATIO", "SPICE_CACERT_OPT", @@ -221,7 +239,10 @@ __all__ = [ "ToStderr", "ToStdout", "FormatError", "FormatQueryResult", - "FormatParameterDict", + "FormatParamsDictInfo", + "FormatPolicyInfo", + "PrintIPolicyCommand", + "PrintGenericInfo", "GenerateTable", "AskUser", "FormatTimestamp", @@ -234,11 +255,13 @@ __all__ = [ "ARGS_MANY_INSTANCES", "ARGS_MANY_NODES", "ARGS_MANY_GROUPS", + "ARGS_MANY_NETWORKS", "ARGS_NONE", "ARGS_ONE_INSTANCE", "ARGS_ONE_NODE", "ARGS_ONE_GROUP", "ARGS_ONE_OS", + "ARGS_ONE_NETWORK", "ArgChoice", "ArgCommand", "ArgFile", @@ -246,8 +269,10 @@ __all__ = [ "ArgHost", "ArgInstance", "ArgJobId", + "ArgNetwork", "ArgNode", "ArgOs", + "ArgExtStorage", "ArgSuggest", "ArgUnknown", "OPT_COMPL_INST_ADD_NODES", @@ -256,8 +281,11 @@ __all__ = [ "OPT_COMPL_ONE_INSTANCE", "OPT_COMPL_ONE_NODE", "OPT_COMPL_ONE_NODEGROUP", + "OPT_COMPL_ONE_NETWORK", "OPT_COMPL_ONE_OS", + "OPT_COMPL_ONE_EXTSTORAGE", "cli_option", + "FixHvParams", "SplitNodeOption", "CalculateOSNames", "ParseFields", @@ -300,6 +328,17 @@ TISPECS_CLUSTER_TYPES = { constants.ISPECS_STD: constants.VTYPE_INT, } +#: User-friendly names for query2 field types +_QFT_NAMES = { + constants.QFT_UNKNOWN: "Unknown", + constants.QFT_TEXT: "Text", + constants.QFT_BOOL: "Boolean", + constants.QFT_NUMBER: "Number", + constants.QFT_UNIT: "Storage size", + constants.QFT_TIMESTAMP: "Timestamp", + constants.QFT_OTHER: "Custom", + } + class _Argument: def __init__(self, min=0, max=None): # pylint: disable=W0622 @@ -354,6 +393,12 @@ class ArgNode(_Argument): """ +class ArgNetwork(_Argument): + """Network argument. + + """ + + class ArgGroup(_Argument): """Node group argument. @@ -390,11 +435,19 @@ class ArgOs(_Argument): """ +class ArgExtStorage(_Argument): + """ExtStorage argument. + + """ + + ARGS_NONE = [] ARGS_MANY_INSTANCES = [ArgInstance()] +ARGS_MANY_NETWORKS = [ArgNetwork()] ARGS_MANY_NODES = [ArgNode()] ARGS_MANY_GROUPS = [ArgGroup()] ARGS_ONE_INSTANCE = [ArgInstance(min=1, max=1)] +ARGS_ONE_NETWORK = [ArgNetwork(min=1, max=1)] ARGS_ONE_NODE = [ArgNode(min=1, max=1)] # TODO ARGS_ONE_GROUP = [ArgGroup(min=1, max=1)] @@ -411,9 +464,10 @@ def _ExtractTagsObject(opts, args): raise errors.ProgrammerError("tag_type not passed to _ExtractTagsObject") kind = opts.tag_type if kind == constants.TAG_CLUSTER: - retval = kind, kind + retval = kind, None elif kind in (constants.TAG_NODEGROUP, constants.TAG_NODE, + constants.TAG_NETWORK, constants.TAG_INSTANCE): if not args: raise errors.OpPrereqError("no arguments passed to the command", @@ -464,7 +518,7 @@ def ListTags(opts, args): """ kind, name = _ExtractTagsObject(opts, args) - cl = GetClient() + cl = GetClient(query=True) result = cl.QueryTags(kind, name) result = list(result) result.sort() @@ -516,12 +570,13 @@ def check_unit(option, opt, value): # pylint: disable=W0613 raise OptionValueError("option %s: %s" % (opt, err)) -def _SplitKeyVal(opt, data): +def _SplitKeyVal(opt, data, parse_prefixes): """Convert a KeyVal string into a dict. This function will convert a key=val[,...] string into a dict. Empty values will be converted specially: keys which have the prefix 'no_' - will have the value=False and the prefix stripped, the others will + will have the value=False and the prefix stripped, keys with the prefix + "-" will have value=None and the prefix stripped, and the others will have value=True. @type opt: string @@ -529,6 +584,8 @@ def _SplitKeyVal(opt, data): data, used in building error messages @type data: string @param data: a string of the format key=val,key=val,... + @type parse_prefixes: bool + @param parse_prefixes: whether to handle prefixes specially @rtype: dict @return: {key=val, key=val} @raises errors.ParameterError: if there are duplicate keys @@ -539,13 +596,16 @@ def _SplitKeyVal(opt, data): for elem in utils.UnescapeAndSplit(data, sep=","): if "=" in elem: key, val = elem.split("=", 1) - else: + elif parse_prefixes: if elem.startswith(NO_PREFIX): key, val = elem[len(NO_PREFIX):], False elif elem.startswith(UN_PREFIX): key, val = elem[len(UN_PREFIX):], None else: key, val = elem, True + else: + raise errors.ParameterError("Missing value for key '%s' in option %s" % + (elem, opt)) if key in kv_dict: raise errors.ParameterError("Duplicate key '%s' in option %s" % (key, opt)) @@ -553,11 +613,19 @@ def _SplitKeyVal(opt, data): return kv_dict -def check_ident_key_val(option, opt, value): # pylint: disable=W0613 - """Custom parser for ident:key=val,key=val options. +def _SplitIdentKeyVal(opt, value, parse_prefixes): + """Helper function to parse "ident:key=val,key=val" options. - This will store the parsed values as a tuple (ident, {key: val}). As such, - multiple uses of this option via action=append is possible. + @type opt: string + @param opt: option name, used in error messages + @type value: string + @param value: expected to be in the format "ident:key=val,key=val,..." + @type parse_prefixes: bool + @param parse_prefixes: whether to handle prefixes specially (see + L{_SplitKeyVal}) + @rtype: tuple + @return: (ident, {key=val, key=val}) + @raises errors.ParameterError: in case of duplicates or other parsing errors """ if ":" not in value: @@ -565,31 +633,67 @@ def check_ident_key_val(option, opt, value): # pylint: disable=W0613 else: ident, rest = value.split(":", 1) - if ident.startswith(NO_PREFIX): + if parse_prefixes and ident.startswith(NO_PREFIX): if rest: msg = "Cannot pass options when removing parameter groups: %s" % value raise errors.ParameterError(msg) retval = (ident[len(NO_PREFIX):], False) - elif (ident.startswith(UN_PREFIX) and - (len(ident) <= len(UN_PREFIX) or - not ident[len(UN_PREFIX)][0].isdigit())): + elif (parse_prefixes and ident.startswith(UN_PREFIX) and + (len(ident) <= len(UN_PREFIX) or not ident[len(UN_PREFIX)].isdigit())): if rest: msg = "Cannot pass options when removing parameter groups: %s" % value raise errors.ParameterError(msg) retval = (ident[len(UN_PREFIX):], None) else: - kv_dict = _SplitKeyVal(opt, rest) + kv_dict = _SplitKeyVal(opt, rest, parse_prefixes) retval = (ident, kv_dict) return retval +def check_ident_key_val(option, opt, value): # pylint: disable=W0613 + """Custom parser for ident:key=val,key=val options. + + This will store the parsed values as a tuple (ident, {key: val}). As such, + multiple uses of this option via action=append is possible. + + """ + return _SplitIdentKeyVal(opt, value, True) + + def check_key_val(option, opt, value): # pylint: disable=W0613 """Custom parser class for key=val,key=val options. This will store the parsed values as a dict {key: val}. """ - return _SplitKeyVal(opt, value) + return _SplitKeyVal(opt, value, True) + + +def _SplitListKeyVal(opt, value): + retval = {} + for elem in value.split("/"): + if not elem: + raise errors.ParameterError("Empty section in option '%s'" % opt) + (ident, valdict) = _SplitIdentKeyVal(opt, elem, False) + if ident in retval: + msg = ("Duplicated parameter '%s' in parsing %s: %s" % + (ident, opt, elem)) + raise errors.ParameterError(msg) + retval[ident] = valdict + return retval + + +def check_multilist_ident_key_val(_, opt, value): + """Custom parser for "ident:key=val,key=val/ident:key=val//ident:.." options. + + @rtype: list of dictionary + @return: [{ident: {key: val, key: val}, ident: {key: val}}, {ident:..}] + + """ + retval = [] + for line in value.split("//"): + retval.append(_SplitListKeyVal(opt, line)) + return retval def check_bool(option, opt, value): # pylint: disable=W0613 @@ -637,16 +741,20 @@ def check_maybefloat(option, opt, value): # pylint: disable=W0613 OPT_COMPL_ONE_NODE, OPT_COMPL_ONE_INSTANCE, OPT_COMPL_ONE_OS, + OPT_COMPL_ONE_EXTSTORAGE, OPT_COMPL_ONE_IALLOCATOR, + OPT_COMPL_ONE_NETWORK, OPT_COMPL_INST_ADD_NODES, - OPT_COMPL_ONE_NODEGROUP) = range(100, 107) + OPT_COMPL_ONE_NODEGROUP) = range(100, 109) -OPT_COMPL_ALL = frozenset([ +OPT_COMPL_ALL = compat.UniqueFrozenset([ OPT_COMPL_MANY_NODES, OPT_COMPL_ONE_NODE, OPT_COMPL_ONE_INSTANCE, OPT_COMPL_ONE_OS, + OPT_COMPL_ONE_EXTSTORAGE, OPT_COMPL_ONE_IALLOCATOR, + OPT_COMPL_ONE_NETWORK, OPT_COMPL_INST_ADD_NODES, OPT_COMPL_ONE_NODEGROUP, ]) @@ -660,6 +768,7 @@ class CliOption(Option): "completion_suggest", ] TYPES = Option.TYPES + ( + "multilistidentkeyval", "identkeyval", "keyval", "unit", @@ -668,6 +777,7 @@ class CliOption(Option): "maybefloat", ) TYPE_CHECKER = Option.TYPE_CHECKER.copy() + TYPE_CHECKER["multilistidentkeyval"] = check_multilist_ident_key_val TYPE_CHECKER["identkeyval"] = check_ident_key_val TYPE_CHECKER["keyval"] = check_key_val TYPE_CHECKER["unit"] = check_unit @@ -733,7 +843,7 @@ SYNC_OPT = cli_option("--sync", dest="do_locking", DRY_RUN_OPT = cli_option("--dry-run", default=False, action="store_true", help=("Do not execute the operation, just run the" - " check steps and verify it it could be" + " check steps and verify if it could be" " executed")) VERBOSE_OPT = cli_option("-v", "--verbose", default=False, @@ -778,7 +888,7 @@ FILESTORE_DIR_OPT = cli_option("--file-storage-dir", dest="file_storage_dir", FILESTORE_DRIVER_OPT = cli_option("--file-driver", dest="file_driver", help="Driver to use for image files", - default="loop", metavar="", + default=None, metavar="", choices=list(constants.FILE_DRIVER)) IALLOCATOR_OPT = cli_option("-I", "--iallocator", metavar="", @@ -788,18 +898,19 @@ IALLOCATOR_OPT = cli_option("-I", "--iallocator", metavar="", completion_suggest=OPT_COMPL_ONE_IALLOCATOR) DEFAULT_IALLOCATOR_OPT = cli_option("-I", "--default-iallocator", - metavar="", - help="Set the default instance allocator plugin", - default=None, type="string", - completion_suggest=OPT_COMPL_ONE_IALLOCATOR) + metavar="", + help="Set the default instance" + " allocator plugin", + default=None, type="string", + completion_suggest=OPT_COMPL_ONE_IALLOCATOR) OS_OPT = cli_option("-o", "--os-type", dest="os", help="What OS to run", metavar="", completion_suggest=OPT_COMPL_ONE_OS) OSPARAMS_OPT = cli_option("-O", "--os-parameters", dest="osparams", - type="keyval", default={}, - help="OS parameters") + type="keyval", default={}, + help="OS parameters") FORCE_VARIANT_OPT = cli_option("--force-variant", dest="force_variant", action="store_true", default=False, @@ -848,7 +959,7 @@ SPECS_DISK_COUNT_OPT = cli_option("--specs-disk-count", SPECS_DISK_SIZE_OPT = cli_option("--specs-disk-size", dest="ispecs_disk_size", type="keyval", default={}, help="Disk size specs: list of key=value," - " where key is one of min, max, std" + " where key is one of min, max, std" " (in MB or using a unit)") SPECS_NIC_COUNT_OPT = cli_option("--specs-nic-count", dest="ispecs_nic_count", @@ -856,11 +967,23 @@ SPECS_NIC_COUNT_OPT = cli_option("--specs-nic-count", dest="ispecs_nic_count", help="NIC count specs: list of key=value," " where key is one of min, max, std") +IPOLICY_BOUNDS_SPECS_STR = "--ipolicy-bounds-specs" +IPOLICY_BOUNDS_SPECS_OPT = cli_option(IPOLICY_BOUNDS_SPECS_STR, + dest="ipolicy_bounds_specs", + type="multilistidentkeyval", default=None, + help="Complete instance specs limits") + +IPOLICY_STD_SPECS_STR = "--ipolicy-std-specs" +IPOLICY_STD_SPECS_OPT = cli_option(IPOLICY_STD_SPECS_STR, + dest="ipolicy_std_specs", + type="keyval", default=None, + help="Complte standard instance specs") + IPOLICY_DISK_TEMPLATES = cli_option("--ipolicy-disk-templates", - dest="ipolicy_disk_templates", - type="list", default=None, - help="Comma-separated list of" - " enabled disk templates") + dest="ipolicy_disk_templates", + type="list", default=None, + help="Comma-separated list of" + " enabled disk templates") IPOLICY_VCPU_RATIO = cli_option("--ipolicy-vcpu-ratio", dest="ipolicy_vcpu_ratio", @@ -967,12 +1090,12 @@ SHOWCMD_OPT = cli_option("--show-cmd", dest="show_command", CLEANUP_OPT = cli_option("--cleanup", dest="cleanup", default=False, action="store_true", - help="Instead of performing the migration, try to" - " recover from a failed cleanup. This is safe" + help="Instead of performing the migration/failover," + " try to recover from a failed cleanup. This is safe" " to run even if the instance is healthy, but it" " will create extra replication traffic and " " disrupt briefly the replication (like during the" - " migration") + " migration/failover") STATIC_OPT = cli_option("-s", "--static", dest="static", action="store_true", default=False, @@ -1016,6 +1139,11 @@ NEW_SECONDARY_OPT = cli_option("-n", "--new-secondary", dest="dst_node", metavar="NODE", default=None, completion_suggest=OPT_COMPL_ONE_NODE) +NEW_PRIMARY_OPT = cli_option("--new-primary", dest="new_primary_node", + help="Specifies the new primary node", + metavar="", default=None, + completion_suggest=OPT_COMPL_ONE_NODE) + ON_PRIMARY_OPT = cli_option("-p", "--on-primary", dest="on_primary", default=False, action="store_true", help="Replace the disk(s) on the primary" @@ -1087,12 +1215,12 @@ DRAINED_OPT = cli_option("-D", "--drained", dest="drained", metavar=_YORNO, " (excluded from allocation operations)")) CAPAB_MASTER_OPT = cli_option("--master-capable", dest="master_capable", - type="bool", default=None, metavar=_YORNO, - help="Set the master_capable flag on the node") + type="bool", default=None, metavar=_YORNO, + help="Set the master_capable flag on the node") CAPAB_VM_OPT = cli_option("--vm-capable", dest="vm_capable", - type="bool", default=None, metavar=_YORNO, - help="Set the vm_capable flag on the node") + type="bool", default=None, metavar=_YORNO, + help="Set the vm_capable flag on the node") ALLOCATABLE_OPT = cli_option("--allocatable", dest="allocatable", type="bool", default=None, metavar=_YORNO, @@ -1108,6 +1236,12 @@ ENABLED_HV_OPT = cli_option("--enabled-hypervisors", help="Comma-separated list of hypervisors", type="string", default=None) +ENABLED_DISK_TEMPLATES_OPT = cli_option("--enabled-disk-templates", + dest="enabled_disk_templates", + help="Comma-separated list of " + "disk templates", + type="string", default=None) + NIC_PARAMS_OPT = cli_option("-N", "--nic-parameters", dest="nicparams", type="keyval", default={}, help="NIC parameters") @@ -1149,32 +1283,38 @@ MASTER_NETMASK_OPT = cli_option("--master-netmask", dest="master_netmask", default=None) USE_EXTERNAL_MIP_SCRIPT = cli_option("--use-external-mip-script", - dest="use_external_mip_script", - help="Specify whether to run a user-provided" - " script for the master IP address turnup and" - " turndown operations", - type="bool", metavar=_YORNO, default=None) + dest="use_external_mip_script", + help="Specify whether to run a" + " user-provided script for the master" + " IP address turnup and" + " turndown operations", + type="bool", metavar=_YORNO, default=None) GLOBAL_FILEDIR_OPT = cli_option("--file-storage-dir", dest="file_storage_dir", help="Specify the default directory (cluster-" "wide) for storing the file-based disks [%s]" % - constants.DEFAULT_FILE_STORAGE_DIR, + pathutils.DEFAULT_FILE_STORAGE_DIR, metavar="DIR", - default=constants.DEFAULT_FILE_STORAGE_DIR) + default=pathutils.DEFAULT_FILE_STORAGE_DIR) -GLOBAL_SHARED_FILEDIR_OPT = cli_option("--shared-file-storage-dir", - dest="shared_file_storage_dir", - help="Specify the default directory (cluster-" - "wide) for storing the shared file-based" - " disks [%s]" % - constants.DEFAULT_SHARED_FILE_STORAGE_DIR, - metavar="SHAREDDIR", - default=constants.DEFAULT_SHARED_FILE_STORAGE_DIR) +GLOBAL_SHARED_FILEDIR_OPT = cli_option( + "--shared-file-storage-dir", + dest="shared_file_storage_dir", + help="Specify the default directory (cluster-wide) for storing the" + " shared file-based disks [%s]" % + pathutils.DEFAULT_SHARED_FILE_STORAGE_DIR, + metavar="SHAREDDIR", default=pathutils.DEFAULT_SHARED_FILE_STORAGE_DIR) NOMODIFY_ETCHOSTS_OPT = cli_option("--no-etc-hosts", dest="modify_etc_hosts", - help="Don't modify /etc/hosts", + help="Don't modify %s" % pathutils.ETC_HOSTS, action="store_false", default=True) +MODIFY_ETCHOSTS_OPT = \ + cli_option("--modify-etc-hosts", dest="modify_etc_hosts", metavar=_YORNO, + default=None, type="bool", + help="Defines whether the cluster should autonomously modify" + " and keep in sync the /etc/hosts file of the nodes") + NOMODIFY_SSH_SETUP_OPT = cli_option("--no-ssh-init", dest="modify_ssh_setup", help="Don't initialize SSH keys", action="store_false", default=True) @@ -1207,9 +1347,10 @@ TIMEOUT_OPT = cli_option("--timeout", dest="timeout", type="int", help="Maximum time to wait") SHUTDOWN_TIMEOUT_OPT = cli_option("--shutdown-timeout", - dest="shutdown_timeout", type="int", - default=constants.DEFAULT_SHUTDOWN_TIMEOUT, - help="Maximum time to wait for instance shutdown") + dest="shutdown_timeout", type="int", + default=constants.DEFAULT_SHUTDOWN_TIMEOUT, + help="Maximum time to wait for instance" + " shutdown") INTERVAL_OPT = cli_option("--interval", dest="interval", type="int", default=None, @@ -1237,19 +1378,19 @@ NEW_RAPI_CERT_OPT = cli_option("--new-rapi-certificate", dest="new_rapi_cert", " certificate")) SPICE_CERT_OPT = cli_option("--spice-certificate", dest="spice_cert", - default=None, - help="File containing new SPICE certificate") + default=None, + help="File containing new SPICE certificate") SPICE_CACERT_OPT = cli_option("--spice-ca-certificate", dest="spice_cacert", - default=None, - help="File containing the certificate of the CA" - " which signed the SPICE certificate") + default=None, + help="File containing the certificate of the CA" + " which signed the SPICE certificate") NEW_SPICE_CERT_OPT = cli_option("--new-spice-certificate", - dest="new_spice_cert", default=None, - action="store_true", - help=("Generate a new self-signed SPICE" - " certificate")) + dest="new_spice_cert", default=None, + action="store_true", + help=("Generate a new self-signed SPICE" + " certificate")) NEW_CONFD_HMAC_KEY_OPT = cli_option("--new-confd-hmac-key", dest="new_confd_hmac_key", @@ -1307,10 +1448,10 @@ REMOVE_UIDS_OPT = cli_option("--remove-uids", default=None, " removed from the user-id pool")) RESERVED_LVS_OPT = cli_option("--reserved-lvs", default=None, - action="store", dest="reserved_lvs", - help=("A comma-separated list of reserved" - " logical volumes names, that will be" - " ignored by cluster verify")) + action="store", dest="reserved_lvs", + help=("A comma-separated list of reserved" + " logical volumes names, that will be" + " ignored by cluster verify")) ROMAN_OPT = cli_option("--roman", dest="roman_integers", default=False, @@ -1332,9 +1473,33 @@ PRIMARY_IP_VERSION_OPT = \ constants.IP6_VERSION), help="Cluster-wide IP version for primary IP") +SHOW_MACHINE_OPT = cli_option("-M", "--show-machine-names", default=False, + action="store_true", + help="Show machine name for every line in output") + +FAILURE_ONLY_OPT = cli_option("--failure-only", default=False, + action="store_true", + help=("Hide successful results and show failures" + " only (determined by the exit code)")) + +REASON_OPT = cli_option("--reason", default=None, + help="The reason for executing the command") + + +def _PriorityOptionCb(option, _, value, parser): + """Callback for processing C{--priority} option. + + """ + value = _PRIONAME_TO_VALUE[value] + + setattr(parser.values, option.dest, value) + + PRIORITY_OPT = cli_option("--priority", default=None, dest="priority", metavar="|".join(name for name, _ in _PRIORITY_NAMES), choices=_PRIONAME_TO_VALUE.keys(), + action="callback", type="choice", + callback=_PriorityOptionCb, help="Priority for opcode processing") HID_OS_OPT = cli_option("--hidden", dest="hidden", @@ -1365,8 +1530,8 @@ NODE_POWERED_OPT = cli_option("--node-powered", default=None, help="Specify if the SoR for node is powered") OOB_TIMEOUT_OPT = cli_option("--oob-timeout", dest="oob_timeout", type="int", - default=constants.OOB_TIMEOUT, - help="Maximum time to wait for out-of-band helper") + default=constants.OOB_TIMEOUT, + help="Maximum time to wait for out-of-band helper") POWER_DELAY_OPT = cli_option("--power-delay", dest="power_delay", type="float", default=constants.OOB_POWER_DELAY, @@ -1437,8 +1602,50 @@ ABSOLUTE_OPT = cli_option("--absolute", dest="absolute", help="Marks the grow as absolute instead of the" " (default) relative mode") +NETWORK_OPT = cli_option("--network", + action="store", default=None, dest="network", + help="IP network in CIDR notation") + +GATEWAY_OPT = cli_option("--gateway", + action="store", default=None, dest="gateway", + help="IP address of the router (gateway)") + +ADD_RESERVED_IPS_OPT = cli_option("--add-reserved-ips", + action="store", default=None, + dest="add_reserved_ips", + help="Comma-separated list of" + " reserved IPs to add") + +REMOVE_RESERVED_IPS_OPT = cli_option("--remove-reserved-ips", + action="store", default=None, + dest="remove_reserved_ips", + help="Comma-delimited list of" + " reserved IPs to remove") + +NETWORK6_OPT = cli_option("--network6", + action="store", default=None, dest="network6", + help="IP network in CIDR notation") + +GATEWAY6_OPT = cli_option("--gateway6", + action="store", default=None, dest="gateway6", + help="IP6 address of the router (gateway)") + +NOCONFLICTSCHECK_OPT = cli_option("--no-conflicts-check", + dest="conflicts_check", + default=True, + action="store_false", + help="Don't check for conflicting IPs") + +INCLUDEDEFAULTS_OPT = cli_option("--include-defaults", dest="include_defaults", + default=False, action="store_true", + help="Include default values") + +HOTPLUG_OPT = cli_option("--hotplug", dest="hotplug", + action="store_true", default=False, + help="Hotplug supported devices (NICs and Disks)") + #: Options provided by all commands -COMMON_OPTS = [DEBUG_OPT] +COMMON_OPTS = [DEBUG_OPT, REASON_OPT] # common options for creating instances. add and import then add their own # specific ones. @@ -1453,6 +1660,7 @@ COMMON_CREATE_OPTS = [ NET_OPT, NODE_PLACEMENT_OPT, NOIPCHECK_OPT, + NOCONFLICTSCHECK_OPT, NONAMECHECK_OPT, NONICS_OPT, NWSYNC_OPT, @@ -1466,79 +1674,76 @@ COMMON_CREATE_OPTS = [ # common instance policy options INSTANCE_POLICY_OPTS = [ + IPOLICY_BOUNDS_SPECS_OPT, + IPOLICY_DISK_TEMPLATES, + IPOLICY_VCPU_RATIO, + IPOLICY_SPINDLE_RATIO, + ] + +# instance policy split specs options +SPLIT_ISPECS_OPTS = [ SPECS_CPU_COUNT_OPT, SPECS_DISK_COUNT_OPT, SPECS_DISK_SIZE_OPT, SPECS_MEM_SIZE_OPT, SPECS_NIC_COUNT_OPT, - IPOLICY_DISK_TEMPLATES, - IPOLICY_VCPU_RATIO, - IPOLICY_SPINDLE_RATIO, ] -def _ParseArgs(argv, commands, aliases, env_override): +class _ShowUsage(Exception): + """Exception class for L{_ParseArgs}. + + """ + def __init__(self, exit_error): + """Initializes instances of this class. + + @type exit_error: bool + @param exit_error: Whether to report failure on exit + + """ + Exception.__init__(self) + self.exit_error = exit_error + + +class _ShowVersion(Exception): + """Exception class for L{_ParseArgs}. + + """ + + +def _ParseArgs(binary, argv, commands, aliases, env_override): """Parser for the command line arguments. This function parses the arguments and returns the function which must be executed together with its (modified) arguments. - @param argv: the command line - @param commands: dictionary with special contents, see the design - doc for cmdline handling - @param aliases: dictionary with command aliases {'alias': 'target, ...} + @param binary: Script name + @param argv: Command line arguments + @param commands: Dictionary containing command definitions + @param aliases: dictionary with command aliases {"alias": "target", ...} @param env_override: list of env variables allowed for default args + @raise _ShowUsage: If usage description should be shown + @raise _ShowVersion: If version should be shown """ assert not (env_override - set(commands)) + assert not (set(aliases.keys()) & set(commands.keys())) - if len(argv) == 0: - binary = "" + if len(argv) > 1: + cmd = argv[1] else: - binary = argv[0].split("/")[-1] - - if len(argv) > 1 and argv[1] == "--version": - ToStdout("%s (ganeti %s) %s", binary, constants.VCS_VERSION, - constants.RELEASE_VERSION) - # Quit right away. That way we don't have to care about this special - # argument. optparse.py does it the same. - sys.exit(0) - - if len(argv) < 2 or not (argv[1] in commands or - argv[1] in aliases): - # let's do a nice thing - sortedcmds = commands.keys() - sortedcmds.sort() - - ToStdout("Usage: %s {command} [options...] [argument...]", binary) - ToStdout("%s --help to see details, or man %s", binary, binary) - ToStdout("") - - # compute the max line length for cmd + usage - mlen = max([len(" %s" % cmd) for cmd in commands]) - mlen = min(60, mlen) # should not get here... - - # and format a nice command list - ToStdout("Commands:") - for cmd in sortedcmds: - cmdstr = " %s" % (cmd,) - help_text = commands[cmd][4] - help_lines = textwrap.wrap(help_text, 79 - 3 - mlen) - ToStdout("%-*s - %s", mlen, cmdstr, help_lines.pop(0)) - for line in help_lines: - ToStdout("%-*s %s", mlen, "", line) - - ToStdout("") + # No option or command given + raise _ShowUsage(exit_error=True) - return None, None, None + if cmd == "--version": + raise _ShowVersion() + elif cmd == "--help": + raise _ShowUsage(exit_error=False) + elif not (cmd in commands or cmd in aliases): + raise _ShowUsage(exit_error=True) # get command, unalias it, and look it up in commands - cmd = argv.pop(1) if cmd in aliases: - if cmd in commands: - raise errors.ProgrammerError("Alias '%s' overrides an existing" - " command" % cmd) - if aliases[cmd] not in commands: raise errors.ProgrammerError("Alias '%s' maps to non-existing" " command '%s'" % (cmd, aliases[cmd])) @@ -1549,7 +1754,7 @@ def _ParseArgs(argv, commands, aliases, env_override): args_env_name = ("%s_%s" % (binary.replace("-", "_"), cmd)).upper() env_args = os.environ.get(args_env_name) if env_args: - argv = utils.InsertAtPos(argv, 1, shlex.split(env_args)) + argv = utils.InsertAtPos(argv, 2, shlex.split(env_args)) func, args_def, parser_opts, usage, description = commands[cmd] parser = OptionParser(option_list=parser_opts + COMMON_OPTS, @@ -1557,7 +1762,7 @@ def _ParseArgs(argv, commands, aliases, env_override): formatter=TitledHelpFormatter(), usage="%%prog %s %s" % (cmd, usage)) parser.disable_interspersed_args() - options, args = parser.parse_args(args=argv[1:]) + options, args = parser.parse_args(args=argv[2:]) if not _CheckArguments(cmd, args_def, args): return None, None, None @@ -1565,6 +1770,31 @@ def _ParseArgs(argv, commands, aliases, env_override): return func, options, args +def _FormatUsage(binary, commands): + """Generates a nice description of all commands. + + @param binary: Script name + @param commands: Dictionary containing command definitions + + """ + # compute the max line length for cmd + usage + mlen = min(60, max(map(len, commands))) + + yield "Usage: %s {command} [options...] [argument...]" % binary + yield "%s --help to see details, or man %s" % (binary, binary) + yield "" + yield "Commands:" + + # and format a nice command list + for (cmd, (_, _, _, _, help_text)) in sorted(commands.items()): + help_lines = textwrap.wrap(help_text, 79 - 3 - mlen) + yield " %-*s - %s" % (mlen, cmd, help_lines.pop(0)) + for line in help_lines: + yield " %-*s %s" % (mlen, "", line) + + yield "" + + def _CheckArguments(cmd, args_def, args): """Verifies the arguments using the argument definition. @@ -2059,6 +2289,31 @@ def SubmitOrSend(op, opts, cl=None, feedback_fn=None): return SubmitOpCode(op, cl=cl, feedback_fn=feedback_fn, opts=opts) +def _InitReasonTrail(op, opts): + """Builds the first part of the reason trail + + Builds the initial part of the reason trail, adding the user provided reason + (if it exists) and the name of the command starting the operation. + + @param op: the opcode the reason trail will be added to + @param opts: the command line options selected by the user + + """ + assert len(sys.argv) >= 2 + trail = [] + + if opts.reason: + trail.append((constants.OPCODE_REASON_SRC_USER, + opts.reason, + utils.EpochNano())) + + binary = os.path.basename(sys.argv[0]) + source = "%s:%s" % (constants.OPCODE_REASON_SRC_CLIENT, binary) + command = sys.argv[1] + trail.append((source, command, utils.EpochNano())) + op.reason = trail + + def SetGenericOpcodeOpts(opcode_list, options): """Processor for generic options. @@ -2077,13 +2332,35 @@ def SetGenericOpcodeOpts(opcode_list, options): if hasattr(options, "dry_run"): op.dry_run = options.dry_run if getattr(options, "priority", None) is not None: - op.priority = _PRIONAME_TO_VALUE[options.priority] + op.priority = options.priority + _InitReasonTrail(op, options) + + +def GetClient(query=False): + """Connects to the a luxi socket and returns a client. + @type query: boolean + @param query: this signifies that the client will only be + used for queries; if the build-time parameter + enable-split-queries is enabled, then the client will be + connected to the query socket instead of the masterd socket -def GetClient(): + """ + override_socket = os.getenv(constants.LUXI_OVERRIDE, "") + if override_socket: + if override_socket == constants.LUXI_OVERRIDE_MASTER: + address = pathutils.MASTER_SOCKET + elif override_socket == constants.LUXI_OVERRIDE_QUERY: + address = pathutils.QUERY_SOCKET + else: + address = override_socket + elif query and constants.ENABLE_SPLIT_QUERY: + address = pathutils.QUERY_SOCKET + else: + address = None # TODO: Cache object? try: - client = luxi.Client() + client = luxi.Client(address=address) except luxi.NoMasterError: ss = ssconf.SimpleStore() @@ -2143,7 +2420,7 @@ def FormatError(err): elif isinstance(err, errors.OpPrereqError): if len(err.args) == 2: obuf.write("Failure: prerequisites not met for this" - " operation:\nerror type: %s, error details:\n%s" % + " operation:\nerror type: %s, error details:\n%s" % (err.args[1], err.args[0])) else: obuf.write("Failure: prerequisites not met for this" @@ -2163,8 +2440,14 @@ def FormatError(err): elif isinstance(err, errors.ParameterError): obuf.write("Failure: unknown/wrong parameter name '%s'" % msg) elif isinstance(err, luxi.NoMasterError): - obuf.write("Cannot communicate with the master daemon.\nIs it running" - " and listening for connections?") + if err.args[0] == pathutils.MASTER_SOCKET: + daemon = "the master daemon" + elif err.args[0] == pathutils.QUERY_SOCKET: + daemon = "the config daemon" + else: + daemon = "socket '%s'" % str(err.args[0]) + obuf.write("Cannot communicate with %s.\nIs the process running" + " and listening for connections?" % daemon) elif isinstance(err, luxi.TimeoutError): obuf.write("Timeout while talking to the master daemon. Jobs might have" " been submitted and will continue to run even if the call" @@ -2226,7 +2509,20 @@ def GenericMain(commands, override=None, aliases=None, aliases = {} try: - func, options, args = _ParseArgs(sys.argv, commands, aliases, env_override) + (func, options, args) = _ParseArgs(binary, sys.argv, commands, aliases, + env_override) + except _ShowVersion: + ToStdout("%s (ganeti %s) %s", binary, constants.VCS_VERSION, + constants.RELEASE_VERSION) + return constants.EXIT_SUCCESS + except _ShowUsage, err: + for line in _FormatUsage(binary, commands): + ToStdout(line) + + if err.exit_error: + return constants.EXIT_FAILURE + else: + return constants.EXIT_SUCCESS except errors.ParameterError, err: result, err_msg = FormatError(err) ToStderr(err_msg) @@ -2239,7 +2535,7 @@ def GenericMain(commands, override=None, aliases=None, for key, val in override.iteritems(): setattr(options, key, val) - utils.SetupLogging(constants.LOG_COMMANDS, logname, debug=options.debug, + utils.SetupLogging(pathutils.LOG_COMMANDS, logname, debug=options.debug, stderr_logging=True) logging.info("Command line: %s", cmdline) @@ -2291,6 +2587,21 @@ def ParseNicOption(optvalue): return nics +def FixHvParams(hvparams): + # In Ganeti 2.8.4 the separator for the usb_devices hvparam was changed from + # comma to space because commas cannot be accepted on the command line + # (they already act as the separator between different hvparams). Still, + # RAPI should be able to accept commas for backwards compatibility. + # Therefore, we convert spaces into commas here, and we keep the old + # parsing logic everywhere else. + try: + new_usb_devices = hvparams[constants.HV_USB_DEVICES].replace(" ", ",") + hvparams[constants.HV_USB_DEVICES] = new_usb_devices + except KeyError: + #No usb_devices, no modification required + pass + + def GenericInstanceCreate(mode, opts, args): """Add an instance to the cluster via either creation or import. @@ -2380,6 +2691,7 @@ def GenericInstanceCreate(mode, opts, args): utils.ForceDictType(opts.beparams, constants.BES_PARAMETER_COMPAT) utils.ForceDictType(hvparams, constants.HVS_PARAMETER_TYPES) + FixHvParams(hvparams) if mode == constants.INSTANCE_CREATE: start = opts.start @@ -2404,6 +2716,7 @@ def GenericInstanceCreate(mode, opts, args): disks=disks, disk_template=opts.disk_template, nics=nics, + conflicts_check=opts.conflicts_check, pnode=pnode, snode=snode, ip_check=opts.ip_check, name_check=opts.name_check, @@ -2472,7 +2785,8 @@ class _RunWhileClusterStoppedHelper: # No need to use SSH result = utils.RunCmd(cmd) else: - result = self.ssh.Run(node_name, "root", utils.ShellQuoteArgs(cmd)) + result = self.ssh.Run(node_name, constants.SSH_LOGIN_USER, + utils.ShellQuoteArgs(cmd)) if result.failed: errmsg = ["Failed to run command %s" % result.cmd] @@ -2491,7 +2805,7 @@ class _RunWhileClusterStoppedHelper: """ # Pause watcher by acquiring an exclusive lock on watcher state file self.feedback_fn("Blocking watcher") - watcher_block = utils.FileLock.Open(constants.WATCHER_LOCK_FILE) + watcher_block = utils.FileLock.Open(pathutils.WATCHER_LOCK_FILE) try: # TODO: Currently, this just blocks. There's no timeout. # TODO: Should it be a shared lock? @@ -2500,12 +2814,12 @@ class _RunWhileClusterStoppedHelper: # Stop master daemons, so that no new jobs can come in and all running # ones are finished self.feedback_fn("Stopping master daemons") - self._RunCmd(None, [constants.DAEMON_UTIL, "stop-master"]) + self._RunCmd(None, [pathutils.DAEMON_UTIL, "stop-master"]) try: # Stop daemons on all nodes for node_name in self.online_nodes: self.feedback_fn("Stopping daemons on %s" % node_name) - self._RunCmd(node_name, [constants.DAEMON_UTIL, "stop-all"]) + self._RunCmd(node_name, [pathutils.DAEMON_UTIL, "stop-all"]) # All daemons are shut down now try: @@ -2519,7 +2833,7 @@ class _RunWhileClusterStoppedHelper: # Start cluster again, master node last for node_name in self.nonmaster_nodes + [self.master_node]: self.feedback_fn("Starting daemons on %s" % node_name) - self._RunCmd(node_name, [constants.DAEMON_UTIL, "start-all"]) + self._RunCmd(node_name, [pathutils.DAEMON_UTIL, "start-all"]) finally: # Resume watcher watcher_block.Close() @@ -2931,6 +3245,21 @@ def GenericList(resource, fields, names, unit, separator, header, cl=None, return constants.EXIT_SUCCESS +def _FieldDescValues(fdef): + """Helper function for L{GenericListFields} to get query field description. + + @type fdef: L{objects.QueryFieldDefinition} + @rtype: list + + """ + return [ + fdef.name, + _QFT_NAMES.get(fdef.kind, fdef.kind), + fdef.title, + fdef.doc, + ] + + def GenericListFields(resource, fields, separator, header, cl=None): """Generic implementation for listing fields for a resource. @@ -2955,11 +3284,12 @@ def GenericListFields(resource, fields, separator, header, cl=None): columns = [ TableColumn("Name", str, False), + TableColumn("Type", str, False), TableColumn("Title", str, False), TableColumn("Description", str, False), ] - rows = [[fdef.name, fdef.title, fdef.doc] for fdef in response.fields] + rows = map(_FieldDescValues, response.fields) for line in FormatTable(rows, columns, header, separator): ToStdout(line) @@ -3392,31 +3722,127 @@ class JobExecutor(object): return [row[1:3] for row in self.jobs] -def FormatParameterDict(buf, param_dict, actual, level=1): +def FormatParamsDictInfo(param_dict, actual): """Formats a parameter dictionary. - @type buf: L{StringIO} - @param buf: the buffer into which to write @type param_dict: dict @param param_dict: the own parameters @type actual: dict @param actual: the current parameter set (including defaults) - @param level: Level of indent + @rtype: dict + @return: dictionary where the value of each parameter is either a fully + formatted string or a dictionary containing formatted strings """ - indent = " " * level - - for key in sorted(actual): - data = actual[key] - buf.write("%s- %s:" % (indent, key)) - + ret = {} + for (key, data) in actual.items(): if isinstance(data, dict) and data: - buf.write("\n") - FormatParameterDict(buf, param_dict.get(key, {}), data, - level=level + 1) + ret[key] = FormatParamsDictInfo(param_dict.get(key, {}), data) else: - val = param_dict.get(key, "default (%s)" % data) - buf.write(" %s\n" % val) + ret[key] = str(param_dict.get(key, "default (%s)" % data)) + return ret + + +def _FormatListInfoDefault(data, def_data): + if data is not None: + ret = utils.CommaJoin(data) + else: + ret = "default (%s)" % utils.CommaJoin(def_data) + return ret + + +def FormatPolicyInfo(custom_ipolicy, eff_ipolicy, iscluster): + """Formats an instance policy. + + @type custom_ipolicy: dict + @param custom_ipolicy: own policy + @type eff_ipolicy: dict + @param eff_ipolicy: effective policy (including defaults); ignored for + cluster + @type iscluster: bool + @param iscluster: the policy is at cluster level + @rtype: list of pairs + @return: formatted data, suitable for L{PrintGenericInfo} + + """ + if iscluster: + eff_ipolicy = custom_ipolicy + + minmax_out = [] + custom_minmax = custom_ipolicy.get(constants.ISPECS_MINMAX) + if custom_minmax: + for (k, minmax) in enumerate(custom_minmax): + minmax_out.append([ + ("%s/%s" % (key, k), + FormatParamsDictInfo(minmax[key], minmax[key])) + for key in constants.ISPECS_MINMAX_KEYS + ]) + else: + for (k, minmax) in enumerate(eff_ipolicy[constants.ISPECS_MINMAX]): + minmax_out.append([ + ("%s/%s" % (key, k), + FormatParamsDictInfo({}, minmax[key])) + for key in constants.ISPECS_MINMAX_KEYS + ]) + ret = [("bounds specs", minmax_out)] + + if iscluster: + stdspecs = custom_ipolicy[constants.ISPECS_STD] + ret.append( + (constants.ISPECS_STD, + FormatParamsDictInfo(stdspecs, stdspecs)) + ) + + ret.append( + ("allowed disk templates", + _FormatListInfoDefault(custom_ipolicy.get(constants.IPOLICY_DTS), + eff_ipolicy[constants.IPOLICY_DTS])) + ) + ret.extend([ + (key, str(custom_ipolicy.get(key, "default (%s)" % eff_ipolicy[key]))) + for key in constants.IPOLICY_PARAMETERS + ]) + return ret + + +def _PrintSpecsParameters(buf, specs): + values = ("%s=%s" % (par, val) for (par, val) in sorted(specs.items())) + buf.write(",".join(values)) + + +def PrintIPolicyCommand(buf, ipolicy, isgroup): + """Print the command option used to generate the given instance policy. + + Currently only the parts dealing with specs are supported. + + @type buf: StringIO + @param buf: stream to write into + @type ipolicy: dict + @param ipolicy: instance policy + @type isgroup: bool + @param isgroup: whether the policy is at group level + + """ + if not isgroup: + stdspecs = ipolicy.get("std") + if stdspecs: + buf.write(" %s " % IPOLICY_STD_SPECS_STR) + _PrintSpecsParameters(buf, stdspecs) + minmaxes = ipolicy.get("minmax", []) + first = True + for minmax in minmaxes: + minspecs = minmax.get("min") + maxspecs = minmax.get("max") + if minspecs and maxspecs: + if first: + buf.write(" %s " % IPOLICY_BOUNDS_SPECS_STR) + first = False + else: + buf.write("//") + buf.write("min:") + _PrintSpecsParameters(buf, minspecs) + buf.write("/max:") + _PrintSpecsParameters(buf, maxspecs) def ConfirmOperation(names, list_type, text, extra=""): @@ -3471,24 +3897,9 @@ def _MaybeParseUnit(elements): return parsed -def CreateIPolicyFromOpts(ispecs_mem_size=None, - ispecs_cpu_count=None, - ispecs_disk_count=None, - ispecs_disk_size=None, - ispecs_nic_count=None, - ipolicy_disk_templates=None, - ipolicy_vcpu_ratio=None, - ipolicy_spindle_ratio=None, - group_ipolicy=False, - allowed_values=None, - fill_all=False): - """Creation of instance policy based on command line options. - - @param fill_all: whether for cluster policies we should ensure that - all values are filled - - - """ +def _InitISpecsFromSplitOpts(ipolicy, ispecs_mem_size, ispecs_cpu_count, + ispecs_disk_count, ispecs_disk_size, + ispecs_nic_count, group_ipolicy, fill_all): try: if ispecs_mem_size: ispecs_mem_size = _MaybeParseUnit(ispecs_mem_size) @@ -3501,7 +3912,7 @@ def CreateIPolicyFromOpts(ispecs_mem_size=None, errors.ECODE_INVAL) # prepare ipolicy dict - ipolicy_transposed = { + ispecs_transposed = { constants.ISPEC_MEM_SIZE: ispecs_mem_size, constants.ISPEC_CPU_COUNT: ispecs_cpu_count, constants.ISPEC_DISK_COUNT: ispecs_disk_count, @@ -3514,29 +3925,136 @@ def CreateIPolicyFromOpts(ispecs_mem_size=None, forced_type = TISPECS_GROUP_TYPES else: forced_type = TISPECS_CLUSTER_TYPES - - for specs in ipolicy_transposed.values(): - utils.ForceDictType(specs, forced_type, allowed_values=allowed_values) + for specs in ispecs_transposed.values(): + assert type(specs) is dict + utils.ForceDictType(specs, forced_type) # then transpose - ipolicy_out = objects.MakeEmptyIPolicy() - for name, specs in ipolicy_transposed.iteritems(): + ispecs = { + constants.ISPECS_MIN: {}, + constants.ISPECS_MAX: {}, + constants.ISPECS_STD: {}, + } + for (name, specs) in ispecs_transposed.iteritems(): assert name in constants.ISPECS_PARAMETERS for key, val in specs.items(): # {min: .. ,max: .., std: ..} - ipolicy_out[key][name] = val + assert key in ispecs + ispecs[key][name] = val + minmax_out = {} + for key in constants.ISPECS_MINMAX_KEYS: + if fill_all: + minmax_out[key] = \ + objects.FillDict(constants.ISPECS_MINMAX_DEFAULTS[key], ispecs[key]) + else: + minmax_out[key] = ispecs[key] + ipolicy[constants.ISPECS_MINMAX] = [minmax_out] + if fill_all: + ipolicy[constants.ISPECS_STD] = \ + objects.FillDict(constants.IPOLICY_DEFAULTS[constants.ISPECS_STD], + ispecs[constants.ISPECS_STD]) + else: + ipolicy[constants.ISPECS_STD] = ispecs[constants.ISPECS_STD] + + +def _ParseSpecUnit(spec, keyname): + ret = spec.copy() + for k in [constants.ISPEC_DISK_SIZE, constants.ISPEC_MEM_SIZE]: + if k in ret: + try: + ret[k] = utils.ParseUnit(ret[k]) + except (TypeError, ValueError, errors.UnitParseError), err: + raise errors.OpPrereqError(("Invalid parameter %s (%s) in %s instance" + " specs: %s" % (k, ret[k], keyname, err)), + errors.ECODE_INVAL) + return ret + + +def _ParseISpec(spec, keyname, required): + ret = _ParseSpecUnit(spec, keyname) + utils.ForceDictType(ret, constants.ISPECS_PARAMETER_TYPES) + missing = constants.ISPECS_PARAMETERS - frozenset(ret.keys()) + if required and missing: + raise errors.OpPrereqError("Missing parameters in ipolicy spec %s: %s" % + (keyname, utils.CommaJoin(missing)), + errors.ECODE_INVAL) + return ret + + +def _GetISpecsInAllowedValues(minmax_ispecs, allowed_values): + ret = None + if (minmax_ispecs and allowed_values and len(minmax_ispecs) == 1 and + len(minmax_ispecs[0]) == 1): + for (key, spec) in minmax_ispecs[0].items(): + # This loop is executed exactly once + if key in allowed_values and not spec: + ret = key + return ret + + +def _InitISpecsFromFullOpts(ipolicy_out, minmax_ispecs, std_ispecs, + group_ipolicy, allowed_values): + found_allowed = _GetISpecsInAllowedValues(minmax_ispecs, allowed_values) + if found_allowed is not None: + ipolicy_out[constants.ISPECS_MINMAX] = found_allowed + elif minmax_ispecs is not None: + minmax_out = [] + for mmpair in minmax_ispecs: + mmpair_out = {} + for (key, spec) in mmpair.items(): + if key not in constants.ISPECS_MINMAX_KEYS: + msg = "Invalid key in bounds instance specifications: %s" % key + raise errors.OpPrereqError(msg, errors.ECODE_INVAL) + mmpair_out[key] = _ParseISpec(spec, key, True) + minmax_out.append(mmpair_out) + ipolicy_out[constants.ISPECS_MINMAX] = minmax_out + if std_ispecs is not None: + assert not group_ipolicy # This is not an option for gnt-group + ipolicy_out[constants.ISPECS_STD] = _ParseISpec(std_ispecs, "std", False) + + +def CreateIPolicyFromOpts(ispecs_mem_size=None, + ispecs_cpu_count=None, + ispecs_disk_count=None, + ispecs_disk_size=None, + ispecs_nic_count=None, + minmax_ispecs=None, + std_ispecs=None, + ipolicy_disk_templates=None, + ipolicy_vcpu_ratio=None, + ipolicy_spindle_ratio=None, + group_ipolicy=False, + allowed_values=None, + fill_all=False): + """Creation of instance policy based on command line options. + + @param fill_all: whether for cluster policies we should ensure that + all values are filled + + """ + assert not (fill_all and allowed_values) + + split_specs = (ispecs_mem_size or ispecs_cpu_count or ispecs_disk_count or + ispecs_disk_size or ispecs_nic_count) + if (split_specs and (minmax_ispecs is not None or std_ispecs is not None)): + raise errors.OpPrereqError("A --specs-xxx option cannot be specified" + " together with any --ipolicy-xxx-specs option", + errors.ECODE_INVAL) + + ipolicy_out = objects.MakeEmptyIPolicy() + if split_specs: + assert fill_all + _InitISpecsFromSplitOpts(ipolicy_out, ispecs_mem_size, ispecs_cpu_count, + ispecs_disk_count, ispecs_disk_size, + ispecs_nic_count, group_ipolicy, fill_all) + elif (minmax_ispecs is not None or std_ispecs is not None): + _InitISpecsFromFullOpts(ipolicy_out, minmax_ispecs, std_ispecs, + group_ipolicy, allowed_values) - # no filldict for non-dicts - if not group_ipolicy and fill_all: - if ipolicy_disk_templates is None: - ipolicy_disk_templates = constants.DISK_TEMPLATES - if ipolicy_vcpu_ratio is None: - ipolicy_vcpu_ratio = \ - constants.IPOLICY_DEFAULTS[constants.IPOLICY_VCPU_RATIO] - if ipolicy_spindle_ratio is None: - ipolicy_spindle_ratio = \ - constants.IPOLICY_DEFAULTS[constants.IPOLICY_SPINDLE_RATIO] if ipolicy_disk_templates is not None: - ipolicy_out[constants.IPOLICY_DTS] = list(ipolicy_disk_templates) + if allowed_values and ipolicy_disk_templates in allowed_values: + ipolicy_out[constants.IPOLICY_DTS] = ipolicy_disk_templates + else: + ipolicy_out[constants.IPOLICY_DTS] = list(ipolicy_disk_templates) if ipolicy_vcpu_ratio is not None: ipolicy_out[constants.IPOLICY_VCPU_RATIO] = ipolicy_vcpu_ratio if ipolicy_spindle_ratio is not None: @@ -3544,4 +4062,97 @@ def CreateIPolicyFromOpts(ispecs_mem_size=None, assert not (frozenset(ipolicy_out.keys()) - constants.IPOLICY_ALL_KEYS) + if not group_ipolicy and fill_all: + ipolicy_out = objects.FillIPolicy(constants.IPOLICY_DEFAULTS, ipolicy_out) + return ipolicy_out + + +def _SerializeGenericInfo(buf, data, level, afterkey=False): + """Formatting core of L{PrintGenericInfo}. + + @param buf: (string) stream to accumulate the result into + @param data: data to format + @type level: int + @param level: depth in the data hierarchy, used for indenting + @type afterkey: bool + @param afterkey: True when we are in the middle of a line after a key (used + to properly add newlines or indentation) + + """ + baseind = " " + if isinstance(data, dict): + if not data: + buf.write("\n") + else: + if afterkey: + buf.write("\n") + doindent = True + else: + doindent = False + for key in sorted(data): + if doindent: + buf.write(baseind * level) + else: + doindent = True + buf.write(key) + buf.write(": ") + _SerializeGenericInfo(buf, data[key], level + 1, afterkey=True) + elif isinstance(data, list) and len(data) > 0 and isinstance(data[0], tuple): + # list of tuples (an ordered dictionary) + if afterkey: + buf.write("\n") + doindent = True + else: + doindent = False + for (key, val) in data: + if doindent: + buf.write(baseind * level) + else: + doindent = True + buf.write(key) + buf.write(": ") + _SerializeGenericInfo(buf, val, level + 1, afterkey=True) + elif isinstance(data, list): + if not data: + buf.write("\n") + else: + if afterkey: + buf.write("\n") + doindent = True + else: + doindent = False + for item in data: + if doindent: + buf.write(baseind * level) + else: + doindent = True + buf.write("-") + buf.write(baseind[1:]) + _SerializeGenericInfo(buf, item, level + 1) + else: + # This branch should be only taken for strings, but it's practically + # impossible to guarantee that no other types are produced somewhere + buf.write(str(data)) + buf.write("\n") + + +def PrintGenericInfo(data): + """Print information formatted according to the hierarchy. + + The output is a valid YAML string. + + @param data: the data to print. It's a hierarchical structure whose elements + can be: + - dictionaries, where keys are strings and values are of any of the + types listed here + - lists of pairs (key, value), where key is a string and value is of + any of the types listed here; it's a way to encode ordered + dictionaries + - lists of any of the types listed here + - strings + + """ + buf = StringIO() + _SerializeGenericInfo(buf, data, 0) + ToStdout(buf.getvalue().rstrip("\n"))