X-Git-Url: https://code.grnet.gr/git/ganeti-local/blobdiff_plain/7071e725928eecaadcad978d7b2f185bf9fb7540..2a6f6ef7ce23bdf1b55627785320d8086dcb77aa:/lib/rpc.py diff --git a/lib/rpc.py b/lib/rpc.py index 44ef760..c934ba2 100644 --- a/lib/rpc.py +++ b/lib/rpc.py @@ -1,7 +1,7 @@ # # -# Copyright (C) 2006, 2007, 2008, 2009, 2010, 2011 Google Inc. +# Copyright (C) 2006, 2007, 2008, 2009, 2010, 2011, 2012, 2013 Google Inc. # # This program is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by @@ -30,12 +30,12 @@ # if they need to start using instance attributes # R0904: Too many public methods -import os import logging import zlib import base64 import pycurl import threading +import copy from ganeti import utils from ganeti import objects @@ -47,37 +47,22 @@ from ganeti import netutils from ganeti import ssconf from ganeti import runtime from ganeti import compat +from ganeti import rpc_defs +from ganeti import pathutils +from ganeti import vcluster + +# Special module generated at build time +from ganeti import _generated_rpc # pylint has a bug here, doesn't see this import import ganeti.http.client # pylint: disable=W0611 -# Timeout for connecting to nodes (seconds) -_RPC_CONNECT_TIMEOUT = 5 - _RPC_CLIENT_HEADERS = [ "Content-type: %s" % http.HTTP_APP_JSON, "Expect:", ] -# Various time constants for the timeout table -_TMO_URGENT = 60 # one minute -_TMO_FAST = 5 * 60 # five minutes -_TMO_NORMAL = 15 * 60 # 15 minutes -_TMO_SLOW = 3600 # one hour -_TMO_4HRS = 4 * 3600 -_TMO_1DAY = 86400 - -# Timeout table that will be built later by decorators -# Guidelines for choosing timeouts: -# - call used during watcher: timeout -> 1min, _TMO_URGENT -# - trivial (but be sure it is trivial) (e.g. reading a file): 5min, _TMO_FAST -# - other calls: 15 min, _TMO_NORMAL -# - special calls (instance add, etc.): either _TMO_SLOW (1h) or huge timeouts - -_TIMEOUTS = { -} - #: Special value to describe an offline host _OFFLINE = object() @@ -111,7 +96,7 @@ def Shutdown(): def _ConfigRpcCurl(curl): - noded_cert = str(constants.NODED_CERT_FILE) + noded_cert = str(pathutils.NODED_CERT_FILE) curl.setopt(pycurl.FOLLOWLOCATION, False) curl.setopt(pycurl.CAINFO, noded_cert) @@ -121,22 +106,7 @@ def _ConfigRpcCurl(curl): curl.setopt(pycurl.SSLCERT, noded_cert) curl.setopt(pycurl.SSLKEYTYPE, "PEM") curl.setopt(pycurl.SSLKEY, noded_cert) - curl.setopt(pycurl.CONNECTTIMEOUT, _RPC_CONNECT_TIMEOUT) - - -def _RpcTimeout(secs): - """Timeout decorator. - - When applied to a rpc call_* function, it updates the global timeout - table with the given function/timeout. - - """ - def decorator(f): - name = f.__name__ - assert name.startswith("call_") - _TIMEOUTS[name[len("call_"):]] = secs - return f - return decorator + curl.setopt(pycurl.CONNECTTIMEOUT, constants.RPC_CONNECT_TIMEOUT) def RunWithRPC(fn): @@ -180,7 +150,7 @@ class RpcResult(object): """RPC Result class. This class holds an RPC result. It is needed since in multi-node - calls we can't raise an exception just because one one out of many + calls we can't raise an exception just because one out of many failed, and therefore we use this class to encapsulate the result. @ivar data: the data payload, for successful results, or None @@ -260,12 +230,27 @@ class RpcResult(object): args = (msg, ) raise ec(*args) # pylint: disable=W0142 + def Warn(self, msg, feedback_fn): + """If the result has failed, call the feedback_fn. + + This is used to in cases were LU wants to warn the + user about a failure, but continue anyway. + + """ + if not self.fail_msg: + return + + msg = "%s: %s" % (msg, self.fail_msg) + feedback_fn(msg) + -def _SsconfResolver(node_list, +def _SsconfResolver(ssconf_ips, node_list, _, ssc=ssconf.SimpleStore, nslookup_fn=netutils.Hostname.GetIP): """Return addresses for given node names. + @type ssconf_ips: bool + @param ssconf_ips: Use the ssconf IPs @type node_list: list @param node_list: List of node names @type ssc: class @@ -277,16 +262,20 @@ def _SsconfResolver(node_list, """ ss = ssc() - iplist = ss.GetNodePrimaryIPList() family = ss.GetPrimaryIPFamily() - ipmap = dict(entry.split() for entry in iplist) + + if ssconf_ips: + iplist = ss.GetNodePrimaryIPList() + ipmap = dict(entry.split() for entry in iplist) + else: + ipmap = {} result = [] for node in node_list: ip = ipmap.get(node) if ip is None: ip = nslookup_fn(node, family=family) - result.append((node, ip)) + result.append((node, ip, node)) return result @@ -298,54 +287,65 @@ class _StaticResolver: """ self._addresses = addresses - def __call__(self, hosts): + def __call__(self, hosts, _): """Returns static addresses for hosts. """ assert len(hosts) == len(self._addresses) - return zip(hosts, self._addresses) + return zip(hosts, self._addresses, hosts) -def _CheckConfigNode(name, node): +def _CheckConfigNode(node_uuid_or_name, node, accept_offline_node): """Checks if a node is online. - @type name: string - @param name: Node name + @type node_uuid_or_name: string + @param node_uuid_or_name: Node UUID @type node: L{objects.Node} or None @param node: Node object """ if node is None: - # Depend on DNS for name resolution - ip = name - elif node.offline: - ip = _OFFLINE + # Assume that the passed parameter was actually a node name, so depend on + # DNS for name resolution + return (node_uuid_or_name, node_uuid_or_name, node_uuid_or_name) else: - ip = node.primary_ip - return (name, ip) + if node.offline and not accept_offline_node: + ip = _OFFLINE + else: + ip = node.primary_ip + return (node.name, ip, node_uuid_or_name) -def _NodeConfigResolver(single_node_fn, all_nodes_fn, hosts): +def _NodeConfigResolver(single_node_fn, all_nodes_fn, node_uuids, opts): """Calculate node addresses using configuration. + Note that strings in node_uuids are treated as node names if the UUID is not + found in the configuration. + """ + accept_offline_node = (opts is rpc_defs.ACCEPT_OFFLINE_NODE) + + assert accept_offline_node or opts is None, "Unknown option" + # Special case for single-host lookups - if len(hosts) == 1: - (name, ) = hosts - return [_CheckConfigNode(name, single_node_fn(name))] + if len(node_uuids) == 1: + (uuid, ) = node_uuids + return [_CheckConfigNode(uuid, single_node_fn(uuid), accept_offline_node)] else: all_nodes = all_nodes_fn() - return [_CheckConfigNode(name, all_nodes.get(name, None)) - for name in hosts] + return [_CheckConfigNode(uuid, all_nodes.get(uuid, None), + accept_offline_node) + for uuid in node_uuids] class _RpcProcessor: def __init__(self, resolver, port, lock_monitor_cb=None): """Initializes this class. - @param resolver: callable accepting a list of hostnames, returning a list - of tuples containing name and IP address (IP address can be the name or - the special value L{_OFFLINE} to mark offline machines) + @param resolver: callable accepting a list of node UUIDs or hostnames, + returning a list of tuples containing name, IP address and original name + of the resolved node. IP address can be the name or the special value + L{_OFFLINE} to mark offline machines. @type port: int @param port: TCP port @param lock_monitor_cb: Callable for registering with lock monitor @@ -359,22 +359,34 @@ class _RpcProcessor: def _PrepareRequests(hosts, port, procedure, body, read_timeout): """Prepares requests by sorting offline hosts into separate list. + @type body: dict + @param body: a dictionary with per-host body data + """ results = {} requests = {} - for (name, ip) in hosts: + assert isinstance(body, dict) + assert len(body) == len(hosts) + assert compat.all(isinstance(v, str) for v in body.values()) + assert frozenset(map(lambda x: x[2], hosts)) == frozenset(body.keys()), \ + "%s != %s" % (hosts, body.keys()) + + for (name, ip, original_name) in hosts: if ip is _OFFLINE: # Node is marked as offline - results[name] = RpcResult(node=name, offline=True, call=procedure) + results[original_name] = RpcResult(node=name, + offline=True, + call=procedure) else: - requests[name] = \ + requests[original_name] = \ http.client.HttpClientRequest(str(ip), port, - http.HTTP_PUT, str("/%s" % procedure), + http.HTTP_POST, str("/%s" % procedure), headers=_RPC_CLIENT_HEADERS, - post_data=body, + post_data=body[original_name], read_timeout=read_timeout, - nicename="%s/%s" % (name, procedure)) + nicename="%s/%s" % (name, procedure), + curl_config_fn=_ConfigRpcCurl) return (results, requests) @@ -402,1224 +414,609 @@ class _RpcProcessor: return results - def __call__(self, hosts, procedure, body, read_timeout=None, http_pool=None): + def __call__(self, nodes, procedure, body, read_timeout, resolver_opts, + _req_process_fn=None): """Makes an RPC request to a number of nodes. - @type hosts: sequence - @param hosts: Hostnames + @type nodes: sequence + @param nodes: node UUIDs or Hostnames @type procedure: string @param procedure: Request path - @type body: string - @param body: Request body + @type body: dictionary + @param body: dictionary with request bodies per host @type read_timeout: int or None @param read_timeout: Read timeout for request + @rtype: dictionary + @return: a dictionary mapping host names to rpc.RpcResult objects """ - assert procedure in _TIMEOUTS, "RPC call not declared in the timeouts table" - - if not http_pool: - http_pool = http.client.HttpClientPool(_ConfigRpcCurl) + assert read_timeout is not None, \ + "Missing RPC read timeout for procedure '%s'" % procedure - if read_timeout is None: - read_timeout = _TIMEOUTS[procedure] + if _req_process_fn is None: + _req_process_fn = http.client.ProcessRequests (results, requests) = \ - self._PrepareRequests(self._resolver(hosts), self._port, procedure, - str(body), read_timeout) + self._PrepareRequests(self._resolver(nodes, resolver_opts), self._port, + procedure, body, read_timeout) - http_pool.ProcessRequests(requests.values(), - lock_monitor_cb=self._lock_monitor_cb) + _req_process_fn(requests.values(), lock_monitor_cb=self._lock_monitor_cb) assert not frozenset(results).intersection(requests) return self._CombineResults(results, requests, procedure) -def _EncodeImportExportIO(ieio, ieioargs): - """Encodes import/export I/O information. - - """ - if ieio == constants.IEIO_RAW_DISK: - assert len(ieioargs) == 1 - return (ieioargs[0].ToDict(), ) - - if ieio == constants.IEIO_SCRIPT: - assert len(ieioargs) == 2 - return (ieioargs[0].ToDict(), ieioargs[1]) - - return ieioargs - - -class RpcRunner(object): - """RPC runner class. - - """ - def __init__(self, context): - """Initialized the RPC runner. - - @type context: C{masterd.GanetiContext} - @param context: Ganeti context - - """ - self._cfg = context.cfg - self._proc = _RpcProcessor(compat.partial(_NodeConfigResolver, - self._cfg.GetNodeInfo, - self._cfg.GetAllNodesInfo), - netutils.GetDaemonPort(constants.NODED), - lock_monitor_cb=context.glm.AddToLockMonitor) - - def _InstDict(self, instance, hvp=None, bep=None, osp=None): - """Convert the given instance to a dict. - - This is done via the instance's ToDict() method and additionally - we fill the hvparams with the cluster defaults. - - @type instance: L{objects.Instance} - @param instance: an Instance object - @type hvp: dict or None - @param hvp: a dictionary with overridden hypervisor parameters - @type bep: dict or None - @param bep: a dictionary with overridden backend parameters - @type osp: dict or None - @param osp: a dictionary with overridden os parameters - @rtype: dict - @return: the instance dict, with the hvparams filled with the - cluster defaults - - """ - idict = instance.ToDict() - cluster = self._cfg.GetClusterInfo() - idict["hvparams"] = cluster.FillHV(instance) - if hvp is not None: - idict["hvparams"].update(hvp) - idict["beparams"] = cluster.FillBE(instance) - if bep is not None: - idict["beparams"].update(bep) - idict["osparams"] = cluster.SimpleFillOS(instance.os, instance.osparams) - if osp is not None: - idict["osparams"].update(osp) - for nic in idict["nics"]: - nic['nicparams'] = objects.FillDict( - cluster.nicparams[constants.PP_DEFAULT], - nic['nicparams']) - return idict - - def _MultiNodeCall(self, node_list, procedure, args, read_timeout=None): - """Helper for making a multi-node call +class _RpcClientBase: + def __init__(self, resolver, encoder_fn, lock_monitor_cb=None, + _req_process_fn=None): + """Initializes this class. """ - body = serializer.DumpJson(args, indent=False) - return self._proc(node_list, procedure, body, read_timeout=read_timeout) + proc = _RpcProcessor(resolver, + netutils.GetDaemonPort(constants.NODED), + lock_monitor_cb=lock_monitor_cb) + self._proc = compat.partial(proc, _req_process_fn=_req_process_fn) + self._encoder = compat.partial(self._EncodeArg, encoder_fn) @staticmethod - def _StaticMultiNodeCall(node_list, procedure, args, - address_list=None, read_timeout=None): - """Helper for making a multi-node static call + def _EncodeArg(encoder_fn, (argkind, value)): + """Encode argument. """ - body = serializer.DumpJson(args, indent=False) - - if address_list is None: - resolver = _SsconfResolver + if argkind is None: + return value else: - # Caller provided an address list - resolver = _StaticResolver(address_list) - - proc = _RpcProcessor(resolver, - netutils.GetDaemonPort(constants.NODED)) - return proc(node_list, procedure, body, read_timeout=read_timeout) - - def _SingleNodeCall(self, node, procedure, args, read_timeout=None): - """Helper for making a single-node call - - """ - body = serializer.DumpJson(args, indent=False) - return self._proc([node], procedure, body, read_timeout=read_timeout)[node] - - @classmethod - def _StaticSingleNodeCall(cls, node, procedure, args, read_timeout=None): - """Helper for making a single-node static call - - """ - body = serializer.DumpJson(args, indent=False) - proc = _RpcProcessor(_SsconfResolver, - netutils.GetDaemonPort(constants.NODED)) - return proc([node], procedure, body, read_timeout=read_timeout)[node] - - # - # Begin RPC calls - # - - @_RpcTimeout(_TMO_URGENT) - def call_bdev_sizes(self, node_list, devices): - """Gets the sizes of requested block devices present on a node - - This is a multi-node call. - - """ - return self._MultiNodeCall(node_list, "bdev_sizes", [devices]) - - @_RpcTimeout(_TMO_URGENT) - def call_lv_list(self, node_list, vg_name): - """Gets the logical volumes present in a given volume group. - - This is a multi-node call. - - """ - return self._MultiNodeCall(node_list, "lv_list", [vg_name]) - - @_RpcTimeout(_TMO_URGENT) - def call_vg_list(self, node_list): - """Gets the volume group list. + return encoder_fn(argkind)(value) - This is a multi-node call. + def _Call(self, cdef, node_list, args): + """Entry point for automatically generated RPC wrappers. """ - return self._MultiNodeCall(node_list, "vg_list", []) - - @_RpcTimeout(_TMO_NORMAL) - def call_storage_list(self, node_list, su_name, su_args, name, fields): - """Get list of storage units. - - This is a multi-node call. - - """ - return self._MultiNodeCall(node_list, "storage_list", - [su_name, su_args, name, fields]) - - @_RpcTimeout(_TMO_NORMAL) - def call_storage_modify(self, node, su_name, su_args, name, changes): - """Modify a storage unit. - - This is a single-node call. - - """ - return self._SingleNodeCall(node, "storage_modify", - [su_name, su_args, name, changes]) - - @_RpcTimeout(_TMO_NORMAL) - def call_storage_execute(self, node, su_name, su_args, name, op): - """Executes an operation on a storage unit. - - This is a single-node call. - - """ - return self._SingleNodeCall(node, "storage_execute", - [su_name, su_args, name, op]) - - @_RpcTimeout(_TMO_URGENT) - def call_bridges_exist(self, node, bridges_list): - """Checks if a node has all the bridges given. - - This method checks if all bridges given in the bridges_list are - present on the remote node, so that an instance that uses interfaces - on those bridges can be started. - - This is a single-node call. - - """ - return self._SingleNodeCall(node, "bridges_exist", [bridges_list]) - - @_RpcTimeout(_TMO_NORMAL) - def call_instance_start(self, node, instance, hvp, bep, startup_paused): - """Starts an instance. - - This is a single-node call. - - """ - idict = self._InstDict(instance, hvp=hvp, bep=bep) - return self._SingleNodeCall(node, "instance_start", [idict, startup_paused]) - - @_RpcTimeout(_TMO_NORMAL) - def call_instance_shutdown(self, node, instance, timeout): - """Stops an instance. - - This is a single-node call. - - """ - return self._SingleNodeCall(node, "instance_shutdown", - [self._InstDict(instance), timeout]) - - @_RpcTimeout(_TMO_NORMAL) - def call_migration_info(self, node, instance): - """Gather the information necessary to prepare an instance migration. - - This is a single-node call. - - @type node: string - @param node: the node on which the instance is currently running - @type instance: C{objects.Instance} - @param instance: the instance definition - - """ - return self._SingleNodeCall(node, "migration_info", - [self._InstDict(instance)]) - - @_RpcTimeout(_TMO_NORMAL) - def call_accept_instance(self, node, instance, info, target): - """Prepare a node to accept an instance. - - This is a single-node call. - - @type node: string - @param node: the target node for the migration - @type instance: C{objects.Instance} - @param instance: the instance definition - @type info: opaque/hypervisor specific (string/data) - @param info: result for the call_migration_info call - @type target: string - @param target: target hostname (usually ip address) (on the node itself) - - """ - return self._SingleNodeCall(node, "accept_instance", - [self._InstDict(instance), info, target]) - - @_RpcTimeout(_TMO_NORMAL) - def call_instance_finalize_migration_dst(self, node, instance, info, success): - """Finalize any target-node migration specific operation. - - This is called both in case of a successful migration and in case of error - (in which case it should abort the migration). - - This is a single-node call. - - @type node: string - @param node: the target node for the migration - @type instance: C{objects.Instance} - @param instance: the instance definition - @type info: opaque/hypervisor specific (string/data) - @param info: result for the call_migration_info call - @type success: boolean - @param success: whether the migration was a success or a failure - - """ - return self._SingleNodeCall(node, "instance_finalize_migration_dst", - [self._InstDict(instance), info, success]) - - @_RpcTimeout(_TMO_SLOW) - def call_instance_migrate(self, node, instance, target, live): - """Migrate an instance. - - This is a single-node call. - - @type node: string - @param node: the node on which the instance is currently running - @type instance: C{objects.Instance} - @param instance: the instance definition - @type target: string - @param target: the target node name - @type live: boolean - @param live: whether the migration should be done live or not (the - interpretation of this parameter is left to the hypervisor) - - """ - return self._SingleNodeCall(node, "instance_migrate", - [self._InstDict(instance), target, live]) - - @_RpcTimeout(_TMO_SLOW) - def call_instance_finalize_migration_src(self, node, instance, success, live): - """Finalize the instance migration on the source node. - - This is a single-node call. - - @type instance: L{objects.Instance} - @param instance: the instance that was migrated - @type success: bool - @param success: whether the migration succeeded or not - @type live: bool - @param live: whether the user requested a live migration or not - - """ - return self._SingleNodeCall(node, "instance_finalize_migration_src", - [self._InstDict(instance), success, live]) - - @_RpcTimeout(_TMO_SLOW) - def call_instance_get_migration_status(self, node, instance): - """Report migration status. - - This is a single-node call that must be executed on the source node. - - @type instance: L{objects.Instance} - @param instance: the instance that is being migrated - @rtype: L{objects.MigrationStatus} - @return: the status of the current migration (one of - L{constants.HV_MIGRATION_VALID_STATUSES}), plus any additional - progress info that can be retrieved from the hypervisor - - """ - result = self._SingleNodeCall(node, "instance_get_migration_status", - [self._InstDict(instance)]) - if not result.fail_msg and result.payload is not None: - result.payload = objects.MigrationStatus.FromDict(result.payload) - return result - - @_RpcTimeout(_TMO_NORMAL) - def call_instance_reboot(self, node, inst, reboot_type, shutdown_timeout): - """Reboots an instance. - - This is a single-node call. - - """ - return self._SingleNodeCall(node, "instance_reboot", - [self._InstDict(inst), reboot_type, - shutdown_timeout]) - - @_RpcTimeout(_TMO_1DAY) - def call_instance_os_add(self, node, inst, reinstall, debug, osparams=None): - """Installs an OS on the given instance. - - This is a single-node call. - - """ - return self._SingleNodeCall(node, "instance_os_add", - [self._InstDict(inst, osp=osparams), - reinstall, debug]) - - @_RpcTimeout(_TMO_SLOW) - def call_instance_run_rename(self, node, inst, old_name, debug): - """Run the OS rename script for an instance. - - This is a single-node call. - - """ - return self._SingleNodeCall(node, "instance_run_rename", - [self._InstDict(inst), old_name, debug]) - - @_RpcTimeout(_TMO_URGENT) - def call_instance_info(self, node, instance, hname): - """Returns information about a single instance. - - This is a single-node call. - - @type node: list - @param node: the list of nodes to query - @type instance: string - @param instance: the instance name - @type hname: string - @param hname: the hypervisor type of the instance - - """ - return self._SingleNodeCall(node, "instance_info", [instance, hname]) - - @_RpcTimeout(_TMO_NORMAL) - def call_instance_migratable(self, node, instance): - """Checks whether the given instance can be migrated. - - This is a single-node call. - - @param node: the node to query - @type instance: L{objects.Instance} - @param instance: the instance to check - - - """ - return self._SingleNodeCall(node, "instance_migratable", - [self._InstDict(instance)]) - - @_RpcTimeout(_TMO_URGENT) - def call_all_instances_info(self, node_list, hypervisor_list): - """Returns information about all instances on the given nodes. - - This is a multi-node call. - - @type node_list: list - @param node_list: the list of nodes to query - @type hypervisor_list: list - @param hypervisor_list: the hypervisors to query for instances - - """ - return self._MultiNodeCall(node_list, "all_instances_info", - [hypervisor_list]) - - @_RpcTimeout(_TMO_URGENT) - def call_instance_list(self, node_list, hypervisor_list): - """Returns the list of running instances on a given node. - - This is a multi-node call. - - @type node_list: list - @param node_list: the list of nodes to query - @type hypervisor_list: list - @param hypervisor_list: the hypervisors to query for instances - - """ - return self._MultiNodeCall(node_list, "instance_list", [hypervisor_list]) - - @_RpcTimeout(_TMO_FAST) - def call_node_tcp_ping(self, node, source, target, port, timeout, - live_port_needed): - """Do a TcpPing on the remote node - - This is a single-node call. - - """ - return self._SingleNodeCall(node, "node_tcp_ping", - [source, target, port, timeout, - live_port_needed]) - - @_RpcTimeout(_TMO_FAST) - def call_node_has_ip_address(self, node, address): - """Checks if a node has the given IP address. - - This is a single-node call. - - """ - return self._SingleNodeCall(node, "node_has_ip_address", [address]) - - @_RpcTimeout(_TMO_URGENT) - def call_node_info(self, node_list, vg_name, hypervisor_type): - """Return node information. - - This will return memory information and volume group size and free - space. - - This is a multi-node call. + (procedure, _, resolver_opts, timeout, argdefs, + prep_fn, postproc_fn, _) = cdef - @type node_list: list - @param node_list: the list of nodes to query - @type vg_name: C{string} - @param vg_name: the name of the volume group to ask for disk space - information - @type hypervisor_type: C{str} - @param hypervisor_type: the name of the hypervisor to ask for - memory information - - """ - return self._MultiNodeCall(node_list, "node_info", - [vg_name, hypervisor_type]) - - @_RpcTimeout(_TMO_NORMAL) - def call_etc_hosts_modify(self, node, mode, name, ip): - """Modify hosts file with name - - @type node: string - @param node: The node to call - @type mode: string - @param mode: The mode to operate. Currently "add" or "remove" - @type name: string - @param name: The host name to be modified - @type ip: string - @param ip: The ip of the entry (just valid if mode is "add") - - """ - return self._SingleNodeCall(node, "etc_hosts_modify", [mode, name, ip]) - - @_RpcTimeout(_TMO_NORMAL) - def call_node_verify(self, node_list, checkdict, cluster_name): - """Request verification of given parameters. - - This is a multi-node call. - - """ - return self._MultiNodeCall(node_list, "node_verify", - [checkdict, cluster_name]) - - @classmethod - @_RpcTimeout(_TMO_FAST) - def call_node_start_master_daemons(cls, node, no_voting): - """Starts master daemons on a node. - - This is a single-node call. - - """ - return cls._StaticSingleNodeCall(node, "node_start_master_daemons", - [no_voting]) - - @classmethod - @_RpcTimeout(_TMO_FAST) - def call_node_activate_master_ip(cls, node): - """Activates master IP on a node. - - This is a single-node call. - - """ - return cls._StaticSingleNodeCall(node, "node_activate_master_ip", []) - - @classmethod - @_RpcTimeout(_TMO_FAST) - def call_node_stop_master(cls, node): - """Deactivates master IP and stops master daemons on a node. - - This is a single-node call. - - """ - return cls._StaticSingleNodeCall(node, "node_stop_master", []) - - @classmethod - @_RpcTimeout(_TMO_FAST) - def call_node_deactivate_master_ip(cls, node): - """Deactivates master IP on a node. - - This is a single-node call. - - """ - return cls._StaticSingleNodeCall(node, "node_deactivate_master_ip", []) - - @classmethod - @_RpcTimeout(_TMO_FAST) - def call_node_change_master_netmask(cls, node, netmask): - """Change master IP netmask. - - This is a single-node call. - - """ - return cls._StaticSingleNodeCall(node, "node_change_master_netmask", - [netmask]) - - @classmethod - @_RpcTimeout(_TMO_URGENT) - def call_master_info(cls, node_list): - """Query master info. - - This is a multi-node call. - - """ - # TODO: should this method query down nodes? - return cls._StaticMultiNodeCall(node_list, "master_info", []) - - @classmethod - @_RpcTimeout(_TMO_URGENT) - def call_version(cls, node_list): - """Query node version. - - This is a multi-node call. - - """ - return cls._StaticMultiNodeCall(node_list, "version", []) - - @_RpcTimeout(_TMO_NORMAL) - def call_blockdev_create(self, node, bdev, size, owner, on_primary, info): - """Request creation of a given block device. - - This is a single-node call. - - """ - return self._SingleNodeCall(node, "blockdev_create", - [bdev.ToDict(), size, owner, on_primary, info]) - - @_RpcTimeout(_TMO_SLOW) - def call_blockdev_wipe(self, node, bdev, offset, size): - """Request wipe at given offset with given size of a block device. - - This is a single-node call. - - """ - return self._SingleNodeCall(node, "blockdev_wipe", - [bdev.ToDict(), offset, size]) - - @_RpcTimeout(_TMO_NORMAL) - def call_blockdev_remove(self, node, bdev): - """Request removal of a given block device. - - This is a single-node call. - - """ - return self._SingleNodeCall(node, "blockdev_remove", [bdev.ToDict()]) - - @_RpcTimeout(_TMO_NORMAL) - def call_blockdev_rename(self, node, devlist): - """Request rename of the given block devices. - - This is a single-node call. - - """ - return self._SingleNodeCall(node, "blockdev_rename", - [(d.ToDict(), uid) for d, uid in devlist]) - - @_RpcTimeout(_TMO_NORMAL) - def call_blockdev_pause_resume_sync(self, node, disks, pause): - """Request a pause/resume of given block device. - - This is a single-node call. - - """ - return self._SingleNodeCall(node, "blockdev_pause_resume_sync", - [[bdev.ToDict() for bdev in disks], pause]) - - @_RpcTimeout(_TMO_NORMAL) - def call_blockdev_assemble(self, node, disk, owner, on_primary, idx): - """Request assembling of a given block device. - - This is a single-node call. - - """ - return self._SingleNodeCall(node, "blockdev_assemble", - [disk.ToDict(), owner, on_primary, idx]) - - @_RpcTimeout(_TMO_NORMAL) - def call_blockdev_shutdown(self, node, disk): - """Request shutdown of a given block device. - - This is a single-node call. - - """ - return self._SingleNodeCall(node, "blockdev_shutdown", [disk.ToDict()]) - - @_RpcTimeout(_TMO_NORMAL) - def call_blockdev_addchildren(self, node, bdev, ndevs): - """Request adding a list of children to a (mirroring) device. - - This is a single-node call. - - """ - return self._SingleNodeCall(node, "blockdev_addchildren", - [bdev.ToDict(), - [disk.ToDict() for disk in ndevs]]) - - @_RpcTimeout(_TMO_NORMAL) - def call_blockdev_removechildren(self, node, bdev, ndevs): - """Request removing a list of children from a (mirroring) device. - - This is a single-node call. - - """ - return self._SingleNodeCall(node, "blockdev_removechildren", - [bdev.ToDict(), - [disk.ToDict() for disk in ndevs]]) - - @_RpcTimeout(_TMO_NORMAL) - def call_blockdev_getmirrorstatus(self, node, disks): - """Request status of a (mirroring) device. - - This is a single-node call. + if callable(timeout): + read_timeout = timeout(args) + else: + read_timeout = timeout - """ - result = self._SingleNodeCall(node, "blockdev_getmirrorstatus", - [dsk.ToDict() for dsk in disks]) - if not result.fail_msg: - result.payload = [objects.BlockDevStatus.FromDict(i) - for i in result.payload] - return result + if callable(resolver_opts): + req_resolver_opts = resolver_opts(args) + else: + req_resolver_opts = resolver_opts - @_RpcTimeout(_TMO_NORMAL) - def call_blockdev_getmirrorstatus_multi(self, node_list, node_disks): - """Request status of (mirroring) devices from multiple nodes. + if len(args) != len(argdefs): + raise errors.ProgrammerError("Number of passed arguments doesn't match") - This is a multi-node call. + enc_args = map(self._encoder, zip(map(compat.snd, argdefs), args)) + if prep_fn is None: + # for a no-op prep_fn, we serialise the body once, and then we + # reuse it in the dictionary values + body = serializer.DumpJson(enc_args) + pnbody = dict((n, body) for n in node_list) + else: + # for a custom prep_fn, we pass the encoded arguments and the + # node name to the prep_fn, and we serialise its return value + assert callable(prep_fn) + pnbody = dict((n, serializer.DumpJson(prep_fn(n, enc_args))) + for n in node_list) + + result = self._proc(node_list, procedure, pnbody, read_timeout, + req_resolver_opts) + + if postproc_fn: + return dict(map(lambda (key, value): (key, postproc_fn(value)), + result.items())) + else: + return result - """ - result = self._MultiNodeCall(node_list, "blockdev_getmirrorstatus_multi", - [dict((name, [dsk.ToDict() for dsk in disks]) - for name, disks in node_disks.items())]) - for nres in result.values(): - if nres.fail_msg: - continue - for idx, (success, status) in enumerate(nres.payload): - if success: - nres.payload[idx] = (success, objects.BlockDevStatus.FromDict(status)) +def _ObjectToDict(value): + """Converts an object to a dictionary. - return result + @note: See L{objects}. - @_RpcTimeout(_TMO_NORMAL) - def call_blockdev_find(self, node, disk): - """Request identification of a given block device. + """ + return value.ToDict() - This is a single-node call. - """ - result = self._SingleNodeCall(node, "blockdev_find", [disk.ToDict()]) - if not result.fail_msg and result.payload is not None: - result.payload = objects.BlockDevStatus.FromDict(result.payload) - return result +def _ObjectListToDict(value): + """Converts a list of L{objects} to dictionaries. - @_RpcTimeout(_TMO_NORMAL) - def call_blockdev_close(self, node, instance_name, disks): - """Closes the given block devices. + """ + return map(_ObjectToDict, value) - This is a single-node call. - """ - params = [instance_name, [cf.ToDict() for cf in disks]] - return self._SingleNodeCall(node, "blockdev_close", params) +def _EncodeNodeToDiskDict(value): + """Encodes a dictionary with node name as key and disk objects as values. - @_RpcTimeout(_TMO_NORMAL) - def call_blockdev_getsize(self, node, disks): - """Returns the size of the given disks. + """ + return dict((name, _ObjectListToDict(disks)) + for name, disks in value.items()) - This is a single-node call. - """ - params = [[cf.ToDict() for cf in disks]] - return self._SingleNodeCall(node, "blockdev_getsize", params) +def _PrepareFileUpload(getents_fn, filename): + """Loads a file and prepares it for an upload to nodes. - @_RpcTimeout(_TMO_NORMAL) - def call_drbd_disconnect_net(self, node_list, nodes_ip, disks): - """Disconnects the network of the given drbd devices. + """ + statcb = utils.FileStatHelper() + data = _Compress(utils.ReadFile(filename, preread=statcb)) + st = statcb.st - This is a multi-node call. + if getents_fn is None: + getents_fn = runtime.GetEnts - """ - return self._MultiNodeCall(node_list, "drbd_disconnect_net", - [nodes_ip, [cf.ToDict() for cf in disks]]) + getents = getents_fn() - @_RpcTimeout(_TMO_NORMAL) - def call_drbd_attach_net(self, node_list, nodes_ip, - disks, instance_name, multimaster): - """Disconnects the given drbd devices. + virt_filename = vcluster.MakeVirtualPath(filename) - This is a multi-node call. + return [virt_filename, data, st.st_mode, getents.LookupUid(st.st_uid), + getents.LookupGid(st.st_gid), st.st_atime, st.st_mtime] - """ - return self._MultiNodeCall(node_list, "drbd_attach_net", - [nodes_ip, [cf.ToDict() for cf in disks], - instance_name, multimaster]) - @_RpcTimeout(_TMO_SLOW) - def call_drbd_wait_sync(self, node_list, nodes_ip, disks): - """Waits for the synchronization of drbd devices is complete. +def _PrepareFinalizeExportDisks(snap_disks): + """Encodes disks for finalizing export. - This is a multi-node call. + """ + flat_disks = [] - """ - return self._MultiNodeCall(node_list, "drbd_wait_sync", - [nodes_ip, [cf.ToDict() for cf in disks]]) + for disk in snap_disks: + if isinstance(disk, bool): + flat_disks.append(disk) + else: + flat_disks.append(disk.ToDict()) - @_RpcTimeout(_TMO_URGENT) - def call_drbd_helper(self, node_list): - """Gets drbd helper. + return flat_disks - This is a multi-node call. - """ - return self._MultiNodeCall(node_list, "drbd_helper", []) +def _EncodeImportExportIO((ieio, ieioargs)): + """Encodes import/export I/O information. - @classmethod - @_RpcTimeout(_TMO_NORMAL) - def call_upload_file(cls, node_list, file_name, address_list=None): - """Upload a file. + """ + if ieio == constants.IEIO_RAW_DISK: + assert len(ieioargs) == 1 + return (ieio, (ieioargs[0].ToDict(), )) - The node will refuse the operation in case the file is not on the - approved file list. + if ieio == constants.IEIO_SCRIPT: + assert len(ieioargs) == 2 + return (ieio, (ieioargs[0].ToDict(), ieioargs[1])) - This is a multi-node call. + return (ieio, ieioargs) - @type node_list: list - @param node_list: the list of node names to upload to - @type file_name: str - @param file_name: the filename to upload - @type address_list: list or None - @keyword address_list: an optional list of node addresses, in order - to optimize the RPC speed - """ - file_contents = utils.ReadFile(file_name) - data = _Compress(file_contents) - st = os.stat(file_name) - getents = runtime.GetEnts() - params = [file_name, data, st.st_mode, getents.LookupUid(st.st_uid), - getents.LookupGid(st.st_gid), st.st_atime, st.st_mtime] - return cls._StaticMultiNodeCall(node_list, "upload_file", params, - address_list=address_list) - - @classmethod - @_RpcTimeout(_TMO_NORMAL) - def call_write_ssconf_files(cls, node_list, values): - """Write ssconf files. - - This is a multi-node call. +def _EncodeBlockdevRename(value): + """Encodes information for renaming block devices. - """ - return cls._StaticMultiNodeCall(node_list, "write_ssconf_files", [values]) + """ + return [(d.ToDict(), uid) for d, uid in value] - @_RpcTimeout(_TMO_NORMAL) - def call_run_oob(self, node, oob_program, command, remote_node, timeout): - """Runs OOB. - This is a single-node call. +def _AddSpindlesToLegacyNodeInfo(result, space_info): + """Extracts the spindle information from the space info and adds + it to the result dictionary. - """ - return self._SingleNodeCall(node, "run_oob", [oob_program, command, - remote_node, timeout]) + @type result: dict of strings + @param result: dictionary holding the result of the legacy node info + @type space_info: list of dicts of strings + @param space_info: list, each row holding space information of one storage + unit + @rtype: None + @return: does not return anything, manipulates the C{result} variable - @_RpcTimeout(_TMO_FAST) - def call_os_diagnose(self, node_list): - """Request a diagnose of OS definitions. + """ + lvm_pv_info = utils.storage.LookupSpaceInfoByStorageType( + space_info, constants.ST_LVM_PV) + if lvm_pv_info: + result["spindles_free"] = lvm_pv_info["storage_free"] + result["spindles_total"] = lvm_pv_info["storage_size"] + else: + raise errors.OpExecError("No spindle storage information available.") - This is a multi-node call. - """ - return self._MultiNodeCall(node_list, "os_diagnose", []) +def _AddDefaultStorageInfoToLegacyNodeInfo(result, space_info): + """Extracts the storage space information of the default storage type from + the space info and adds it to the result dictionary. - @_RpcTimeout(_TMO_FAST) - def call_os_get(self, node, name): - """Returns an OS definition. + @see: C{_AddSpindlesToLegacyNodeInfo} for parameter information. - This is a single-node call. + """ + # Check if there is at least one row for non-spindle storage info. + no_defaults = (len(space_info) < 1) or \ + (space_info[0]["type"] == constants.ST_LVM_PV and len(space_info) == 1) - """ - result = self._SingleNodeCall(node, "os_get", [name]) - if not result.fail_msg and isinstance(result.payload, dict): - result.payload = objects.OS.FromDict(result.payload) - return result + default_space_info = None + if no_defaults: + logging.warning("No storage info provided for default storage type.") + else: + default_space_info = space_info[0] - @_RpcTimeout(_TMO_FAST) - def call_os_validate(self, required, nodes, name, checks, params): - """Run a validation routine for a given OS. + if default_space_info: + result["name"] = default_space_info["name"] + result["storage_free"] = default_space_info["storage_free"] + result["storage_size"] = default_space_info["storage_size"] - This is a multi-node call. - """ - return self._MultiNodeCall(nodes, "os_validate", - [required, name, checks, params]) +def MakeLegacyNodeInfo(data, require_spindles=False): + """Formats the data returned by L{rpc.RpcRunner.call_node_info}. - @_RpcTimeout(_TMO_NORMAL) - def call_hooks_runner(self, node_list, hpath, phase, env): - """Call the hooks runner. + Converts the data into a single dictionary. This is fine for most use cases, + but some require information from more than one volume group or hypervisor. - Args: - - op: the OpCode instance - - env: a dictionary with the environment + @param require_spindles: add spindle storage information to the legacy node + info - This is a multi-node call. + """ + (bootid, space_info, (hv_info, )) = data - """ - params = [hpath, phase, env] - return self._MultiNodeCall(node_list, "hooks_runner", params) + ret = utils.JoinDisjointDicts(hv_info, {"bootid": bootid}) - @_RpcTimeout(_TMO_NORMAL) - def call_iallocator_runner(self, node, name, idata): - """Call an iallocator on a remote node + if require_spindles: + _AddSpindlesToLegacyNodeInfo(ret, space_info) + _AddDefaultStorageInfoToLegacyNodeInfo(ret, space_info) - Args: - - name: the iallocator name - - input: the json-encoded input string + return ret - This is a single-node call. - """ - return self._SingleNodeCall(node, "iallocator_runner", [name, idata]) +def _AnnotateDParamsDRBD(disk, (drbd_params, data_params, meta_params)): + """Annotates just DRBD disks layouts. - @_RpcTimeout(_TMO_NORMAL) - def call_blockdev_grow(self, node, cf_bdev, amount, dryrun): - """Request a snapshot of the given block device. + """ + assert disk.dev_type == constants.DT_DRBD8 - This is a single-node call. + disk.params = objects.FillDict(drbd_params, disk.params) + (dev_data, dev_meta) = disk.children + dev_data.params = objects.FillDict(data_params, dev_data.params) + dev_meta.params = objects.FillDict(meta_params, dev_meta.params) - """ - return self._SingleNodeCall(node, "blockdev_grow", - [cf_bdev.ToDict(), amount, dryrun]) + return disk - @_RpcTimeout(_TMO_1DAY) - def call_blockdev_export(self, node, cf_bdev, - dest_node, dest_path, cluster_name): - """Export a given disk to another node. - This is a single-node call. +def _AnnotateDParamsGeneric(disk, (params, )): + """Generic disk parameter annotation routine. - """ - return self._SingleNodeCall(node, "blockdev_export", - [cf_bdev.ToDict(), dest_node, dest_path, - cluster_name]) + """ + assert disk.dev_type != constants.DT_DRBD8 - @_RpcTimeout(_TMO_NORMAL) - def call_blockdev_snapshot(self, node, cf_bdev): - """Request a snapshot of the given block device. + disk.params = objects.FillDict(params, disk.params) - This is a single-node call. + return disk - """ - return self._SingleNodeCall(node, "blockdev_snapshot", [cf_bdev.ToDict()]) - @_RpcTimeout(_TMO_NORMAL) - def call_finalize_export(self, node, instance, snap_disks): - """Request the completion of an export operation. +def AnnotateDiskParams(template, disks, disk_params): + """Annotates the disk objects with the disk parameters. - This writes the export config file, etc. + @param template: The disk template used + @param disks: The list of disks objects to annotate + @param disk_params: The disk paramaters for annotation + @returns: A list of disk objects annotated - This is a single-node call. + """ + ld_params = objects.Disk.ComputeLDParams(template, disk_params) - """ - flat_disks = [] - for disk in snap_disks: - if isinstance(disk, bool): - flat_disks.append(disk) - else: - flat_disks.append(disk.ToDict()) + if template == constants.DT_DRBD8: + annotation_fn = _AnnotateDParamsDRBD + elif template == constants.DT_DISKLESS: + annotation_fn = lambda disk, _: disk + else: + annotation_fn = _AnnotateDParamsGeneric - return self._SingleNodeCall(node, "finalize_export", - [self._InstDict(instance), flat_disks]) + return [annotation_fn(disk.Copy(), ld_params) for disk in disks] - @_RpcTimeout(_TMO_FAST) - def call_export_info(self, node, path): - """Queries the export information in a given path. - This is a single-node call. +def _GetExclusiveStorageFlag(cfg, node_uuid): + ni = cfg.GetNodeInfo(node_uuid) + if ni is None: + raise errors.OpPrereqError("Invalid node name %s" % node_uuid, + errors.ECODE_NOENT) + return cfg.GetNdParams(ni)[constants.ND_EXCLUSIVE_STORAGE] - """ - return self._SingleNodeCall(node, "export_info", [path]) - @_RpcTimeout(_TMO_FAST) - def call_export_list(self, node_list): - """Gets the stored exports list. +def _AddExclusiveStorageFlagToLvmStorageUnits(storage_units, es_flag): + """Adds the exclusive storage flag to lvm units. - This is a multi-node call. + This function creates a copy of the storage_units lists, with the + es_flag being added to all lvm storage units. - """ - return self._MultiNodeCall(node_list, "export_list", []) + @type storage_units: list of pairs (string, string) + @param storage_units: list of 'raw' storage units, consisting only of + (storage_type, storage_key) + @type es_flag: boolean + @param es_flag: exclusive storage flag + @rtype: list of tuples (string, string, list) + @return: list of storage units (storage_type, storage_key, params) with + the params containing the es_flag for lvm-vg storage units - @_RpcTimeout(_TMO_FAST) - def call_export_remove(self, node, export): - """Requests removal of a given export. + """ + result = [] + for (storage_type, storage_key) in storage_units: + if storage_type in [constants.ST_LVM_VG, constants.ST_LVM_PV]: + result.append((storage_type, storage_key, [es_flag])) + else: + result.append((storage_type, storage_key, [])) + return result - This is a single-node call. - """ - return self._SingleNodeCall(node, "export_remove", [export]) +def GetExclusiveStorageForNodes(cfg, node_uuids): + """Return the exclusive storage flag for all the given nodes. - @classmethod - @_RpcTimeout(_TMO_NORMAL) - def call_node_leave_cluster(cls, node, modify_ssh_setup): - """Requests a node to clean the cluster information it has. + @type cfg: L{config.ConfigWriter} + @param cfg: cluster configuration + @type node_uuids: list or tuple + @param node_uuids: node UUIDs for which to read the flag + @rtype: dict + @return: mapping from node uuids to exclusive storage flags + @raise errors.OpPrereqError: if any given node name has no corresponding + node - This will remove the configuration information from the ganeti data - dir. - - This is a single-node call. + """ + getflag = lambda n: _GetExclusiveStorageFlag(cfg, n) + flags = map(getflag, node_uuids) + return dict(zip(node_uuids, flags)) + + +def PrepareStorageUnitsForNodes(cfg, storage_units, node_uuids): + """Return the lvm storage unit for all the given nodes. + + Main purpose of this function is to map the exclusive storage flag, which + can be different for each node, to the default LVM storage unit. + + @type cfg: L{config.ConfigWriter} + @param cfg: cluster configuration + @type storage_units: list of pairs (string, string) + @param storage_units: list of 'raw' storage units, e.g. pairs of + (storage_type, storage_key) + @type node_uuids: list or tuple + @param node_uuids: node UUIDs for which to read the flag + @rtype: dict + @return: mapping from node uuids to a list of storage units which include + the exclusive storage flag for lvm storage + @raise errors.OpPrereqError: if any given node name has no corresponding + node - """ - return cls._StaticSingleNodeCall(node, "node_leave_cluster", - [modify_ssh_setup]) + """ + getunit = lambda n: _AddExclusiveStorageFlagToLvmStorageUnits( + storage_units, _GetExclusiveStorageFlag(cfg, n)) + flags = map(getunit, node_uuids) + return dict(zip(node_uuids, flags)) + + +#: Generic encoders +_ENCODERS = { + rpc_defs.ED_OBJECT_DICT: _ObjectToDict, + rpc_defs.ED_OBJECT_DICT_LIST: _ObjectListToDict, + rpc_defs.ED_NODE_TO_DISK_DICT: _EncodeNodeToDiskDict, + rpc_defs.ED_COMPRESS: _Compress, + rpc_defs.ED_FINALIZE_EXPORT_DISKS: _PrepareFinalizeExportDisks, + rpc_defs.ED_IMPEXP_IO: _EncodeImportExportIO, + rpc_defs.ED_BLOCKDEV_RENAME: _EncodeBlockdevRename, + } + + +class RpcRunner(_RpcClientBase, + _generated_rpc.RpcClientDefault, + _generated_rpc.RpcClientBootstrap, + _generated_rpc.RpcClientDnsOnly, + _generated_rpc.RpcClientConfig): + """RPC runner class. - @_RpcTimeout(_TMO_FAST) - def call_node_volumes(self, node_list): - """Gets all volumes on node(s). + """ + def __init__(self, cfg, lock_monitor_cb, _req_process_fn=None, _getents=None): + """Initialized the RPC runner. - This is a multi-node call. + @type cfg: L{config.ConfigWriter} + @param cfg: Configuration + @type lock_monitor_cb: callable + @param lock_monitor_cb: Lock monitor callback """ - return self._MultiNodeCall(node_list, "node_volumes", []) + self._cfg = cfg - @_RpcTimeout(_TMO_FAST) - def call_node_demote_from_mc(self, node): - """Demote a node from the master candidate role. + encoders = _ENCODERS.copy() - This is a single-node call. + encoders.update({ + # Encoders requiring configuration object + rpc_defs.ED_INST_DICT: self._InstDict, + rpc_defs.ED_INST_DICT_HVP_BEP_DP: self._InstDictHvpBepDp, + rpc_defs.ED_INST_DICT_OSP_DP: self._InstDictOspDp, + rpc_defs.ED_NIC_DICT: self._NicDict, - """ - return self._SingleNodeCall(node, "node_demote_from_mc", []) - - @_RpcTimeout(_TMO_NORMAL) - def call_node_powercycle(self, node, hypervisor): - """Tries to powercycle a node. + # Encoders annotating disk parameters + rpc_defs.ED_DISKS_DICT_DP: self._DisksDictDP, + rpc_defs.ED_MULTI_DISKS_DICT_DP: self._MultiDiskDictDP, + rpc_defs.ED_SINGLE_DISK_DICT_DP: self._SingleDiskDictDP, - This is a single-node call. + # Encoders with special requirements + rpc_defs.ED_FILE_DETAILS: compat.partial(_PrepareFileUpload, _getents), + }) - """ - return self._SingleNodeCall(node, "node_powercycle", [hypervisor]) + # Resolver using configuration + resolver = compat.partial(_NodeConfigResolver, cfg.GetNodeInfo, + cfg.GetAllNodesInfo) - @_RpcTimeout(None) - def call_test_delay(self, node_list, duration): - """Sleep for a fixed time on given node(s). + # Pylint doesn't recognize multiple inheritance properly, see + # and + # + # pylint: disable=W0233 + _RpcClientBase.__init__(self, resolver, encoders.get, + lock_monitor_cb=lock_monitor_cb, + _req_process_fn=_req_process_fn) + _generated_rpc.RpcClientConfig.__init__(self) + _generated_rpc.RpcClientBootstrap.__init__(self) + _generated_rpc.RpcClientDnsOnly.__init__(self) + _generated_rpc.RpcClientDefault.__init__(self) - This is a multi-node call. + def _NicDict(self, nic): + """Convert the given nic to a dict and encapsulate netinfo """ - return self._MultiNodeCall(node_list, "test_delay", [duration], - read_timeout=int(duration + 5)) + n = copy.deepcopy(nic) + if n.network: + net_uuid = self._cfg.LookupNetwork(n.network) + if net_uuid: + nobj = self._cfg.GetNetwork(net_uuid) + n.netinfo = objects.Network.ToDict(nobj) + return n.ToDict() - @_RpcTimeout(_TMO_FAST) - def call_file_storage_dir_create(self, node, file_storage_dir): - """Create the given file storage directory. - - This is a single-node call. - - """ - return self._SingleNodeCall(node, "file_storage_dir_create", - [file_storage_dir]) + def _InstDict(self, instance, hvp=None, bep=None, osp=None): + """Convert the given instance to a dict. - @_RpcTimeout(_TMO_FAST) - def call_file_storage_dir_remove(self, node, file_storage_dir): - """Remove the given file storage directory. + This is done via the instance's ToDict() method and additionally + we fill the hvparams with the cluster defaults. - This is a single-node call. + @type instance: L{objects.Instance} + @param instance: an Instance object + @type hvp: dict or None + @param hvp: a dictionary with overridden hypervisor parameters + @type bep: dict or None + @param bep: a dictionary with overridden backend parameters + @type osp: dict or None + @param osp: a dictionary with overridden os parameters + @rtype: dict + @return: the instance dict, with the hvparams filled with the + cluster defaults """ - return self._SingleNodeCall(node, "file_storage_dir_remove", - [file_storage_dir]) - - @_RpcTimeout(_TMO_FAST) - def call_file_storage_dir_rename(self, node, old_file_storage_dir, - new_file_storage_dir): - """Rename file storage directory. + idict = instance.ToDict() + cluster = self._cfg.GetClusterInfo() + idict["hvparams"] = cluster.FillHV(instance) + if hvp is not None: + idict["hvparams"].update(hvp) + idict["beparams"] = cluster.FillBE(instance) + if bep is not None: + idict["beparams"].update(bep) + idict["osparams"] = cluster.SimpleFillOS(instance.os, instance.osparams) + if osp is not None: + idict["osparams"].update(osp) + idict["disks"] = self._DisksDictDP((instance.disks, instance)) + for nic in idict["nics"]: + nic["nicparams"] = objects.FillDict( + cluster.nicparams[constants.PP_DEFAULT], + nic["nicparams"]) + network = nic.get("network", None) + if network: + net_uuid = self._cfg.LookupNetwork(network) + if net_uuid: + nobj = self._cfg.GetNetwork(net_uuid) + nic["netinfo"] = objects.Network.ToDict(nobj) + return idict - This is a single-node call. + def _InstDictHvpBepDp(self, (instance, hvp, bep)): + """Wrapper for L{_InstDict}. """ - return self._SingleNodeCall(node, "file_storage_dir_rename", - [old_file_storage_dir, new_file_storage_dir]) - - @classmethod - @_RpcTimeout(_TMO_URGENT) - def call_jobqueue_update(cls, node_list, address_list, file_name, content): - """Update job queue. + return self._InstDict(instance, hvp=hvp, bep=bep) - This is a multi-node call. + def _InstDictOspDp(self, (instance, osparams)): + """Wrapper for L{_InstDict}. """ - return cls._StaticMultiNodeCall(node_list, "jobqueue_update", - [file_name, _Compress(content)], - address_list=address_list) - - @classmethod - @_RpcTimeout(_TMO_NORMAL) - def call_jobqueue_purge(cls, node): - """Purge job queue. + return self._InstDict(instance, osp=osparams) - This is a single-node call. + def _DisksDictDP(self, (disks, instance)): + """Wrapper for L{AnnotateDiskParams}. """ - return cls._StaticSingleNodeCall(node, "jobqueue_purge", []) + diskparams = self._cfg.GetInstanceDiskParams(instance) + return [disk.ToDict() + for disk in AnnotateDiskParams(instance.disk_template, + disks, diskparams)] - @classmethod - @_RpcTimeout(_TMO_URGENT) - def call_jobqueue_rename(cls, node_list, address_list, rename): - """Rename a job queue file. - - This is a multi-node call. + def _MultiDiskDictDP(self, disks_insts): + """Wrapper for L{AnnotateDiskParams}. + Supports a list of (disk, instance) tuples. """ - return cls._StaticMultiNodeCall(node_list, "jobqueue_rename", rename, - address_list=address_list) - - @_RpcTimeout(_TMO_NORMAL) - def call_hypervisor_validate_params(self, node_list, hvname, hvparams): - """Validate the hypervisor params. - - This is a multi-node call. + return [disk for disk_inst in disks_insts + for disk in self._DisksDictDP(disk_inst)] - @type node_list: list - @param node_list: the list of nodes to query - @type hvname: string - @param hvname: the hypervisor name - @type hvparams: dict - @param hvparams: the hypervisor parameters to be validated + def _SingleDiskDictDP(self, (disk, instance)): + """Wrapper for L{AnnotateDiskParams}. """ - cluster = self._cfg.GetClusterInfo() - hv_full = objects.FillDict(cluster.hvparams.get(hvname, {}), hvparams) - return self._MultiNodeCall(node_list, "hypervisor_validate_params", - [hvname, hv_full]) + (anno_disk,) = self._DisksDictDP(([disk], instance)) + return anno_disk - @_RpcTimeout(_TMO_NORMAL) - def call_x509_cert_create(self, node, validity): - """Creates a new X509 certificate for SSL/TLS. - This is a single-node call. +class JobQueueRunner(_RpcClientBase, _generated_rpc.RpcClientJobQueue): + """RPC wrappers for job queue. - @type validity: int - @param validity: Validity in seconds + """ + def __init__(self, context, address_list): + """Initializes this class. """ - return self._SingleNodeCall(node, "x509_cert_create", [validity]) - - @_RpcTimeout(_TMO_NORMAL) - def call_x509_cert_remove(self, node, name): - """Removes a X509 certificate. - - This is a single-node call. + if address_list is None: + resolver = compat.partial(_SsconfResolver, True) + else: + # Caller provided an address list + resolver = _StaticResolver(address_list) - @type name: string - @param name: Certificate name + _RpcClientBase.__init__(self, resolver, _ENCODERS.get, + lock_monitor_cb=context.glm.AddToLockMonitor) + _generated_rpc.RpcClientJobQueue.__init__(self) - """ - return self._SingleNodeCall(node, "x509_cert_remove", [name]) - @_RpcTimeout(_TMO_NORMAL) - def call_import_start(self, node, opts, instance, component, - dest, dest_args): - """Starts a listener for an import. +class BootstrapRunner(_RpcClientBase, + _generated_rpc.RpcClientBootstrap, + _generated_rpc.RpcClientDnsOnly): + """RPC wrappers for bootstrapping. - This is a single-node call. - - @type node: string - @param node: Node name - @type instance: C{objects.Instance} - @param instance: Instance object - @type component: string - @param component: which part of the instance is being imported + """ + def __init__(self): + """Initializes this class. """ - return self._SingleNodeCall(node, "import_start", - [opts.ToDict(), - self._InstDict(instance), component, dest, - _EncodeImportExportIO(dest, dest_args)]) + # Pylint doesn't recognize multiple inheritance properly, see + # and + # + # pylint: disable=W0233 + _RpcClientBase.__init__(self, compat.partial(_SsconfResolver, True), + _ENCODERS.get) + _generated_rpc.RpcClientBootstrap.__init__(self) + _generated_rpc.RpcClientDnsOnly.__init__(self) - @_RpcTimeout(_TMO_NORMAL) - def call_export_start(self, node, opts, host, port, - instance, component, source, source_args): - """Starts an export daemon. - This is a single-node call. +class DnsOnlyRunner(_RpcClientBase, _generated_rpc.RpcClientDnsOnly): + """RPC wrappers for calls using only DNS. - @type node: string - @param node: Node name - @type instance: C{objects.Instance} - @param instance: Instance object - @type component: string - @param component: which part of the instance is being imported - - """ - return self._SingleNodeCall(node, "export_start", - [opts.ToDict(), host, port, - self._InstDict(instance), - component, source, - _EncodeImportExportIO(source, source_args)]) - - @_RpcTimeout(_TMO_FAST) - def call_impexp_status(self, node, names): - """Gets the status of an import or export. - - This is a single-node call. - - @type node: string - @param node: Node name - @type names: List of strings - @param names: Import/export names - @rtype: List of L{objects.ImportExportStatus} instances - @return: Returns a list of the state of each named import/export or None if - a status couldn't be retrieved + """ + def __init__(self): + """Initialize this class. """ - result = self._SingleNodeCall(node, "impexp_status", [names]) + _RpcClientBase.__init__(self, compat.partial(_SsconfResolver, False), + _ENCODERS.get) + _generated_rpc.RpcClientDnsOnly.__init__(self) - if not result.fail_msg: - decoded = [] - for i in result.payload: - if i is None: - decoded.append(None) - continue - decoded.append(objects.ImportExportStatus.FromDict(i)) +class ConfigRunner(_RpcClientBase, _generated_rpc.RpcClientConfig): + """RPC wrappers for L{config}. - result.payload = decoded - - return result - - @_RpcTimeout(_TMO_NORMAL) - def call_impexp_abort(self, node, name): - """Aborts an import or export. - - This is a single-node call. - - @type node: string - @param node: Node name - @type name: string - @param name: Import/export name + """ + def __init__(self, context, address_list, _req_process_fn=None, + _getents=None): + """Initializes this class. """ - return self._SingleNodeCall(node, "impexp_abort", [name]) + if context: + lock_monitor_cb = context.glm.AddToLockMonitor + else: + lock_monitor_cb = None - @_RpcTimeout(_TMO_NORMAL) - def call_impexp_cleanup(self, node, name): - """Cleans up after an import or export. + if address_list is None: + resolver = compat.partial(_SsconfResolver, True) + else: + # Caller provided an address list + resolver = _StaticResolver(address_list) - This is a single-node call. + encoders = _ENCODERS.copy() - @type node: string - @param node: Node name - @type name: string - @param name: Import/export name + encoders.update({ + rpc_defs.ED_FILE_DETAILS: compat.partial(_PrepareFileUpload, _getents), + }) - """ - return self._SingleNodeCall(node, "impexp_cleanup", [name]) + _RpcClientBase.__init__(self, resolver, encoders.get, + lock_monitor_cb=lock_monitor_cb, + _req_process_fn=_req_process_fn) + _generated_rpc.RpcClientConfig.__init__(self)