#
#
-# Copyright (C) 2006, 2007 Google Inc.
+# Copyright (C) 2006, 2007, 2010 Google Inc.
#
# This program is free software; you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
# R0904: Too many public methods
import os
-import socket
-import httplib
import logging
-
-import simplejson
+import zlib
+import base64
+import pycurl
+import threading
from ganeti import utils
from ganeti import objects
from ganeti import http
+from ganeti import serializer
+from ganeti import constants
+from ganeti import errors
+from ganeti import netutils
+from ganeti import ssconf
+
+# pylint has a bug here, doesn't see this import
+import ganeti.http.client # pylint: disable-msg=W0611
+
+
+# Timeout for connecting to nodes (seconds)
+_RPC_CONNECT_TIMEOUT = 5
+
+_RPC_CLIENT_HEADERS = [
+ "Content-type: %s" % http.HTTP_APP_JSON,
+ ]
+
+# Various time constants for the timeout table
+_TMO_URGENT = 60 # one minute
+_TMO_FAST = 5 * 60 # five minutes
+_TMO_NORMAL = 15 * 60 # 15 minutes
+_TMO_SLOW = 3600 # one hour
+_TMO_4HRS = 4 * 3600
+_TMO_1DAY = 86400
+
+# Timeout table that will be built later by decorators
+# Guidelines for choosing timeouts:
+# - call used during watcher: timeout -> 1min, _TMO_URGENT
+# - trivial (but be sure it is trivial) (e.g. reading a file): 5min, _TMO_FAST
+# - other calls: 15 min, _TMO_NORMAL
+# - special calls (instance add, etc.): either _TMO_SLOW (1h) or huge timeouts
+
+_TIMEOUTS = {
+}
+
+
+def Init():
+ """Initializes the module-global HTTP client manager.
+
+ Must be called before using any RPC function and while exactly one thread is
+ running.
+
+ """
+ # curl_global_init(3) and curl_global_cleanup(3) must be called with only
+ # one thread running. This check is just a safety measure -- it doesn't
+ # cover all cases.
+ assert threading.activeCount() == 1, \
+ "Found more than one active thread when initializing pycURL"
+
+ logging.info("Using PycURL %s", pycurl.version)
+
+ pycurl.global_init(pycurl.GLOBAL_ALL)
+
+
+def Shutdown():
+ """Stops the module-global HTTP client manager.
+
+ Must be called before quitting the program and while exactly one thread is
+ running.
+
+ """
+ pycurl.global_cleanup()
+
+
+def _ConfigRpcCurl(curl):
+ noded_cert = str(constants.NODED_CERT_FILE)
+
+ curl.setopt(pycurl.FOLLOWLOCATION, False)
+ curl.setopt(pycurl.CAINFO, noded_cert)
+ curl.setopt(pycurl.SSL_VERIFYHOST, 0)
+ curl.setopt(pycurl.SSL_VERIFYPEER, True)
+ curl.setopt(pycurl.SSLCERTTYPE, "PEM")
+ curl.setopt(pycurl.SSLCERT, noded_cert)
+ curl.setopt(pycurl.SSLKEYTYPE, "PEM")
+ curl.setopt(pycurl.SSLKEY, noded_cert)
+ curl.setopt(pycurl.CONNECTTIMEOUT, _RPC_CONNECT_TIMEOUT)
+
+
+class _RpcThreadLocal(threading.local):
+ def GetHttpClientPool(self):
+ """Returns a per-thread HTTP client pool.
+
+ @rtype: L{http.client.HttpClientPool}
+
+ """
+ try:
+ pool = self.hcp
+ except AttributeError:
+ pool = http.client.HttpClientPool(_ConfigRpcCurl)
+ self.hcp = pool
+
+ return pool
+
+
+_thread_local = _RpcThreadLocal()
+
+
+def _RpcTimeout(secs):
+ """Timeout decorator.
+
+ When applied to a rpc call_* function, it updates the global timeout
+ table with the given function/timeout.
+
+ """
+ def decorator(f):
+ name = f.__name__
+ assert name.startswith("call_")
+ _TIMEOUTS[name[len("call_"):]] = secs
+ return f
+ return decorator
+
+
+def RunWithRPC(fn):
+ """RPC-wrapper decorator.
+
+ When applied to a function, it runs it with the RPC system
+ initialized, and it shutsdown the system afterwards. This means the
+ function must be called without RPC being initialized.
+
+ """
+ def wrapper(*args, **kwargs):
+ Init()
+ try:
+ return fn(*args, **kwargs)
+ finally:
+ Shutdown()
+ return wrapper
+
+
+class RpcResult(object):
+ """RPC Result class.
+
+ This class holds an RPC result. It is needed since in multi-node
+ calls we can't raise an exception just because one one out of many
+ failed, and therefore we use this class to encapsulate the result.
+
+ @ivar data: the data payload, for successful results, or None
+ @ivar call: the name of the RPC call
+ @ivar node: the name of the node to which we made the call
+ @ivar offline: whether the operation failed because the node was
+ offline, as opposed to actual failure; offline=True will always
+ imply failed=True, in order to allow simpler checking if
+ the user doesn't care about the exact failure mode
+ @ivar fail_msg: the error message if the call failed
+
+ """
+ def __init__(self, data=None, failed=False, offline=False,
+ call=None, node=None):
+ self.offline = offline
+ self.call = call
+ self.node = node
+
+ if offline:
+ self.fail_msg = "Node is marked offline"
+ self.data = self.payload = None
+ elif failed:
+ self.fail_msg = self._EnsureErr(data)
+ self.data = self.payload = None
+ else:
+ self.data = data
+ if not isinstance(self.data, (tuple, list)):
+ self.fail_msg = ("RPC layer error: invalid result type (%s)" %
+ type(self.data))
+ self.payload = None
+ elif len(data) != 2:
+ self.fail_msg = ("RPC layer error: invalid result length (%d), "
+ "expected 2" % len(self.data))
+ self.payload = None
+ elif not self.data[0]:
+ self.fail_msg = self._EnsureErr(self.data[1])
+ self.payload = None
+ else:
+ # finally success
+ self.fail_msg = None
+ self.payload = data[1]
+
+ assert hasattr(self, "call")
+ assert hasattr(self, "data")
+ assert hasattr(self, "fail_msg")
+ assert hasattr(self, "node")
+ assert hasattr(self, "offline")
+ assert hasattr(self, "payload")
+
+ @staticmethod
+ def _EnsureErr(val):
+ """Helper to ensure we return a 'True' value for error."""
+ if val:
+ return val
+ else:
+ return "No error information"
+
+ def Raise(self, msg, prereq=False, ecode=None):
+ """If the result has failed, raise an OpExecError.
+
+ This is used so that LU code doesn't have to check for each
+ result, but instead can call this function.
+
+ """
+ if not self.fail_msg:
+ return
+
+ if not msg: # one could pass None for default message
+ msg = ("Call '%s' to node '%s' has failed: %s" %
+ (self.call, self.node, self.fail_msg))
+ else:
+ msg = "%s: %s" % (msg, self.fail_msg)
+ if prereq:
+ ec = errors.OpPrereqError
+ else:
+ ec = errors.OpExecError
+ if ecode is not None:
+ args = (msg, ecode)
+ else:
+ args = (msg, )
+ raise ec(*args) # pylint: disable-msg=W0142
+
+
+def _AddressLookup(node_list,
+ ssc=ssconf.SimpleStore,
+ nslookup_fn=netutils.Hostname.GetIP):
+ """Return addresses for given node names.
+
+ @type node_list: list
+ @param node_list: List of node names
+ @type ssc: class
+ @param ssc: SimpleStore class that is used to obtain node->ip mappings
+ @type lookup_fn: callable
+ @param lookup_fn: function use to do NS lookup
+ @rtype: list of addresses and/or None's
+ @returns: List of corresponding addresses, if found
+
+ """
+ iplist = ssc().GetNodePrimaryIPList()
+ addresses = []
+ ipmap = dict(entry.split() for entry in iplist)
+ for node in node_list:
+ address = ipmap.get(node)
+ if address is None:
+ address = nslookup_fn(node)
+ addresses.append(address)
+
+ return addresses
class Client:
list of nodes, will contact (in parallel) all nodes, and return a
dict of results (key: node name, value: result).
- One current bug is that generic failure is still signalled by
+ One current bug is that generic failure is still signaled by
'False' result, which is not good. This overloading of values can
cause bugs.
"""
- def __init__(self, procedure, args):
+ def __init__(self, procedure, body, port, address_lookup_fn=_AddressLookup):
+ assert procedure in _TIMEOUTS, ("New RPC call not declared in the"
+ " timeouts table")
self.procedure = procedure
- self.args = args
- self.body = simplejson.dumps(args)
-
- self.port = utils.GetNodeDaemonPort()
- self.nodepw = utils.GetNodeDaemonPassword()
- self.nc = {}
+ self.body = body
+ self.port = port
+ self._request = {}
+ self._address_lookup_fn = address_lookup_fn
- def ConnectList(self, node_list, address_list=None):
+ def ConnectList(self, node_list, address_list=None, read_timeout=None):
"""Add a list of nodes to the target nodes.
@type node_list: list
@type address_list: list or None
@keyword address_list: either None or a list with node addresses,
which must have the same length as the node list
+ @type read_timeout: int
+ @param read_timeout: overwrites default timeout for operation
"""
if address_list is None:
- address_list = [None for _ in node_list]
- else:
- assert len(node_list) == len(address_list), \
- "Name and address lists should have the same length"
+ # Always use IP address instead of node name
+ address_list = self._address_lookup_fn(node_list)
+
+ assert len(node_list) == len(address_list), \
+ "Name and address lists must have the same length"
+
for node, address in zip(node_list, address_list):
- self.ConnectNode(node, address)
+ self.ConnectNode(node, address, read_timeout=read_timeout)
- def ConnectNode(self, name, address=None):
+ def ConnectNode(self, name, address=None, read_timeout=None):
"""Add a node to the target list.
@type name: str
@param name: the node name
@type address: str
- @keyword address: the node address, if known
+ @param address: the node address, if known
+ @type read_timeout: int
+ @param read_timeout: overwrites default timeout for operation
"""
if address is None:
- address = name
+ # Always use IP address instead of node name
+ address = self._address_lookup_fn([name])[0]
- self.nc[name] = http.HttpClientRequest(address, self.port, http.HTTP_PUT,
- "/%s" % self.procedure,
- post_data=self.body)
+ assert(address is not None)
- def GetResults(self):
+ if read_timeout is None:
+ read_timeout = _TIMEOUTS[self.procedure]
+
+ self._request[name] = \
+ http.client.HttpClientRequest(str(address), self.port,
+ http.HTTP_PUT, str("/%s" % self.procedure),
+ headers=_RPC_CLIENT_HEADERS,
+ post_data=str(self.body),
+ read_timeout=read_timeout)
+
+ def GetResults(self, http_pool=None):
"""Call nodes and return results.
@rtype: list
- @returns: List of RPC results
+ @return: List of RPC results
"""
- # TODO: Shared and reused manager
- mgr = http.HttpClientManager()
- try:
- mgr.ExecRequests(self.nc.values())
- finally:
- mgr.Shutdown()
+ if not http_pool:
+ http_pool = _thread_local.GetHttpClientPool()
+
+ http_pool.ProcessRequests(self._request.values())
results = {}
- for name, req in self.nc.iteritems():
- if req.success and req.resp_status == http.HTTP_OK:
- results[name] = simplejson.loads(req.resp_body)
+ for name, req in self._request.iteritems():
+ if req.success and req.resp_status_code == http.HTTP_OK:
+ results[name] = RpcResult(data=serializer.LoadJson(req.resp_body),
+ node=name, call=self.procedure)
continue
+ # TODO: Better error reporting
if req.error:
msg = req.error
else:
msg = req.resp_body
- logging.error("RPC error from node %s: %s", name, msg)
- results[name] = False
+ logging.error("RPC error in %s from node %s: %s",
+ self.procedure, name, msg)
+ results[name] = RpcResult(data=msg, failed=True, node=name,
+ call=self.procedure)
return results
+def _EncodeImportExportIO(ieio, ieioargs):
+ """Encodes import/export I/O information.
+
+ """
+ if ieio == constants.IEIO_RAW_DISK:
+ assert len(ieioargs) == 1
+ return (ieioargs[0].ToDict(), )
+
+ if ieio == constants.IEIO_SCRIPT:
+ assert len(ieioargs) == 2
+ return (ieioargs[0].ToDict(), ieioargs[1])
+
+ return ieioargs
+
+
class RpcRunner(object):
"""RPC runner class"""
"""
self._cfg = cfg
+ self.port = netutils.GetDaemonPort(constants.NODED)
- def _InstDict(self, instance):
+ def _InstDict(self, instance, hvp=None, bep=None, osp=None):
"""Convert the given instance to a dict.
This is done via the instance's ToDict() method and additionally
@type instance: L{objects.Instance}
@param instance: an Instance object
+ @type hvp: dict or None
+ @param hvp: a dictionary with overridden hypervisor parameters
+ @type bep: dict or None
+ @param bep: a dictionary with overridden backend parameters
+ @type osp: dict or None
+ @param osp: a dictionary with overriden os parameters
@rtype: dict
@return: the instance dict, with the hvparams filled with the
cluster defaults
idict = instance.ToDict()
cluster = self._cfg.GetClusterInfo()
idict["hvparams"] = cluster.FillHV(instance)
+ if hvp is not None:
+ idict["hvparams"].update(hvp)
idict["beparams"] = cluster.FillBE(instance)
+ if bep is not None:
+ idict["beparams"].update(bep)
+ idict["osparams"] = cluster.SimpleFillOS(instance.os, instance.osparams)
+ if osp is not None:
+ idict["osparams"].update(osp)
+ for nic in idict["nics"]:
+ nic['nicparams'] = objects.FillDict(
+ cluster.nicparams[constants.PP_DEFAULT],
+ nic['nicparams'])
return idict
- def _ConnectList(self, client, node_list):
+ def _ConnectList(self, client, node_list, call, read_timeout=None):
"""Helper for computing node addresses.
- @type client: L{Client}
+ @type client: L{ganeti.rpc.Client}
@param client: a C{Client} instance
@type node_list: list
@param node_list: the node list we should connect
+ @type call: string
+ @param call: the name of the remote procedure call, for filling in
+ correctly any eventual offline nodes' results
+ @type read_timeout: int
+ @param read_timeout: overwrites the default read timeout for the
+ given operation
"""
all_nodes = self._cfg.GetAllNodesInfo()
+ name_list = []
addr_list = []
+ skip_dict = {}
for node in node_list:
if node in all_nodes:
+ if all_nodes[node].offline:
+ skip_dict[node] = RpcResult(node=node, offline=True, call=call)
+ continue
val = all_nodes[node].primary_ip
else:
val = None
addr_list.append(val)
- client.ConnectList(node_list, address_list=addr_list)
+ name_list.append(node)
+ if name_list:
+ client.ConnectList(name_list, address_list=addr_list,
+ read_timeout=read_timeout)
+ return skip_dict
- def _ConnectNode(self, client, node):
+ def _ConnectNode(self, client, node, call, read_timeout=None):
"""Helper for computing one node's address.
- @type client: L{Client}
+ @type client: L{ganeti.rpc.Client}
@param client: a C{Client} instance
@type node: str
@param node: the node we should connect
+ @type call: string
+ @param call: the name of the remote procedure call, for filling in
+ correctly any eventual offline nodes' results
+ @type read_timeout: int
+ @param read_timeout: overwrites the default read timeout for the
+ given operation
"""
node_info = self._cfg.GetNodeInfo(node)
if node_info is not None:
+ if node_info.offline:
+ return RpcResult(node=node, offline=True, call=call)
addr = node_info.primary_ip
else:
addr = None
- client.ConnectNode(node, address=addr)
+ client.ConnectNode(node, address=addr, read_timeout=read_timeout)
+
+ def _MultiNodeCall(self, node_list, procedure, args, read_timeout=None):
+ """Helper for making a multi-node call
+
+ """
+ body = serializer.DumpJson(args, indent=False)
+ c = Client(procedure, body, self.port)
+ skip_dict = self._ConnectList(c, node_list, procedure,
+ read_timeout=read_timeout)
+ skip_dict.update(c.GetResults())
+ return skip_dict
+
+ @classmethod
+ def _StaticMultiNodeCall(cls, node_list, procedure, args,
+ address_list=None, read_timeout=None):
+ """Helper for making a multi-node static call
+
+ """
+ body = serializer.DumpJson(args, indent=False)
+ c = Client(procedure, body, netutils.GetDaemonPort(constants.NODED))
+ c.ConnectList(node_list, address_list=address_list,
+ read_timeout=read_timeout)
+ return c.GetResults()
+
+ def _SingleNodeCall(self, node, procedure, args, read_timeout=None):
+ """Helper for making a single-node call
+
+ """
+ body = serializer.DumpJson(args, indent=False)
+ c = Client(procedure, body, self.port)
+ result = self._ConnectNode(c, node, procedure, read_timeout=read_timeout)
+ if result is None:
+ # we did connect, node is not offline
+ result = c.GetResults()[node]
+ return result
+
+ @classmethod
+ def _StaticSingleNodeCall(cls, node, procedure, args, read_timeout=None):
+ """Helper for making a single-node static call
+
+ """
+ body = serializer.DumpJson(args, indent=False)
+ c = Client(procedure, body, netutils.GetDaemonPort(constants.NODED))
+ c.ConnectNode(node, read_timeout=read_timeout)
+ return c.GetResults()[node]
- def call_volume_list(self, node_list, vg_name):
+ @staticmethod
+ def _Compress(data):
+ """Compresses a string for transport over RPC.
+
+ Small amounts of data are not compressed.
+
+ @type data: str
+ @param data: Data
+ @rtype: tuple
+ @return: Encoded data to send
+
+ """
+ # Small amounts of data are not compressed
+ if len(data) < 512:
+ return (constants.RPC_ENCODING_NONE, data)
+
+ # Compress with zlib and encode in base64
+ return (constants.RPC_ENCODING_ZLIB_BASE64,
+ base64.b64encode(zlib.compress(data, 3)))
+
+ #
+ # Begin RPC calls
+ #
+
+ @_RpcTimeout(_TMO_URGENT)
+ def call_lv_list(self, node_list, vg_name):
"""Gets the logical volumes present in a given volume group.
This is a multi-node call.
"""
- c = Client("volume_list", [vg_name])
- self._ConnectList(c, node_list)
- return c.GetResults()
+ return self._MultiNodeCall(node_list, "lv_list", [vg_name])
+ @_RpcTimeout(_TMO_URGENT)
def call_vg_list(self, node_list):
"""Gets the volume group list.
This is a multi-node call.
"""
- c = Client("vg_list", [])
- self._ConnectList(c, node_list)
- return c.GetResults()
+ return self._MultiNodeCall(node_list, "vg_list", [])
+
+ @_RpcTimeout(_TMO_NORMAL)
+ def call_storage_list(self, node_list, su_name, su_args, name, fields):
+ """Get list of storage units.
+
+ This is a multi-node call.
+
+ """
+ return self._MultiNodeCall(node_list, "storage_list",
+ [su_name, su_args, name, fields])
+
+ @_RpcTimeout(_TMO_NORMAL)
+ def call_storage_modify(self, node, su_name, su_args, name, changes):
+ """Modify a storage unit.
+
+ This is a single-node call.
+
+ """
+ return self._SingleNodeCall(node, "storage_modify",
+ [su_name, su_args, name, changes])
+ @_RpcTimeout(_TMO_NORMAL)
+ def call_storage_execute(self, node, su_name, su_args, name, op):
+ """Executes an operation on a storage unit.
+
+ This is a single-node call.
+
+ """
+ return self._SingleNodeCall(node, "storage_execute",
+ [su_name, su_args, name, op])
+
+ @_RpcTimeout(_TMO_URGENT)
def call_bridges_exist(self, node, bridges_list):
"""Checks if a node has all the bridges given.
This is a single-node call.
"""
- c = Client("bridges_exist", [bridges_list])
- self._ConnectNode(c, node)
- return c.GetResults().get(node, False)
+ return self._SingleNodeCall(node, "bridges_exist", [bridges_list])
- def call_instance_start(self, node, instance, extra_args):
+ @_RpcTimeout(_TMO_NORMAL)
+ def call_instance_start(self, node, instance, hvp, bep):
"""Starts an instance.
This is a single-node call.
"""
- c = Client("instance_start", [self._InstDict(instance), extra_args])
- self._ConnectNode(c, node)
- return c.GetResults().get(node, False)
+ idict = self._InstDict(instance, hvp=hvp, bep=bep)
+ return self._SingleNodeCall(node, "instance_start", [idict])
- def call_instance_shutdown(self, node, instance):
+ @_RpcTimeout(_TMO_NORMAL)
+ def call_instance_shutdown(self, node, instance, timeout):
"""Stops an instance.
This is a single-node call.
"""
- c = Client("instance_shutdown", [self._InstDict(instance)])
- self._ConnectNode(c, node)
- return c.GetResults().get(node, False)
+ return self._SingleNodeCall(node, "instance_shutdown",
+ [self._InstDict(instance), timeout])
+
+ @_RpcTimeout(_TMO_NORMAL)
+ def call_migration_info(self, node, instance):
+ """Gather the information necessary to prepare an instance migration.
+
+ This is a single-node call.
+
+ @type node: string
+ @param node: the node on which the instance is currently running
+ @type instance: C{objects.Instance}
+ @param instance: the instance definition
+
+ """
+ return self._SingleNodeCall(node, "migration_info",
+ [self._InstDict(instance)])
+ @_RpcTimeout(_TMO_NORMAL)
+ def call_accept_instance(self, node, instance, info, target):
+ """Prepare a node to accept an instance.
+
+ This is a single-node call.
+
+ @type node: string
+ @param node: the target node for the migration
+ @type instance: C{objects.Instance}
+ @param instance: the instance definition
+ @type info: opaque/hypervisor specific (string/data)
+ @param info: result for the call_migration_info call
+ @type target: string
+ @param target: target hostname (usually ip address) (on the node itself)
+
+ """
+ return self._SingleNodeCall(node, "accept_instance",
+ [self._InstDict(instance), info, target])
+
+ @_RpcTimeout(_TMO_NORMAL)
+ def call_finalize_migration(self, node, instance, info, success):
+ """Finalize any target-node migration specific operation.
+
+ This is called both in case of a successful migration and in case of error
+ (in which case it should abort the migration).
+
+ This is a single-node call.
+
+ @type node: string
+ @param node: the target node for the migration
+ @type instance: C{objects.Instance}
+ @param instance: the instance definition
+ @type info: opaque/hypervisor specific (string/data)
+ @param info: result for the call_migration_info call
+ @type success: boolean
+ @param success: whether the migration was a success or a failure
+
+ """
+ return self._SingleNodeCall(node, "finalize_migration",
+ [self._InstDict(instance), info, success])
+
+ @_RpcTimeout(_TMO_SLOW)
def call_instance_migrate(self, node, instance, target, live):
"""Migrate an instance.
interpretation of this parameter is left to the hypervisor)
"""
- c = Client("instance_migrate", [self._InstDict(instance), target, live])
- self._ConnectNode(c, node)
- return c.GetResults().get(node, False)
+ return self._SingleNodeCall(node, "instance_migrate",
+ [self._InstDict(instance), target, live])
- def call_instance_reboot(self, node, instance, reboot_type, extra_args):
+ @_RpcTimeout(_TMO_NORMAL)
+ def call_instance_reboot(self, node, inst, reboot_type, shutdown_timeout):
"""Reboots an instance.
This is a single-node call.
"""
- c = Client("instance_reboot", [self._InstDict(instance),
- reboot_type, extra_args])
- self._ConnectNode(c, node)
- return c.GetResults().get(node, False)
+ return self._SingleNodeCall(node, "instance_reboot",
+ [self._InstDict(inst), reboot_type,
+ shutdown_timeout])
- def call_instance_os_add(self, node, inst):
+ @_RpcTimeout(_TMO_1DAY)
+ def call_instance_os_add(self, node, inst, reinstall, debug):
"""Installs an OS on the given instance.
This is a single-node call.
"""
- params = [self._InstDict(inst)]
- c = Client("instance_os_add", params)
- self._ConnectNode(c, node)
- return c.GetResults().get(node, False)
+ return self._SingleNodeCall(node, "instance_os_add",
+ [self._InstDict(inst), reinstall, debug])
- def call_instance_run_rename(self, node, inst, old_name):
+ @_RpcTimeout(_TMO_SLOW)
+ def call_instance_run_rename(self, node, inst, old_name, debug):
"""Run the OS rename script for an instance.
This is a single-node call.
"""
- params = [self._InstDict(inst), old_name]
- c = Client("instance_run_rename", params)
- self._ConnectNode(c, node)
- return c.GetResults().get(node, False)
+ return self._SingleNodeCall(node, "instance_run_rename",
+ [self._InstDict(inst), old_name, debug])
+ @_RpcTimeout(_TMO_URGENT)
def call_instance_info(self, node, instance, hname):
"""Returns information about a single instance.
This is a single-node call.
- @type node_list: list
- @param node_list: the list of nodes to query
+ @type node: list
+ @param node: the list of nodes to query
@type instance: string
@param instance: the instance name
@type hname: string
@param hname: the hypervisor type of the instance
"""
- c = Client("instance_info", [instance, hname])
- self._ConnectNode(c, node)
- return c.GetResults().get(node, False)
+ return self._SingleNodeCall(node, "instance_info", [instance, hname])
+ @_RpcTimeout(_TMO_NORMAL)
+ def call_instance_migratable(self, node, instance):
+ """Checks whether the given instance can be migrated.
+
+ This is a single-node call.
+
+ @param node: the node to query
+ @type instance: L{objects.Instance}
+ @param instance: the instance to check
+
+
+ """
+ return self._SingleNodeCall(node, "instance_migratable",
+ [self._InstDict(instance)])
+
+ @_RpcTimeout(_TMO_URGENT)
def call_all_instances_info(self, node_list, hypervisor_list):
"""Returns information about all instances on the given nodes.
@param hypervisor_list: the hypervisors to query for instances
"""
- c = Client("all_instances_info", [hypervisor_list])
- self._ConnectList(c, node_list)
- return c.GetResults()
+ return self._MultiNodeCall(node_list, "all_instances_info",
+ [hypervisor_list])
+ @_RpcTimeout(_TMO_URGENT)
def call_instance_list(self, node_list, hypervisor_list):
"""Returns the list of running instances on a given node.
@param hypervisor_list: the hypervisors to query for instances
"""
- c = Client("instance_list", [hypervisor_list])
- self._ConnectList(c, node_list)
- return c.GetResults()
+ return self._MultiNodeCall(node_list, "instance_list", [hypervisor_list])
+ @_RpcTimeout(_TMO_FAST)
def call_node_tcp_ping(self, node, source, target, port, timeout,
live_port_needed):
"""Do a TcpPing on the remote node
This is a single-node call.
"""
- c = Client("node_tcp_ping", [source, target, port, timeout,
+ return self._SingleNodeCall(node, "node_tcp_ping",
+ [source, target, port, timeout,
live_port_needed])
- self._ConnectNode(c, node)
- return c.GetResults().get(node, False)
+ @_RpcTimeout(_TMO_FAST)
def call_node_has_ip_address(self, node, address):
"""Checks if a node has the given IP address.
This is a single-node call.
"""
- c = Client("node_has_ip_address", [address])
- self._ConnectNode(c, node)
- return c.GetResults().get(node, False)
+ return self._SingleNodeCall(node, "node_has_ip_address", [address])
+ @_RpcTimeout(_TMO_URGENT)
def call_node_info(self, node_list, vg_name, hypervisor_type):
"""Return node information.
@type node_list: list
@param node_list: the list of nodes to query
- @type vgname: C{string}
- @param vgname: the name of the volume group to ask for disk space
+ @type vg_name: C{string}
+ @param vg_name: the name of the volume group to ask for disk space
information
@type hypervisor_type: C{str}
@param hypervisor_type: the name of the hypervisor to ask for
memory information
"""
- c = Client("node_info", [vg_name, hypervisor_type])
- self._ConnectList(c, node_list)
- retux = c.GetResults()
-
- for node_name in retux:
- ret = retux.get(node_name, False)
- if type(ret) != dict:
- logging.error("could not connect to node %s", node_name)
- ret = {}
-
- utils.CheckDict(ret,
- { 'memory_total' : '-',
- 'memory_dom0' : '-',
- 'memory_free' : '-',
- 'vg_size' : 'node_unreachable',
- 'vg_free' : '-' },
- "call_node_info",
- )
- return retux
+ return self._MultiNodeCall(node_list, "node_info",
+ [vg_name, hypervisor_type])
+ @_RpcTimeout(_TMO_NORMAL)
def call_node_add(self, node, dsa, dsapub, rsa, rsapub, ssh, sshpub):
"""Add a node to the cluster.
This is a single-node call.
"""
- params = [dsa, dsapub, rsa, rsapub, ssh, sshpub]
- c = Client("node_add", params)
- self._ConnectNode(c, node)
- return c.GetResults().get(node, False)
+ return self._SingleNodeCall(node, "node_add",
+ [dsa, dsapub, rsa, rsapub, ssh, sshpub])
+ @_RpcTimeout(_TMO_NORMAL)
def call_node_verify(self, node_list, checkdict, cluster_name):
"""Request verification of given parameters.
This is a multi-node call.
"""
- c = Client("node_verify", [checkdict, cluster_name])
- self._ConnectList(c, node_list)
- return c.GetResults()
+ return self._MultiNodeCall(node_list, "node_verify",
+ [checkdict, cluster_name])
- @staticmethod
- def call_node_start_master(node, start_daemons):
+ @classmethod
+ @_RpcTimeout(_TMO_FAST)
+ def call_node_start_master(cls, node, start_daemons, no_voting):
"""Tells a node to activate itself as a master.
This is a single-node call.
"""
- c = Client("node_start_master", [start_daemons])
- c.ConnectNode(node)
- return c.GetResults().get(node, False)
+ return cls._StaticSingleNodeCall(node, "node_start_master",
+ [start_daemons, no_voting])
- @staticmethod
- def call_node_stop_master(node, stop_daemons):
+ @classmethod
+ @_RpcTimeout(_TMO_FAST)
+ def call_node_stop_master(cls, node, stop_daemons):
"""Tells a node to demote itself from master status.
This is a single-node call.
"""
- c = Client("node_stop_master", [stop_daemons])
- c.ConnectNode(node)
- return c.GetResults().get(node, False)
+ return cls._StaticSingleNodeCall(node, "node_stop_master", [stop_daemons])
- @staticmethod
- def call_master_info(node_list):
+ @classmethod
+ @_RpcTimeout(_TMO_URGENT)
+ def call_master_info(cls, node_list):
"""Query master info.
This is a multi-node call.
"""
# TODO: should this method query down nodes?
- c = Client("master_info", [])
- c.ConnectList(node_list)
- return c.GetResults()
+ return cls._StaticMultiNodeCall(node_list, "master_info", [])
- def call_version(self, node_list):
+ @classmethod
+ @_RpcTimeout(_TMO_URGENT)
+ def call_version(cls, node_list):
"""Query node version.
This is a multi-node call.
"""
- c = Client("version", [])
- self._ConnectList(c, node_list)
- return c.GetResults()
+ return cls._StaticMultiNodeCall(node_list, "version", [])
+ @_RpcTimeout(_TMO_NORMAL)
def call_blockdev_create(self, node, bdev, size, owner, on_primary, info):
"""Request creation of a given block device.
This is a single-node call.
"""
- params = [bdev.ToDict(), size, owner, on_primary, info]
- c = Client("blockdev_create", params)
- self._ConnectNode(c, node)
- return c.GetResults().get(node, False)
+ return self._SingleNodeCall(node, "blockdev_create",
+ [bdev.ToDict(), size, owner, on_primary, info])
+ @_RpcTimeout(_TMO_NORMAL)
def call_blockdev_remove(self, node, bdev):
"""Request removal of a given block device.
This is a single-node call.
"""
- c = Client("blockdev_remove", [bdev.ToDict()])
- self._ConnectNode(c, node)
- return c.GetResults().get(node, False)
+ return self._SingleNodeCall(node, "blockdev_remove", [bdev.ToDict()])
+ @_RpcTimeout(_TMO_NORMAL)
def call_blockdev_rename(self, node, devlist):
"""Request rename of the given block devices.
This is a single-node call.
"""
- params = [(d.ToDict(), uid) for d, uid in devlist]
- c = Client("blockdev_rename", params)
- self._ConnectNode(c, node)
- return c.GetResults().get(node, False)
+ return self._SingleNodeCall(node, "blockdev_rename",
+ [(d.ToDict(), uid) for d, uid in devlist])
+ @_RpcTimeout(_TMO_NORMAL)
def call_blockdev_assemble(self, node, disk, owner, on_primary):
"""Request assembling of a given block device.
This is a single-node call.
"""
- params = [disk.ToDict(), owner, on_primary]
- c = Client("blockdev_assemble", params)
- self._ConnectNode(c, node)
- return c.GetResults().get(node, False)
+ return self._SingleNodeCall(node, "blockdev_assemble",
+ [disk.ToDict(), owner, on_primary])
+ @_RpcTimeout(_TMO_NORMAL)
def call_blockdev_shutdown(self, node, disk):
"""Request shutdown of a given block device.
This is a single-node call.
"""
- c = Client("blockdev_shutdown", [disk.ToDict()])
- self._ConnectNode(c, node)
- return c.GetResults().get(node, False)
+ return self._SingleNodeCall(node, "blockdev_shutdown", [disk.ToDict()])
+ @_RpcTimeout(_TMO_NORMAL)
def call_blockdev_addchildren(self, node, bdev, ndevs):
"""Request adding a list of children to a (mirroring) device.
This is a single-node call.
"""
- params = [bdev.ToDict(), [disk.ToDict() for disk in ndevs]]
- c = Client("blockdev_addchildren", params)
- self._ConnectNode(c, node)
- return c.GetResults().get(node, False)
+ return self._SingleNodeCall(node, "blockdev_addchildren",
+ [bdev.ToDict(),
+ [disk.ToDict() for disk in ndevs]])
+ @_RpcTimeout(_TMO_NORMAL)
def call_blockdev_removechildren(self, node, bdev, ndevs):
"""Request removing a list of children from a (mirroring) device.
This is a single-node call.
"""
- params = [bdev.ToDict(), [disk.ToDict() for disk in ndevs]]
- c = Client("blockdev_removechildren", params)
- self._ConnectNode(c, node)
- return c.GetResults().get(node, False)
+ return self._SingleNodeCall(node, "blockdev_removechildren",
+ [bdev.ToDict(),
+ [disk.ToDict() for disk in ndevs]])
+ @_RpcTimeout(_TMO_NORMAL)
def call_blockdev_getmirrorstatus(self, node, disks):
"""Request status of a (mirroring) device.
This is a single-node call.
"""
- params = [dsk.ToDict() for dsk in disks]
- c = Client("blockdev_getmirrorstatus", params)
- self._ConnectNode(c, node)
- return c.GetResults().get(node, False)
+ result = self._SingleNodeCall(node, "blockdev_getmirrorstatus",
+ [dsk.ToDict() for dsk in disks])
+ if not result.fail_msg:
+ result.payload = [objects.BlockDevStatus.FromDict(i)
+ for i in result.payload]
+ return result
+ @_RpcTimeout(_TMO_NORMAL)
def call_blockdev_find(self, node, disk):
"""Request identification of a given block device.
This is a single-node call.
"""
- c = Client("blockdev_find", [disk.ToDict()])
- self._ConnectNode(c, node)
- return c.GetResults().get(node, False)
+ result = self._SingleNodeCall(node, "blockdev_find", [disk.ToDict()])
+ if not result.fail_msg and result.payload is not None:
+ result.payload = objects.BlockDevStatus.FromDict(result.payload)
+ return result
- def call_blockdev_close(self, node, disks):
+ @_RpcTimeout(_TMO_NORMAL)
+ def call_blockdev_close(self, node, instance_name, disks):
"""Closes the given block devices.
This is a single-node call.
"""
- params = [cf.ToDict() for cf in disks]
- c = Client("blockdev_close", params)
- self._ConnectNode(c, node)
- return c.GetResults().get(node, False)
+ params = [instance_name, [cf.ToDict() for cf in disks]]
+ return self._SingleNodeCall(node, "blockdev_close", params)
- @staticmethod
- def call_upload_file(node_list, file_name, address_list=None):
+ @_RpcTimeout(_TMO_NORMAL)
+ def call_blockdev_getsizes(self, node, disks):
+ """Returns the size of the given disks.
+
+ This is a single-node call.
+
+ """
+ params = [[cf.ToDict() for cf in disks]]
+ return self._SingleNodeCall(node, "blockdev_getsize", params)
+
+ @_RpcTimeout(_TMO_NORMAL)
+ def call_drbd_disconnect_net(self, node_list, nodes_ip, disks):
+ """Disconnects the network of the given drbd devices.
+
+ This is a multi-node call.
+
+ """
+ return self._MultiNodeCall(node_list, "drbd_disconnect_net",
+ [nodes_ip, [cf.ToDict() for cf in disks]])
+
+ @_RpcTimeout(_TMO_NORMAL)
+ def call_drbd_attach_net(self, node_list, nodes_ip,
+ disks, instance_name, multimaster):
+ """Disconnects the given drbd devices.
+
+ This is a multi-node call.
+
+ """
+ return self._MultiNodeCall(node_list, "drbd_attach_net",
+ [nodes_ip, [cf.ToDict() for cf in disks],
+ instance_name, multimaster])
+
+ @_RpcTimeout(_TMO_SLOW)
+ def call_drbd_wait_sync(self, node_list, nodes_ip, disks):
+ """Waits for the synchronization of drbd devices is complete.
+
+ This is a multi-node call.
+
+ """
+ return self._MultiNodeCall(node_list, "drbd_wait_sync",
+ [nodes_ip, [cf.ToDict() for cf in disks]])
+
+ @_RpcTimeout(_TMO_URGENT)
+ def call_drbd_helper(self, node_list):
+ """Gets drbd helper.
+
+ This is a multi-node call.
+
+ """
+ return self._MultiNodeCall(node_list, "drbd_helper", [])
+
+ @classmethod
+ @_RpcTimeout(_TMO_NORMAL)
+ def call_upload_file(cls, node_list, file_name, address_list=None):
"""Upload a file.
The node will refuse the operation in case the file is not on the
to optimize the RPC speed
"""
- fh = file(file_name)
- try:
- data = fh.read()
- finally:
- fh.close()
+ file_contents = utils.ReadFile(file_name)
+ data = cls._Compress(file_contents)
st = os.stat(file_name)
params = [file_name, data, st.st_mode, st.st_uid, st.st_gid,
st.st_atime, st.st_mtime]
- c = Client("upload_file", params)
- c.ConnectList(node_list, address_list=address_list)
- return c.GetResults()
+ return cls._StaticMultiNodeCall(node_list, "upload_file", params,
+ address_list=address_list)
+
+ @classmethod
+ @_RpcTimeout(_TMO_NORMAL)
+ def call_write_ssconf_files(cls, node_list, values):
+ """Write ssconf files.
+
+ This is a multi-node call.
+ """
+ return cls._StaticMultiNodeCall(node_list, "write_ssconf_files", [values])
+
+ @_RpcTimeout(_TMO_FAST)
def call_os_diagnose(self, node_list):
"""Request a diagnose of OS definitions.
This is a multi-node call.
"""
- c = Client("os_diagnose", [])
- self._ConnectList(c, node_list)
- result = c.GetResults()
- new_result = {}
- for node_name in result:
- if result[node_name]:
- nr = [objects.OS.FromDict(oss) for oss in result[node_name]]
- else:
- nr = []
- new_result[node_name] = nr
- return new_result
+ return self._MultiNodeCall(node_list, "os_diagnose", [])
+ @_RpcTimeout(_TMO_FAST)
def call_os_get(self, node, name):
"""Returns an OS definition.
This is a single-node call.
"""
- c = Client("os_get", [name])
- self._ConnectNode(c, node)
- result = c.GetResults().get(node, False)
- if isinstance(result, dict):
- return objects.OS.FromDict(result)
- else:
- return result
+ result = self._SingleNodeCall(node, "os_get", [name])
+ if not result.fail_msg and isinstance(result.payload, dict):
+ result.payload = objects.OS.FromDict(result.payload)
+ return result
+
+ @_RpcTimeout(_TMO_FAST)
+ def call_os_validate(self, required, nodes, name, checks, params):
+ """Run a validation routine for a given OS.
+
+ This is a multi-node call.
+
+ """
+ return self._MultiNodeCall(nodes, "os_validate",
+ [required, name, checks, params])
+ @_RpcTimeout(_TMO_NORMAL)
def call_hooks_runner(self, node_list, hpath, phase, env):
"""Call the hooks runner.
"""
params = [hpath, phase, env]
- c = Client("hooks_runner", params)
- self._ConnectList(c, node_list)
- result = c.GetResults()
- return result
+ return self._MultiNodeCall(node_list, "hooks_runner", params)
+ @_RpcTimeout(_TMO_NORMAL)
def call_iallocator_runner(self, node, name, idata):
"""Call an iallocator on a remote node
This is a single-node call.
"""
- params = [name, idata]
- c = Client("iallocator_runner", params)
- self._ConnectNode(c, node)
- result = c.GetResults().get(node, False)
- return result
+ return self._SingleNodeCall(node, "iallocator_runner", [name, idata])
+ @_RpcTimeout(_TMO_NORMAL)
def call_blockdev_grow(self, node, cf_bdev, amount):
"""Request a snapshot of the given block device.
This is a single-node call.
"""
- c = Client("blockdev_grow", [cf_bdev.ToDict(), amount])
- self._ConnectNode(c, node)
- return c.GetResults().get(node, False)
+ return self._SingleNodeCall(node, "blockdev_grow",
+ [cf_bdev.ToDict(), amount])
- def call_blockdev_snapshot(self, node, cf_bdev):
- """Request a snapshot of the given block device.
+ @_RpcTimeout(_TMO_1DAY)
+ def call_blockdev_export(self, node, cf_bdev,
+ dest_node, dest_path, cluster_name):
+ """Export a given disk to another node.
This is a single-node call.
"""
- c = Client("blockdev_snapshot", [cf_bdev.ToDict()])
- self._ConnectNode(c, node)
- return c.GetResults().get(node, False)
+ return self._SingleNodeCall(node, "blockdev_export",
+ [cf_bdev.ToDict(), dest_node, dest_path,
+ cluster_name])
- def call_snapshot_export(self, node, snap_bdev, dest_node, instance,
- cluster_name, idx):
- """Request the export of a given snapshot.
+ @_RpcTimeout(_TMO_NORMAL)
+ def call_blockdev_snapshot(self, node, cf_bdev):
+ """Request a snapshot of the given block device.
This is a single-node call.
"""
- params = [snap_bdev.ToDict(), dest_node,
- self._InstDict(instance), cluster_name, idx]
- c = Client("snapshot_export", params)
- self._ConnectNode(c, node)
- return c.GetResults().get(node, False)
+ return self._SingleNodeCall(node, "blockdev_snapshot", [cf_bdev.ToDict()])
+ @_RpcTimeout(_TMO_NORMAL)
def call_finalize_export(self, node, instance, snap_disks):
"""Request the completion of an export operation.
"""
flat_disks = []
for disk in snap_disks:
- flat_disks.append(disk.ToDict())
- params = [self._InstDict(instance), flat_disks]
- c = Client("finalize_export", params)
- self._ConnectNode(c, node)
- return c.GetResults().get(node, False)
+ if isinstance(disk, bool):
+ flat_disks.append(disk)
+ else:
+ flat_disks.append(disk.ToDict())
+
+ return self._SingleNodeCall(node, "finalize_export",
+ [self._InstDict(instance), flat_disks])
+ @_RpcTimeout(_TMO_FAST)
def call_export_info(self, node, path):
"""Queries the export information in a given path.
This is a single-node call.
"""
- c = Client("export_info", [path])
- self._ConnectNode(c, node)
- result = c.GetResults().get(node, False)
- if not result:
- return result
- return objects.SerializableConfigParser.Loads(str(result))
-
- def call_instance_os_import(self, node, inst, src_node, src_images,
- cluster_name):
- """Request the import of a backup into an instance.
-
- This is a single-node call.
-
- """
- params = [self._InstDict(inst), src_node, src_images, cluster_name]
- c = Client("instance_os_import", params)
- self._ConnectNode(c, node)
- return c.GetResults().get(node, False)
+ return self._SingleNodeCall(node, "export_info", [path])
+ @_RpcTimeout(_TMO_FAST)
def call_export_list(self, node_list):
"""Gets the stored exports list.
This is a multi-node call.
"""
- c = Client("export_list", [])
- self._ConnectList(c, node_list)
- result = c.GetResults()
- return result
+ return self._MultiNodeCall(node_list, "export_list", [])
+ @_RpcTimeout(_TMO_FAST)
def call_export_remove(self, node, export):
"""Requests removal of a given export.
This is a single-node call.
"""
- c = Client("export_remove", [export])
- self._ConnectNode(c, node)
- return c.GetResults().get(node, False)
+ return self._SingleNodeCall(node, "export_remove", [export])
- @staticmethod
- def call_node_leave_cluster(node):
+ @classmethod
+ @_RpcTimeout(_TMO_NORMAL)
+ def call_node_leave_cluster(cls, node, modify_ssh_setup):
"""Requests a node to clean the cluster information it has.
This will remove the configuration information from the ganeti data
This is a single-node call.
"""
- c = Client("node_leave_cluster", [])
- c.ConnectNode(node)
- return c.GetResults().get(node, False)
+ return cls._StaticSingleNodeCall(node, "node_leave_cluster",
+ [modify_ssh_setup])
+ @_RpcTimeout(_TMO_FAST)
def call_node_volumes(self, node_list):
"""Gets all volumes on node(s).
This is a multi-node call.
"""
- c = Client("node_volumes", [])
- self._ConnectList(c, node_list)
- return c.GetResults()
+ return self._MultiNodeCall(node_list, "node_volumes", [])
+
+ @_RpcTimeout(_TMO_FAST)
+ def call_node_demote_from_mc(self, node):
+ """Demote a node from the master candidate role.
+
+ This is a single-node call.
+ """
+ return self._SingleNodeCall(node, "node_demote_from_mc", [])
+
+ @_RpcTimeout(_TMO_NORMAL)
+ def call_node_powercycle(self, node, hypervisor):
+ """Tries to powercycle a node.
+
+ This is a single-node call.
+
+ """
+ return self._SingleNodeCall(node, "node_powercycle", [hypervisor])
+
+ @_RpcTimeout(None)
def call_test_delay(self, node_list, duration):
"""Sleep for a fixed time on given node(s).
This is a multi-node call.
"""
- c = Client("test_delay", [duration])
- self._ConnectList(c, node_list)
- return c.GetResults()
+ return self._MultiNodeCall(node_list, "test_delay", [duration],
+ read_timeout=int(duration + 5))
+ @_RpcTimeout(_TMO_FAST)
def call_file_storage_dir_create(self, node, file_storage_dir):
"""Create the given file storage directory.
This is a single-node call.
"""
- c = Client("file_storage_dir_create", [file_storage_dir])
- self._ConnectNode(c, node)
- return c.GetResults().get(node, False)
+ return self._SingleNodeCall(node, "file_storage_dir_create",
+ [file_storage_dir])
+ @_RpcTimeout(_TMO_FAST)
def call_file_storage_dir_remove(self, node, file_storage_dir):
"""Remove the given file storage directory.
This is a single-node call.
"""
- c = Client("file_storage_dir_remove", [file_storage_dir])
- self._ConnectNode(c, node)
- return c.GetResults().get(node, False)
+ return self._SingleNodeCall(node, "file_storage_dir_remove",
+ [file_storage_dir])
+ @_RpcTimeout(_TMO_FAST)
def call_file_storage_dir_rename(self, node, old_file_storage_dir,
new_file_storage_dir):
"""Rename file storage directory.
This is a single-node call.
"""
- c = Client("file_storage_dir_rename",
- [old_file_storage_dir, new_file_storage_dir])
- self._ConnectNode(c, node)
- return c.GetResults().get(node, False)
+ return self._SingleNodeCall(node, "file_storage_dir_rename",
+ [old_file_storage_dir, new_file_storage_dir])
- @staticmethod
- def call_jobqueue_update(node_list, address_list, file_name, content):
+ @classmethod
+ @_RpcTimeout(_TMO_FAST)
+ def call_jobqueue_update(cls, node_list, address_list, file_name, content):
"""Update job queue.
This is a multi-node call.
"""
- c = Client("jobqueue_update", [file_name, content])
- c.ConnectList(node_list, address_list=address_list)
- result = c.GetResults()
- return result
+ return cls._StaticMultiNodeCall(node_list, "jobqueue_update",
+ [file_name, cls._Compress(content)],
+ address_list=address_list)
- @staticmethod
- def call_jobqueue_purge(node):
+ @classmethod
+ @_RpcTimeout(_TMO_NORMAL)
+ def call_jobqueue_purge(cls, node):
"""Purge job queue.
This is a single-node call.
"""
- c = Client("jobqueue_purge", [])
- c.ConnectNode(node)
- return c.GetResults().get(node, False)
+ return cls._StaticSingleNodeCall(node, "jobqueue_purge", [])
- @staticmethod
- def call_jobqueue_rename(node_list, address_list, old, new):
+ @classmethod
+ @_RpcTimeout(_TMO_FAST)
+ def call_jobqueue_rename(cls, node_list, address_list, rename):
"""Rename a job queue file.
This is a multi-node call.
"""
- c = Client("jobqueue_rename", [old, new])
- c.ConnectList(node_list, address_list=address_list)
- result = c.GetResults()
- return result
-
-
- @staticmethod
- def call_jobqueue_set_drain(node_list, drain_flag):
- """Set the drain flag on the queue.
-
- This is a multi-node call.
-
- @type node_list: list
- @param node_list: the list of nodes to query
- @type drain_flag: bool
- @param drain_flag: if True, will set the drain flag, otherwise reset it.
-
- """
- c = Client("jobqueue_set_drain", [drain_flag])
- c.ConnectList(node_list)
- result = c.GetResults()
- return result
-
+ return cls._StaticMultiNodeCall(node_list, "jobqueue_rename", rename,
+ address_list=address_list)
+ @_RpcTimeout(_TMO_NORMAL)
def call_hypervisor_validate_params(self, node_list, hvname, hvparams):
"""Validate the hypervisor params.
"""
cluster = self._cfg.GetClusterInfo()
- hv_full = cluster.FillDict(cluster.hvparams.get(hvname, {}), hvparams)
- c = Client("hypervisor_validate_params", [hvname, hv_full])
- self._ConnectList(c, node_list)
- result = c.GetResults()
+ hv_full = objects.FillDict(cluster.hvparams.get(hvname, {}), hvparams)
+ return self._MultiNodeCall(node_list, "hypervisor_validate_params",
+ [hvname, hv_full])
+
+ @_RpcTimeout(_TMO_NORMAL)
+ def call_x509_cert_create(self, node, validity):
+ """Creates a new X509 certificate for SSL/TLS.
+
+ This is a single-node call.
+
+ @type validity: int
+ @param validity: Validity in seconds
+
+ """
+ return self._SingleNodeCall(node, "x509_cert_create", [validity])
+
+ @_RpcTimeout(_TMO_NORMAL)
+ def call_x509_cert_remove(self, node, name):
+ """Removes a X509 certificate.
+
+ This is a single-node call.
+
+ @type name: string
+ @param name: Certificate name
+
+ """
+ return self._SingleNodeCall(node, "x509_cert_remove", [name])
+
+ @_RpcTimeout(_TMO_NORMAL)
+ def call_import_start(self, node, opts, instance, dest, dest_args):
+ """Starts a listener for an import.
+
+ This is a single-node call.
+
+ @type node: string
+ @param node: Node name
+ @type instance: C{objects.Instance}
+ @param instance: Instance object
+
+ """
+ return self._SingleNodeCall(node, "import_start",
+ [opts.ToDict(),
+ self._InstDict(instance), dest,
+ _EncodeImportExportIO(dest, dest_args)])
+
+ @_RpcTimeout(_TMO_NORMAL)
+ def call_export_start(self, node, opts, host, port,
+ instance, source, source_args):
+ """Starts an export daemon.
+
+ This is a single-node call.
+
+ @type node: string
+ @param node: Node name
+ @type instance: C{objects.Instance}
+ @param instance: Instance object
+
+ """
+ return self._SingleNodeCall(node, "export_start",
+ [opts.ToDict(), host, port,
+ self._InstDict(instance), source,
+ _EncodeImportExportIO(source, source_args)])
+
+ @_RpcTimeout(_TMO_FAST)
+ def call_impexp_status(self, node, names):
+ """Gets the status of an import or export.
+
+ This is a single-node call.
+
+ @type node: string
+ @param node: Node name
+ @type names: List of strings
+ @param names: Import/export names
+ @rtype: List of L{objects.ImportExportStatus} instances
+ @return: Returns a list of the state of each named import/export or None if
+ a status couldn't be retrieved
+
+ """
+ result = self._SingleNodeCall(node, "impexp_status", [names])
+
+ if not result.fail_msg:
+ decoded = []
+
+ for i in result.payload:
+ if i is None:
+ decoded.append(None)
+ continue
+ decoded.append(objects.ImportExportStatus.FromDict(i))
+
+ result.payload = decoded
+
return result
+
+ @_RpcTimeout(_TMO_NORMAL)
+ def call_impexp_abort(self, node, name):
+ """Aborts an import or export.
+
+ This is a single-node call.
+
+ @type node: string
+ @param node: Node name
+ @type name: string
+ @param name: Import/export name
+
+ """
+ return self._SingleNodeCall(node, "impexp_abort", [name])
+
+ @_RpcTimeout(_TMO_NORMAL)
+ def call_impexp_cleanup(self, node, name):
+ """Cleans up after an import or export.
+
+ This is a single-node call.
+
+ @type node: string
+ @param node: Node name
+ @type name: string
+ @param name: Import/export name
+
+ """
+ return self._SingleNodeCall(node, "impexp_cleanup", [name])