import errno
import re
import subprocess
+import random
from ganeti import logger
from ganeti import errors
def StopMaster():
"""Deactivate this node as master.
- This does two things:
- - run the master stop script
- - remove link to master cron script.
+ This runs the master stop script.
"""
result = utils.RunCmd([constants.MASTER_SCRIPT, "-d", "stop"])
if os.path.isfile(full_name) and not os.path.islink(full_name):
utils.RemoveFile(full_name)
-
try:
priv_key, pub_key, auth_keys = ssh.GetUserFiles(constants.GANETI_RUNAS)
except errors.OpExecError, err:
if 'nodelist' in what:
result['nodelist'] = {}
+ random.shuffle(what['nodelist'])
for node in what['nodelist']:
success, message = ssh.VerifyNodeHostname(node)
if not success:
result['nodelist'][node] = message
+ if 'node-net-test' in what:
+ result['node-net-test'] = {}
+ my_name = utils.HostInfo().name
+ my_pip = my_sip = None
+ for name, pip, sip in what['node-net-test']:
+ if name == my_name:
+ my_pip = pip
+ my_sip = sip
+ break
+ if not my_pip:
+ result['node-net-test'][my_name] = ("Can't find my own"
+ " primary/secondary IP"
+ " in the node list")
+ else:
+ port = ssconf.SimpleStore().GetNodeDaemonPort()
+ for name, pip, sip in what['node-net-test']:
+ fail = []
+ if not utils.TcpPing(pip, port, source=my_pip):
+ fail.append("primary")
+ if sip != pip:
+ if not utils.TcpPing(sip, port, source=my_sip):
+ fail.append("secondary")
+ if fail:
+ result['node-net-test'][name] = ("failure using the %s"
+ " interface(s)" %
+ " and ".join(fail))
+
return result
"""Compute list of logical volumes and their size.
Returns:
- dictionary of all partions (key) with their size:
- test1: 20.06MiB
+ dictionary of all partions (key) with their size (in MiB), inactive
+ and online status:
+ {'test1': ('20.06', True, True)}
"""
- result = utils.RunCmd(["lvs", "--noheadings", "--units=m",
- "-oname,size", vg_name])
+ lvs = {}
+ sep = '|'
+ result = utils.RunCmd(["lvs", "--noheadings", "--units=m", "--nosuffix",
+ "--separator=%s" % sep,
+ "-olv_name,lv_size,lv_attr", vg_name])
if result.failed:
logger.Error("Failed to list logical volumes, lvs output: %s" %
result.output)
- return {}
+ return result.output
- lvlist = [line.split() for line in result.output.splitlines()]
- return dict(lvlist)
+ for line in result.stdout.splitlines():
+ line = line.strip().rstrip(sep)
+ name, size, attr = line.split(sep)
+ if len(attr) != 6:
+ attr = '------'
+ inactive = attr[4] == '-'
+ online = attr[5] == 'o'
+ lvs[name] = (size, inactive, online)
+
+ return lvs
def ListVolumeGroups():
'vg': line[3].strip(),
}
- return [map_line(line.split('|')) for line in result.output.splitlines()]
+ return [map_line(line.split('|')) for line in result.stdout.splitlines()]
def BridgesExist(bridges_list):
logfile)
result = utils.RunCmd(command)
-
if result.failed:
- logger.Error("os create command '%s' returned error: %s"
+ logger.Error("os create command '%s' returned error: %s, logfile: %s,"
" output: %s" %
- (command, result.fail_reason, result.output))
+ (command, result.fail_reason, logfile, result.output))
return False
return True
vg_free is the free size of the volume group in MiB
pv_count are the number of physical disks in that vg
+ If an error occurs during gathering of data, we return the same dict
+ with keys all set to None.
+
"""
+ retdic = dict.fromkeys(["vg_size", "vg_free", "pv_count"])
+
retval = utils.RunCmd(["vgs", "-ovg_size,vg_free,pv_count", "--noheadings",
"--nosuffix", "--units=m", "--separator=:", vg_name])
if retval.failed:
errmsg = "volume group %s not present" % vg_name
logger.Error(errmsg)
- raise errors.LVMError(errmsg)
- valarr = retval.stdout.strip().split(':')
- retdic = {
- "vg_size": int(round(float(valarr[0]), 0)),
- "vg_free": int(round(float(valarr[1]), 0)),
- "pv_count": int(valarr[2]),
- }
+ return retdic
+ valarr = retval.stdout.strip().rstrip(':').split(':')
+ if len(valarr) == 3:
+ try:
+ retdic = {
+ "vg_size": int(round(float(valarr[0]), 0)),
+ "vg_free": int(round(float(valarr[1]), 0)),
+ "pv_count": int(valarr[2]),
+ }
+ except ValueError, err:
+ logger.Error("Fail to parse vgs output: %s" % str(err))
+ else:
+ logger.Error("vgs output has the wrong number of fields (expected"
+ " three): %s" % str(valarr))
return retdic
"""Creates a block device for an instance.
Args:
- bdev: a ganeti.objects.Disk object
- size: the size of the physical underlying devices
- do_open: if the device should be `Assemble()`-d and
- `Open()`-ed after creation
+ disk: a ganeti.objects.Disk object
+ size: the size of the physical underlying device
+ owner: a string with the name of the instance
+ on_primary: a boolean indicating if it is the primary node or not
+ info: string that will be sent to the physical device creation
Returns:
the new unique_id of the device (this can sometime be
# we need the children open in case the device itself has to
# be assembled
crdev.Open()
- else:
- crdev.Close()
clist.append(crdev)
try:
device = bdev.FindDevice(disk.dev_type, disk.physical_id, clist)
(disk, size))
if on_primary or disk.AssembleOnSecondary():
if not device.Assemble():
- raise errors.BlockDeviceError("Can't assemble device after creation,"
- " very unusual event - check the node"
- " daemon logs")
+ errorstring = "Can't assemble device after creation"
+ logger.Error(errorstring)
+ raise errors.BlockDeviceError("%s, very unusual event - check the node"
+ " daemon logs" % errorstring)
device.SetSyncSpeed(constants.SYNC_SPEED)
if on_primary or disk.OpenOnSecondary():
device.Open(force=True)
try:
cdev = _RecursiveAssembleBD(chld_disk, owner, as_primary)
except errors.BlockDeviceError, err:
- if children.count(None) > mcn:
+ if children.count(None) >= mcn:
raise
cdev = None
logger.Debug("Error in child activation: %s" % str(err))
result = r_dev
if as_primary or disk.OpenOnSecondary():
r_dev.Open()
- else:
- r_dev.Close()
DevCacheManager.UpdateCache(r_dev.dev_path, owner,
as_primary, disk.iv_name)
rbd = _RecursiveFindBD(disk)
if rbd is None:
return rbd
- sync_p, est_t, is_degr = rbd.GetSyncStatus()
- return rbd.dev_path, rbd.major, rbd.minor, sync_p, est_t, is_degr
+ return (rbd.dev_path, rbd.major, rbd.minor) + rbd.GetSyncStatus()
def UploadFile(file_name, data, mode, uid, gid, atime, mtime):
file_name)
return False
- allowed_files = [constants.CLUSTER_CONF_FILE, "/etc/hosts",
- constants.SSH_KNOWN_HOSTS_FILE]
+ allowed_files = [
+ constants.CLUSTER_CONF_FILE,
+ constants.ETC_HOSTS,
+ constants.SSH_KNOWN_HOSTS_FILE,
+ ]
allowed_files.extend(ssconf.SimpleStore().GetFileList())
if file_name not in allowed_files:
logger.Error("Filename passed to UploadFile not in allowed"
return detail
-def _OSSearch(name, search_path=None):
- """Search for OSes with the given name in the search_path.
-
- Args:
- name: The name of the OS to look for
- search_path: List of dirs to search (defaults to constants.OS_SEARCH_PATH)
-
- Returns:
- The base_dir the OS resides in
-
- """
- if search_path is None:
- search_path = constants.OS_SEARCH_PATH
-
- for dir_name in search_path:
- t_os_dir = os.path.sep.join([dir_name, name])
- if os.path.isdir(t_os_dir):
- return dir_name
-
- return None
-
-
def _OSOndiskVersion(name, os_dir):
"""Compute and return the API version of a given OS.
"""
if base_dir is None:
- base_dir = _OSSearch(name)
-
- if base_dir is None:
- raise errors.InvalidOS(name, None, "OS dir not found in search path")
+ os_dir = utils.FindFile(name, constants.OS_SEARCH_PATH, os.path.isdir)
+ if os_dir is None:
+ raise errors.InvalidOS(name, None, "OS dir not found in search path")
+ else:
+ os_dir = os.path.sep.join([base_dir, name])
- os_dir = os.path.sep.join([base_dir, name])
api_version = _OSOndiskVersion(name, os_dir)
if api_version != constants.OS_API_VERSION:
return None
else:
raise errors.ProgrammerError("Cannot snapshot non-lvm block device"
- "'%s' of type '%s'" %
+ " '%s' of type '%s'" %
(disk.unique_id, disk.dev_type))
config.set(constants.INISECT_INS, 'nic%d_mac' %
nic_count, '%s' % nic.mac)
config.set(constants.INISECT_INS, 'nic%d_ip' % nic_count, '%s' % nic.ip)
+ config.set(constants.INISECT_INS, 'nic%d_bridge' % nic_count,
+ '%s' % nic.bridge)
# TODO: redundant: on load can read nics until it doesn't exist
config.set(constants.INISECT_INS, 'nic_count' , '%d' % nic_count)
Args:
- hooks_base_dir: if not None, this overrides the
constants.HOOKS_BASE_DIR (useful for unittests)
- - logs_base_dir: if not None, this overrides the
- constants.LOG_HOOKS_DIR (useful for unittests)
- - logging: enable or disable logging of script output
"""
if hooks_base_dir is None:
"""Exec one hook script.
Args:
- - phase: the phase
- script: the full path to the script
- env: the environment with which to exec the script
fdstdin = open("/dev/null", "r")
child = subprocess.Popen([script], stdin=fdstdin, stdout=subprocess.PIPE,
stderr=subprocess.STDOUT, close_fds=True,
- shell=False, cwd="/",env=env)
+ shell=False, cwd="/", env=env)
output = ""
try:
output = child.stdout.read(4096)
return rr
+class IAllocatorRunner(object):
+ """IAllocator runner.
+
+ This class is instantiated on the node side (ganeti-noded) and not on
+ the master side.
+
+ """
+ def Run(self, name, idata):
+ """Run an iallocator script.
+
+ Return value: tuple of:
+ - run status (one of the IARUN_ constants)
+ - stdout
+ - stderr
+ - fail reason (as from utils.RunResult)
+
+ """
+ alloc_script = utils.FindFile(name, constants.IALLOCATOR_SEARCH_PATH,
+ os.path.isfile)
+ if alloc_script is None:
+ return (constants.IARUN_NOTFOUND, None, None, None)
+
+ fd, fin_name = tempfile.mkstemp(prefix="ganeti-iallocator.")
+ try:
+ os.write(fd, idata)
+ os.close(fd)
+ result = utils.RunCmd([alloc_script, fin_name])
+ if result.failed:
+ return (constants.IARUN_FAILURE, result.stdout, result.stderr,
+ result.fail_reason)
+ finally:
+ os.unlink(fin_name)
+
+ return (constants.IARUN_SUCCESS, result.stdout, result.stderr, None)
+
+
class DevCacheManager(object):
- """Simple class for managing a chache of block device information.
+ """Simple class for managing a cache of block device information.
"""
_DEV_PREFIX = "/dev/"