#!/usr/bin/python # # Copyright (C) 2006, 2007 Google Inc. # # This program is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation; either version 2 of the License, or # (at your option) any later version. # # This program is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU # General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program; if not, write to the Free Software # Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA # 02110-1301, USA. """Ganeti node daemon""" # functions in this module need to have a given name structure, so: # pylint: disable-msg=C0103 import os import sys import traceback import SocketServer import errno import logging import signal from optparse import OptionParser from ganeti import backend from ganeti import constants from ganeti import objects from ganeti import errors from ganeti import jstore from ganeti import daemon from ganeti import http from ganeti import utils import ganeti.http.server queue_lock = None def _RequireJobQueueLock(fn): """Decorator for job queue manipulating functions. """ QUEUE_LOCK_TIMEOUT = 10 def wrapper(*args, **kwargs): # Locking in exclusive, blocking mode because there could be several # children running at the same time. Waiting up to 10 seconds. queue_lock.Exclusive(blocking=True, timeout=QUEUE_LOCK_TIMEOUT) try: return fn(*args, **kwargs) finally: queue_lock.Unlock() return wrapper class NodeHttpServer(http.server.HttpServer): """The server implementation. This class holds all methods exposed over the RPC interface. """ def __init__(self, *args, **kwargs): http.server.HttpServer.__init__(self, *args, **kwargs) self.noded_pid = os.getpid() def HandleRequest(self, req): """Handle a request. """ if req.request_method.upper() != http.HTTP_PUT: raise http.HttpBadRequest() path = req.request_path if path.startswith("/"): path = path[1:] method = getattr(self, "perspective_%s" % path, None) if method is None: raise http.HttpNotFound() try: try: return method(req.request_body) except: logging.exception("Error in RPC call") raise except errors.QuitGanetiException, err: # Tell parent to quit os.kill(self.noded_pid, signal.SIGTERM) # the new block devices -------------------------- @staticmethod def perspective_blockdev_create(params): """Create a block device. """ bdev_s, size, owner, on_primary, info = params bdev = objects.Disk.FromDict(bdev_s) if bdev is None: raise ValueError("can't unserialize data!") return backend.BlockdevCreate(bdev, size, owner, on_primary, info) @staticmethod def perspective_blockdev_remove(params): """Remove a block device. """ bdev_s = params[0] bdev = objects.Disk.FromDict(bdev_s) return backend.BlockdevRemove(bdev) @staticmethod def perspective_blockdev_rename(params): """Remove a block device. """ devlist = [(objects.Disk.FromDict(ds), uid) for ds, uid in params] return backend.BlockdevRename(devlist) @staticmethod def perspective_blockdev_assemble(params): """Assemble a block device. """ bdev_s, owner, on_primary = params bdev = objects.Disk.FromDict(bdev_s) if bdev is None: raise ValueError("can't unserialize data!") return backend.BlockdevAssemble(bdev, owner, on_primary) @staticmethod def perspective_blockdev_shutdown(params): """Shutdown a block device. """ bdev_s = params[0] bdev = objects.Disk.FromDict(bdev_s) if bdev is None: raise ValueError("can't unserialize data!") return backend.BlockdevShutdown(bdev) @staticmethod def perspective_blockdev_addchildren(params): """Add a child to a mirror device. Note: this is only valid for mirror devices. It's the caller's duty to send a correct disk, otherwise we raise an error. """ bdev_s, ndev_s = params bdev = objects.Disk.FromDict(bdev_s) ndevs = [objects.Disk.FromDict(disk_s) for disk_s in ndev_s] if bdev is None or ndevs.count(None) > 0: raise ValueError("can't unserialize data!") return backend.BlockdevAddchildren(bdev, ndevs) @staticmethod def perspective_blockdev_removechildren(params): """Remove a child from a mirror device. This is only valid for mirror devices, of course. It's the callers duty to send a correct disk, otherwise we raise an error. """ bdev_s, ndev_s = params bdev = objects.Disk.FromDict(bdev_s) ndevs = [objects.Disk.FromDict(disk_s) for disk_s in ndev_s] if bdev is None or ndevs.count(None) > 0: raise ValueError("can't unserialize data!") return backend.BlockdevRemovechildren(bdev, ndevs) @staticmethod def perspective_blockdev_getmirrorstatus(params): """Return the mirror status for a list of disks. """ disks = [objects.Disk.FromDict(dsk_s) for dsk_s in params] return backend.BlockdevGetmirrorstatus(disks) @staticmethod def perspective_blockdev_find(params): """Expose the FindBlockDevice functionality for a disk. This will try to find but not activate a disk. """ disk = objects.Disk.FromDict(params[0]) return backend.BlockdevFind(disk) @staticmethod def perspective_blockdev_snapshot(params): """Create a snapshot device. Note that this is only valid for LVM disks, if we get passed something else we raise an exception. The snapshot device can be remove by calling the generic block device remove call. """ cfbd = objects.Disk.FromDict(params[0]) return backend.BlockdevSnapshot(cfbd) @staticmethod def perspective_blockdev_grow(params): """Grow a stack of devices. """ cfbd = objects.Disk.FromDict(params[0]) amount = params[1] return backend.BlockdevGrow(cfbd, amount) @staticmethod def perspective_blockdev_close(params): """Closes the given block devices. """ disks = [objects.Disk.FromDict(cf) for cf in params[1]] return backend.BlockdevClose(params[0], disks) # blockdev/drbd specific methods ---------- @staticmethod def perspective_drbd_disconnect_net(params): """Disconnects the network connection of drbd disks. Note that this is only valid for drbd disks, so the members of the disk list must all be drbd devices. """ nodes_ip, disks = params disks = [objects.Disk.FromDict(cf) for cf in disks] return backend.DrbdDisconnectNet(nodes_ip, disks) @staticmethod def perspective_drbd_attach_net(params): """Attaches the network connection of drbd disks. Note that this is only valid for drbd disks, so the members of the disk list must all be drbd devices. """ nodes_ip, disks, instance_name, multimaster = params disks = [objects.Disk.FromDict(cf) for cf in disks] return backend.DrbdAttachNet(nodes_ip, disks, instance_name, multimaster) @staticmethod def perspective_drbd_wait_sync(params): """Wait until DRBD disks are synched. Note that this is only valid for drbd disks, so the members of the disk list must all be drbd devices. """ nodes_ip, disks = params disks = [objects.Disk.FromDict(cf) for cf in disks] return backend.DrbdWaitSync(nodes_ip, disks) # export/import -------------------------- @staticmethod def perspective_snapshot_export(params): """Export a given snapshot. """ disk = objects.Disk.FromDict(params[0]) dest_node = params[1] instance = objects.Instance.FromDict(params[2]) cluster_name = params[3] dev_idx = params[4] return backend.ExportSnapshot(disk, dest_node, instance, cluster_name, dev_idx) @staticmethod def perspective_finalize_export(params): """Expose the finalize export functionality. """ instance = objects.Instance.FromDict(params[0]) snap_disks = [objects.Disk.FromDict(str_data) for str_data in params[1]] return backend.FinalizeExport(instance, snap_disks) @staticmethod def perspective_export_info(params): """Query information about an existing export on this node. The given path may not contain an export, in which case we return None. """ path = params[0] einfo = backend.ExportInfo(path) if einfo is None: return einfo return einfo.Dumps() @staticmethod def perspective_export_list(params): """List the available exports on this node. Note that as opposed to export_info, which may query data about an export in any path, this only queries the standard Ganeti path (constants.EXPORT_DIR). """ return backend.ListExports() @staticmethod def perspective_export_remove(params): """Remove an export. """ export = params[0] return backend.RemoveExport(export) # volume -------------------------- @staticmethod def perspective_volume_list(params): """Query the list of logical volumes in a given volume group. """ vgname = params[0] return backend.GetVolumeList(vgname) @staticmethod def perspective_vg_list(params): """Query the list of volume groups. """ return backend.ListVolumeGroups() # bridge -------------------------- @staticmethod def perspective_bridges_exist(params): """Check if all bridges given exist on this node. """ bridges_list = params[0] return backend.BridgesExist(bridges_list) # instance -------------------------- @staticmethod def perspective_instance_os_add(params): """Install an OS on a given instance. """ inst_s = params[0] inst = objects.Instance.FromDict(inst_s) return backend.InstanceOsAdd(inst) @staticmethod def perspective_instance_run_rename(params): """Runs the OS rename script for an instance. """ inst_s, old_name = params inst = objects.Instance.FromDict(inst_s) return backend.RunRenameInstance(inst, old_name) @staticmethod def perspective_instance_os_import(params): """Run the import function of an OS onto a given instance. """ inst_s, src_node, src_images, cluster_name = params inst = objects.Instance.FromDict(inst_s) return backend.ImportOSIntoInstance(inst, src_node, src_images, cluster_name) @staticmethod def perspective_instance_shutdown(params): """Shutdown an instance. """ instance = objects.Instance.FromDict(params[0]) return backend.InstanceShutdown(instance) @staticmethod def perspective_instance_start(params): """Start an instance. """ instance = objects.Instance.FromDict(params[0]) return backend.StartInstance(instance) @staticmethod def perspective_migration_info(params): """Gather information about an instance to be migrated. """ instance = objects.Instance.FromDict(params[0]) return backend.MigrationInfo(instance) @staticmethod def perspective_accept_instance(params): """Prepare the node to accept an instance. """ instance, info, target = params instance = objects.Instance.FromDict(instance) return backend.AcceptInstance(instance, info, target) @staticmethod def perspective_finalize_migration(params): """Finalize the instance migration. """ instance, info, success = params instance = objects.Instance.FromDict(instance) return backend.FinalizeMigration(instance, info, success) @staticmethod def perspective_instance_migrate(params): """Migrates an instance. """ instance, target, live = params instance = objects.Instance.FromDict(instance) return backend.MigrateInstance(instance, target, live) @staticmethod def perspective_instance_reboot(params): """Reboot an instance. """ instance = objects.Instance.FromDict(params[0]) reboot_type = params[1] return backend.InstanceReboot(instance, reboot_type) @staticmethod def perspective_instance_info(params): """Query instance information. """ return backend.GetInstanceInfo(params[0], params[1]) @staticmethod def perspective_instance_migratable(params): """Query whether the specified instance can be migrated. """ instance = objects.Instance.FromDict(params[0]) return backend.GetInstanceMigratable(instance) @staticmethod def perspective_all_instances_info(params): """Query information about all instances. """ return backend.GetAllInstancesInfo(params[0]) @staticmethod def perspective_instance_list(params): """Query the list of running instances. """ return backend.GetInstanceList(params[0]) # node -------------------------- @staticmethod def perspective_node_tcp_ping(params): """Do a TcpPing on the remote node. """ return utils.TcpPing(params[1], params[2], timeout=params[3], live_port_needed=params[4], source=params[0]) @staticmethod def perspective_node_has_ip_address(params): """Checks if a node has the given ip address. """ return utils.OwnIpAddress(params[0]) @staticmethod def perspective_node_info(params): """Query node information. """ vgname, hypervisor_type = params return backend.GetNodeInfo(vgname, hypervisor_type) @staticmethod def perspective_node_add(params): """Complete the registration of this node in the cluster. """ return backend.AddNode(params[0], params[1], params[2], params[3], params[4], params[5]) @staticmethod def perspective_node_verify(params): """Run a verify sequence on this node. """ return backend.VerifyNode(params[0], params[1]) @staticmethod def perspective_node_start_master(params): """Promote this node to master status. """ return backend.StartMaster(params[0]) @staticmethod def perspective_node_stop_master(params): """Demote this node from master status. """ return backend.StopMaster(params[0]) @staticmethod def perspective_node_leave_cluster(params): """Cleanup after leaving a cluster. """ return backend.LeaveCluster() @staticmethod def perspective_node_volumes(params): """Query the list of all logical volume groups. """ return backend.NodeVolumes() @staticmethod def perspective_node_demote_from_mc(params): """Demote a node from the master candidate role. """ return backend.DemoteFromMC() # cluster -------------------------- @staticmethod def perspective_version(params): """Query version information. """ return constants.PROTOCOL_VERSION @staticmethod def perspective_upload_file(params): """Upload a file. Note that the backend implementation imposes strict rules on which files are accepted. """ return backend.UploadFile(*params) @staticmethod def perspective_master_info(params): """Query master information. """ return backend.GetMasterInfo() @staticmethod def perspective_write_ssconf_files(params): """Write ssconf files. """ (values,) = params return backend.WriteSsconfFiles(values) # os ----------------------- @staticmethod def perspective_os_diagnose(params): """Query detailed information about existing OSes. """ return [os_obj.ToDict() for os_obj in backend.DiagnoseOS()] @staticmethod def perspective_os_get(params): """Query information about a given OS. """ name = params[0] try: os_obj = backend.OSFromDisk(name) except errors.InvalidOS, err: os_obj = objects.OS.FromInvalidOS(err) return os_obj.ToDict() # hooks ----------------------- @staticmethod def perspective_hooks_runner(params): """Run hook scripts. """ hpath, phase, env = params hr = backend.HooksRunner() return hr.RunHooks(hpath, phase, env) # iallocator ----------------- @staticmethod def perspective_iallocator_runner(params): """Run an iallocator script. """ name, idata = params iar = backend.IAllocatorRunner() return iar.Run(name, idata) # test ----------------------- @staticmethod def perspective_test_delay(params): """Run test delay. """ duration = params[0] return utils.TestDelay(duration) # file storage --------------- @staticmethod def perspective_file_storage_dir_create(params): """Create the file storage directory. """ file_storage_dir = params[0] return backend.CreateFileStorageDir(file_storage_dir) @staticmethod def perspective_file_storage_dir_remove(params): """Remove the file storage directory. """ file_storage_dir = params[0] return backend.RemoveFileStorageDir(file_storage_dir) @staticmethod def perspective_file_storage_dir_rename(params): """Rename the file storage directory. """ old_file_storage_dir = params[0] new_file_storage_dir = params[1] return backend.RenameFileStorageDir(old_file_storage_dir, new_file_storage_dir) # jobs ------------------------ @staticmethod @_RequireJobQueueLock def perspective_jobqueue_update(params): """Update job queue. """ (file_name, content) = params return backend.JobQueueUpdate(file_name, content) @staticmethod @_RequireJobQueueLock def perspective_jobqueue_purge(params): """Purge job queue. """ return backend.JobQueuePurge() @staticmethod @_RequireJobQueueLock def perspective_jobqueue_rename(params): """Rename a job queue file. """ # TODO: What if a file fails to rename? return [backend.JobQueueRename(old, new) for old, new in params] @staticmethod def perspective_jobqueue_set_drain(params): """Set/unset the queue drain flag. """ drain_flag = params[0] return backend.JobQueueSetDrainFlag(drain_flag) # hypervisor --------------- @staticmethod def perspective_hypervisor_validate_params(params): """Validate the hypervisor parameters. """ (hvname, hvparams) = params return backend.ValidateHVParams(hvname, hvparams) def ParseOptions(): """Parse the command line options. @return: (options, args) as from OptionParser.parse_args() """ parser = OptionParser(description="Ganeti node daemon", usage="%prog [-f] [-d] [-b ADDRESS]", version="%%prog (ganeti) %s" % constants.RELEASE_VERSION) parser.add_option("-f", "--foreground", dest="fork", help="Don't detach from the current terminal", default=True, action="store_false") parser.add_option("-d", "--debug", dest="debug", help="Enable some debug messages", default=False, action="store_true") parser.add_option("-b", "--bind", dest="bind_address", help="Bind address", default="", metavar="ADDRESS") options, args = parser.parse_args() return options, args def main(): """Main function for the node daemon. """ global queue_lock options, args = ParseOptions() if options.fork: utils.CloseFDs() for fname in (constants.SSL_CERT_FILE,): if not os.path.isfile(fname): print "config %s not there, will not run." % fname sys.exit(constants.EXIT_NOTCLUSTER) port = utils.GetNodeDaemonPort() dirs = [(val, constants.RUN_DIRS_MODE) for val in constants.SUB_RUN_DIRS] dirs.append((constants.LOG_OS_DIR, 0750)) dirs.append((constants.LOCK_DIR, 1777)) utils.EnsureDirs(dirs) # become a daemon if options.fork: utils.Daemonize(logfile=constants.LOG_NODESERVER) utils.WritePidFile(constants.NODED_PID) try: utils.SetupLogging(logfile=constants.LOG_NODESERVER, debug=options.debug, stderr_logging=not options.fork) logging.info("ganeti node daemon startup") # Read SSL certificate ssl_params = http.HttpSslParams(ssl_key_path=constants.SSL_CERT_FILE, ssl_cert_path=constants.SSL_CERT_FILE) # Prepare job queue queue_lock = jstore.InitAndVerifyQueue(must_lock=False) mainloop = daemon.Mainloop() server = NodeHttpServer(mainloop, options.bind_address, port, ssl_params=ssl_params, ssl_verify_peer=True) server.Start() try: mainloop.Run() finally: server.Stop() finally: utils.RemovePidFile(constants.NODED_PID) if __name__ == '__main__': main()