X-Git-Url: https://code.grnet.gr/git/ganeti-local/blobdiff_plain/8342c325c7ea6ed3aa7f380dad78c629757a089d..b990eedd100be28469c65af95935f90e2583cafe:/lib/daemon.py diff --git a/lib/daemon.py b/lib/daemon.py index 99f7dd3..1e41b80 100644 --- a/lib/daemon.py +++ b/lib/daemon.py @@ -1,7 +1,7 @@ # # -# Copyright (C) 2006, 2007, 2008, 2010 Google Inc. +# Copyright (C) 2006, 2007, 2008, 2010, 2011 Google Inc. # # This program is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by @@ -40,6 +40,7 @@ from ganeti import errors from ganeti import netutils from ganeti import ssconf from ganeti import runtime +from ganeti import compat class SchedulerBreakout(Exception): @@ -74,7 +75,43 @@ class AsyncoreScheduler(sched.scheduler): """ def __init__(self, timefunc): - sched.scheduler.__init__(self, timefunc, AsyncoreDelayFunction) + """Initializes this class. + + """ + sched.scheduler.__init__(self, timefunc, self._LimitedDelay) + self._max_delay = None + + def run(self, max_delay=None): # pylint: disable=W0221 + """Run any pending events. + + @type max_delay: None or number + @param max_delay: Maximum delay (useful if caller has timeouts running) + + """ + assert self._max_delay is None + + # The delay function used by the scheduler can't be different on each run, + # hence an instance variable must be used. + if max_delay is None: + self._max_delay = None + else: + self._max_delay = utils.RunningTimeout(max_delay, False) + + try: + return sched.scheduler.run(self) + finally: + self._max_delay = None + + def _LimitedDelay(self, duration): + """Custom delay function for C{sched.scheduler}. + + """ + if self._max_delay is None: + timeout = duration + else: + timeout = min(duration, self._max_delay.Remaining()) + + return AsyncoreDelayFunction(timeout) class GanetiBaseAsyncoreDispatcher(asyncore.dispatcher): @@ -420,6 +457,47 @@ class AsyncAwaker(GanetiBaseAsyncoreDispatcher): self.out_socket.send("\0") +class _ShutdownCheck: + """Logic for L{Mainloop} shutdown. + + """ + def __init__(self, fn): + """Initializes this class. + + @type fn: callable + @param fn: Function returning C{None} if mainloop can be stopped or a + duration in seconds after which the function should be called again + @see: L{Mainloop.Run} + + """ + assert callable(fn) + + self._fn = fn + self._defer = None + + def CanShutdown(self): + """Checks whether mainloop can be stopped. + + @rtype: bool + + """ + if self._defer and self._defer.Remaining() > 0: + # A deferred check has already been scheduled + return False + + # Ask mainloop driver whether we can stop or should check again + timeout = self._fn() + + if timeout is None: + # Yes, can stop mainloop + return True + + # Schedule another check in the future + self._defer = utils.RunningTimeout(timeout, True) + + return False + + class Mainloop(object): """Generic mainloop for daemons @@ -427,6 +505,8 @@ class Mainloop(object): timed events """ + _SHUTDOWN_TIMEOUT_PRIORITY = -(sys.maxint - 1) + def __init__(self): """Constructs a new Mainloop instance. @@ -434,12 +514,19 @@ class Mainloop(object): self._signal_wait = [] self.scheduler = AsyncoreScheduler(time.time) + # Resolve uid/gids used + runtime.GetEnts() + @utils.SignalHandled([signal.SIGCHLD]) @utils.SignalHandled([signal.SIGTERM]) @utils.SignalHandled([signal.SIGINT]) - def Run(self, signal_handlers=None): + def Run(self, shutdown_wait_fn=None, signal_handlers=None): """Runs the mainloop. + @type shutdown_wait_fn: callable + @param shutdown_wait_fn: Function to check whether loop can be terminated; + B{important}: function must be idempotent and must return either None + for shutting down or a timeout for another call @type signal_handlers: dict @param signal_handlers: signal->L{utils.SignalHandler} passed by decorator @@ -447,23 +534,50 @@ class Mainloop(object): assert isinstance(signal_handlers, dict) and \ len(signal_handlers) > 0, \ "Broken SignalHandled decorator" - running = True + + # Counter for received signals + shutdown_signals = 0 + + # Logic to wait for shutdown + shutdown_waiter = None + # Start actual main loop - while running: - if not self.scheduler.empty(): + while True: + if shutdown_signals == 1 and shutdown_wait_fn is not None: + if shutdown_waiter is None: + shutdown_waiter = _ShutdownCheck(shutdown_wait_fn) + + # Let mainloop driver decide if we can already abort + if shutdown_waiter.CanShutdown(): + break + + # Re-evaluate in a second + timeout = 1.0 + + elif shutdown_signals >= 1: + # Abort loop if more than one signal has been sent or no callback has + # been given + break + + else: + # Wait forever on I/O events + timeout = None + + if self.scheduler.empty(): + asyncore.loop(count=1, timeout=timeout, use_poll=True) + else: try: - self.scheduler.run() + self.scheduler.run(max_delay=timeout) except SchedulerBreakout: pass - else: - asyncore.loop(count=1, use_poll=True) # Check whether a signal was raised - for sig in signal_handlers: - handler = signal_handlers[sig] + for (sig, handler) in signal_handlers.items(): if handler.called: self._CallSignalWaiters(sig) - running = sig not in (signal.SIGTERM, signal.SIGINT) + if sig in (signal.SIGTERM, signal.SIGINT): + logging.info("Received signal %s asking for shutdown", sig) + shutdown_signals += 1 handler.Clear() def _CallSignalWaiters(self, signum): @@ -534,11 +648,24 @@ def _BeautifyError(err): err.errno) else: return str(err) - except Exception: # pylint: disable-msg=W0703 + except Exception: # pylint: disable=W0703 logging.exception("Error while handling existing error %s", err) return "%s" % str(err) +def _HandleSigHup(reopen_fn, signum, frame): # pylint: disable=W0613 + """Handler for SIGHUP. + + @param reopen_fn: List of callback functions for reopening log files + + """ + logging.info("Reopening log files after receiving SIGHUP") + + for fn in reopen_fn: + if fn: + fn() + + def GenericMain(daemon_name, optionparser, check_fn, prepare_fn, exec_fn, multithreaded=False, console_logging=False, @@ -620,7 +747,8 @@ def GenericMain(daemon_name, optionparser, metavar="SSL_CERT_PATH") # Disable the use of fork(2) if the daemon uses threads - utils.no_fork = multithreaded + if multithreaded: + utils.DisableFork() options, args = optionparser.parse_args() @@ -651,25 +779,36 @@ def GenericMain(daemon_name, optionparser, if options.fork: utils.CloseFDs() - wpipe = utils.Daemonize(logfile=constants.DAEMONS_LOGFILES[daemon_name]) + (wpipe, stdio_reopen_fn) = \ + utils.Daemonize(logfile=constants.DAEMONS_LOGFILES[daemon_name]) else: - wpipe = None + (wpipe, stdio_reopen_fn) = (None, None) + + log_reopen_fn = \ + utils.SetupLogging(constants.DAEMONS_LOGFILES[daemon_name], daemon_name, + debug=options.debug, + stderr_logging=not options.fork, + multithreaded=multithreaded, + syslog=options.syslog, + console_logging=console_logging) + + # Reopen log file(s) on SIGHUP + signal.signal(signal.SIGHUP, + compat.partial(_HandleSigHup, [log_reopen_fn, stdio_reopen_fn])) + + try: + utils.WritePidFile(utils.DaemonPidFileName(daemon_name)) + except errors.PidFileLockError, err: + print >> sys.stderr, "Error while locking PID file:\n%s" % err + sys.exit(constants.EXIT_FAILURE) - utils.WritePidFile(utils.DaemonPidFileName(daemon_name)) try: try: - utils.SetupLogging(logfile=constants.DAEMONS_LOGFILES[daemon_name], - debug=options.debug, - stderr_logging=not options.fork, - multithreaded=multithreaded, - program=daemon_name, - syslog=options.syslog, - console_logging=console_logging) + logging.info("%s daemon startup", daemon_name) if callable(prepare_fn): prep_results = prepare_fn(options, args) else: prep_results = None - logging.info("%s daemon startup", daemon_name) except Exception, err: utils.WriteErrorToFD(wpipe, _BeautifyError(err)) raise @@ -681,4 +820,4 @@ def GenericMain(daemon_name, optionparser, exec_fn(options, args, prep_results) finally: - utils.RemovePidFile(utils.DaemonPidFileName(daemon_name)) + utils.RemoveFile(utils.DaemonPidFileName(daemon_name))