May 29, 2008"> 8"> gnt-instance"> Debian"> GNU"> GPL"> ]> 2006 2007 2008 Google Inc. &dhdate; &dhucpackage; &dhsection; ganeti 1.2 &dhpackage; ganeti instance administration &dhpackage; command arguments... DESCRIPTION The &dhpackage; is used for instance administration in the ganeti system. COMMANDS Creation/removal/querying ADD add -s disksize --swap-size disksize -m memsize -b bridge --mac MAC-address --hvm-boot-order boot-order --hvm-acpi ACPI-support --hvm-pae PAE-support --hvm-cdrom-image-path cdrom-image-path --hvm-nic-type NICTYPE --hvm-disk-type DISKTYPE --vnc-bind-address vnc-bind-address --kernel default kernel_path --initrd default none initrd_path --file-storage-dir dir_path --file-driver loop blktap -t diskless file plain drbd -n node:secondary-node --iallocator name -o os-type instance Creates a new instance on the specified host. instance must be in DNS and resolve to a IP in the same network as the nodes in the cluster. The option specifies the disk size for the instance, in mebibytes (defaults to 20480MiB = 20GiB). You can also use one of the suffixes m, g or t to specificy the exact the units used; these suffixes map to mebibytes, gibibytes and tebibytes. The option specifies the swap disk size (in mebibytes) for the instance (the one presented as /dev/sdb). The default is 4096MiB. As for the disk size, you can specify other suffixes. The option specifies the memory size for the instance, in mebibytes (defaults to 128 MiB). Again, you can use other suffixes (e.g. 2g). The options specifies the operating system to be installed. The available operating systems can be listed with gnt-os list. The option specifies the bridge to which the instance will be connected. (defaults to the cluster-wide default bridge specified at cluster initialization time). The option specifies the MAC address of the ethernet interface for the instance. If this option is not specified, a new MAC address is generated randomly with the configured MAC prefix. The randomly generated MAC address is guaranteed to be unique among the instances of this cluster. The option specifies the boot device order for Xen HVM instances. The boot order is a string of letters listing the boot devices, with valid device letters being: a floppy drive c hard disk d CDROM drive n network boot (PXE) The default is not to set an HVM boot order which is interpreted as 'dc'. This option, like all options starting with 'hvm', is only relevant for Xen HVM instances and ignored by all other instance types. The option specifies if Xen should enable ACPI support for this HVM instance. Valid values are true or false. The default value is false, disabling ACPI support for this instance. The option specifies if Xen should enabled PAE support for this HVM instance. Valid values are true or false. The default is false, disabling PAE support for this instance. The option specifies the path to the file Xen uses to emulate a virtual CDROM drive for this HVM instance. Valid values are either an absolute path to an existing file or None, which disables virtual CDROM support for this instance. The default is None, disabling virtual CDROM support. The specifies the NIC type Xen should use for this HVM instance. Valid choices are rtl8139, ne2k_pci, ne2k_isa and paravirtual with rtl8139 as the default. The paravirtual setting is intended for use with the GPL PV drivers inside HVM Windows instances. The specifies the disk type Xen should use for the HVM instance. Valid choices are ioemu and paravirtual with ioemu as the default. The paravirtual setting is intended for use with the GPL PV drivers inside HVM Windows instances. The option specifies the address that the VNC listener for this instance should bind to. Valid values are IPv4 addresses. Use the address 0.0.0.0 to bind to all available interfaces (this is the default) or specify the address of one of the interfaces on the node to restrict listening to that interface. The option specifies the instance allocator plugin to use. If you pass in this option the allocator will select nodes for this instance automatically, so you don't need to pass them with the option. For more information please refer to the instance allocator documentation. The option allows the instance to use a custom kernel (if a filename is passed) or to use the default kernel (@CUSTOM_XEN_KERNEL@), if the string default is passed. The option is similar: it allows the instance to use a custom initrd (if a filename is passed) or to use the default initrd (@CUSTOM_XEN_INITRD@), if the string default is passed, or to disable the use of an initrd, if the string none is passed. Note that in the case the instance is set to use the default initrd and it doesn't exist, it will be silently ignored; if the instance is set to use a custom initrd and it doesn't exist, this will be treated as an error and will prevent the startup of the instance. The options specifies the disk layout type for the instance. The available choices are: diskless This creates an instance with no disks. Its useful for testing only (or other special cases). file Disk devices will be regular files. plain Disk devices will be logical volumes. drbd Disk devices will be drbd (version 8.x) on top of lvm volumes. The optional second value of the is used for the drbd template type and specifies the remote node. If you do not want gnt-instance to wait for the disk mirror to be synced, use the option. The specifies the relative path under the cluster-wide file storage directory to store file-based disks. It is useful for having different subdirectories for different instances. The full path of the directory where the disk files are stored will consist of cluster-wide file storage directory + optional subdirectory + instance name. Example: /srv/ganeti/file-storage/mysubdir/instance1.example.com. This option is only relevant for instances using the file storage backend. The specifies the driver to use for file-based disks. Note that currently these drivers work with the xen hypervisor only. This option is only relevant for instances using the file storage backend. The available choices are: loop Kernel loopback driver. blktap blktap driver. The loop driver uses loopback devices to access the filesystem within the file. However, running I/O intensive applications in your instance using the loop driver might result in slowdowns. Furthermore, if you use the loopback driver consider increasing the maximum amount of loopback devices (on most systems it's 8) using the max_loop param. In order to be able to use the blktap driver you should check if the 'blktapctrl' user space disk agent is running (usually automatically started via xend). This user-level disk I/O interface has the advantage of better performance. Especially if you use a network file system (e.g. NFS) to store your instances this is the recommended choice. Example: # gnt-instance add -t file -s 30g -m 512 -o debian-etch \ -n node1.example.com --file-storage-dir=mysubdir instance1.example.com # gnt-instance add -t plain -s 30g -m 512 -o debian-etch \ -n node1.example.com instance1.example.com # gnt-instance add -t drbd -s 30g -m 512 -o debian-etch \ -n node1.example.com:node2.example.com instance2.example.com REMOVE remove --ignore-failures instance Remove an instance. This will remove all data from the instance and there is no way back. If you are not sure if you use an instance again, use shutdown first and leave it in the shutdown state for a while. The option will cause the removal to proceed even in the presence of errors during the removal of the instance (e.g. during the shutdown or the disk removal). If this option is not given, the command will stop at the first error. Example: # gnt-instance remove instance1.example.com LIST list --no-headers --separator=SEPARATOR -o [+]FIELD,... Shows the currently configured instances with memory usage, disk usage, the node they are running on, and the CPU time, counted in seconds, used by each instance since its latest restart. The option will skip the initial header line. The option takes an argument which denotes what will be used between the output fields. Both these options are to help scripting. The option takes a comma-separated list of output fields. The available fields and their meaning are: name the instance name os the OS of the instance pnode the primary node of the instance snodes comma-separated list of secondary nodes for the instance; usually this will be just one node admin_state the desired state of the instance (either "yes" or "no" denoting the instance should run or not) admin_ram the desired memory for the instance disk_template the disk template of the instance oper_state the actual state of the instance; can be one of the values "running", "stopped", "(node down)" status combined form of admin_state and oper_stat; this can be one of: ERROR_nodedown if the node of the instance is down, ERROR_down if the instance should run but is down, ERROR_up if the instance should be stopped but is actually running, ADMIN_down if the instance has been stopped (and is stopped) and running if the instance is set to be running (and is running) oper_ram the actual memory usage of the instance as seen by the hypervisor ip the ip address ganeti recognizes as associated with the instance interface mac the instance interface MAC address bridge bridge the instance is connected to sda_size the size of the instance's first disk sdb_size the size of the instance's second disk vcpus the number of VCPUs allocated to the instance tags comma-separated list of the instances's tags serial_no the so called 'serial number' of the instance; this is a numeric field that is incremented each time the instance is modified, and it can be used to detect modifications If the value of the option starts with the character +, the new fields will be added to the default list. This allows to quickly see the default list plus a few other fields, instead of retyping the entire list of fields. There is a subtle grouping about the available output fields: all fields except for , and are configuration value and not run-time values. So if you don't select any of the these fields, the query will be satisfied instantly from the cluster configuration, without having to ask the remote nodes for the data. This can be helpful for big clusters when you only want some data and it makes sense to specify a reduced set of output fields. The default output field list is: name os pnode admin_state oper_state oper_ram . INFO info -s --static instance Show detailed information about the (given) instances. This is different from list as it shows detailed data about the instance's disks (especially useful for drbd disk template). If the option is used, only information available in the configuration file is returned, without querying nodes, making the operation faster. MODIFY modify -m memsize -p vcpus -i ip -b bridge --mac MAC-address --hvm-boot-order boot-order --hvm-acpi ACPI-support --hvm-pae PAE-support --hvm-cdrom-image-path cdrom-image-path --hvm-nic-type NICTYPE --hvm-disk-type DISKTYPE --vnc-bind-address vnc-bind-address --kernel default kernel_path --initrd default none initrd_path instance Modify the memory size, number of vcpus, ip address, MAC address and/or bridge for an instance. The memory size is given in MiB. Note that you need to give at least one of the arguments, otherwise the command complains. The , and options are described in the add command. Additionally, the HVM boot order can be reset to the default values by using . The option specifies if Xen should enable ACPI support for this HVM instance. Valid values are true or false. The option specifies if Xen should enabled PAE support for this HVM instance. Valid values are true or false. The specifies the path to the file xen uses to emulate a virtual CDROM drive for this HVM instance. Valid values are either an absolute path to an existing file or None, which disables virtual CDROM support for this instance. The specifies the NIC type Xen should use for this HVM instance. Valid choices are rtl8139, ne2k_pci, ne2k_isa and paravirtual with rtl8139 as the default. The paravirtual setting is intended for use with the GPL PV drivers inside HVM Windows instances. The specifies the disk type Xen should use for the HVM instance. Valid choices are ioemu and paravirtual with ioemu as the default. The paravirtual setting is intended for use with the GPL PV drivers inside HVM Windows instances. The specifies the address that the VNC listener for this instance should bind to. Valid values are IPv4 addresses. Use the address 0.0.0.0 to bind to all available interfaces. All the changes take effect at the next restart. If the instance is running, there is no effect on the instance. REINSTALL reinstall -o os-type -f force --select-os instance Reinstalls the operating system on the given instance. The instance must be stopped when running this command. If the is specified, the operating system is changed. The option switches to an interactive OS reinstall. The user is prompted to select the OS template from the list of available OS templates. RENAME rename --no-ip-check instance new_name Renames the given instance. The instance must be stopped when running this command. The requirements for the new name are the same as for adding an instance: the new name must be resolvable and the IP it resolves to must not be reachable (in order to prevent duplicate IPs the next time the instance is started). The IP test can be skipped if the option is passed. Starting/stopping/connecting to console STARTUP startup --extra=PARAMS --force --instance --node --primary --secondary --all name Starts one or more instances, depending on the following options. The four available modes are: will start the instances given as arguments (at least one argument required); this is the default selection --node will start the instances who have the given node as either primary or secondary will start all instances whose primary node is in the list of nodes passed as arguments (at least one node required) will start all instances whose secondary node is in the list of nodes passed as arguments (at least one node required) --all will start all instances in the cluster (no arguments accepted) Note that although you can pass more than one selection option, the last one wins, so in order to guarantee the desired result, don't pass more than one such option. The option is used to pass additional argument to the instance's kernel for this start only. Currently there is no way to specify a persistent set of arguments (beside the one hardcoded). Note that this may not apply to all virtualization types. Use to start even if secondary disks are failing. Example: # gnt-instance start instance1.example.com # gnt-instance start --extra single test1.example.com # gnt-instance start --node node1.example.com node2.example.com # gnt-instance start --all SHUTDOWN shutdown --instance --node --primary --secondary --all name Stops one or more instances. If the instance cannot be cleanly stopped during a hardcoded interval (currently 2 minutes), it will forcibly stop the instance (equivalent to switching off the power on a physical machine). The , , , and options are similar as for the startup command and they influence the actual instances being shutdown. Example: # gnt-instance shutdown instance1.example.com # gnt-instance shutdown --all REBOOT reboot --extra=PARAMS --type=REBOOT-TYPE --ignore-secondaries --force-multiple --instance --node --primary --secondary --all name Reboots one or more instances. The type of reboot depends on the value of . A soft reboot does a hypervisor reboot, a hard reboot does a instance stop, recreates the hypervisor config for the instance and starts the instance. A full reboot does the equivalent of gnt-instance shutdown && gnt-instance startup. The default is hard reboot. For the hard reboot the option ignores errors for the secondary node while re-assembling the instance disks. The , , , and options are similar as for the startup command and they influence the actual instances being rebooted. Use the option to keep gnt-instance from asking for confirmation when more than one instance is affected. Example: # gnt-instance reboot instance1.example.com # gnt-instance reboot --type=full instance1.example.com CONSOLE console --show-cmd instance Connects to the console of the given instance. If the instance is not up, an error is returned. Use the option to display the command instead of executing it. For HVM instances, this will attempt to connect to the serial console of the instance. To connect to the virtualized "physical" console of a HVM instance, use a VNC client with the connection info from gnt-instance info. Example: # gnt-instance console instance1.example.com Disk management REPLACE-DISKS replace-disks -p instance replace-disks -s instance replace-disks --iallocator name --new-secondary NODE instance This command is a generalized form for adding and replacing disks. It is currently only valid for the mirrored (DRBD) disk template. The first form (when passing the option) will replace the disks on the primary, while the second form (when passing the option will replace the disks on the secondary node. The third form (when passing either the or the option) is designed to change secondary node of the instance. Specifying makes the new secondary be selected automatically by the specified allocator plugin, otherwise the new secondary node will be the one chosen manually via the option. ACTIVATE-DISKS activate-disks instance Activates the block devices of the given instance. If successful, the command will show the location and name of the block devices: node1.example.com:sda:/dev/drbd0 node1.example.com:sdb:/dev/drbd1 In this example, node1.example.com is the name of the node on which the devices have been activated. The sda and sdb are the names of the block devices inside the instance. /dev/drbd0 and /dev/drbd1 are the names of the block devices as visible on the node. Note that it is safe to run this command while the instance is already running. DEACTIVATE-DISKS deactivate-disks instance De-activates the block devices of the given instance. Note that if you run this command for an instance with a drbd disk template, while it is running, it will not be able to shutdown the block devices on the primary node, but it will shutdown the block devices on the secondary nodes, thus breaking the replication. GROW-DISK grow-disk --no-wait-for-sync instance disk amount Grows an instance's disk. This is only possible for instances having a plain or drbd disk template. Note that this command only change the block device size; it will not grow the actual filesystems, partitions, etc. that live on that disk. Usually, you will need to: use gnt-instance grow-disk reboot the instance (later, at a convenient time) use a filesystem resizer, such as ext2online 8 or xfs_growfs 8 to resize the filesystem, or use fdisk 8 to change the partition table on the disk The disk argument is either sda or sdb. The amount argument is given either as a number (and it represents the amount to increase the disk with in mebibytes) or can be given similar to the arguments in the create instance operation, with a suffix denoting the unit. Note that the disk grow operation might complete on one node but fail on the other; this will leave the instance with different-sized LVs on the two nodes, but this will not create problems (except for unused space). If you do not want gnt-instance to wait for the new disk region to be synced, use the option. Example (increase sda for instance1 by 16GiB): # gnt-instance grow-disk instance1.example.com sda 16g Also note that disk shrinking will not be supported; use gnt-backup export and then gnt-backup import to reduce the disk size of an instance. Recovery FAILOVER failover -f --ignore-consistency instance Failover will fail the instance over its secondary node. This works only for instances having a drbd disk template. Normally the failover will check the consistency of the disks before failing over the instance. If you are trying to migrate instances off a dead node, this will fail. Use the option for this purpose. Note that this option can be dangerous as errors in shutting down the instance will be ignored, resulting in possibly having the instance running on two machines in parallel (on disconnected DRBD drives). Example: # gnt-instance failover instance1.example.com TAGS ADD-TAGS add-tags --from file instancename tag Add tags to the given instance. If any of the tags contains invalid characters, the entire operation will abort. If the option is given, the list of tags will be extended with the contents of that file (each line becomes a tag). In this case, there is not need to pass tags on the command line (if you do, both sources will be used). A file name of - will be interpreted as stdin. LIST-TAGS list-tags instancename List the tags of the given instance. REMOVE-TAGS remove-tags --from file instancename tag Remove tags from the given instance. If any of the tags are not existing on the node, the entire operation will abort. If the option is given, the list of tags will be extended with the contents of that file (each line becomes a tag). In this case, there is not need to pass tags on the command line (if you do, both sources will be used). A file name of - will be interpreted as stdin. &footer;