1 <!DOCTYPE article PUBLIC "-//OASIS//DTD DocBook V4.2//EN" [
3 <article class="specification">
5 <title>Ganeti installation tutorial</title>
7 <para>Documents Ganeti version 1.2</para>
10 <title>Introduction</title>
13 Ganeti is a cluster virtualization management system based on
14 Xen. This document explains how to bootstrap a Ganeti node (Xen
15 <literal>dom0</literal>), create a running cluster and install
16 virtual instance (Xen <literal>domU</literal>). You need to
17 repeat most of the steps in this document for every node you
18 want to install, but of course we recommend creating some
19 semi-automatic procedure if you plan to deploy Ganeti on a
24 A basic Ganeti terminology glossary is provided in the
25 introductory section of the <emphasis>Ganeti administrator's
26 guide</emphasis>. Please refer to that document if you are
27 uncertain about the terms we are using.
31 Ganeti has been developed for Linux and is
32 distribution-agnostic. This documentation will use Debian Etch
33 as an example system but the examples can easily be translated
34 to any other distribution. You are expected to be familiar with
35 your distribution, its package management system, and Xen before
39 <para>This document is divided into two main sections:
43 <simpara>Installation of the base system and base
47 <simpara>Configuration of the environment for
52 Each of these is divided into sub-sections. While a full Ganeti system
53 will need all of the steps specified, some are not strictly required for
54 every environment. Which ones they are, and why, is specified in the
55 corresponding sections.
61 <title>Installing the base system and base components</title>
64 <title>Hardware requirements</title>
67 Any system supported by your Linux distribution is fine. 64-bit
68 systems are better as they can support more memory.
72 Any disk drive recognized by Linux
73 (<literal>IDE</literal>/<literal>SCSI</literal>/<literal>SATA</literal>/etc.)
74 is supported in Ganeti. Note that no shared storage (e.g.
75 <literal>SAN</literal>) is needed to get high-availability features. It
76 is highly recommended to use more than one disk drive to improve speed.
77 But Ganeti also works with one disk per machine.
81 <title>Installing the base system</title>
84 <emphasis role="strong">Mandatory</emphasis> on all nodes.
88 It is advised to start with a clean, minimal install of the
89 operating system. The only requirement you need to be aware of
90 at this stage is to partition leaving enough space for a big
91 (<emphasis role="strong">minimum
92 <constant>20GiB</constant></emphasis>) LVM volume group which
93 will then host your instance filesystems. The volume group
94 name Ganeti 1.2 uses (by default) is
95 <emphasis>xenvg</emphasis>.
99 While you can use an existing system, please note that the
100 Ganeti installation is intrusive in terms of changes to the
101 system configuration, and it's best to use a newly-installed
102 system without important data on it.
106 Also, for best results, it's advised that the nodes have as
107 much as possible the same hardware and software
108 configuration. This will make administration much easier.
112 <title>Hostname issues</title>
114 Note that Ganeti requires the hostnames of the systems
115 (i.e. what the <computeroutput>hostname</computeroutput>
116 command outputs to be a fully-qualified name, not a short
117 name. In other words, you should use
118 <literal>node1.example.com</literal> as a hostname and not
119 just <literal>node1</literal>.
123 <title>Debian</title>
125 Note that Debian Etch configures the hostname differently
126 than you need it for Ganeti. For example, this is what
127 Etch puts in <filename>/etc/hosts</filename> in certain
131 127.0.1.1 node1.example.com node1
134 but for Ganeti you need to have:
137 192.168.1.1 node1.example.com node1
139 replacing <literal>192.168.1.1</literal> with your node's
140 address. Also, the file <filename>/etc/hostname</filename>
141 which configures the hostname of the system should contain
142 <literal>node1.example.com</literal> and not just
143 <literal>node1</literal> (you need to run the command
144 <computeroutput>/etc/init.d/hostname.sh
145 start</computeroutput> after changing the file).
153 <title>Installing Xen</title>
156 <emphasis role="strong">Mandatory</emphasis> on all nodes.
160 While Ganeti is developed with the ability to modularly run on
161 different virtualization environments in mind the only one
162 currently useable on a live system is <ulink
163 url="http://xen.xensource.com/">Xen</ulink>. Supported
164 versions are: <simplelist type="inline">
165 <member><literal>3.0.3</literal></member>
166 <member><literal>3.0.4</literal></member>
167 <member><literal>3.1</literal></member> </simplelist>.
171 Please follow your distribution's recommended way to install
172 and set up Xen, or install Xen from the upstream source, if
173 you wish, following their manual.
177 After installing Xen you need to reboot into your Xen-ified
178 dom0 system. On some distributions this might involve
179 configuring GRUB appropriately, whereas others will configure
180 it automatically when you install Xen from a package.
183 <formalpara><title>Debian</title>
185 Under Debian Etch or Sarge+backports you can install the
186 relevant <literal>xen-linux-system</literal> package, which
187 will pull in both the hypervisor and the relevant
188 kernel. Also, if you are installing a 32-bit Etch, you should
189 install the <computeroutput>libc6-xen</computeroutput> package
190 (run <computeroutput>apt-get install
191 libc6-xen</computeroutput>).
196 <title>Xen settings</title>
199 It's recommended that dom0 is restricted to a low amount of
200 memory (<constant>512MiB</constant> is reasonable) and that
201 memory ballooning is disabled in the file
202 <filename>/etc/xen/xend-config.sxp</filename> by setting the
203 value <literal>dom0-min-mem</literal> to
204 <constant>0</constant>, like this:
205 <computeroutput>(dom0-min-mem 0)</computeroutput>
209 For optimum performance when running both CPU and I/O
210 intensive instances, it's also recommended that the dom0 is
211 restricted to one CPU only, for example by booting with the
212 kernel parameter <literal>nosmp</literal>.
216 It is recommended that you disable xen's automatic save of virtual
217 machines at system shutdown and subsequent restore of them at reboot.
218 To obtain this make sure the variable
219 <literal>XENDOMAINS_SAVE</literal> in the file
220 <literal>/etc/default/xendomains</literal> is set to an empty value.
224 <title>Debian</title>
226 Besides the ballooning change which you need to set in
227 <filename>/etc/xen/xend-config.sxp</filename>, you need to
228 set the memory and nosmp parameters in the file
229 <filename>/boot/grub/menu.lst</filename>. You need to
230 modify the variable <literal>xenhopt</literal> to add
231 <userinput>dom0_mem=512M</userinput> like this:
233 ## Xen hypervisor options to use with the default Xen boot option
234 # xenhopt=dom0_mem=512M
236 and the <literal>xenkopt</literal> needs to include the
237 <userinput>nosmp</userinput> option like this:
239 ## Xen Linux kernel options to use with the default Xen boot option
243 Any existing parameters can be left in place: it's ok to
244 have <computeroutput>xenkopt=console=tty0
245 nosmp</computeroutput>, for example. After modifying the
246 files, you need to run:
253 If you want to test the HVM support
254 with Ganeti and want VNC access to the console of your
255 instances, set the following two entries in
256 <filename>/etc/xen/xend-config.sxp</filename>:
258 (vnc-listen '0.0.0.0')
261 You need to restart the Xen daemon for these settings to
264 /etc/init.d/xend restart
271 <title>Selecting the instance kernel</title>
274 After you have installed Xen, you need to tell Ganeti
275 exactly what kernel to use for the instances it will
276 create. This is done by creating a
277 <emphasis>symlink</emphasis> from your actual kernel to
278 <filename>/boot/vmlinuz-2.6-xenU</filename>, and one from
280 <filename>/boot/initrd-2.6-xenU</filename>. Note that if you
281 don't use an initrd for the <literal>domU</literal> kernel,
282 you don't need to create the initrd symlink.
286 <title>Debian</title>
288 After installation of the
289 <literal>xen-linux-system</literal> package, you need to
290 run (replace the exact version number with the one you
294 ln -s vmlinuz-2.6.18-5-xen-686 vmlinuz-2.6-xenU
295 ln -s initrd.img-2.6.18-5-xen-686 initrd-2.6-xenU
304 <title>Installing DRBD</title>
307 Recommended on all nodes: <ulink
308 url="http://www.drbd.org/">DRBD</ulink> is required if you
309 want to use the high availability (HA) features of Ganeti, but
310 optional if you don't require HA or only run Ganeti on
311 single-node clusters. You can upgrade a non-HA cluster to an
312 HA one later, but you might need to export and re-import all
313 your instances to take advantage of the new features.
317 Supported DRBD versions: <literal>8.0.x</literal>.
318 It's recommended to have at least version <literal>8.0.7</literal>.
322 Now the bad news: unless your distribution already provides it
323 installing DRBD might involve recompiling your kernel or
324 anyway fiddling with it. Hopefully at least the Xen-ified
325 kernel source to start from will be provided.
329 The good news is that you don't need to configure DRBD at all.
330 Ganeti will do it for you for every instance you set up. If
331 you have the DRBD utils installed and the module in your
332 kernel you're fine. Please check that your system is
333 configured to load the module at every boot, and that it
334 passes the following option to the module
335 <computeroutput>minor_count=255</computeroutput>. This will
336 allow you to use up to 128 instances per node (for most clusters
337 <constant>128 </constant> should be enough, though).
340 <formalpara><title>Debian</title>
342 You can just install (build) the DRBD 8.0.x module with the
343 following commands (make sure you are running the Xen
349 apt-get install -t etch-backports drbd8-source drbd8-utils
352 echo drbd minor_count=128 >> /etc/modules
354 modprobe drbd minor_count=128
358 It is also recommended that you comment out the default
359 resources in the <filename>/etc/drbd.conf</filename> file, so
360 that the init script doesn't try to configure any drbd
361 devices. You can do this by prefixing all
362 <literal>resource</literal> lines in the file with the keyword
363 <literal>skip</literal>, like this:
379 <title>Other required software</title>
381 <para>Besides Xen and DRBD, you will need to install the
382 following (on all nodes):</para>
386 <simpara><ulink url="http://sourceware.org/lvm2/">LVM
387 version 2</ulink></simpara>
391 url="http://www.openssl.org/">OpenSSL</ulink></simpara>
395 url="http://www.openssh.com/portable.html">OpenSSH</ulink></simpara>
398 <simpara><ulink url="http://bridge.sourceforge.net/">Bridge
399 utilities</ulink></simpara>
403 url="http://developer.osdl.org/dev/iproute2">iproute2</ulink></simpara>
407 url="ftp://ftp.inr.ac.ru/ip-routing/iputils-current.tar.gz">arping</ulink>
408 (part of iputils package)</simpara>
411 <simpara><ulink url="http://www.python.org">Python 2.4</ulink></simpara>
415 url="http://pyopenssl.sourceforge.net/">Python OpenSSL
416 bindings</ulink></simpara>
420 url="http://www.undefined.org/python/#simplejson">simplejson Python
421 module</ulink></simpara>
425 url="http://pyparsing.wikispaces.com/">pyparsing Python
426 module</ulink></simpara>
431 These programs are supplied as part of most Linux
432 distributions, so usually they can be installed via apt or
433 similar methods. Also many of them will already be installed
434 on a standard machine.
438 <formalpara><title>Debian</title>
440 <para>You can use this command line to install all of them:</para>
444 # apt-get install lvm2 ssh bridge-utils iproute iputils-arping \
445 python2.4 python-pyopenssl openssl python-pyparsing python-simplejson
454 <title>Setting up the environment for Ganeti</title>
457 <title>Configuring the network</title>
459 <para><emphasis role="strong">Mandatory</emphasis> on all nodes.</para>
462 Ganeti relies on Xen running in "bridge mode", which means the
463 instances network interfaces will be attached to a software bridge
464 running in dom0. Xen by default creates such a bridge at startup, but
465 your distribution might have a different way to do things.
469 Beware that the default name Ganeti uses is
470 <hardware>xen-br0</hardware> (which was used in Xen 2.0)
471 while Xen 3.0 uses <hardware>xenbr0</hardware> by
472 default. The default bridge your Ganeti cluster will use for new
473 instances can be specified at cluster initialization time.
476 <formalpara><title>Debian</title>
478 The recommended Debian way to configure the Xen bridge is to
479 edit your <filename>/etc/network/interfaces</filename> file
480 and substitute your normal Ethernet stanza with the
485 iface xen-br0 inet static
486 address <replaceable>YOUR_IP_ADDRESS</replaceable>
487 netmask <replaceable>YOUR_NETMASK</replaceable>
488 network <replaceable>YOUR_NETWORK</replaceable>
489 broadcast <replaceable>YOUR_BROADCAST_ADDRESS</replaceable>
490 gateway <replaceable>YOUR_GATEWAY</replaceable>
499 The following commands need to be executed on the local console
507 To check if the bridge is setup, use <command>ip</command>
508 and <command>brctl show</command>:
513 9: xen-br0: <BROADCAST,MULTICAST,UP,10000> mtu 1500 qdisc noqueue
514 link/ether 00:20:fc:1e:d5:5d brd ff:ff:ff:ff:ff:ff
515 inet 10.1.1.200/24 brd 10.1.1.255 scope global xen-br0
516 inet6 fe80::220:fcff:fe1e:d55d/64 scope link
517 valid_lft forever preferred_lft forever
520 bridge name bridge id STP enabled interfaces
521 xen-br0 8000.0020fc1ed55d no eth0
528 <title>Configuring LVM</title>
531 <para><emphasis role="strong">Mandatory</emphasis> on all nodes.</para>
534 <simpara>The volume group is required to be at least
535 <constant>20GiB</constant>.</simpara>
538 If you haven't configured your LVM volume group at install
539 time you need to do it before trying to initialize the Ganeti
540 cluster. This is done by formatting the devices/partitions you
541 want to use for it and then adding them to the relevant volume
546 vgcreate xenvg /dev/sda3
552 vgcreate xenvg /dev/sdb1 /dev/sdc1
557 If you want to add a device later you can do so with the
558 <citerefentry><refentrytitle>vgextend</refentrytitle>
559 <manvolnum>8</manvolnum></citerefentry> command:
564 vgextend xenvg /dev/sdd1
568 <title>Optional</title>
570 It is recommended to configure LVM not to scan the DRBD
571 devices for physical volumes. This can be accomplished by
572 editing <filename>/etc/lvm/lvm.conf</filename> and adding
573 the <literal>/dev/drbd[0-9]+</literal> regular expression to
574 the <literal>filter</literal> variable, like this:
576 filter = [ "r|/dev/cdrom|", "r|/dev/drbd[0-9]+|" ]
584 <title>Installing Ganeti</title>
586 <para><emphasis role="strong">Mandatory</emphasis> on all nodes.</para>
589 It's now time to install the Ganeti software itself. Download
590 the source from <ulink
591 url="http://code.google.com/p/ganeti/"></ulink>.
595 tar xvzf ganeti-@GANETI_VERSION@.tar.gz
596 cd ganeti-@GANETI_VERSION@
597 ./configure --localstatedir=/var --sysconfdir=/etc
600 mkdir /srv/ganeti/ /srv/ganeti/os /srv/ganeti/export
604 You also need to copy the file
605 <filename>doc/examples/ganeti.initd</filename>
606 from the source archive to
607 <filename>/etc/init.d/ganeti</filename> and register it with
608 your distribution's startup scripts, for example in Debian:
610 <screen>update-rc.d ganeti defaults 20 80</screen>
613 In order to automatically restart failed instances, you need
614 to setup a cron job run the
615 <computeroutput>ganeti-watcher</computeroutput> program. A
616 sample cron file is provided in the source at
617 <filename>doc/examples/ganeti.cron</filename> and you can
618 copy that (eventually altering the path) to
619 <filename>/etc/cron.d/ganeti</filename>
625 <title>Installing the Operating System support packages</title>
627 <para><emphasis role="strong">Mandatory</emphasis> on all nodes.</para>
630 To be able to install instances you need to have an Operating
631 System installation script. An example for Debian Etch is
632 provided on the project web site. Download it from <ulink
633 url="http://code.google.com/p/ganeti/"></ulink> and follow the
634 instructions in the <filename>README</filename> file. Here is
635 the installation procedure (replace <constant>0.2</constant>
636 with the latest version that is compatible with your ganeti
642 tar xvf ganeti-instance-debian-etch-0.4.tar
643 mv ganeti-instance-debian-etch-0.4 debian-etch
647 In order to use this OS definition, you need to have internet
648 access from your nodes and have the <citerefentry>
649 <refentrytitle>debootstrap</refentrytitle>
650 <manvolnum>8</manvolnum></citerefentry>, <citerefentry>
651 <refentrytitle>dump</refentrytitle><manvolnum>8</manvolnum>
652 </citerefentry> and <citerefentry>
653 <refentrytitle>restore</refentrytitle>
654 <manvolnum>8</manvolnum> </citerefentry> commands installed on
658 <title>Debian</title>
660 Use this command on all nodes to install the required
663 <screen>apt-get install debootstrap dump</screen>
668 Alternatively, you can create your own OS definitions. See the
671 <refentrytitle>ganeti-os-interface</refentrytitle>
672 <manvolnum>8</manvolnum>
679 <title>Initializing the cluster</title>
681 <para><emphasis role="strong">Mandatory:</emphasis> only on one
682 node per cluster.</para>
685 <para>The last step is to initialize the cluster. After you've repeated
686 the above process on all of your nodes, choose one as the master, and execute:
690 gnt-cluster init <replaceable>CLUSTERNAME</replaceable>
694 The <replaceable>CLUSTERNAME</replaceable> is a hostname,
695 which must be resolvable (e.g. it must exist in DNS or in
696 <filename>/etc/hosts</filename>) by all the nodes in the
697 cluster. You must choose a name different from any of the
698 nodes names for a multi-node cluster. In general the best
699 choice is to have a unique name for a cluster, even if it
700 consists of only one machine, as you will be able to expand it
701 later without any problems. Please note that the hostname used
702 for this must resolve to an IP address reserved exclusively
707 If the bridge name you are using is not
708 <literal>xen-br0</literal>, use the <option>-b
709 <replaceable>BRIDGENAME</replaceable></option> option to
710 specify the bridge name. In this case, you should also use the
711 <option>--master-netdev
712 <replaceable>BRIDGENAME</replaceable></option> option with the
713 same <replaceable>BRIDGENAME</replaceable> argument.
717 You can use a different name than <literal>xenvg</literal> for
718 the volume group (but note that the name must be identical on
719 all nodes). In this case you need to specify it by passing the
720 <option>-g <replaceable>VGNAME</replaceable></option> option
721 to <computeroutput>gnt-cluster init</computeroutput>.
725 To set up the cluster as an HVM cluster, use the
726 <option>--hypervisor=xen-hvm3.1</option> option to use
727 the Xen 3.1 HVM hypervisor. Note that with the
728 HVM support, you will only be able to create
729 HVM instances in a cluster set to this hypervisor type. Mixed
730 PVM/HVM clusters are not supported by the Ganeti 1.2
731 HVM support. You will also need to create the VNC
732 cluster password file
733 <filename>/etc/ganeti/vnc-cluster-password</filename>
734 which contains one line with the default VNC password for the
739 You can also invoke the command with the
740 <option>--help</option> option in order to see all the
747 <title>Joining the nodes to the cluster</title>
750 <emphasis role="strong">Mandatory:</emphasis> for all the
755 After you have initialized your cluster you need to join the
756 other nodes to it. You can do so by executing the following
757 command on the master node:
760 gnt-node add <replaceable>NODENAME</replaceable>
765 <title>Separate replication network</title>
767 <para><emphasis role="strong">Optional</emphasis></para>
769 Ganeti uses DRBD to mirror the disk of the virtual instances
770 between nodes. To use a dedicated network interface for this
771 (in order to improve performance or to enhance security) you
772 need to configure an additional interface for each node. Use
773 the <option>-s</option> option with
774 <computeroutput>gnt-cluster init</computeroutput> and
775 <computeroutput>gnt-node add</computeroutput> to specify the
776 IP address of this secondary interface to use for each
777 node. Note that if you specified this option at cluster setup
778 time, you must afterwards use it for every node add operation.
783 <title>Testing the setup</title>
786 Execute the <computeroutput>gnt-node list</computeroutput>
787 command to see all nodes in the cluster:
790 Node DTotal DFree MTotal MNode MFree Pinst Sinst
791 node1.example.com 197404 197404 2047 1896 125 0 0
797 <title>Setting up and managing virtual instances</title>
799 <title>Setting up virtual instances</title>
801 This step shows how to setup a virtual instance with either
802 non-mirrored disks (<computeroutput>plain</computeroutput>) or
803 with network mirrored disks
804 (<computeroutput>drbd</computeroutput>). All
805 commands need to be executed on the Ganeti master node (the
806 one on which <computeroutput>gnt-cluster init</computeroutput>
807 was run). Verify that the OS scripts are present on all
808 cluster nodes with <computeroutput>gnt-os
809 list</computeroutput>.
812 To create a virtual instance, you need a hostname which is
813 resolvable (DNS or <filename>/etc/hosts</filename> on all
814 nodes). The following command will create a non-mirrored
818 gnt-instance add --node=node1 -o debian-etch -t plain inst1.example.com
819 * creating instance disks...
820 adding instance inst1.example.com to cluster config
821 Waiting for instance inst1.example.com to sync disks.
822 Instance inst1.example.com's disks are in sync.
823 creating os for instance inst1.example.com on node node1.example.com
824 * running the instance OS create scripts...
828 The above instance will have no network interface enabled.
829 You can access it over the virtual console with
830 <computeroutput>gnt-instance console
831 <literal>inst1</literal></computeroutput>. There is no
832 password for root. As this is a Debian instance, you can
833 modify the <filename>/etc/network/interfaces</filename> file
834 to setup the network interface (<literal>eth0</literal> is the
835 name of the interface provided to the instance).
839 To create a network mirrored instance, change the argument to
840 the <option>-t</option> option from <literal>plain</literal>
841 to <literal>drbd</literal> and specify the node on
842 which the mirror should reside with the second value of the
843 <option>--node</option> option, like this:
847 # gnt-instance add -t drbd -n node1:node2 -o debian-etch instance2
848 * creating instance disks...
849 adding instance instance2 to cluster config
850 Waiting for instance instance1 to sync disks.
851 - device sdb: 3.50% done, 304 estimated seconds remaining
852 - device sdb: 21.70% done, 270 estimated seconds remaining
853 - device sdb: 39.80% done, 247 estimated seconds remaining
854 - device sdb: 58.10% done, 121 estimated seconds remaining
855 - device sdb: 76.30% done, 72 estimated seconds remaining
856 - device sdb: 94.80% done, 18 estimated seconds remaining
857 Instance instance2's disks are in sync.
858 creating os for instance instance2 on node node1.example.com
859 * running the instance OS create scripts...
860 * starting instance...
866 <title>Managing virtual instances</title>
868 All commands need to be executed on the Ganeti master node
872 To access the console of an instance, use
873 <computeroutput>gnt-instance console
874 <replaceable>INSTANCENAME</replaceable></computeroutput>.
878 To shutdown an instance, use <computeroutput>gnt-instance
880 <replaceable>INSTANCENAME</replaceable></computeroutput>. To
881 startup an instance, use <computeroutput>gnt-instance startup
882 <replaceable>INSTANCENAME</replaceable></computeroutput>.
886 To failover an instance to its secondary node (only possible
887 with <literal>drbd</literal> disk templates), use
888 <computeroutput>gnt-instance failover
889 <replaceable>INSTANCENAME</replaceable></computeroutput>.
893 For more instance and cluster administration details, see the
894 <emphasis>Ganeti administrator's guide</emphasis>.