File:  [NetBSD Developer Wiki] / wikisrc / ports / xen / howto.mdwn
Revision 1.46: download - view: text, annotated - select for diffs
Fri Dec 26 16:43:51 2014 UTC (4 years, 11 months ago) by gdt
Branches: MAIN
CVS tags: HEAD
explain networking

add a domU example config

    1: Introduction
    2: ============
    3: 
    4: [![[Xen
    5: screenshot]](http://www.netbsd.org/gallery/in-Action/hubertf-xens.png)](../../gallery/in-Action/hubertf-xen.png)
    6: 
    7: Xen is a virtual machine monitor or hypervisor for x86 hardware
    8: (i686-class or higher), which supports running multiple guest
    9: operating systems on a single physical machine.  With Xen, one uses
   10: the Xen kernel to control the CPU, memory and console, a dom0
   11: operating system which mediates access to other hardware (e.g., disks,
   12: network, USB), and one or more domU operating systems which operate in
   13: an unprivileged virtualized environment.  IO requests from the domU
   14: systems are forwarded by the hypervisor (Xen) to the dom0 to be
   15: fulfilled.
   16: 
   17: Xen supports two styles of guests.  The original is Para-Virtualized
   18: (PV) which means that the guest OS does not attempt to access hardware
   19: directly, but instead makes hypercalls to the hypervisor.  This is
   20: analogous to a user-space program making system calls.  (The dom0
   21: operating system uses PV calls for some functions, such as updating
   22: memory mapping page tables, but has direct hardware access for disk
   23: and network.)   PV guests must be specifically coded for Xen.
   24: 
   25: The more recent style is HVM, which means that the guest does not have
   26: code for Xen and need not be aware that it is running under Xen.
   27: Attempts to access hardware registers are trapped and emulated.  This
   28: style is less efficient but can run unmodified guests.
   29: 
   30: Generally any amd64 machine will work with Xen and PV guests.  In
   31: theory i386 computers without amd64 support can be used for Xen <=
   32: 4.2, but we have no recent reports of this working (this is a hint).
   33: For HVM guests, the VT or VMX cpu feature (Intel) or SVM/HVM/VT
   34: (amd64) is needed; "cpuctl identify 0" will show this.  TODO: Clean up
   35: and check the above features.
   36: 
   37: At boot, the dom0 kernel is loaded as a module with Xen as the kernel.
   38: The dom0 can start one or more domUs.  (Booting is explained in detail
   39: in the dom0 section.)
   40: 
   41: NetBSD supports Xen in that it can serve as dom0, be used as a domU,
   42: and that Xen kernels and tools are available in pkgsrc.  This HOWTO
   43: attempts to address both the case of running a NetBSD dom0 on hardware
   44: and running domUs under it (NetBSD and other), and also running NetBSD
   45: as a domU in a VPS.
   46: 
   47: Some versions of Xen support "PCI passthrough", which means that
   48: specific PCI devices can be made available to a specific domU instead
   49: of the dom0.  This can be useful to let a domU run X11, or access some
   50: network interface or other peripheral.
   51: 
   52: Prerequisites
   53: -------------
   54: 
   55: Installing NetBSD/Xen is not extremely difficult, but it is more
   56: complex than a normal installation of NetBSD.
   57: In general, this HOWTO is occasionally overly restrictive about how
   58: things must be done, guiding the reader to stay on the established
   59: path when there are no known good reasons to stray.
   60: 
   61: This HOWTO presumes a basic familiarity with the Xen system
   62: architecture.  This HOWTO presumes familiarity with installing NetBSD
   63: on i386/amd64 hardware and installing software from pkgsrc.
   64: See also the [Xen website](http://www.xenproject.org/).
   65: 
   66: History
   67: -------
   68: 
   69: NetBSD used to support Xen2; this has been removed.
   70: 
   71: Before NetBSD's native bootloader could support Xen, the use of
   72: grub was recommended.  If necessary, see the
   73: [old grub information](/ports/xen/howto-grub/).
   74: 
   75: Versions of Xen and NetBSD
   76: ==========================
   77: 
   78: Most of the installation concepts and instructions are independent
   79: of Xen version and NetBSD version.  This section gives advice on
   80: which version to choose.  Versions not in pkgsrc and older unsupported
   81: versions of NetBSD are intentionally ignored.
   82: 
   83: Xen
   84: ---
   85: 
   86: In NetBSD, xen is provided in pkgsrc, via matching pairs of packages
   87: xenkernel and xentools.  We will refer only to the kernel versions,
   88: but note that both packages must be installed together and must have
   89: matching versions.
   90: 
   91: xenkernel3 and xenkernel33 provide Xen 3.1 and 3.3.  These no longer
   92: receive security patches and should not be used.  Xen 3.1 supports PCI
   93: passthrough.  Xen 3.1 supports non-PAE on i386.
   94: 
   95: xenkernel41 provides Xen 4.1.  This is no longer maintained by Xen,
   96: but as of 2014-12 receives backported security patches.  It is a
   97: reasonable although trailing-edge choice.
   98: 
   99: xenkernel42 provides Xen 4.2.  This is maintained by Xen, but old as
  100: of 2014-12.
  101: 
  102: Ideally newer versions of Xen will be added to pkgsrc.
  103: 
  104: Note that NetBSD support is called XEN3.  It works with 3.1 through
  105: 4.2 because the hypercall interface has been stable.
  106: 
  107: Xen command program
  108: -------------------
  109: 
  110: Early Xen used a program called "xm" to manipulate the system from the
  111: dom0.  Starting in 4.1, a replacement program with similar behavior
  112: called "xl" is provided.  In 4.2 and later, "xl" is preferred.  4.4 is
  113: the last version that has "xm".
  114: 
  115: NetBSD
  116: ------
  117: 
  118: The netbsd-5, netbsd-6, netbsd-7, and -current branches are all
  119: reasonable choices, with more or less the same considerations for
  120: non-Xen use.  Therefore, netbsd-6 is recommended as the stable version
  121: of the most recent release for production use.  For those wanting to
  122: learn Xen or without production stability concerns, netbsd-7 is likely
  123: most appropriate.
  124: 
  125: As of NetBSD 6, a NetBSD domU will support multiple vcpus.  There is
  126: no SMP support for NetBSD as dom0.  (The dom0 itself doesn't really
  127: need SMP; the lack of support is really a problem when using a dom0 as
  128: a normal computer.)
  129: 
  130: Architecture
  131: ------------
  132: 
  133: Xen itself can run on i386 or amd64 machines.  (Practically, almost
  134: any computer where one would want to run Xen supports amd64.)  If
  135: using an i386 NetBSD kernel for the dom0, PAE is required (PAE
  136: versions are built by default).  While i386 dom0 works fine, amd64 is
  137: recommended as more normal.
  138: 
  139: Xen 4.2 is the last version to support i386 as a host.  TODO: Clarify
  140: if this is about the CPU having to be amd64, or about the dom0 kernel
  141: having to be amd64.
  142: 
  143: One can then run i386 domUs and amd64 domUs, in any combination.  If
  144: running an i386 NetBSD kernel as a domU, the PAE version is required.
  145: (Note that emacs (at least) fails if run on i386 with PAE when built
  146: without, and vice versa, presumably due to bugs in the undump code.)
  147: 
  148: Recommendation
  149: --------------
  150: 
  151: Therefore, this HOWTO recommends running xenkernel42 (and xentools42),
  152: xl, the NetBSD 6 stable branch, and to use an amd64 kernel as the
  153: dom0.  Either the i386 or amd64 of NetBSD may be used as domUs.
  154: 
  155: Build problems
  156: --------------
  157: 
  158: Ideally, all versions of Xen in pkgsrc would build on all versions of
  159: NetBSD on both i386 and amd64.  However, that isn't the case.  Besides
  160: aging code and aging compilers, qemu (included in xentools for HVM
  161: support) is difficult to build.  The following are known to fail:
  162: 
  163:         xenkernel3 netbsd-6 i386
  164:         xentools42 netbsd-6 i386 
  165: 
  166: The following are known to work:
  167: 
  168:         xenkernel41 netbsd-5 amd64
  169:         xentools41 netbsd-5 amd64
  170:         xenkernel41 netbsd-6 i386
  171:         xentools41 netbsd-6 i386
  172: 
  173: NetBSD as a dom0
  174: ================
  175: 
  176: NetBSD can be used as a dom0 and works very well.  The following
  177: sections address installation, updating NetBSD, and updating Xen.
  178: Note that it doesn't make sense to talk about installing a dom0 OS
  179: without also installing Xen itself.  We first address installing
  180: NetBSD, which is not yet a dom0, and then adding Xen, pivoting the
  181: NetBSD install to a dom0 install by just changing the kernel and boot
  182: configuration.
  183: 
  184: For experimenting with Xen, a machine with as little as 1G of RAM and
  185: 100G of disk can work.  For running many domUs in productions, far
  186: more will be needed.
  187: 
  188: Styles of dom0 operation
  189: ------------------------
  190: 
  191: There are two basic ways to use Xen.  The traditional method is for
  192: the dom0 to do absolutely nothing other than providing support to some
  193: number of domUs.  Such a system was probably installed for the sole
  194: purpose of hosting domUs, and sits in a server room on a UPS.
  195: 
  196: The other way is to put Xen under a normal-usage computer, so that the
  197: dom0 is what the computer would have been without Xen, perhaps a
  198: desktop or laptop.  Then, one can run domUs at will.  Purists will
  199: deride this as less secure than the previous approach, and for a
  200: computer whose purpose is to run domUs, they are right.  But Xen and a
  201: dom0 (without domUs) is not meaingfully less secure than the same
  202: things running without Xen.  One can boot Xen or boot regular NetBSD
  203: alternately with little problems, simply refraining from starting the
  204: Xen daemons when not running Xen.
  205: 
  206: Note that NetBSD as dom0 does not support multiple CPUs.  This will
  207: limit the performance of the Xen/dom0 workstation approach.
  208: 
  209: Installation of NetBSD
  210: ----------------------
  211: 
  212: First,
  213: [install NetBSD/amd64](/guide/inst/)
  214: just as you would if you were not using Xen.
  215: However, the partitioning approach is very important.
  216: 
  217: If you want to use RAIDframe for the dom0, there are no special issues
  218: for Xen.  Typically one provides RAID storage for the dom0, and the
  219: domU systems are unaware of RAID.  The 2nd-stage loader bootxx_* skips
  220: over a RAID1 header to find /boot from a filesystem within a RAID
  221: partition; this is no different when booting Xen.
  222: 
  223: There are 4 styles of providing backing storage for the virtual disks
  224: used by domUs: raw partitions, LVM, file-backed vnd(4), and SAN,
  225: 
  226: With raw partitions, one has a disklabel (or gpt) partition sized for
  227: each virtual disk to be used by the domU.  (If you are able to predict
  228: how domU usage will evolve, please add an explanation to the HOWTO.
  229: Seriously, needs tend to change over time.)
  230: 
  231: One can use [lvm(8)](/guide/lvm/) to create logical devices to use
  232: for domU disks.  This is almost as efficient as raw disk partitions
  233: and more flexible.  Hence raw disk partitions should typically not
  234: be used.
  235: 
  236: One can use files in the dom0 filesystem, typically created by dd'ing
  237: /dev/zero to create a specific size.  This is somewhat less efficient,
  238: but very convenient, as one can cp the files for backup, or move them
  239: between dom0 hosts.
  240: 
  241: Finally, in theory one can place the files backing the domU disks in a
  242: SAN.  (This is an invitation for someone who has done this to add a
  243: HOWTO page.)
  244: 
  245: Installation of Xen
  246: -------------------
  247: 
  248: In the dom0, install sysutils/xenkernel42 and sysutils/xentools42 from
  249: pkgsrc (or another matching pair).
  250: See [the pkgsrc
  251: documentation](http://www.NetBSD.org/docs/pkgsrc/) for help with pkgsrc.
  252: 
  253: For Xen 3.1, support for HVM guests is in sysutils/xentool3-hvm.  More
  254: recent versions have HVM support integrated in the main xentools
  255: package.  It is entirely reasonable to run only PV guests.
  256: 
  257: Next you need to install the selected Xen kernel itself, which is
  258: installed by pkgsrc as "/usr/pkg/xen*-kernel/xen.gz".  Copy it to /.
  259: For debugging, one may copy xen-debug.gz; this is conceptually similar
  260: to DIAGNOSTIC and DEBUG in NetBSD.  xen-debug.gz is basically only
  261: useful with a serial console.  Then, place a NetBSD XEN3_DOM0 kernel
  262: in /, copied from releasedir/amd64/binary/kernel/netbsd-XEN3_DOM0.gz
  263: of a NetBSD build.  Both xen and NetBSD may be left compressed.  (If
  264: using i386, use releasedir/i386/binary/kernel/netbsd-XEN3PAE_DOM0.gz.)
  265: 
  266: In a dom0 kernel, kernfs is mandatory for xend to comunicate with the
  267: kernel, so ensure that /kern is in fstab.
  268: 
  269: Because you already installed NetBSD, you have a working boot setup
  270: with an MBR bootblock, either bootxx_ffsv1 or bootxx_ffsv2 at the
  271: beginning of your root filesystem, /boot present, and likely
  272: /boot.cfg.  (If not, fix before continuing!)
  273: 
  274: See boot.cfg(5) for an example.  The basic line is
  275: 
  276:         menu=Xen:load /netbsd-XEN3_DOM0.gz console=pc;multiboot /xen.gz dom0_mem=256M
  277: 
  278: which specifies that the dom0 should have 256M, leaving the rest to be
  279: allocated for domUs.  In an attempt to add performance, one can also
  280: add
  281: 
  282:         dom0_max_vcpus=1 dom0_vcpus_pin
  283: 
  284: to force only one vcpu to be provided (since NetBSD dom0 can't use
  285: more) and to pin that vcpu to a physical cpu.  TODO: benchmark this.
  286: 
  287: As with non-Xen systems, you should have a line to boot /netbsd (a
  288: kernel that works without Xen) and fallback versions of the non-Xen
  289: kernel, Xen, and the dom0 kernel.
  290: 
  291: The [HowTo on Installing into
  292: RAID-1](http://mail-index.NetBSD.org/port-xen/2006/03/01/0010.html)
  293: explains how to set up booting a dom0 with Xen using grub with
  294: NetBSD's RAIDframe.  (This is obsolete with the use of NetBSD's native
  295: boot.)
  296: 
  297: Configuring Xen
  298: ---------------
  299: 
  300: Now, you have a system that will boot Xen and the dom0 kernel, and
  301: just run the dom0 kernel.  There will be no domUs, and none can be
  302: started because you still have to configure the dom0 tools.  The
  303: daemons which should be run vary with Xen version and with whether one
  304: is using xm or xl.  Note that xend is for supporting "xm", and should
  305: only be used if you plan on using "xm".  Do NOT enable xend if you
  306: plan on using "xl" as it will cause problems.
  307: 
  308: The installation of NetBSD should already have created devices for xen
  309: (xencons, xenevt), but if they are not present, create them:
  310: 
  311:         cd /dev && sh MAKEDEV xen
  312: 
  313: TODO: Give 3.1 advice (or remove it from pkgsrc).
  314: 
  315: For 3.3 (and thus xm), add to rc.conf (but note that you should have
  316: installed 4.1 or 4.2):
  317: 
  318:         xend=YES
  319:         xenbackendd=YES
  320: 
  321: For 4.1 (and thus xm; xl is believed not to work well), add to rc.conf:
  322: 
  323:         xend=YES
  324:         xencommons=YES
  325: 
  326: TODO: Explain why if xm is preferred on 4.1, rc.d/xendomains has xl.
  327: Or fix the package.
  328: 
  329: For 4.2 with xm, add to rc.conf
  330: 
  331:         xend=YES
  332:         xencommons=YES
  333: 
  334: For 4.2 with xl (preferred), add to rc.conf:
  335: 
  336:         TODO: explain if there is a xend replacement
  337:         xencommons=YES
  338: 
  339: TODO: Recommend for/against xen-watchdog.
  340: 
  341: After you have configured the daemons and either started them or
  342: rebooted, run the following (or use xl) to inspect Xen's boot
  343: messages, available resources, and running domains:
  344: 
  345:         # xm dmesg
  346: 	[xen's boot info]
  347:         # xm info
  348: 	[available memory, etc.]
  349:         # xm list
  350:         Name              Id  Mem(MB)  CPU  State  Time(s)  Console
  351:         Domain-0           0       64    0  r----     58.1
  352: 
  353: anita (for testing NetBSD)
  354: --------------------------
  355: 
  356: With the setup so far, one should be able to run anita (see
  357: pkgsrc/sysutils/py-anita) to test NetBSD releases, by doing (as root,
  358: because anita must create a domU):
  359: 
  360:         anita --vmm=xm test file:///usr/obj/i386/
  361: 
  362: Alternatively, one can use --vmm=xl to use xl-based domU creation instead.
  363: TODO: check this.
  364: 
  365: Xen-specific NetBSD issues
  366: --------------------------
  367: 
  368: There are (at least) two additional things different about NetBSD as a
  369: dom0 kernel compared to hardware.
  370: 
  371: One is that modules are not usable in DOM0 kernels, so one must
  372: compile in what's needed.  It's not really that modules cannot work,
  373: but that modules must be built for XEN3_DOM0 because some of the
  374: defines change and the normal module builds don't do this.  Basically,
  375: enabling Xen changes the kernel ABI, and the module build system
  376: doesn't cope with this.
  377: 
  378: The other difference is that XEN3_DOM0 does not have exactly the same
  379: options as GENERIC.  While it is debatable whether or not this is a
  380: bug, users should be aware of this and can simply add missing config
  381: items if desired.
  382: 
  383: Updating NetBSD in a dom0
  384: -------------------------
  385: 
  386: This is just like updating NetBSD on bare hardware, assuming the new
  387: version supports the version of Xen you are running.  Generally, one
  388: replaces the kernel and reboots, and then overlays userland binaries
  389: and adjusts /etc.
  390: 
  391: Note that one must update both the non-Xen kernel typically used for
  392: rescue purposes and the DOM0 kernel used with Xen.
  393: 
  394: To convert from grub to /boot, install an mbr bootblock with fdisk,
  395: bootxx_ with installboot, /boot and /boot.cfg.  This really should be
  396: no different than completely reinstalling boot blocks on a non-Xen
  397: system.
  398: 
  399: Updating Xen versions
  400: ---------------------
  401: 
  402: Updating Xen is conceptually not difficult, but can run into all the
  403: issues found when installing Xen.  Assuming migration from 4.1 to 4.2,
  404: remove the xenkernel41 and xentools41 packages and install the
  405: xenkernel42 and xentools42 packages.  Copy the 4.2 xen.gz to /.
  406: 
  407: Ensure that the contents of /etc/rc.d/xen* are correct.  Enable the
  408: correct set of daemons.  Ensure that the domU config files are valid
  409: for the new version.
  410: 
  411: 
  412: Unprivileged domains (domU)
  413: ===========================
  414: 
  415: This section describes general concepts about domUs.  It does not
  416: address specific domU operating systems or how to install them.  The
  417: config files for domUs are typically in /usr/pkg/etc/xen, and are
  418: typically named so that the file anme, domU name and the domU's host
  419: name match.
  420: 
  421: The domU is provided with cpu and memory by Xen, configured by the
  422: dom0.  The domU is provided with disk and network by the dom0,
  423: mediated by Xen, and configured in the dom0.
  424: 
  425: Entropy in domUs can be an issue; physical disks and network are on
  426: the dom0.  NetBSD's /dev/random system works, but is often challenged.
  427: 
  428: CPU and memory
  429: --------------
  430: 
  431: A domain is provided with some number of vcpus, less than the
  432: number of cpus seen by the hypervisor.  For a dom0, this is controlled
  433: by the boot argument "dom0_max_vcpus=1".  For a domU, it is controlled
  434: from the config file.
  435: 
  436: A domain is provided with memory, In the straightforward case, the sum
  437: of the the memory allocated to the dom0 and all domUs must be less
  438: than the available memory.
  439: 
  440: Xen also provides a "balloon" driver, which can be used to let domains
  441: use more memory temporarily.  TODO: Explain better, and explain how
  442: well it works with NetBSD.
  443: 
  444: Virtual disks
  445: -------------
  446: 
  447: With the file/vnd style, typically one creates a directory,
  448: e.g. /u0/xen, on a disk large enough to hold virtual disks for all
  449: domUs.  Then, for each domU disk, one writes zeros to a file that then
  450: serves to hold the virtual disk's bits; a suggested name is foo-xbd0
  451: for the first virtual disk for the domU called foo.  Writing zeros to
  452: the file serves two purposes.  One is that preallocating the contents
  453: improves performance.  The other is that vnd on sparse files has
  454: failed to work.  TODO: give working/notworking NetBSD versions for
  455: sparse vnd.  Note that the use of file/vnd for Xen is not really
  456: different than creating a file-backed virtual disk for some other
  457: purpose, except that xentools handles the vnconfig commands.  To
  458: create an empty 4G virtual disk, simply do
  459: 
  460:         dd if=/dev/zero of=foo-xbd0 bs=1m count=4096
  461: 
  462: With the lvm style, one creates logical devices.  They are then used
  463: similarly to vnds.
  464: 
  465: Virtual Networking
  466: ------------------
  467: 
  468: Xen provides virtual ethernets, each of which connects the dom0 and a
  469: domU.  For each virtual network, there is an interface "xvifN.M" in
  470: the dom0, and in domU index N, a matching interface xennetM (NetBSD
  471: name).  The interfaces behave as if there is an Ethernet with two
  472: adaptors connected.  From this primitive, one can construct various
  473: configurations.  We focus on two common and useful cases for which
  474: there are existing scripts: bridging and NAT.
  475: 
  476: With bridging, the domU perceives itself to be on the same network as
  477: the dom0.  For server virtualization, this is usually best.  Bridging
  478: is accomplished by creating a bridge(4) device and adding the dom0's
  479: physical interface and the various xvifN.0 interfaces to the bridge.
  480: One specifies "bridge=bridge0" in the domU config file.  The bridge
  481: must be set up already in the dom0; an example /etc/ifconfig.bridge0
  482: is:
  483: 
  484:         create
  485:         up
  486:         !brconfig bridge0 add wm0
  487: 
  488: With NAT, the domU perceives itself to be behind a NAT running on the
  489: dom0.  This is often appropriate when running Xen on a workstation.
  490: 
  491: Sizing domains
  492: --------------
  493: 
  494: Modern x86 hardware has vast amounts of resources.  However, many
  495: virtual servers can function just fine on far less.  A system with
  496: 256M of RAM and a 4G disk can be a reasonable choice.  Note that it is
  497: far easier to adjust virtual resources than physical ones.  For
  498: memory, it's just a config file edit and a reboot.  For disk, one can
  499: create a new file and vnconfig it (or lvm), and then dump/restore,
  500: just like updating physical disks, but without having to be there and
  501: without those pesky connectors.
  502: 
  503: domU kernels
  504: ------------
  505: 
  506: On a physical computer, the BIOS reads sector 0, and a chain of boot
  507: loaders finds and loads a kernel.  Normally this comes from the root
  508: filesystem.  With Xen domUs, the process is totally different.  The
  509: normal path is for the domU kernel to be a file in the dom0's
  510: filesystem.  At the request of the dom0, Xen loads that kernel into a
  511: new domU instance and starts execution.  While domU kernels can be
  512: anyplace, reasonable places to store domU kernels on the dom0 are in /
  513: (so they are near the dom0 kernel), in /usr/pkg/etc/xen (near the
  514: config files), or in /u0/xen (where the vdisks are).
  515: 
  516: See the VPS section near the end for discussion of alternate ways to
  517: obtain domU kernels.
  518: 
  519: Config files
  520: ------------
  521: 
  522: The following is an example domain configuration file, lightly
  523: sanitized from a known working on Xen 4.1 (NetBSD 5 amd64 dom0 and
  524: NetBSD 6 i386 domU):
  525: 
  526:         # -*- mode: python; -*-
  527: 
  528:         kernel = "/netbsd-XEN3PAE_DOMU-i386-foo.gz"
  529: 
  530:         memory = 1024
  531: 
  532:         name = "foo"
  533: 
  534:         #cpu = -1
  535: 
  536:         vif = [ 'mac=aa:00:00:d1:00:09,bridge=bridge0' ]
  537: 
  538:         disk = [ 'file:/n0/xen/foo-wd0,0x1,w',
  539:                  'file:/n0/xen/foo-wd1,0x2,w' ]
  540: 
  541:         root = "xbd0"
  542: 
  543:         autorestart = True
  544: 
  545: The kernel has the host/domU name in it, so that on the dom0 one can
  546: update the various domUs independently.  The vif line causes an
  547: interface to be provided, with a specific mac address (do not reuse
  548: MAC addresses!), in bridge mode.  Two disks are provided, and they are
  549: writable.
  550: 
  551: TODO: explain if the root line is really necessary.
  552: TODO: explain or remove autorestart.
  553: 
  554: TODO: Add an example with lvm
  555: 
  556: TODO: explain, someplace the mess with 3 arguments for disks and how to cope (0x1).
  557: 
  558: Starting domains
  559: ----------------
  560: 
  561: TODO: Explain "xm start" and "xl start".  Explain rc.d/xendomains.
  562: 
  563: TODO: Explain why 4.1 rc.d/xendomains has xl, when one should use xm
  564: on 4.1.
  565: 
  566: Creating specific unprivileged domains (domU)
  567: =============================================
  568: 
  569: Creating domUs is almost entirely independent of operating system.  We
  570: first explain NetBSD, and then differences for Linux and Solaris.
  571: Note that you must have already completed the dom0 setup so that "xm
  572: list" (or "xl list") works.
  573: 
  574: Creating an unprivileged NetBSD domain (domU)
  575: ---------------------------------------------
  576: 
  577: 'xm create' allows you to create a new domain. It uses a config file in
  578: PKG\_SYSCONFDIR for its parameters. By default, this file will be in
  579: `/usr/pkg/etc/xen/`. On creation, a kernel has to be specified, which
  580: will be executed in the new domain (this kernel is in the *domain0* file
  581: system, not on the new domain virtual disk; but please note, you should
  582: install the same kernel into *domainU* as `/netbsd` in order to make
  583: your system tools, like savecore(8), work). A suitable kernel is
  584: provided as part of the i386 and amd64 binary sets: XEN3\_DOMU.
  585: 
  586: Here is an /usr/pkg/etc/xen/nbsd example config file:
  587: 
  588:     #  -*- mode: python; -*-
  589:     #============================================================================
  590:     # Python defaults setup for 'xm create'.
  591:     # Edit this file to reflect the configuration of your system.
  592:     #============================================================================
  593: 
  594:     #----------------------------------------------------------------------------
  595:     # Kernel image file. This kernel will be loaded in the new domain.
  596:     kernel = "/home/bouyer/netbsd-XEN3_DOMU"
  597:     #kernel = "/home/bouyer/netbsd-INSTALL_XEN3_DOMU"
  598: 
  599:     # Memory allocation (in megabytes) for the new domain.
  600:     memory = 128
  601: 
  602:     # A handy name for your new domain. This will appear in 'xm list',
  603:     # and you can use this as parameters for xm in place of the domain
  604:     # number. All domains must have different names.
  605:     #
  606:     name = "nbsd"
  607: 
  608:     # The number of virtual CPUs this domain has.
  609:     #
  610:     vcpus = 1
  611: 
  612:     #----------------------------------------------------------------------------
  613:     # Define network interfaces for the new domain.
  614: 
  615:     # Number of network interfaces (must be at least 1). Default is 1.
  616:     nics = 1
  617: 
  618:     # Define MAC and/or bridge for the network interfaces.
  619:     #
  620:     # The MAC address specified in ``mac'' is the one used for the interface
  621:     # in the new domain. The interface in domain0 will use this address XOR'd
  622:     # with 00:00:00:01:00:00 (i.e. aa:00:00:51:02:f0 in our example). Random
  623:     # MACs are assigned if not given.
  624:     #
  625:     # ``bridge'' is a required parameter, which will be passed to the
  626:     # vif-script called by xend(8) when a new domain is created to configure
  627:     # the new xvif interface in domain0.
  628:     #
  629:     # In this example, the xvif is added to bridge0, which should have been
  630:     # set up prior to the new domain being created -- either in the
  631:     # ``network'' script or using a /etc/ifconfig.bridge0 file.
  632:     #
  633:     vif = [ 'mac=aa:00:00:50:02:f0, bridge=bridge0' ]
  634: 
  635:     #----------------------------------------------------------------------------
  636:     # Define the disk devices you want the domain to have access to, and
  637:     # what you want them accessible as.
  638:     #
  639:     # Each disk entry is of the form:
  640:     #
  641:     #   phy:DEV,VDEV,MODE
  642:     #
  643:     # where DEV is the device, VDEV is the device name the domain will see,
  644:     # and MODE is r for read-only, w for read-write.  You can also create
  645:     # file-backed domains using disk entries of the form:
  646:     #
  647:     #   file:PATH,VDEV,MODE
  648:     #
  649:     # where PATH is the path to the file used as the virtual disk, and VDEV
  650:     # and MODE have the same meaning as for ``phy'' devices.
  651:     #
  652:     # VDEV doesn't really matter for a NetBSD guest OS (it's just used as an index),
  653:     # but it does for Linux.
  654:     # Worse, the device has to exist in /dev/ of domain0, because xm will
  655:     # try to stat() it. This means that in order to load a Linux guest OS
  656:     # from a NetBSD domain0, you'll have to create /dev/hda1, /dev/hda2, ...
  657:     # on domain0, with the major/minor from Linux :(
  658:     # Alternatively it's possible to specify the device number in hex,
  659:     # e.g. 0x301 for /dev/hda1, 0x302 for /dev/hda2, etc ...
  660: 
  661:     disk = [ 'phy:/dev/wd0e,0x1,w' ]
  662:     #disk = [ 'file:/var/xen/nbsd-disk,0x01,w' ]
  663:     #disk = [ 'file:/var/xen/nbsd-disk,0x301,w' ]
  664: 
  665:     #----------------------------------------------------------------------------
  666:     # Set the kernel command line for the new domain.
  667: 
  668:     # Set root device. This one does matter for NetBSD
  669:     root = "xbd0"
  670:     # extra parameters passed to the kernel
  671:     # this is where you can set boot flags like -s, -a, etc ...
  672:     #extra = ""
  673: 
  674:     #----------------------------------------------------------------------------
  675:     # Set according to whether you want the domain restarted when it exits.
  676:     # The default is False.
  677:     #autorestart = True
  678: 
  679:     # end of nbsd config file ====================================================
  680: 
  681: When a new domain is created, xen calls the
  682: `/usr/pkg/etc/xen/vif-bridge` script for each virtual network interface
  683: created in *domain0*. This can be used to automatically configure the
  684: xvif?.? interfaces in *domain0*. In our example, these will be bridged
  685: with the bridge0 device in *domain0*, but the bridge has to exist first.
  686: To do this, create the file `/etc/ifconfig.bridge0` and make it look
  687: like this:
  688: 
  689:     create
  690:     !brconfig $int add ex0 up
  691: 
  692: (replace `ex0` with the name of your physical interface). Then bridge0
  693: will be created on boot. See the bridge(4) man page for details.
  694: 
  695: So, here is a suitable `/usr/pkg/etc/xen/vif-bridge` for xvif?.? (a
  696: working vif-bridge is also provided with xentools20) configuring:
  697: 
  698:     #!/bin/sh
  699:     #============================================================================
  700:     # $NetBSD: howto.mdwn,v 1.46 2014/12/26 16:43:51 gdt Exp $
  701:     #
  702:     # /usr/pkg/etc/xen/vif-bridge
  703:     #
  704:     # Script for configuring a vif in bridged mode with a dom0 interface.
  705:     # The xend(8) daemon calls a vif script when bringing a vif up or down.
  706:     # The script name to use is defined in /usr/pkg/etc/xen/xend-config.sxp
  707:     # in the ``vif-script'' field.
  708:     #
  709:     # Usage: vif-bridge up|down [var=value ...]
  710:     #
  711:     # Actions:
  712:     #    up     Adds the vif interface to the bridge.
  713:     #    down   Removes the vif interface from the bridge.
  714:     #
  715:     # Variables:
  716:     #    domain name of the domain the interface is on (required).
  717:     #    vifq   vif interface name (required).
  718:     #    mac    vif MAC address (required).
  719:     #    bridge bridge to add the vif to (required).
  720:     #
  721:     # Example invocation:
  722:     #
  723:     # vif-bridge up domain=VM1 vif=xvif1.0 mac="ee:14:01:d0:ec:af" bridge=bridge0
  724:     #
  725:     #============================================================================
  726: 
  727:     # Exit if anything goes wrong
  728:     set -e
  729: 
  730:     echo "vif-bridge $*"
  731: 
  732:     # Operation name.
  733:     OP=$1; shift
  734: 
  735:     # Pull variables in args into environment
  736:     for arg ; do export "${arg}" ; done
  737: 
  738:     # Required parameters. Fail if not set.
  739:     domain=${domain:?}
  740:     vif=${vif:?}
  741:     mac=${mac:?}
  742:     bridge=${bridge:?}
  743: 
  744:     # Optional parameters. Set defaults.
  745:     ip=${ip:-''}   # default to null (do nothing)
  746: 
  747:     # Are we going up or down?
  748:     case $OP in
  749:     up) brcmd='add' ;;
  750:     down)   brcmd='delete' ;;
  751:     *)
  752:         echo 'Invalid command: ' $OP
  753:         echo 'Valid commands are: up, down'
  754:         exit 1
  755:         ;;
  756:     esac
  757: 
  758:     # Don't do anything if the bridge is "null".
  759:     if [ "${bridge}" = "null" ] ; then
  760:         exit
  761:     fi
  762: 
  763:     # Don't do anything if the bridge doesn't exist.
  764:     if ! ifconfig -l | grep "${bridge}" >/dev/null; then
  765:         exit
  766:     fi
  767: 
  768:     # Add/remove vif to/from bridge.
  769:     ifconfig x${vif} $OP
  770:     brconfig ${bridge} ${brcmd} x${vif}
  771: 
  772: Now, running
  773: 
  774:     xm create -c /usr/pkg/etc/xen/nbsd
  775: 
  776: should create a domain and load a NetBSD kernel in it. (Note: `-c`
  777: causes xm to connect to the domain's console once created.) The kernel
  778: will try to find its root file system on xbd0 (i.e., wd0e) which hasn't
  779: been created yet. wd0e will be seen as a disk device in the new domain,
  780: so it will be 'sub-partitioned'. We could attach a ccd to wd0e in
  781: *domain0* and partition it, newfs and extract the NetBSD/i386 or amd64
  782: tarballs there, but there's an easier way: load the
  783: `netbsd-INSTALL_XEN3_DOMU` kernel provided in the NetBSD binary sets.
  784: Like other install kernels, it contains a ramdisk with sysinst, so you
  785: can install NetBSD using sysinst on your new domain.
  786: 
  787: If you want to install NetBSD/Xen with a CDROM image, the following line
  788: should be used in the `/usr/pkg/etc/xen/nbsd` file:
  789: 
  790:     disk = [ 'phy:/dev/wd0e,0x1,w', 'phy:/dev/cd0a,0x2,r' ]
  791: 
  792: After booting the domain, the option to install via CDROM may be
  793: selected. The CDROM device should be changed to `xbd1d`.
  794: 
  795: Once done installing, `halt -p` the new domain (don't reboot or halt, it
  796: would reload the INSTALL\_XEN3\_DOMU kernel even if you changed the
  797: config file), switch the config file back to the XEN3\_DOMU kernel, and
  798: start the new domain again. Now it should be able to use `root on xbd0a`
  799: and you should have a second, functional NetBSD system on your xen
  800: installation.
  801: 
  802: When the new domain is booting you'll see some warnings about *wscons*
  803: and the pseudo-terminals. These can be fixed by editing the files
  804: `/etc/ttys` and `/etc/wscons.conf`. You must disable all terminals in
  805: `/etc/ttys`, except *console*, like this:
  806: 
  807:     console "/usr/libexec/getty Pc"         vt100   on secure
  808:     ttyE0   "/usr/libexec/getty Pc"         vt220   off secure
  809:     ttyE1   "/usr/libexec/getty Pc"         vt220   off secure
  810:     ttyE2   "/usr/libexec/getty Pc"         vt220   off secure
  811:     ttyE3   "/usr/libexec/getty Pc"         vt220   off secure
  812: 
  813: Finally, all screens must be commented out from `/etc/wscons.conf`.
  814: 
  815: It is also desirable to add
  816: 
  817:     powerd=YES
  818: 
  819: in rc.conf. This way, the domain will be properly shut down if
  820: `xm shutdown -R` or `xm shutdown -H` is used on the domain0.
  821: 
  822: Your domain should be now ready to work, enjoy.
  823: 
  824: Creating an unprivileged Linux domain (domU)
  825: --------------------------------------------
  826: 
  827: Creating unprivileged Linux domains isn't much different from
  828: unprivileged NetBSD domains, but there are some details to know.
  829: 
  830: First, the second parameter passed to the disk declaration (the '0x1' in
  831: the example below)
  832: 
  833:     disk = [ 'phy:/dev/wd0e,0x1,w' ]
  834: 
  835: does matter to Linux. It wants a Linux device number here (e.g. 0x300
  836: for hda). Linux builds device numbers as: (major \<\< 8 + minor). So,
  837: hda1 which has major 3 and minor 1 on a Linux system will have device
  838: number 0x301. Alternatively, devices names can be used (hda, hdb, ...)
  839: as xentools has a table to map these names to devices numbers. To export
  840: a partition to a Linux guest we can use:
  841: 
  842:     disk = [ 'phy:/dev/wd0e,0x300,w' ]
  843:     root = "/dev/hda1 ro"
  844: 
  845: and it will appear as /dev/hda on the Linux system, and be used as root
  846: partition.
  847: 
  848: To install the Linux system on the partition to be exported to the guest
  849: domain, the following method can be used: install sysutils/e2fsprogs
  850: from pkgsrc. Use mke2fs to format the partition that will be the root
  851: partition of your Linux domain, and mount it. Then copy the files from a
  852: working Linux system, make adjustments in `/etc` (fstab, network
  853: config). It should also be possible to extract binary packages such as
  854: .rpm or .deb directly to the mounted partition using the appropriate
  855: tool, possibly running under NetBSD's Linux emulation. Once the
  856: filesystem has been populated, umount it. If desirable, the filesystem
  857: can be converted to ext3 using tune2fs -j. It should now be possible to
  858: boot the Linux guest domain, using one of the vmlinuz-\*-xenU kernels
  859: available in the Xen binary distribution.
  860: 
  861: To get the linux console right, you need to add:
  862: 
  863:     extra = "xencons=tty1"
  864: 
  865: to your configuration since not all linux distributions auto-attach a
  866: tty to the xen console.
  867: 
  868: Creating an unprivileged Solaris domain (domU)
  869: ----------------------------------------------
  870: 
  871: Download an Opensolaris [release](http://opensolaris.org/os/downloads/)
  872: or [development snapshot](http://genunix.org/) DVD image. Attach the DVD
  873: image to a MAN.VND.4 device. Copy the kernel and ramdisk filesystem
  874: image to your dom0 filesystem.
  875: 
  876:     dom0# mkdir /root/solaris
  877:     dom0# vnconfig vnd0 osol-1002-124-x86.iso
  878:     dom0# mount /dev/vnd0a /mnt
  879: 
  880:     ## for a 64-bit guest
  881:     dom0# cp /mnt/boot/amd64/x86.microroot /root/solaris
  882:     dom0# cp /mnt/platform/i86xpv/kernel/amd64/unix /root/solaris
  883: 
  884:     ## for a 32-bit guest
  885:     dom0# cp /mnt/boot/x86.microroot /root/solaris
  886:     dom0# cp /mnt/platform/i86xpv/kernel/unix /root/solaris
  887: 
  888:     dom0# umount /mnt
  889:           
  890: 
  891: Keep the MAN.VND.4 configured. For some reason the boot process stalls
  892: unless the DVD image is attached to the guest as a "phy" device. Create
  893: an initial configuration file with the following contents. Substitute
  894: */dev/wd0k* with an empty partition at least 8 GB large.
  895: 
  896:     memory = 640
  897:     name = 'solaris'
  898:     disk = [ 'phy:/dev/wd0k,0,w' ]
  899:     disk += [ 'phy:/dev/vnd0d,6:cdrom,r' ]
  900:     vif = [ 'bridge=bridge0' ]
  901:     kernel = '/root/solaris/unix'
  902:     ramdisk = '/root/solaris/x86.microroot'
  903:     # for a 64-bit guest
  904:     extra = '/platform/i86xpv/kernel/amd64/unix - nowin -B install_media=cdrom'
  905:     # for a 32-bit guest
  906:     #extra = '/platform/i86xpv/kernel/unix - nowin -B install_media=cdrom'
  907:           
  908: 
  909: Start the guest.
  910: 
  911:     dom0# xm create -c solaris.cfg
  912:     Started domain solaris
  913:                           v3.3.2 chgset 'unavailable'
  914:     SunOS Release 5.11 Version snv_124 64-bit
  915:     Copyright 1983-2009 Sun Microsystems, Inc.  All rights reserved.
  916:     Use is subject to license terms.
  917:     Hostname: opensolaris
  918:     Remounting root read/write
  919:     Probing for device nodes ...
  920:     WARNING: emlxs: ddi_modopen drv/fct failed: err 2
  921:     Preparing live image for use
  922:     Done mounting Live image
  923:           
  924: 
  925: Make sure the network is configured. Note that it can take a minute for
  926: the xnf0 interface to appear.
  927: 
  928:     opensolaris console login: jack
  929:     Password: jack
  930:     Sun Microsystems Inc.   SunOS 5.11      snv_124 November 2008
  931:     jack@opensolaris:~$ pfexec sh
  932:     sh-3.2# ifconfig -a
  933:     sh-3.2# exit
  934:           
  935: 
  936: Set a password for VNC and start the VNC server which provides the X11
  937: display where the installation program runs.
  938: 
  939:     jack@opensolaris:~$ vncpasswd
  940:     Password: solaris
  941:     Verify: solaris
  942:     jack@opensolaris:~$ cp .Xclients .vnc/xstartup
  943:     jack@opensolaris:~$ vncserver :1
  944:           
  945: 
  946: From a remote machine connect to the VNC server. Use `ifconfig xnf0` on
  947: the guest to find the correct IP address to use.
  948: 
  949:     remote$ vncviewer 172.18.2.99:1
  950:           
  951: 
  952: It is also possible to launch the installation on a remote X11 display.
  953: 
  954:     jack@opensolaris:~$ export DISPLAY=172.18.1.1:0
  955:     jack@opensolaris:~$ pfexec gui-install
  956:            
  957: 
  958: After the GUI installation is complete you will be asked to reboot.
  959: Before that you need to determine the ZFS ID for the new boot filesystem
  960: and update the configuration file accordingly. Return to the guest
  961: console.
  962: 
  963:     jack@opensolaris:~$ pfexec zdb -vvv rpool | grep bootfs
  964:                     bootfs = 43
  965:     ^C
  966:     jack@opensolaris:~$
  967:            
  968: 
  969: The final configuration file should look like this. Note in particular
  970: the last line.
  971: 
  972:     memory = 640
  973:     name = 'solaris'
  974:     disk = [ 'phy:/dev/wd0k,0,w' ]
  975:     vif = [ 'bridge=bridge0' ]
  976:     kernel = '/root/solaris/unix'
  977:     ramdisk = '/root/solaris/x86.microroot'
  978:     extra = '/platform/i86xpv/kernel/amd64/unix -B zfs-bootfs=rpool/43,bootpath="/xpvd/xdf@0:a"'
  979:            
  980: 
  981: Restart the guest to verify it works correctly.
  982: 
  983:     dom0# xm destroy solaris
  984:     dom0# xm create -c solaris.cfg
  985:     Using config file "./solaris.cfg".
  986:     v3.3.2 chgset 'unavailable'
  987:     Started domain solaris
  988:     SunOS Release 5.11 Version snv_124 64-bit
  989:     Copyright 1983-2009 Sun Microsystems, Inc.  All rights reserved.
  990:     Use is subject to license terms.
  991:     WARNING: emlxs: ddi_modopen drv/fct failed: err 2
  992:     Hostname: osol
  993:     Configuring devices.
  994:     Loading smf(5) service descriptions: 160/160
  995:     svccfg import warnings. See /var/svc/log/system-manifest-import:default.log .
  996:     Reading ZFS config: done.
  997:     Mounting ZFS filesystems: (6/6)
  998:     Creating new rsa public/private host key pair
  999:     Creating new dsa public/private host key pair
 1000: 
 1001:     osol console login:
 1002:            
 1003: 
 1004: Using PCI devices in guest domains
 1005: ----------------------------------
 1006: 
 1007: The domain0 can give other domains access to selected PCI devices. This
 1008: can allow, for example, a non-privileged domain to have access to a
 1009: physical network interface or disk controller. However, keep in mind
 1010: that giving a domain access to a PCI device most likely will give the
 1011: domain read/write access to the whole physical memory, as PCs don't have
 1012: an IOMMU to restrict memory access to DMA-capable device. Also, it's not
 1013: possible to export ISA devices to non-domain0 domains (which means that
 1014: the primary VGA adapter can't be exported. A guest domain trying to
 1015: access the VGA registers will panic).
 1016: 
 1017: This functionality is only available in NetBSD-5.1 (and later) domain0
 1018: and domU. If the domain0 is NetBSD, it has to be running Xen 3.1, as
 1019: support has not been ported to later versions at this time.
 1020: 
 1021: For a PCI device to be exported to a domU, is has to be attached to the
 1022: `pciback` driver in domain0. Devices passed to the domain0 via the
 1023: pciback.hide boot parameter will attach to `pciback` instead of the
 1024: usual driver. The list of devices is specified as `(bus:dev.func)`,
 1025: where bus and dev are 2-digit hexadecimal numbers, and func a
 1026: single-digit number:
 1027: 
 1028:     pciback.hide=(00:0a.0)(00:06.0)
 1029: 
 1030: pciback devices should show up in the domain0's boot messages, and the
 1031: devices should be listed in the `/kern/xen/pci` directory.
 1032: 
 1033: PCI devices to be exported to a domU are listed in the `pci` array of
 1034: the domU's config file, with the format `'0000:bus:dev.func'`
 1035: 
 1036:     pci = [ '0000:00:06.0', '0000:00:0a.0' ]
 1037: 
 1038: In the domU an `xpci` device will show up, to which one or more pci
 1039: busses will attach. Then the PCI drivers will attach to PCI busses as
 1040: usual. Note that the default NetBSD DOMU kernels do not have `xpci` or
 1041: any PCI drivers built in by default; you have to build your own kernel
 1042: to use PCI devices in a domU. Here's a kernel config example:
 1043: 
 1044:     include         "arch/i386/conf/XEN3_DOMU"
 1045:     #include         "arch/i386/conf/XENU"           # in NetBSD 3.0
 1046: 
 1047:     # Add support for PCI busses to the XEN3_DOMU kernel
 1048:     xpci* at xenbus ?
 1049:     pci* at xpci ?
 1050: 
 1051:     # Now add PCI and related devices to be used by this domain
 1052:     # USB Controller and Devices
 1053: 
 1054:     # PCI USB controllers
 1055:     uhci*   at pci? dev ? function ?        # Universal Host Controller (Intel)
 1056: 
 1057:     # USB bus support
 1058:     usb*    at uhci?
 1059: 
 1060:     # USB Hubs
 1061:     uhub*   at usb?
 1062:     uhub*   at uhub? port ? configuration ? interface ?
 1063: 
 1064:     # USB Mass Storage
 1065:     umass*  at uhub? port ? configuration ? interface ?
 1066:     wd*     at umass?
 1067:     # SCSI controllers
 1068:     ahc*    at pci? dev ? function ?        # Adaptec [23]94x, aic78x0 SCSI
 1069: 
 1070:     # SCSI bus support (for both ahc and umass)
 1071:     scsibus* at scsi?
 1072: 
 1073:     # SCSI devices
 1074:     sd*     at scsibus? target ? lun ?      # SCSI disk drives
 1075:     cd*     at scsibus? target ? lun ?      # SCSI CD-ROM drives
 1076: 
 1077: 
 1078: NetBSD as a domU in a VPS
 1079: =========================
 1080: 
 1081: The bulk of the HOWTO is about using NetBSD as a dom0 on your own
 1082: hardware.  This section explains how to deal with Xen in a domU as a
 1083: virtual private server where you do not control or have access to the
 1084: dom0.
 1085: 
 1086: TODO: Perhaps reference panix, prmgr, amazon as interesting examples.
 1087: 
 1088: TODO: Somewhere, discuss pvgrub and py-grub to load the domU kernel
 1089: from the domU filesystem.
 1090: 
 1091: Using npf
 1092: ---------
 1093: 
 1094: In standard kernels, npf is a module, and thus cannot be loadeed in a
 1095: DOMU kernel.
 1096: 
 1097: TODO: explain how to compile npf into a custom kernel, answering:
 1098: http://mail-index.netbsd.org/netbsd-users/2014/12/26/msg015576.html

CVSweb for NetBSD wikisrc <wikimaster@NetBSD.org> software: FreeBSD-CVSweb