Install and Configure Hypervisor¶
Obtain a PowerKVM ISO and create PowerKVM osimages with it:copycds ibm-powerkvm-188.8.131.52-39.0-ppc64le-gold-201511041419.iso
The following PowerKVM osimage will be created# lsdef -t osimage -o pkvm3.1-ppc64le-install-compute Object name: pkvm3.1-ppc64le-install-compute imagetype=linux osarch=ppc64le osdistroname=pkvm3.1-ppc64le osname=Linux osvers=pkvm3.1 otherpkgdir=/install/post/otherpkgs/pkvm3.1/ppc64le pkgdir=/install/pkvm3.1/ppc64le profile=compute provmethod=install template=/opt/xcat/share/xcat/install/pkvm/compute.pkvm3.ppc64le.tmpl
Red Hat Virtualization (formally known as RHEV or Red Hat Enterprise Virtualization) is a virtualization solution provided by Red Hat.
At the time of this writing there is no RHEV-H prebuilt hypervisor image on Power LE. The method for creating a Red Hat Hypervisor on Power LE is to first install RHEL and apply the KVM support on top with the provided RPMs.
Obtain and download the RHEV RPM packages from the Red Hat download site.
In the following example, the RPMs are downloaded to
Create a yum repository for the downloaded RPMscreaterepo /install/post/otherpkgs/rhels7.3/ppc64le/RHEV4/4.0-GA
Create new osimage definition based on an existing RHEL7 osimage definitionmkdef -t osimage -o rhels7.3-ppc64le-RHEV4-install-compute \ --template rhels7.3-ppc64le-install-compute
otherpkgdirattribute to point to the package directory with downloaded RPMschdef -t osimage rhels7.3-ppc64le-RHEV4-install-compute \ otherpkgdir=/install/post/otherpkgs/rhels7.3/ppc64le/RHEV4/4.0-GA
Create a new package list file
/install/custom/rhels7.3/ppc64le/rhelv4.pkglistto include necessary packages provided from the OS.#INCLUDE:/opt/xcat/share/xcat/install/rh/compute.rhels7.pkglist# libvirt screen bridge-utils
pkglistattribute to point to the package list file from the step abovechdef -t osimage rhels7.3-snap3-ppc64le-RHEV4-install-compute \ pkglist=/install/custom/rhels7.3/ppc64le/rhelv4.pkglist
Create a new package list file
/install/custom/rhels7.3/ppc64le/rhev4.otherpkgs.pkglistto list required packagesqemu-kvm-rhev qemu-kvm-tools-rhev virt-manager-common virt-install
otherpkglistattribute to point to the package list file from the step abovechdef -t osimage rhels7.3-snap3-ppc64le-RHEV4-install-compute \ otherpkglist=/install/custom/rhels7.3/ppc64le/rhev4.otherpkgs.pkglist
The RHEV osimage should look similar to:Object name: rhels7.3-ppc64le-RHEV4-install-compute imagetype=linux osarch=ppc64le osdistroname=rhels7.3-ppc64le osname=Linux osvers=rhels7.3 otherpkgdir=/install/post/otherpkgs/rhels7.3/ppc64le/RHEV4/4.0-GA otherpkglist=/install/custom/rhels7.3/ppc64le/rhev4.otherpkgs.pkglist pkgdir=/install/rhels7.3/ppc64le pkglist=/install/custom/rhels7.3/ppc64le/rhelv4.pkglist profile=compute provmethod=install template=/opt/xcat/share/xcat/install/rh/compute.rhels7.tmpl
Customize the hypervisor node definition to create network bridge
xCAT ships a postscript xHRM to create a network bridge on kvm host during installation/netbooting. Specify the xHRM with appropriate parameters in postscripts attribute. For example:
To create a bridge named ‘br0’ against the installation network device specified by installnic:
chdef kvmhost1 -p postscripts="xHRM bridgeprereq br0"
To create a bridge with default name ‘default’ against the installation network device specified by installnic:
chdef kvmhost1 -p postscripts="xHRM bridgeprereq"
To create a bridge named ‘br0’ against the network device ‘eth0’:
chdef kvmhost1 -p postscripts="xHRM bridgeprereq eth0:br0"
Note: The network bridge name you use should not be the virtual bridges (vbrX) created by libvirt installation .
Customize the hypervisor node definition to mount the shared kvm storage directory on management node (optional)
If the shared kvm storage directory on the management node has been exported, it can be mounted on PowerKVM hypervisor for virtual machines hosting.
An easy way to do this is to create another postscript named “mountvms” which creates a directory /install/vms on hypervisor and then mounts /install/vms from the management node, the content of “mountvms” can be:
logger -t xcat "Install: setting vms mount in fstab" mkdir -p /install/vms echo "$MASTER:/install/vms /install/vms nfs \ rsize=8192,wsize=8192,timeo=14,intr,nfsvers=2 1 2" >> /etc/fstab
Then set the file permission and specify the script in postscripts attribute of hypervisor node definition:
chmod 755 /install/postscripts/mountvms chdef kvmhost1 -p postscripts=mountvms
Provision the hypervisor node with the osimage
nodeset kvmhost1 osimage=<osimage_name> rpower kvmhost1 boot
Create network bridge on hypervisor¶
To launch VMs, a network bridge must be created on the KVM hypervisor.
If the hypervisor is provisioned successfully according to the steps described above, a network bridge will be created and attached to a physical interface. This can be checked by running
brctl show on the hypervisor to show the network bridge information, please make sure a network bridge has been created and configured according to the parameters passed to postscript “xHRM”
# brctl show bridge name bridge id STP enabled interfaces br0 8000.000000000000 no eth0
If the network bridge is not created or configured successfully, run “xHRM” with updatenode on management node to create it manually::
updatenode kvmhost1 -P "xHRM bridgeprereq eth0:br0"
Start libvirtd service¶
Verify libvirtd service is running:
systemctl status libvirtd
If service is not running, it can be started with:
systemctl start libvirtd
|||Every standard libvirt installation provides NAT based connectivity to virtual machines out of the box using the “virtual bridge” interfaces (virbr0, virbr1, etc) Those will be created by default.|