Much of this developer documentation provides historical context but may not reflect the current state of the project.
If you see outdated content please navigate to the page footer and click "Report an issue on GitHub".

It is not user documentation and should not be treated as such.

User Documentation is available here.

Phoenix Lab Storage Hosts

NOTE: for the latest version of this doc, see

Currently we have two storage servers, both of them have a CentOS 6.5 installation on them.

Disk configuration

The storage servers have a set of 6 disks in a RAID5

Storage replication

For the storage replication we are using DRBD, it was required to install drbd84, and to do that on centos we had to use some special repos as it’s been discontinued on the official repos. Here are the specific ones:

[root@ovirt-storage01 ~]# cat /etc/yum.repos.d/hacluster.repo
name=HA Clustering

You can check specifically the current status using the command:

[root@ovirt-storage01 ~]# drbd-overview
0:ovirt_storage/0  Connected Primary/Secondary UpToDate/UpToDate C r----- /srv/ovirt_storage ext4 11T 563G 9.7T 6% 

The DRBD cluster is started/stopped by the pacemaker cluster, so no need to handle it, but sometimes when the cluster degenerates is required to manually choose which node has to be master and start the replication between the nodes. You can check the cdocumentation on how to fix that type of issues here.


The clustering has been configured using crm and pacemaker. Here are a few tips on managing it:

To enter the management shell you can just type:


From there you can see a list of available commands using tab completion.

To see the current status of the cluster you can use:

[root@ovirt-storage01 ~]# crm status
Last updated: Sat Nov  8 03:59:18 2014
Last change: Thu Jul 31 02:41:35 2014 via cibadmin on ovirt-storage01
Stack: cman
Current DC: ovirt-storage02 - partition with quorum
Version: 1.1.10-14.el6_5.3-368c726
2 Nodes configured
7 Resources configured

Online: [ ovirt-storage01 ovirt-storage02 ]

Master/Slave Set: ms_drbd_ovirt_storage [p_drbd_ovirt_storage]
    Masters: [ ovirt-storage01 ]
    Slaves: [ ovirt-storage02 ]
Resource Group: g_ovirt_storage
    p_fs_ovirt_storage  (ocf::heartbeat:Filesystem):    Started ovirt-storage01 
    p_ip_ovirt_storage  (ocf::heartbeat:IPaddr2):   Started ovirt-storage01 
    p_nfs_ovirt_storage (lsb:nfs):  Started ovirt-storage01
Clone Set: cl_exportfs_ovirt_storage [p_exportfs_ovirt_storage]
    Started: [ ovirt-storage01 ovirt-storage02 ]

Showing/editing the config

To see and edit the configuration you have to enter the configuration space from the crm shell, for future reference here’s the output form the current config:

crm(live)# cd configure
crm(lise)configure# show

node ovirt-storage01
node ovirt-storage02
primitive p_drbd_ovirt_storage ocf:linbit:drbd \
    params drbd_resource=ovirt_storage \
    op monitor interval=15 role=Master \
    op monitor interval=30 role=Slave
primitive p_exportfs_ovirt_storage exportfs \
    params fsid=0 directory="/srv/ovirt_storage" options="rw,mountpoint,no_root_squash" clientspec="" \
    op monitor interval=30s \
    meta target-role=Started
primitive p_fs_ovirt_storage Filesystem \
    params device="/dev/drbd0" directory="/srv/ovirt_storage" fstype=ext4 \
    op monitor interval=10s \
    meta target-role=Started
primitive p_ip_ovirt_storage IPaddr2 \
    params ip= cidr_netmask=26 \
    op monitor interval=30s \
    meta target-role=Started
primitive p_nfs_ovirt_storage lsb:nfs \
    op monitor interval=30s \
    meta target-role=Started
group g_ovirt_storage p_fs_ovirt_storage p_ip_ovirt_storage \
    meta target-role=Started
ms ms_drbd_ovirt_storage p_drbd_ovirt_storage \
    meta master-max=1 master-node-max=1 clone-max=2 clone-node-max=1 notify=true target-role=Started
clone cl_exportfs_ovirt_storage p_exportfs_ovirt_storage
location cli-prefer-ms_drbd_ovirt_storage ms_drbd_ovirt_storage role=Started inf: ovirt-storage01
colocation c_all_on_drbd inf: g_ovirt_storage ms_drbd_ovirt_storage:Master
colocation c_nfs_on_drbd inf: p_nfs_ovirt_storage ms_drbd_ovirt_storage:Master
colocation c_nfs_on_exportfs inf: g_ovirt_storage cl_exportfs_ovirt_storage
order o_drbd_first inf: ms_drbd_ovirt_storage:promote g_ovirt_storage:start
order o_exportfs_before_nfs inf: cl_exportfs_ovirt_storage g_ovirt_storage:start
property cib-bootstrap-options: \
    dc-version=1.1.10-14.el6_5.3-368c726 \
    cluster-infrastructure=cman \
    expected-quorum-votes=2 \
    stonith-enabled=false \
    no-quorum-policy=ignore \


The network is configured to use bonding on all interfaces using 802.3ad bonding protocol (requires special configuration on the switches).

Here’s the current configuration files:

[root@ovirt-storage01 ~]# cat /etc/modprobe.d/bonding.conf
alias bond0 bonding
##mode=4 - 802.3ad   mode=6 - alb
options bond0 mode=4 miimon=100 lacp_rate=1

[root@ovirt-storage01 ~]# cat /etc/sysconfig/network-scripts/ifcfg-em1

[root@ovirt-storage01 ~]# cat /etc/sysconfig/network-scripts/ifcfg-bond0