Feature pages are design documents that developers have created while collaborating on oVirt.

Most of them are outdated, but provide historical design context.

They are not user documentation and should not be treated as such.

Documentation is available here.

GlusterFS Hyperconvergence

Hyperconvergence is a type of infrastructure system with a software-centric architecture that tightly integrates compute, storage, networking and virtualization resources and other technologies from scratch in a commodity hardware box supported by a single vendor[1].


Current status

  • Version: 3.6.7 & above - support for HC with known caveats
  • Target Release: 4.1
  • Status: In Progress


This feature page tracks the effort of integrating oVirt virtualization and GlusterFS storage resources in single commodity boxes that can scale horizontally.


Gluster and compute resources run on the same node as shown below. This setup requires a minimum of 3 hosts. The 3 host requirement is because all gluster volumes MUST be 3-way replica to avoid issues with storage (mainly files split-brain). Hosted engine and data storage domains are on gluster volumes created on these 3 hosts.

HC arch

What’s available now

Since 3.6.7, users can set up a hyperconverged deployment using Gluster and oVirt. The steps to do this are covered in a separate blog post

  • Gluster sharding available from glusterfs 3.7.11. Sharding eliminates the issues seen while healing large files as files are now split into shards. Users need to enable sharding on gluster volumes in hyperconverged enviroments. For further information on sharding, refer Glusterfs Sharding
  • Hosted engine deployment supports using glusterfs for storage domain Feature Page
  • Separate storage network for glusterfs traffic Feature Page - Bugzilla
  • Host disks management and device provisioning Feature Page
  • Simplified provisioning of the 3 hosts using gdeploy. See gdeploy and a Sample hc.conf
  • Disaster recovery solution using gluster geo-replication (TBD: Feature page). Refer DR script
  • Monitoring gluster storage and cluster using nagios

Since 4.0, following features are available in oVirt

  • Self-heal monitoring of gluster volumes Bugzilla
  • Replacing bricks from UI when a host or disk has failed Bugzilla
  • ovirt-node includes vdsm-gluster and gluster-server

Planned features for 4.1

  • QEMU libfapi support Bugzilla
  • Data-center power policies should be aware of hyperconverged nodes to avoid automatic shutdown for power saving

Future plans

  • Improve UX when creating a Gluster Volume to be used as Storage Domain Feature Page
  • Data locality Buzilla-1 Bugzilla-2
    • Scheduling of disk creation: disks of the same VM on the same replica set
    • Scheduling of the VM based on disk locality: start VM on the hosts of disks replica set

Feature pages TBD:

  • Brick scaling and balancing when adding storage Bugzilla
    • Best practices on how to scale by adding new hosts
  • Best practices to handle an entire host failure and replacement