Note: < 5 minutes read

When running a virtualization workload on oVirt, a VM disk is ‘natively’ a disk somewhere on your network-storage.
Entering containers world, on Kubernetes(k8s) or OpenShift, there are many options specifically because the workload can be totally stateless, i.e
they are stored on a host supplied disk and can be removed when the container is terminated. The more interesting case is stateful workloads i.e apps that persist data (think DBs, web servers/services, etc). k8s/OpenShift designed an API to dynamically provision the container storage (volume in k8s terminology).

See the resources section for more details.

In this post I want to cover how oVirt can provide volumes for containers running on k8s/OpenShift cluster.


Consider this: you want to deploy wikimedia as a container, with all its content served from /opt.
For that you will create a persistent volume for the container – when we have state to keep and server
creating a volume makes sense. It is persistent, it exists regardless the container state,
and you can choose which directory exactly you serve that volume, and that is the most important
part, k8s/OpenShift gives you an API to determine who will provide the volume that you need.

There are many options, Cinder, AWS, NFS and more. And in case the node that your pod is running on
is a VM in oVirt, you can use ovirt-flexdriver to attach an oVirt disk and that will
appear as a device in the node, and will be mounted with filesystem to your request. If you want to know more see the documentation about kubernetes-incubator/external-storage

    k8s/OpenShift Node          +-------> oVirt Vm
|                      |                                  +----------------+
|   mediawiki pod      |                                  |                |
| +---------------+    |                                  |                |
| |               |    |                                  |                |
| |               |    |                                  |     oVirt      |
| |               |    |                                  |                |
| |/srv/mediawiki |    |                                  |                |
| +---------------+    |                                  |                |
|                      |                                  +----------------+
|                      |
|                      |
|  /dev/pv001 (/srv/mediawiki)  +-------> oVirt Disk
|                      |


Checkout this youtube video, that demonstrate how it looks like in oVirt admin UI, kubernetes UI in cockpit, and some cli:

External Storage Provisioner and Flexvolume driver

OpenShift is able to request oVirt these special volumes by deploying ovirt-flexdriver and ovirt-provisioner and following this steps:

  • Create a storage class
  • Create a storage claim
  • Create a pod with a volume that refernce the storage claim
  • Run the pod

A storage class will can describe slow or fast data storage that maps to data domains in oVirt

kind: StorageClass
  name: ovirt-ssd-domain
provisioner: external/ovirt 
  type: io1
  iopsPerGB: "10"
  ovirtStorageDomain: "prod-ssd-domain"
  fsType: ext4
  ovirtDiskFormat: "cow"

When you create a storage claim, ovirt-provisioner will create an oVirt disk for you on the
specified domain – notice the reference to the storage class:

# storage claim
kind: PersistentVolumeClaim
apiVersion: v1
  name: mediawiki-data-ssd-disk
  annotations: ovirt-ssd-domain
  storageClassName: ovirt-ssd-domain
    - ReadWriteOnce
      storage: 1Gi

Once the claim is created, oVirt is creating a 1Gb disk which is not attached to any node yet.

Run a mediawiki pod with so-called flex volume:

apiVersion: v1 
kind: Pod 
  name: mediawiki
    app: ovirt 
  - image: mediawiki 
    name: mediawiki 
    - name: mediawiki-storage
      mountPath: "/data/"
  - name: mediawiki-storage
      claimName: mediawiki-data-ssd-disk

And now it is the flexvolume driver job to tell oVirt to attach the disk into the node this
pod is running on, and creat file system on it, as described in the storage class, and to mount
it onto the node. When this is done, the volume is ready and the container can start, with
the mount set into the /data directory as set by mountPath.

Want to give it a try? Want to get updated about this?

This work as for today (Feb 20th 2018) is in progress and all of it can be found at the ovirt-flexdriver project page
To deploy ovirt-flexdriver and ovirt-provisioner I created a container with Ansible playbook that takes an inventory
that has the k8s nodes and k8s master specified, along with the ovirt-engine connection details. The playbook will copy and
configure both component and get you up and running with just few keystrokes. Find more on deployment in the of project
and in an up-coming short video demonstrating the deployment.

Note on versions: this should be working against kubernetes 1.9 and oVirt 4.2 but 4.1 should work as well (because the API in use is the same).