Create and access NFS server from Kubernetes on CREODIAS

In order to enable simultaneous read-write storage to multiple pods running on a Kubernetes cluster, we can use an NFS server.

In this guide we will create an NFS server on a virtual machine, create file share on this server and demonstrate accessing it from a Kubernetes pod.

What We Are Going To Cover

  • Set up an NFS server on a VM

  • Set up a share folder on the NFS server

  • Make the share available

  • Deploy a test pod on the cluster


No. 1 Hosting

You need a CREODIAS hosting account with Horizon interface

The resources that you require and use will reflect on the state of your account wallet. Check your account statistics at

No. 2 Familiarity with Linux and cloud management

We assume you know the basics of Linux and CREODIAS cloud management:

No. 3 A running Kubernetes cluster

You will also need a Kubernetes cluster to try out the commands. To create one from scratch, see How to Create a Kubernetes Cluster Using CREODIAS OpenStack Magnum

No. 4 kubectl access to the Kubernetes cloud

As usual when working with Kubernetes clusters, you will need to use the kubectl command: How To Access Kubernetes Cluster Post Deployment Using Kubectl On CREODIAS OpenStack Magnum

1. Set up NFS server on a VM

As a prerequisite to create an NFS server on a VM, first from the Network tab in Horizon create a security group allowing ingress traffic from port 2049.

Then create an Ubuntu VM from Horizon. During the Network selection dialog, connect the VM to the network of your Kubernetes cluster (not project network nor EODATA network). This ensures that cluster nodes have access to the NFS server over private network. Then add that security group with port 2049 open.


When the VM is created, you can see that it has private address assigned. For this occasion, let the private address be Take note of this address to later use it in NFS configuration.

Set up floating IP on the VM server, just to enable SSH to this VM.

2. Set up a share folder on the NFS server

SSH to the VM, then run:

sudo apt-get update
sudo apt-get install nfs-kernel-server

In the NFS server VM create a share folder:

sudo mkdir /mnt/myshare

Change the owner of the share so that nobody is owner. Thus any user on the client can access the share folder. More restrictive settings can be applied.

sudo chown nobody:nogroup /mnt/myshare

Also change the permissions of the folder, so that anyone can modify the files:

sudo chmod 777 /mnt/myshare

Edit the /etc/exports file and add the following line:


This indicates that all nodes on the cluster network can access this share, with subfolders, in read-write mode.

3. Make the share available

Run the below command to make the share available:

sudo exportfs -a

Then restart the NFS server with:

sudo systemctl restart nfs-kernel-server

Exit from the NFS server VM.

4. Deploy a test pod on the cluster

Ensure you can access your cluster with kubectl. Have a file test-pod.yaml with the following contents:


apiVersion: v1
kind: Pod
  name: test-pod
  namespace: default
  - image: nginx
    name: test-container
    - mountPath: /my-nfs-data
      name: test-volume
  - name: test-volume
      path: /mnt/myshare

The NFS server block refers to private IP address of the NFS server machine, which is on our cluster network. Apply the yaml manifest with:

kubectl apply -f test-pod.yaml

We can then enter the shell of the test-pod with the below command:

kubectl exec -it test-pod -- sh

and see that the my-nfs-data folder got mounted properly:


To verify, create a file testfile in this folder, then exit the container. You can then SSH back to the NFS server and verify that testfile is available in /mnt/myshare folder.