Ceph NFS Gateway CRD
Rook allows exporting NFS shares of the filesystem or object store through the CephNFS custom resource definition. This will spin up a cluster of NFS Ganesha servers that coordinate with one another via shared RADOS objects. The servers will be configured for NFSv4.1+ access, as serving earlier protocols can inhibit responsiveness after a server restart.
The following sample will create a two-node active-active cluster of NFS Ganesha gateways. A CephFS named
myfs is used, and the recovery objects are stored in a RADOS pool named
myfs-data0 with a RADOS namespace of
This example requires the filesystem to first be configured by the Filesystem.
NOTE: For an RGW object store, a data pool of
my-store.rgw.buckets.datacan be used after configuring the Object Store.
apiVersion: ceph.rook.io/v1 kind: CephNFS metadata: name: my-nfs namespace: rook-ceph spec: rados: # RADOS pool where NFS client recovery data is stored. pool: myfs-data0 # RADOS namespace where NFS client recovery data is stored in the pool. namespace: nfs-ns # Settings for the NFS server server: # the number of active NFS servers active: 2 # A key/value list of annotations annotations: # key: value # where to run the NFS server placement: # nodeAffinity: # requiredDuringSchedulingIgnoredDuringExecution: # nodeSelectorTerms: # - matchExpressions: # - key: role # operator: In # values: # - mds-node # tolerations: # - key: mds-node # operator: Exists # podAffinity: # podAntiAffinity: # topologySpreadConstraints: # The requests and limits set here allow the ganesha pod(s) to use half of one CPU core and 1 gigabyte of memory resources: # limits: # cpu: "500m" # memory: "1024Mi" # requests: # cpu: "500m" # memory: "1024Mi" # the priority class to set to influence the scheduler's pod preemption priorityClassName:
pool: The pool where ganesha recovery backend and supplemental configuration objects will be stored
namespace: The namespace in
poolwhere ganesha recovery backend and supplemental configuration objects will be stored
EXPORT Block Configuration
All daemons within a cluster will share configuration with no exports defined, and that includes a RADOS object via:
NOTE: This format of nfs-ganesha config object name was introduced in Ceph Octopus Version. In older versions, each daemon has it’s own config object and with the name as *conf-
. *. The nodeid is a value automatically assigned internally by rook. Nodeids start with "a" and go through "z", at which point they become two letters ("aa" to "az").
The pool and namespace are configured via the spec’s RADOS block.
When a server is started, it will create the included object if it does not already exist. It is possible to prepopulate the included objects prior to starting the server. The format for these objects is documented in the NFS Ganesha project.
Scaling the active server count
It is possible to scale the size of the cluster up or down by modifying
spec.server.active field. Scaling the cluster size up can be done at
will. Once the new server comes up, clients can be assigned to it
The CRD always eliminates the highest index servers first, in reverse order from how they were started. Scaling down the cluster requires that clients be migrated from servers that will be eliminated to others. That process is currently a manual one and should be performed before reducing the size of the cluster.