Filesystem Mirror CRD
This guide assumes you have created a Rook cluster as explained in the main Quickstart guide
Rook allows creation and updating the fs-mirror daemon through the custom resource definitions (CRDs). CephFS will support asynchronous replication of snapshots to a remote (different Ceph cluster) CephFS file system via cephfs-mirror tool. Snapshots are synchronized by mirroring snapshot data followed by creating a snapshot with the same name (for a given directory on the remote file system) as the snapshot being synchronized. For more information about user management and capabilities see the Ceph docs.
To get you started, here is a simple example of a CRD to deploy an cephfs-mirror daemon.
If any setting is unspecified, a suitable default will be used automatically.
name: The name that will be used for the Ceph cephfs-mirror daemon.
namespace: The Kubernetes namespace that will be created for the Rook cluster. The services, pods, and other resources created by the operator will be added to this namespace.
placement: The cephfs-mirror pods can be given standard Kubernetes placement restrictions with
podAntiAffinitysimilar to placement defined for daemons configured by the cluster CRD.
annotations: Key value pair list of annotations to add.
labels: Key value pair list of labels to add.
resources: The resource requirements for the cephfs-mirror pods.
priorityClassName: The priority class to set on the cephfs-mirror pods.
Configuring mirroring peers¶
In order to configure mirroring peers, please refer to the CephFilesystem documentation.