Ceph

    PLEASE NOTE: This document applies to v1.0 version and not to the latest stable release v1.9

    Ceph Examples

    Configuration for Rook and Ceph can be configured in multiple ways to provide block devices, shared file system volumes or object storage in a kubernetes namespace. We have provided several examples to simplify storage setup, but remember there are many tunables and you will need to decide what settings work for your use case and environment.

    See the example yaml files folder for all the rook/ceph setup example spec files.

    Common Resources

    The first step to deploy Rook is to create the common resources. The configuration for these resources will be the same for most deployments. The common.yaml sets these resources up.

    kubectl create -f common.yaml
    

    The examples all assume the operator and all Ceph daemons will be started in the same namespace. If you want to deploy the operator in a separate namespace, see the comments throughout common.yaml.

    Operator

    After the common resources are created, the next step is to create the Operator deployment. Several spec file examples are provided in this directory:

    • operator.yaml: The most common settings for production deployments
      • kubectl create -f operator.yaml
    • operator-openshift.yaml: Includes all of the operator settings for running a basic Rook cluster in an OpenShift environment. You will also want to review the OpenShift Prerequisites to confirm the settings.
      • oc create -f operator-openshift.yaml
    • operator-with-csi.yaml: Includes configuration for testing ceph-csi while the integration is still in progress. See the CSI Drivers topic for more details.
      • kubectl create -f operator-with-csi.yaml
    • operator-openshift-with-csi.yaml: Includes configuration for testing ceph-csi in an OpenShift environment
      • oc create -f operator-openshift-with-csi.yaml

    Settings for the operator are configured through environment variables on the operator deployment. The individual settings are documented in operator.yaml.

    Cluster CRD

    Now that your operator is running, let’s create your Ceph storage cluster:

    • cluster.yaml: This file contains common settings for a production storage cluster. Requires at least three nodes.
    • cluster-test.yaml: Settings for a test cluster where redundancy is not configured. Requires only a single node.
    • cluster-minimal.yaml: Brings up a cluster with only one ceph-mon and a ceph-mgr so the Ceph dashboard can be used for the remaining cluster configuration.

    See the Cluster CRD topic for more details on the settings.

    Setting up consumable storage

    Now we are ready to setup block, shared filesystem or object storage in the rook ceph cluster. These kinds of storage are respectively referred to as CephBlockPool, CephFilesystem and CephObjectStore in the spec files.

    Block Devices

    Ceph can provide raw block device volumes to pods. Each example below sets up a storage class which can then be used to provision a block device in kubernetes pods. The storage class is defined with a pool which defines the level of data redundancy in ceph:

    • storageclass.yaml: This example illustrates replication of 3 for production scenarios and requires at least three nodes. Your data is replicated on three different kubernetes worker nodes and intermittent or long-lasting single node failures will not result in data unavailability or loss.
    • storageclass-ec.yaml: Configures erasure coding for data durability rather than replication. Ceph’s erasure coding is more efficient than replication so you can get high reliability without the 3x replication cost of the preceding example (but at the cost of higher computational encoding and decoding costs on the worker nodes). Erasure coding requires at least three nodes. See the Erasure coding documentation for more details.
    • storageclass-test.yaml: Replication of 1 for test scenarios and it requires only a single node. Do not use this for applications that store valuable data or have high-availability storage requirements, since a single node failure can result in data loss.

    See the Ceph Pool CRD topic for more details on the settings.

    Shared File System

    Ceph file system (CephFS) allows the user to ‘mount’ a shared posix-compliant folder into one or more hosts (pods in the container world). This storage is similar to NFS shared storage or CIFS shared folders, as explained here.

    File storage contains multiple pools that can be configured for different scenarios:

    • filesystem.yaml: Replication of 3 for production scenarios. Requires at least three nodes.
    • filesystem-ec.yaml: Erasure coding for production scenarios. Requires at least three nodes.
    • filesystem-test.yaml: Replication of 1 for test scenarios. Requires only a single node.

    See the Shared File System CRD topic for more details on the settings.

    Object Storage

    Ceph supports storing blobs of data called objects that support HTTP(s)-type get/put/post and delete semantics. This storage is similar to AWS S3 storage, for example.

    Object storage contains multiple pools that can be configured for different scenarios:

    • object.yaml: Replication of 3 for production scenarios. Requires at least three nodes.
    • object-openshift.yaml: Replication of 3 with rgw in a port range valid for OpenShift. Requires at least three nodes.
    • object-ec.yaml: Erasure coding rather than replication for production scenarios. Requires at least three nodes.
    • object-test.yaml: Replication of 1 for test scenarios. Requires only a single node.

    See the Object Store CRD topic for more details on the settings.

    Object Storage User

    • object-user.yaml: Creates a simple object storage user and generates creds for the S3 API