PLEASE NOTE: This document applies to an unreleased version of Rook. It is strongly recommended that you only use official releases of Rook, as unreleased versions are subject to changes and incompatibilities that will not be supported in the official releases.
If you are using an official release version of Rook, you should refer to the documentation for your specific version.Documentation for other releases can be found by using the version selector in the bottom left of any doc page.
Rook can be installed on any existing Kubernetes clusters as long as it meets the minimum version and have the required privilege to run in the cluster (see below for more information). If you dont have a Kubernetes cluster, you can quickly set one up using Minikube, Kubeadm or CoreOS/Vagrant.
Kubernetes v1.10 or higher is supported by Rook.
Privileges and RBAC
Rook requires privileges to manage the storage in your cluster. See the details here for setting up Rook in a Kubernetes cluster with Pod Security Policies enabled.
The Rook agent requires setup as a Flex volume plugin to manage the storage attachments in your cluster. See the Flex Volume Configuration topic to configure your Kubernetes deployment to load the Rook volume plugin.
Kernel with RBD module
Rook Ceph requires a Linux kernel built with the RBD module. Many distributions of Linux have this module but some don’t, e.g. the GKE Container-Optimised OS (COS) does not have RBD. You can test your Kubernetes nodes by running
modprobe rbd. If it says ‘not found’, you may have to rebuild your kernel or choose a different Linux distribution.
Kernel modules directory configuration
Normally, on Linux, kernel modules can be found in
/lib/modules. However, there are some distributions that put them elsewhere. In that case the environment variable
LIB_MODULES_DIR_PATH can be used to override the default. Also see the documentation in helm-operator on the parameter
agent.libModulesDirPath. One notable distribution where this setting is useful would be NixOS.
Extra agent mounts
On certain distributions it may be necessary to mount additional directories into the agent container. That is what the environment variable
AGENT_MOUNTS is for. Also see the documentation in helm-operator on the parameter
agent.mounts. The format of the variable content should be
Some Linux distributions do not ship with the
lvm2 package. This package is required on all storage nodes in your k8s cluster. Please install it using your Linux distribution’s package manager; for example:
# Centos sudo yum install -y lvm2 # Ubuntu sudo apt-get install -y lvm2
Rook will run wherever Kubernetes is running. Here are some simple environments to help you get started with Rook.
minikube, refer to this page. Once you have
minikube installed, start a cluster by doing the following:
$ minikube start Starting local Kubernetes cluster... Starting VM... SSH-ing files into VM... Setting up certs... Starting cluster components... Connecting to cluster... Setting up kubeconfig... Kubectl is now configured to use the cluster.
After these steps, your minikube cluster is ready to install Rook on.
You can easily spin up Rook on top of a
You can find the instructions on how to install kubeadm in the Install
kubeadm, you can use Rook in just a few minutes!
New local Kubernetes cluster with Vagrant
For a quick start with a new local cluster, use the Rook fork of coreos-kubernetes. This will bring up a multi-node Kubernetes cluster with
vagrant and CoreOS virtual machines ready to use Rook immediately.
git clone https://github.com/rook/coreos-kubernetes.git cd coreos-kubernetes/multi-node/vagrant vagrant up export KUBECONFIG="$(pwd)/kubeconfig" kubectl config use-context vagrant-multi
Then wait for the cluster to come up and verify that kubernetes is done initializing (be patient, it takes a bit):
Once you see a url response, your cluster is ready for use by Rook.
Support for authenticated docker registries
If you want to use an image from authenticated docker registry (e.g. for image cache/mirror), you’ll need to
imagePullSecret to all relevant service accounts. This way all pods created by the operator (for service account:
rook-ceph-system) or all new pods in the namespace (for service account:
default) will have the
to their spec.
The whole process is described in the official kubernetes documentation.
Example setup for a ceph cluster
To get you started, here’s a quick rundown for the ceph example from the quickstart guide.
First, we’ll create the secret for our registry as described here:
# for namespace rook-ceph kubectl -n rook-ceph create secret docker-registry my-registry-secret --docker-server=DOCKER_REGISTRY_SERVER --docker-username=DOCKER_USER --docker-password=DOCKER_PASSWORD --docker-email=DOCKER_EMAIL # and for namespace rook-ceph (cluster) kubectl -n rook-ceph create secret docker-registry my-registry-secret --docker-server=DOCKER_REGISTRY_SERVER --docker-username=DOCKER_USER --docker-password=DOCKER_PASSWORD --docker-email=DOCKER_EMAIL
Next we’ll add the following snippet to all relevant service accounts as described here:
imagePullSecrets: - name: my-registry-secret
The service accounts are:
rook-ceph): Will affect all pods created by the rook operator in the
rook-ceph): Will affect most pods in the
rook-ceph): Will affect the MGR pods in the
rook-ceph): Will affect the OSD pods in the
Since it’s the same procedure for all service accounts, here is just one example:
kubectl -n rook-ceph edit serviceaccount default
apiVersion: v1 kind: ServiceAccount metadata: name: default namespace: rook-ceph secrets: - name: default-token-12345 imagePullSecrets: # here are the new - name: my-registry-secret # parts
After doing this for all service accounts all pods should be able to pull the image from your registry.
Using Rook in Kubernetes
Now that you have a Kubernetes cluster running, you can start using Rook with these steps.
Using Rook on Tectonic Bare Metal
Follow these instructions to run Rook on Tectonic Kubernetes