Documentation

    PLEASE NOTE: This document applies to v0.7 version and not to the latest stable release v1.3

    Documentation for other releases can be found by using the version selector in the left bottom of any doc page.

    Monitoring

    Each Rook cluster has some built in metrics collectors/exporters for monitoring with Prometheus. To enable monitoring of Rook in your Kubernetes cluster, you can follow the steps below.

    Note that these steps work only for Kubernetes 1.7.x or higher. For Kubernetes 1.6.x or older, refer to the Rook 0.5.0 documentation and use these manifest files.

    Prometheus Operator

    First the Prometheus operator needs to be started in the cluster so it can watch for our requests to start monitoring Rook and respond by deploying the correct Prometheus pods and configuration. A full explanation can be found in the Prometheus operator repository on github, but the quick instructions can be found here:

    kubectl apply -f https://raw.githubusercontent.com/coreos/prometheus-operator/release-0.15/bundle.yaml
    

    This will start the Prometheus operator, but before moving on, wait until the operator is in the Running state:

    kubectl get pod
    

    Once the Prometheus operator is in the Running state, proceed to the next section.

    Prometheus Instances

    With the Prometheus operator running, we can create a service monitor that will watch the Rook cluster and collect metrics regularly. From the root of your locally cloned Rook repo, go the monitoring directory:

    cd cluster/examples/kubernetes/monitoring
    

    Create the service monitor as well as the Prometheus server pod and service:

    kubectl create -f service-monitor.yaml
    kubectl create -f prometheus.yaml
    kubectl create -f prometheus-service.yaml
    

    Ensure that the Prometheus server pod gets created and advances to the Running state before moving on:

    kubectl -n rook get pod prometheus-rook-prometheus-0
    

    Prometheus Web Console

    Once the Prometheus server is running, you can open a web browser and go to the URL that is output from this command:

    echo "http://$(kubectl -n rook -o jsonpath={.status.hostIP} get pod prometheus-rook-prometheus-0):30900"
    

    You should now see the Prometheus monitoring website. Click on Graph in the top navigation bar. In the dropdown that says insert metric at cursor, select any metric you would like to see, for example ceph_cluster_used_bytes, followed by clicking on the Execute button. Below the Execute button, ensure the Graph tab is selected and you should now see a graph of your chosen metric over time.

    Prometheus Consoles

    You can find Prometheus Consoles here: https://github.com/ceph/cephmetrics/tree/master/dashboards/current. A guide to how you can write your own Prometheus consoles can be found on the official Prometheus site here: https://prometheus.io/docs/visualization/consoles/

    Grafana Dashboards

    Currently there are no official Rook dashboards, which will change soon, but you can find some here: https://github.com/SUSE/DeepSea/tree/master/srv/salt/ceph/monitoring/grafana/files To use the dashboards, just download the JSON file(s) and in Grafan import them by uploading the JSON file.

    Teardown

    To clean up all the artifacts created by the monitoring walkthrough, copy/paste the entire block below (note that errors about resources “not found” can be ignored):

    kubectl delete -f service-monitor.yaml
    kubectl delete -f prometheus.yaml
    kubectl delete -f prometheus-service.yaml
    kubectl -n rook delete statefulset prometheus-rook-prometheus
    kubectl delete -f https://raw.githubusercontent.com/coreos/prometheus-operator/release-0.8/bundle.yaml
    

    Then the rest of the instructions in the Prometheus Operator docs can be followed to finish cleaning up.

    Special Cases

    Tectonic Bare Metal

    Tectonic strongly discourages the tectonic-system Prometheus instance to be used outside their intentions, so you need to create a new Prometheus Operator yourself. After this you only need to create the service monitor as stated above.