Ceph

    PLEASE NOTE: This document applies to v1.8 version and not to the latest stable release v1.9

    Rook-Ceph Upgrades

    This guide will walk you through the steps to upgrade the software in a Rook-Ceph cluster from one version to the next. This includes both the Rook-Ceph operator software itself as well as the Ceph cluster software.

    Upgrades for both the operator and for Ceph are nearly entirely automated save for where Rook’s permissions need to be explicitly updated by an admin or when incompatibilities need to be addressed manually due to customizations.

    We welcome feedback and opening issues!

    Supported Versions

    This guide is for upgrading from Rook v1.7.x to Rook v1.8.x.

    Please refer to the upgrade guides from previous releases for supported upgrade paths. Rook upgrades are only supported between official releases. Upgrades to and from master are not supported.

    For a guide to upgrade previous versions of Rook, please refer to the version of documentation for those releases.

    Breaking changes in this release

    • The minimum Kubernetes version has changed to v1.16. You must update to at least Kubernetes version v1.16 before upgrading Rook from v1.7 to v1.8.

    • Rook v1.8 no longer supports Ceph Nautilus (14.2.x). Nautilus users must upgrade Ceph to Octopus (15.2.x) or Pacific (16.2.x) before upgrading to Rook v1.8.

    • Rook’s FlexVolume driver has been deprecated and removed in Rook v1.8. FlexVolume users must migrate Rook-Ceph block storage PVCs to CSI before upgrading. A migration tool has been created and is documented here.

    • The location of example manifests has changed to reduce the amount of user typing needed and to be easier to discover for new Rook users. cluster/examples/kubernetes/ceph manifests can now be found in deploy/examples.

    Considerations

    With this upgrade guide, there are a few notes to consider:

    • WARNING: Upgrading a Rook cluster is not without risk. There may be unexpected issues or obstacles that damage the integrity and health of your storage cluster, including data loss.
    • The Rook cluster’s storage may be unavailable for short periods during the upgrade process for both Rook operator updates and for Ceph version updates.
    • We recommend that you read this document in full before you undertake a Rook cluster upgrade.

    Patch Release Upgrades

    Unless otherwise noted due to extenuating requirements, upgrades from one patch release of Rook to another are as simple as updating the common resources and the image of the Rook operator. For example, when Rook v1.8.9 is released, the process of updating from v1.8.0 is as simple as running the following:

    First get the latest common resources manifests that contain the latest changes for Rook v1.8.

    git clone --single-branch --depth=1 --branch v1.8.9 https://github.com/rook/rook.git
    cd rook/deploy/examples
    

    If you have deployed the Rook Operator or the Ceph cluster into a different namespace than rook-ceph, see the Update common resources and CRDs section for instructions on how to change the default namespaces in common.yaml.

    Then apply the latest changes from v1.8 and update the Rook Operator image.

    kubectl apply -f common.yaml -f crds.yaml
    kubectl -n rook-ceph set image deploy/rook-ceph-operator rook-ceph-operator=rook/ceph:v1.8.9
    

    As exemplified above, it is a good practice to update Rook-Ceph common resources from the example manifests before any update. The common resources and CRDs might not be updated with every release, but K8s will only apply updates to the ones that changed.

    Also update optional resources like Prometheus monitoring noted more fully in the upgrade section below.

    Helm

    • The minimum supported Helm version is v3.2.0

    If you have installed Rook via the Helm chart, Helm will handle some details of the upgrade for you. The upgrade steps in this guide will clarify if Helm manages the step for you.

    The rook-ceph helm chart upgrade performs the Rook upgrade. The rook-ceph-cluster helm chart upgrade performs a Ceph upgrade if the Ceph image is updated.

    Upgrading from v1.7 to v1.8

    Rook releases from master are expressly unsupported. It is strongly recommended that you use official releases of Rook. Unreleased versions from the master branch are subject to changes and incompatibilities that will not be supported in the official releases. Builds from the master branch can have functionality changed or removed at any time without compatibility support and without prior notice.

    Prerequisites

    We will do all our work in the Ceph example manifests directory.

    $ cd $YOUR_ROOK_REPO/deploy/examples/
    

    Unless your Rook cluster was created with customized namespaces, namespaces for Rook clusters are likely to be:

    • Clusters created by v0.7 or earlier: rook-system and rook
    • Clusters created in v0.8 or v0.9: rook-ceph-system and rook-ceph
    • Clusters created in v1.0 or newer: only rook-ceph

    With this guide, we do our best not to assume the namespaces in your cluster. To make things as easy as possible, modify and use the below snippet to configure your environment. We will use these environment variables throughout this document.

    # Parameterize the environment
    export ROOK_OPERATOR_NAMESPACE="rook-ceph"
    export ROOK_CLUSTER_NAMESPACE="rook-ceph"
    

    In order to successfully upgrade a Rook cluster, the following prerequisites must be met:

    • The cluster should be in a healthy state with full functionality. Review the health verification section in order to verify your cluster is in a good starting state.
    • All pods consuming Rook storage should be created, running, and in a steady state. No Rook persistent volumes should be in the act of being created or deleted.

    Health Verification

    Before we begin the upgrade process, let’s first review some ways that you can verify the health of your cluster, ensuring that the upgrade is going smoothly after each step. Most of the health verification checks for your cluster during the upgrade process can be performed with the Rook toolbox. For more information about how to run the toolbox, please visit the Rook toolbox readme.

    See the common issues pages for troubleshooting and correcting health issues:

    Pods all Running

    In a healthy Rook cluster, the operator, the agents and all Rook namespace pods should be in the Running state and have few, if any, pod restarts. To verify this, run the following commands:

    kubectl -n $ROOK_CLUSTER_NAMESPACE get pods
    

    Status Output

    The Rook toolbox contains the Ceph tools that can give you status details of the cluster with the ceph status command. Let’s look at an output sample and review some of the details:

    TOOLS_POD=$(kubectl -n $ROOK_CLUSTER_NAMESPACE get pod -l "app=rook-ceph-tools" -o jsonpath='{.items[*].metadata.name}')
    kubectl -n $ROOK_CLUSTER_NAMESPACE exec -it $TOOLS_POD -- ceph status
    
     cluster:
       id:     a3f4d647-9538-4aff-9fd1-b845873c3fe9
       health: HEALTH_OK
    
     services:
       mon: 3 daemons, quorum b,c,a
       mgr: a(active)
       mds: myfs-1/1/1 up  {0=myfs-a=up:active}, 1 up:standby-replay
       osd: 6 osds: 6 up, 6 in
       rgw: 1 daemon active
    
     data:
       pools:   9 pools, 900 pgs
       objects: 67  objects, 11 KiB
       usage:   6.1 GiB used, 54 GiB / 60 GiB avail
       pgs:     900 active+clean
    
     io:
       client:   7.4 KiB/s rd, 681 B/s wr, 11 op/s rd, 4 op/s wr
       recovery: 164 B/s, 1 objects/s
    

    In the output above, note the following indications that the cluster is in a healthy state:

    • Cluster health: The overall cluster status is HEALTH_OK and there are no warning or error status messages displayed.
    • Monitors (mon): All of the monitors are included in the quorum list.
    • Manager (mgr): The Ceph manager is in the active state.
    • OSDs (osd): All OSDs are up and in.
    • Placement groups (pgs): All PGs are in the active+clean state.
    • (If applicable) Ceph filesystem metadata server (mds): all MDSes are active for all filesystems
    • (If applicable) Ceph object store RADOS gateways (rgw): all daemons are active

    If your ceph status output has deviations from the general good health described above, there may be an issue that needs to be investigated further. There are other commands you may run for more details on the health of the system, such as ceph osd status. See the Ceph troubleshooting docs for help.

    Rook will prevent the upgrade of the Ceph daemons if the health is in a HEALTH_ERR state. If you desired to proceed with the upgrade anyway, you will need to set either skipUpgradeChecks: true or continueUpgradeAfterChecksEvenIfNotHealthy: true as described in the cluster CR settings.

    Container Versions

    The container version running in a specific pod in the Rook cluster can be verified in its pod spec output. For example for the monitor pod mon-b, we can verify the container version it is running with the below commands:

    POD_NAME=$(kubectl -n $ROOK_CLUSTER_NAMESPACE get pod -o custom-columns=name:.metadata.name --no-headers | grep rook-ceph-mon-b)
    kubectl -n $ROOK_CLUSTER_NAMESPACE get pod ${POD_NAME} -o jsonpath='{.spec.containers[0].image}'
    

    The status and container versions for all Rook pods can be collected all at once with the following commands:

    kubectl -n $ROOK_OPERATOR_NAMESPACE get pod -o jsonpath='{range .items[*]}{.metadata.name}{"\n\t"}{.status.phase}{"\t\t"}{.spec.containers[0].image}{"\t"}{.spec.initContainers[0]}{"\n"}{end}' && \
    kubectl -n $ROOK_CLUSTER_NAMESPACE get pod -o jsonpath='{range .items[*]}{.metadata.name}{"\n\t"}{.status.phase}{"\t\t"}{.spec.containers[0].image}{"\t"}{.spec.initContainers[0].image}{"\n"}{end}'
    

    The rook-version label exists on Ceph controller resources. For various resource controllers, a summary of the resource controllers can be gained with the commands below. These will report the requested, updated, and currently available replicas for various Rook-Ceph resources in addition to the version of Rook for resources managed by the updated Rook-Ceph operator. Note that the operator and toolbox deployments do not have a rook-version label set.

    kubectl -n $ROOK_CLUSTER_NAMESPACE get deployments -o jsonpath='{range .items[*]}{.metadata.name}{"  \treq/upd/avl: "}{.spec.replicas}{"/"}{.status.updatedReplicas}{"/"}{.status.readyReplicas}{"  \trook-version="}{.metadata.labels.rook-version}{"\n"}{end}'
    
    kubectl -n $ROOK_CLUSTER_NAMESPACE get jobs -o jsonpath='{range .items[*]}{.metadata.name}{"  \tsucceeded: "}{.status.succeeded}{"      \trook-version="}{.metadata.labels.rook-version}{"\n"}{end}'
    

    Rook Volume Health

    Any pod that is using a Rook volume should also remain healthy:

    • The pod should be in the Running state with few, if any, restarts
    • There should be no errors in its logs
    • The pod should still be able to read and write to the attached Rook volume.

    Rook Operator Upgrade Process

    In the examples given in this guide, we will be upgrading a live Rook cluster running v1.7.11 to the version v1.8.9. This upgrade should work from any official patch release of Rook v1.7 to any official patch release of v1.8.

    Rook release from master are expressly unsupported. It is strongly recommended that you use official releases of Rook. Unreleased versions from the master branch are subject to changes and incompatibilities that will not be supported in the official releases. Builds from the master branch can have functionality changed or removed at any time without compatibility support and without prior notice.

    These methods should work for any number of Rook-Ceph clusters and Rook Operators as long as you parameterize the environment correctly. Merely repeat these steps for each Rook-Ceph cluster (ROOK_CLUSTER_NAMESPACE), and be sure to update the ROOK_OPERATOR_NAMESPACE parameter each time if applicable.

    Let’s get started!

    1. Update common resources and CRDs

    Automatically updated if you are upgrading via the helm chart

    First apply updates to Rook-Ceph common resources. This includes modified privileges (RBAC) needed by the Operator. Also update the Custom Resource Definitions (CRDs).

    Get the latest common resources manifests that contain the latest changes.

    git clone --single-branch --depth=1 --branch v1.8.9 https://github.com/rook/rook.git
    cd rook/deploy/examples
    

    If you have deployed the Rook Operator or the Ceph cluster into a different namespace than rook-ceph, update the common resource manifests to use your ROOK_OPERATOR_NAMESPACE and ROOK_CLUSTER_NAMESPACE using sed.

    sed -i.bak \
        -e "s/\(.*\):.*# namespace:operator/\1: $ROOK_OPERATOR_NAMESPACE # namespace:operator/g" \
        -e "s/\(.*\):.*# namespace:cluster/\1: $ROOK_CLUSTER_NAMESPACE # namespace:cluster/g" \
      common.yaml
    

    Then apply the latest changes.

    kubectl apply -f common.yaml -f crds.yaml
    

    Updates for optional resources

    If you have Prometheus monitoring enabled, follow the step to upgrade the Prometheus RBAC resources as well.

    kubectl apply -f deploy/examples/monitoring/rbac.yaml
    

    If you use the rook-ceph operator Helm chart, you should also add monitoring.enabled to your Helm values with two caveats:

    • this is unnecessary if you deploy monitoring RBAC from deploy/examples/monitoring/rbac.yaml
    • this is unnecessary if you use rook-ceph-cluster charts exclusively outside of the rook-ceph operator namespace.

    2. Update Ceph CSI versions

    Automatically updated if you are upgrading via the helm chart

    If you have specified custom CSI images in the Rook-Ceph Operator deployment, we recommended you update to use the latest Ceph-CSI drivers. See the CSI Version section for more details.

    Note: If using snapshots, refer to the Upgrade Snapshot API guide.

    3. Update the Rook Operator

    Automatically updated if you are upgrading via the helm chart

    The largest portion of the upgrade is triggered when the operator’s image is updated to v1.8.x. When the operator is updated, it will proceed to update all of the Ceph daemons.

    kubectl -n $ROOK_OPERATOR_NAMESPACE set image deploy/rook-ceph-operator rook-ceph-operator=rook/ceph:v1.8.9
    

    Admission controller

    If you use the optional Admission controller, there are additional updates during this step. The admission controller has been integrated inside the operator instead of a separate deployment. This means that the webhook server certificates are now stored in the operator, and the operator manifest must be updated to use the one provided in deploy/examples/operator.yaml. If you are using Helm to manage the deployment, this is handled automatically.

    When updating the operator deployment with the latest example from Rook, there is risk of overwriting changes if you have customized the operator deployment or to the rook-ceph-operator-config ConfigMap. We suggest that you remove the ConfigMap from operator.yaml before moving on. Additionally, we encourage you to diff the current deployment and the latest one to be sure any changes you may have made don’t get overwritten. Required changes include the webhook-cert volume/mount and https-webhook port, though there are some smaller changes as well.

    Once you are sure any custom modifications to your operator deployment won’t be overwritten, apply the new operator.yaml with kubectl apply -f deploy/examples/operator.yaml.

    4. Wait for the upgrade to complete

    Watch now in amazement as the Ceph mons, mgrs, OSDs, rbd-mirrors, MDSes and RGWs are terminated and replaced with updated versions in sequence. The cluster may be offline very briefly as mons update, and the Ceph Filesystem may fall offline a few times while the MDSes are upgrading. This is normal.

    The versions of the components can be viewed as they are updated:

    watch --exec kubectl -n $ROOK_CLUSTER_NAMESPACE get deployments -l rook_cluster=$ROOK_CLUSTER_NAMESPACE -o jsonpath='{range .items[*]}{.metadata.name}{"  \treq/upd/avl: "}{.spec.replicas}{"/"}{.status.updatedReplicas}{"/"}{.status.readyReplicas}{"  \trook-version="}{.metadata.labels.rook-version}{"\n"}{end}'
    

    As an example, this cluster is midway through updating the OSDs. When all deployments report 1/1/1 availability and rook-version=v1.8.9, the Ceph cluster’s core components are fully updated.

    Every 2.0s: kubectl -n rook-ceph get deployment -o j...
    
    rook-ceph-mgr-a         req/upd/avl: 1/1/1      rook-version=v1.8.9
    rook-ceph-mon-a         req/upd/avl: 1/1/1      rook-version=v1.8.9
    rook-ceph-mon-b         req/upd/avl: 1/1/1      rook-version=v1.8.9
    rook-ceph-mon-c         req/upd/avl: 1/1/1      rook-version=v1.8.9
    rook-ceph-osd-0         req/upd/avl: 1//        rook-version=v1.8.9
    rook-ceph-osd-1         req/upd/avl: 1/1/1      rook-version=v1.7.11
    rook-ceph-osd-2         req/upd/avl: 1/1/1      rook-version=v1.7.11
    

    An easy check to see if the upgrade is totally finished is to check that there is only one rook-version reported across the cluster.

    # kubectl -n $ROOK_CLUSTER_NAMESPACE get deployment -l rook_cluster=$ROOK_CLUSTER_NAMESPACE -o jsonpath='{range .items[*]}{"rook-version="}{.metadata.labels.rook-version}{"\n"}{end}' | sort | uniq
    This cluster is not yet finished:
      rook-version=v1.7.11
      rook-version=v1.8.9
    This cluster is finished:
      rook-version=v1.8.9
    

    5. Verify the updated cluster

    At this point, your Rook operator should be running version rook/ceph:v1.8.9.

    Verify the Ceph cluster’s health using the health verification section.

    Ceph Version Upgrades

    Rook v1.8 supports the following Ceph versions:

    • Ceph Pacific 16.2.0 or newer
    • Ceph Octopus v15.2.0 or newer

    These are the only supported versions of Ceph. Rook v1.8 no longer supports Ceph Nautilus (14.2.x). Nautilus users must upgrade Ceph to Octopus (15.2.x) or Pacific (16.2.x) before upgrading to Rook v1.8.

    IMPORTANT: When an update is requested, the operator will check Ceph’s status, if it is in HEALTH_ERR it will refuse to do the upgrade.

    Rook is cautious when performing upgrades. When an upgrade is requested (the Ceph image has been updated in the CR), Rook will go through all the daemons one by one and will individually perform checks on them. It will make sure a particular daemon can be stopped before performing the upgrade. Once the deployment has been updated, it checks if this is ok to continue. After each daemon is updated we wait for things to settle (monitors to be in a quorum, PGs to be clean for OSDs, up for MDSes, etc.), then only when the condition is met we move to the next daemon. We repeat this process until all the daemons have been updated.

    Disable bluestore_fsck_quick_fix_on_mount

    WARNING: There is a notice from Ceph for users upgrading to Ceph Pacific v16.2.6 or lower from an earlier major version of Ceph. If you are upgrading to Ceph Pacific (v16), please upgrade to v16.2.7 or higher if possible.

    If you must upgrade to a version lower than v16.2.7, ensure that all instances of bluestore_fsck_quick_fix_on_mount in Rook-Ceph configs are removed.

    First, Ensure no references to bluestore_fsck_quick_fix_on_mount are present in the rook-config-override ConfigMap. Remove them if they exist.

    Finally, ensure no references to bluestore_fsck_quick_fix_on_mount are present in Ceph’s internal configuration. Run all commands below from the toolbox.

    In the example below, two instances of bluestore_fsck_quick_fix_on_mount are present and are commented, and some output text has been removed for brevity.

    ceph config-key dump
    
    {
        "config/global/bluestore_fsck_quick_fix_on_mount": "false",       # <-- FALSE
        "config/global/osd_scrub_auto_repair": "true",
        "config/mgr.a/mgr/dashboard/server_port": "7000",
        "config/mgr/mgr/balancer/active": "true",
        "config/osd/bluestore_fsck_quick_fix_on_mount": "true",           # <-- TRUE
    }
    

    Remove the configs for both with the commands below. Note how the config/... paths correspond to the output above.

    ceph config-key rm config/global/bluestore_fsck_quick_fix_on_mount
    ceph config-key rm config/osd/bluestore_fsck_quick_fix_on_mount
    

    It’s best to run ceph config-key dump again to verify references to bluestore_fsck_quick_fix_on_mount are gone after this.

    See for more information, see here: https://github.com/rook/rook/issues/9185

    Important consideration for CephNFS users

    Users of CephNFS need to take additional steps to upgrade Ceph versions. Please see the NFS documentation for full details.

    Ceph images

    Official Ceph container images can be found on Quay. Prior to August 2021, official images were on docker.io. While those images will remain on Docker Hub, all new images are being pushed to Quay.

    These images are tagged in a few ways:

    • The most explicit form of tags are full-ceph-version-and-build tags (e.g., v16.2.7-20220216). These tags are recommended for production clusters, as there is no possibility for the cluster to be heterogeneous with respect to the version of Ceph running in containers.
    • Ceph major version tags (e.g., v16) are useful for development and test clusters so that the latest version of Ceph is always available.

    Ceph containers other than the official images from the registry above will not be supported.

    Example upgrade to Ceph Pacific

    1. Update the main Ceph daemons

    The majority of the upgrade will be handled by the Rook operator. Begin the upgrade by changing the Ceph image field in the cluster CRD (spec.cephVersion.image).

    NEW_CEPH_IMAGE='quay.io/ceph/ceph:v16.2.7-20220216'
    CLUSTER_NAME="$ROOK_CLUSTER_NAMESPACE"  # change if your cluster name is not the Rook namespace
    kubectl -n $ROOK_CLUSTER_NAMESPACE patch CephCluster $CLUSTER_NAME --type=merge -p "{\"spec\": {\"cephVersion\": {\"image\": \"$NEW_CEPH_IMAGE\"}}}"
    

    2. Wait for the daemon pod updates to complete

    As with upgrading Rook, you must now wait for the upgrade to complete. Status can be determined in a similar way to the Rook upgrade as well.

    watch --exec kubectl -n $ROOK_CLUSTER_NAMESPACE get deployments -l rook_cluster=$ROOK_CLUSTER_NAMESPACE -o jsonpath='{range .items[*]}{.metadata.name}{"  \treq/upd/avl: "}{.spec.replicas}{"/"}{.status.updatedReplicas}{"/"}{.status.readyReplicas}{"  \tceph-version="}{.metadata.labels.ceph-version}{"\n"}{end}'
    

    Determining when the Ceph has fully updated is rather simple.

    kubectl -n $ROOK_CLUSTER_NAMESPACE get deployment -l rook_cluster=$ROOK_CLUSTER_NAMESPACE -o jsonpath='{range .items[*]}{"ceph-version="}{.metadata.labels.ceph-version}{"\n"}{end}' | sort | uniq
    This cluster is not yet finished:
        ceph-version=15.2.13-0
        ceph-version=16.2.6-0
    This cluster is finished:
        ceph-version=16.2.6-0
    

    3. Verify the updated cluster

    Verify the Ceph cluster’s health using the health verification section.

    CSI Version

    If you have a cluster running with CSI drivers enabled and you want to configure Rook to use non-default CSI images, the following settings will need to be applied for the desired version of CSI.

    The operator configuration variables have recently moved from the operator deployment to the rook-ceph-operator-config ConfigMap. The values in the operator deployment can still be set, but if the ConfigMap settings are applied, they will override the operator deployment settings. The ConfigMap must exist for the operator to work correctly, even if all configuration is supplied through the environment.

    kubectl -n $ROOK_OPERATOR_NAMESPACE edit configmap rook-ceph-operator-config
    

    The default upstream images are included below, which you can change to your desired images.

    ROOK_CSI_CEPH_IMAGE: "quay.io/cephcsi/cephcsi:v3.5.1"
    ROOK_CSI_REGISTRAR_IMAGE: "k8s.gcr.io/sig-storage/csi-node-driver-registrar:v2.5.0"
    ROOK_CSI_PROVISIONER_IMAGE: "k8s.gcr.io/sig-storage/csi-provisioner:v3.1.0"
    ROOK_CSI_ATTACHER_IMAGE: "k8s.gcr.io/sig-storage/csi-attacher:v3.4.0"
    ROOK_CSI_RESIZER_IMAGE: "k8s.gcr.io/sig-storage/csi-resizer:v1.4.0"
    ROOK_CSI_SNAPSHOTTER_IMAGE: "k8s.gcr.io/sig-storage/csi-snapshotter:v5.0.1"
    CSI_VOLUME_REPLICATION_IMAGE: "quay.io/csiaddons/volumereplication-operator:v0.3.0"
    ROOK_CSIADDONS_IMAGE: "quay.io/csiaddons/k8s-sidecar:v0.2.1"
    

    Use default images

    If you would like Rook to use the inbuilt default upstream images, then you may simply remove all variables matching ROOK_CSI_*_IMAGE from the above ConfigMap and/or the operator deployment.

    Verifying updates

    You can use the below command to see the CSI images currently being used in the cluster. Note that not all images (like volumereplication-operator) may be present in every cluster depending on which CSI features are enabled.

    kubectl --namespace rook-ceph get pod -o jsonpath='{range .items[*]}{range .spec.containers[*]}{.image}{"\n"}' -l 'app in (csi-rbdplugin,csi-rbdplugin-provisioner,csi-cephfsplugin,csi-cephfsplugin-provisioner)' | sort | uniq
    
    k8s.gcr.io/sig-storage/csi-attacher:v3.4.0
    k8s.gcr.io/sig-storage/csi-node-driver-registrar:v2.5.0
    k8s.gcr.io/sig-storage/csi-provisioner:v3.1.0
    k8s.gcr.io/sig-storage/csi-resizer:v1.4.0
    k8s.gcr.io/sig-storage/csi-snapshotter:v5.0.1
    quay.io/cephcsi/cephcsi:v3.5.1
    quay.io/csiaddons/volumereplication-operator:v0.3.0
    quay.io/csiaddons/k8s-sidecar:v0.2.1