From 29b936ab9d52e187ac77dad349c8f07d6b5dc2a5 Mon Sep 17 00:00:00 2001 From: Blaine Gardner Date: Thu, 2 Dec 2021 14:32:13 -0700 Subject: [PATCH] build: set the release version to v1.7.9 Update examples and documentation to v1.7.9. Signed-off-by: Blaine Gardner --- Documentation/ceph-monitoring.md | 2 +- Documentation/ceph-toolbox.md | 6 ++-- Documentation/ceph-upgrade.md | 30 +++++++++---------- Documentation/flex-to-csi-migration.md | 2 +- Documentation/quickstart.md | 2 +- .../kubernetes/ceph/direct-mount.yaml | 2 +- .../kubernetes/ceph/flex-migrator.yaml | 2 +- cluster/examples/kubernetes/ceph/images.txt | 2 +- .../kubernetes/ceph/operator-openshift.yaml | 2 +- .../examples/kubernetes/ceph/operator.yaml | 2 +- .../examples/kubernetes/ceph/osd-purge.yaml | 2 +- .../examples/kubernetes/ceph/toolbox-job.yaml | 4 +-- cluster/examples/kubernetes/ceph/toolbox.yaml | 2 +- tests/scripts/github-action-helper.sh | 2 +- 14 files changed, 31 insertions(+), 31 deletions(-) diff --git a/Documentation/ceph-monitoring.md b/Documentation/ceph-monitoring.md index 41078bf51cfc..6510ce5a9d09 100644 --- a/Documentation/ceph-monitoring.md +++ b/Documentation/ceph-monitoring.md @@ -38,7 +38,7 @@ With the Prometheus operator running, we can create a service monitor that will From the root of your locally cloned Rook repo, go the monitoring directory: ```console -$ git clone --single-branch --branch v1.7.8 https://github.com/rook/rook.git +$ git clone --single-branch --branch v1.7.9 https://github.com/rook/rook.git cd rook/cluster/examples/kubernetes/ceph/monitoring ``` diff --git a/Documentation/ceph-toolbox.md b/Documentation/ceph-toolbox.md index 3ae509c3dbc3..4775dc7cdc49 100644 --- a/Documentation/ceph-toolbox.md +++ b/Documentation/ceph-toolbox.md @@ -43,7 +43,7 @@ spec: dnsPolicy: ClusterFirstWithHostNet containers: - name: rook-ceph-tools - image: rook/ceph:v1.7.8 + image: rook/ceph:v1.7.9 command: ["/tini"] args: ["-g", "--", "/usr/local/bin/toolbox.sh"] imagePullPolicy: IfNotPresent @@ -133,7 +133,7 @@ spec: spec: initContainers: - name: config-init - image: rook/ceph:v1.7.8 + image: rook/ceph:v1.7.9 command: ["/usr/local/bin/toolbox.sh"] args: ["--skip-watch"] imagePullPolicy: IfNotPresent @@ -155,7 +155,7 @@ spec: mountPath: /etc/rook containers: - name: script - image: rook/ceph:v1.7.8 + image: rook/ceph:v1.7.9 volumeMounts: - mountPath: /etc/ceph name: ceph-config diff --git a/Documentation/ceph-upgrade.md b/Documentation/ceph-upgrade.md index 232f2772bbe5..b367f93dced5 100644 --- a/Documentation/ceph-upgrade.md +++ b/Documentation/ceph-upgrade.md @@ -53,12 +53,12 @@ With this upgrade guide, there are a few notes to consider: Unless otherwise noted due to extenuating requirements, upgrades from one patch release of Rook to another are as simple as updating the common resources and the image of the Rook operator. For -example, when Rook v1.7.8 is released, the process of updating from v1.7.0 is as simple as running +example, when Rook v1.7.9 is released, the process of updating from v1.7.0 is as simple as running the following: First get the latest common resources manifests that contain the latest changes for Rook v1.7. ```sh -git clone --single-branch --depth=1 --branch v1.7.8 https://github.com/rook/rook.git +git clone --single-branch --depth=1 --branch v1.7.9 https://github.com/rook/rook.git cd rook/cluster/examples/kubernetes/ceph ``` @@ -75,7 +75,7 @@ section for instructions on how to change the default namespaces in `common.yaml Then apply the latest changes from v1.7 and update the Rook Operator image. ```console kubectl apply -f common.yaml -f crds.yaml -kubectl -n rook-ceph set image deploy/rook-ceph-operator rook-ceph-operator=rook/ceph:v1.7.8 +kubectl -n rook-ceph set image deploy/rook-ceph-operator rook-ceph-operator=rook/ceph:v1.7.9 ``` As exemplified above, it is a good practice to update Rook-Ceph common resources from the example @@ -261,7 +261,7 @@ Any pod that is using a Rook volume should also remain healthy: ## Rook Operator Upgrade Process In the examples given in this guide, we will be upgrading a live Rook cluster running `v1.6.8` to -the version `v1.7.8`. This upgrade should work from any official patch release of Rook v1.6 to any +the version `v1.7.9`. This upgrade should work from any official patch release of Rook v1.6 to any official patch release of v1.7. **Rook release from `master` are expressly unsupported.** It is strongly recommended that you use @@ -291,7 +291,7 @@ needed by the Operator. Also update the Custom Resource Definitions (CRDs). First get the latest common resources manifests that contain the latest changes. ```sh -git clone --single-branch --depth=1 --branch v1.7.8 https://github.com/rook/rook.git +git clone --single-branch --depth=1 --branch v1.7.9 https://github.com/rook/rook.git cd rook/cluster/examples/kubernetes/ceph ``` @@ -337,7 +337,7 @@ The largest portion of the upgrade is triggered when the operator's image is upd When the operator is updated, it will proceed to update all of the Ceph daemons. ```sh -kubectl -n $ROOK_OPERATOR_NAMESPACE set image deploy/rook-ceph-operator rook-ceph-operator=rook/ceph:v1.7.8 +kubectl -n $ROOK_OPERATOR_NAMESPACE set image deploy/rook-ceph-operator rook-ceph-operator=rook/ceph:v1.7.9 ``` ### **4. Wait for the upgrade to complete** @@ -353,16 +353,16 @@ watch --exec kubectl -n $ROOK_CLUSTER_NAMESPACE get deployments -l rook_cluster= ``` As an example, this cluster is midway through updating the OSDs. When all deployments report `1/1/1` -availability and `rook-version=v1.7.8`, the Ceph cluster's core components are fully updated. +availability and `rook-version=v1.7.9`, the Ceph cluster's core components are fully updated. >``` >Every 2.0s: kubectl -n rook-ceph get deployment -o j... > ->rook-ceph-mgr-a req/upd/avl: 1/1/1 rook-version=v1.7.8 ->rook-ceph-mon-a req/upd/avl: 1/1/1 rook-version=v1.7.8 ->rook-ceph-mon-b req/upd/avl: 1/1/1 rook-version=v1.7.8 ->rook-ceph-mon-c req/upd/avl: 1/1/1 rook-version=v1.7.8 ->rook-ceph-osd-0 req/upd/avl: 1// rook-version=v1.7.8 +>rook-ceph-mgr-a req/upd/avl: 1/1/1 rook-version=v1.7.9 +>rook-ceph-mon-a req/upd/avl: 1/1/1 rook-version=v1.7.9 +>rook-ceph-mon-b req/upd/avl: 1/1/1 rook-version=v1.7.9 +>rook-ceph-mon-c req/upd/avl: 1/1/1 rook-version=v1.7.9 +>rook-ceph-osd-0 req/upd/avl: 1// rook-version=v1.7.9 >rook-ceph-osd-1 req/upd/avl: 1/1/1 rook-version=v1.6.8 >rook-ceph-osd-2 req/upd/avl: 1/1/1 rook-version=v1.6.8 >``` @@ -374,14 +374,14 @@ An easy check to see if the upgrade is totally finished is to check that there i # kubectl -n $ROOK_CLUSTER_NAMESPACE get deployment -l rook_cluster=$ROOK_CLUSTER_NAMESPACE -o jsonpath='{range .items[*]}{"rook-version="}{.metadata.labels.rook-version}{"\n"}{end}' | sort | uniq This cluster is not yet finished: rook-version=v1.6.8 - rook-version=v1.7.8 + rook-version=v1.7.9 This cluster is finished: - rook-version=v1.7.8 + rook-version=v1.7.9 ``` ### **5. Verify the updated cluster** -At this point, your Rook operator should be running version `rook/ceph:v1.7.8`. +At this point, your Rook operator should be running version `rook/ceph:v1.7.9`. Verify the Ceph cluster's health using the [health verification section](#health-verification). diff --git a/Documentation/flex-to-csi-migration.md b/Documentation/flex-to-csi-migration.md index 6ff18caca5f2..c26cee39cf94 100644 --- a/Documentation/flex-to-csi-migration.md +++ b/Documentation/flex-to-csi-migration.md @@ -14,7 +14,7 @@ The tool [persistent-volume-migrator](https://github.com/ceph/persistent-volume- ## Migration Preparation -1. Rook v1.7.8 is required. If you have a previous version of Rook running, follow the [upgrade guide](https://rook.io/docs/rook/v1.7/ceph-upgrade.html) to upgrade from previous releases until on v1.7.8. +1. Rook v1.7.9 is required. If you have a previous version of Rook running, follow the [upgrade guide](https://rook.io/docs/rook/v1.7/ceph-upgrade.html) to upgrade from previous releases until on v1.7.9. 2. Enable the CSI driver if not already enabled. See the [operator settings](https://github.com/rook/rook/blob/release-1.7/cluster/examples/kubernetes/ceph/operator.yaml#L29-L32) such as `ROOK_CSI_ENABLE_RBD`. 3. Confirm the Rook-Ceph cluster is healthy (`ceph status` shows `health: OK`) 4. Create the CSI storage class to which you want to migrate diff --git a/Documentation/quickstart.md b/Documentation/quickstart.md index 51910b1ff496..8273535f5891 100644 --- a/Documentation/quickstart.md +++ b/Documentation/quickstart.md @@ -36,7 +36,7 @@ In order to configure the Ceph storage cluster, at least one of these local stor A simple Rook cluster can be created with the following kubectl commands and [example manifests](https://github.com/rook/rook/blob/{{ branchName }}/cluster/examples/kubernetes/ceph). ```console -$ git clone --single-branch --branch v1.7.8 https://github.com/rook/rook.git +$ git clone --single-branch --branch v1.7.9 https://github.com/rook/rook.git cd rook/cluster/examples/kubernetes/ceph kubectl create -f crds.yaml -f common.yaml -f operator.yaml kubectl create -f cluster.yaml diff --git a/cluster/examples/kubernetes/ceph/direct-mount.yaml b/cluster/examples/kubernetes/ceph/direct-mount.yaml index 64a466b8ea8d..da2dc16f592a 100644 --- a/cluster/examples/kubernetes/ceph/direct-mount.yaml +++ b/cluster/examples/kubernetes/ceph/direct-mount.yaml @@ -18,7 +18,7 @@ spec: dnsPolicy: ClusterFirstWithHostNet containers: - name: rook-direct-mount - image: rook/ceph:v1.7.8 + image: rook/ceph:v1.7.9 command: ["/tini"] args: ["-g", "--", "/usr/local/bin/toolbox.sh"] imagePullPolicy: IfNotPresent diff --git a/cluster/examples/kubernetes/ceph/flex-migrator.yaml b/cluster/examples/kubernetes/ceph/flex-migrator.yaml index 285c03db7ad7..6cabdef0a43b 100644 --- a/cluster/examples/kubernetes/ceph/flex-migrator.yaml +++ b/cluster/examples/kubernetes/ceph/flex-migrator.yaml @@ -116,7 +116,7 @@ spec: dnsPolicy: ClusterFirstWithHostNet containers: - name: rook-ceph-migrator - image: rook/ceph:v1.7.8 + image: rook/ceph:v1.7.9 command: ["/tini"] args: ["-g", "--", "/usr/local/bin/toolbox.sh"] imagePullPolicy: IfNotPresent diff --git a/cluster/examples/kubernetes/ceph/images.txt b/cluster/examples/kubernetes/ceph/images.txt index 0a54ce1a6225..12c43e426ed8 100644 --- a/cluster/examples/kubernetes/ceph/images.txt +++ b/cluster/examples/kubernetes/ceph/images.txt @@ -6,4 +6,4 @@ quay.io/ceph/ceph:v16.2.6 quay.io/cephcsi/cephcsi:v3.4.0 quay.io/csiaddons/volumereplication-operator:v0.1.0 - rook/ceph:v1.7.8 + rook/ceph:v1.7.9 diff --git a/cluster/examples/kubernetes/ceph/operator-openshift.yaml b/cluster/examples/kubernetes/ceph/operator-openshift.yaml index ab55a3fab517..6df6f94f698a 100644 --- a/cluster/examples/kubernetes/ceph/operator-openshift.yaml +++ b/cluster/examples/kubernetes/ceph/operator-openshift.yaml @@ -446,7 +446,7 @@ spec: serviceAccountName: rook-ceph-system containers: - name: rook-ceph-operator - image: rook/ceph:v1.7.8 + image: rook/ceph:v1.7.9 args: ["ceph", "operator"] volumeMounts: - mountPath: /var/lib/rook diff --git a/cluster/examples/kubernetes/ceph/operator.yaml b/cluster/examples/kubernetes/ceph/operator.yaml index a80dfef688d5..2e77bfeb58e1 100644 --- a/cluster/examples/kubernetes/ceph/operator.yaml +++ b/cluster/examples/kubernetes/ceph/operator.yaml @@ -369,7 +369,7 @@ spec: serviceAccountName: rook-ceph-system containers: - name: rook-ceph-operator - image: rook/ceph:v1.7.8 + image: rook/ceph:v1.7.9 args: ["ceph", "operator"] volumeMounts: - mountPath: /var/lib/rook diff --git a/cluster/examples/kubernetes/ceph/osd-purge.yaml b/cluster/examples/kubernetes/ceph/osd-purge.yaml index ed9e121ba3b7..2d967c19289b 100644 --- a/cluster/examples/kubernetes/ceph/osd-purge.yaml +++ b/cluster/examples/kubernetes/ceph/osd-purge.yaml @@ -25,7 +25,7 @@ spec: serviceAccountName: rook-ceph-purge-osd containers: - name: osd-removal - image: rook/ceph:v1.7.8 + image: rook/ceph:v1.7.9 # TODO: Insert the OSD ID in the last parameter that is to be removed # The OSD IDs are a comma-separated list. For example: "0" or "0,2". # If you want to preserve the OSD PVCs, set `--preserve-pvc true`. diff --git a/cluster/examples/kubernetes/ceph/toolbox-job.yaml b/cluster/examples/kubernetes/ceph/toolbox-job.yaml index 8b9efb297bfc..7521cb71f66f 100644 --- a/cluster/examples/kubernetes/ceph/toolbox-job.yaml +++ b/cluster/examples/kubernetes/ceph/toolbox-job.yaml @@ -10,7 +10,7 @@ spec: spec: initContainers: - name: config-init - image: rook/ceph:v1.7.8 + image: rook/ceph:v1.7.9 command: ["/usr/local/bin/toolbox.sh"] args: ["--skip-watch"] imagePullPolicy: IfNotPresent @@ -32,7 +32,7 @@ spec: mountPath: /etc/rook containers: - name: script - image: rook/ceph:v1.7.8 + image: rook/ceph:v1.7.9 volumeMounts: - mountPath: /etc/ceph name: ceph-config diff --git a/cluster/examples/kubernetes/ceph/toolbox.yaml b/cluster/examples/kubernetes/ceph/toolbox.yaml index 463504766a7d..01083b8c120f 100644 --- a/cluster/examples/kubernetes/ceph/toolbox.yaml +++ b/cluster/examples/kubernetes/ceph/toolbox.yaml @@ -18,7 +18,7 @@ spec: dnsPolicy: ClusterFirstWithHostNet containers: - name: rook-ceph-tools - image: rook/ceph:v1.7.8 + image: rook/ceph:v1.7.9 command: ["/tini"] args: ["-g", "--", "/usr/local/bin/toolbox.sh"] imagePullPolicy: IfNotPresent diff --git a/tests/scripts/github-action-helper.sh b/tests/scripts/github-action-helper.sh index 3b76261a5112..afce08b08d33 100755 --- a/tests/scripts/github-action-helper.sh +++ b/tests/scripts/github-action-helper.sh @@ -190,7 +190,7 @@ function create_cluster_prerequisites() { } function deploy_manifest_with_local_build() { - sed -i "s|image: rook/ceph:v1.7.8|image: rook/ceph:local-build|g" $1 + sed -i "s|image: rook/ceph:v1.7.9|image: rook/ceph:local-build|g" $1 kubectl create -f $1 }