From f4baa8f8fc388e7df49543717d3cec9271676036 Mon Sep 17 00:00:00 2001 From: Travis Nielsen Date: Wed, 3 Apr 2024 13:44:30 -0600 Subject: [PATCH] build: update release version to v1.13.8 For the patch release update the examples and docs to v1.13.8 Signed-off-by: Travis Nielsen --- Documentation/Getting-Started/quickstart.md | 2 +- .../Monitoring/ceph-monitoring.md | 2 +- Documentation/Upgrade/rook-upgrade.md | 30 +++++++++---------- deploy/charts/rook-ceph/values.yaml | 2 +- deploy/examples/direct-mount.yaml | 2 +- deploy/examples/images.txt | 2 +- deploy/examples/multus-validation.yaml | 2 +- deploy/examples/operator-openshift.yaml | 2 +- deploy/examples/operator.yaml | 2 +- deploy/examples/osd-purge.yaml | 2 +- deploy/examples/toolbox-job.yaml | 4 +-- deploy/examples/toolbox-operator-image.yaml | 2 +- 12 files changed, 27 insertions(+), 27 deletions(-) diff --git a/Documentation/Getting-Started/quickstart.md b/Documentation/Getting-Started/quickstart.md index ba1fae80707a..eaac478243d3 100644 --- a/Documentation/Getting-Started/quickstart.md +++ b/Documentation/Getting-Started/quickstart.md @@ -36,7 +36,7 @@ To configure the Ceph storage cluster, at least one of these local storage optio A simple Rook cluster is created for Kubernetes with the following `kubectl` commands and [example manifests](https://github.com/rook/rook/blob/master/deploy/examples). ```console -$ git clone --single-branch --branch v1.13.7 https://github.com/rook/rook.git +$ git clone --single-branch --branch v1.13.8 https://github.com/rook/rook.git cd rook/deploy/examples kubectl create -f crds.yaml -f common.yaml -f operator.yaml kubectl create -f cluster.yaml diff --git a/Documentation/Storage-Configuration/Monitoring/ceph-monitoring.md b/Documentation/Storage-Configuration/Monitoring/ceph-monitoring.md index 55b9c74e4e96..c3f82d8e3bd9 100644 --- a/Documentation/Storage-Configuration/Monitoring/ceph-monitoring.md +++ b/Documentation/Storage-Configuration/Monitoring/ceph-monitoring.md @@ -44,7 +44,7 @@ There are two sources for metrics collection: From the root of your locally cloned Rook repo, go the monitoring directory: ```console -$ git clone --single-branch --branch v1.13.7 https://github.com/rook/rook.git +$ git clone --single-branch --branch v1.13.8 https://github.com/rook/rook.git cd rook/deploy/examples/monitoring ``` diff --git a/Documentation/Upgrade/rook-upgrade.md b/Documentation/Upgrade/rook-upgrade.md index f533af1414c2..244d28d455ab 100644 --- a/Documentation/Upgrade/rook-upgrade.md +++ b/Documentation/Upgrade/rook-upgrade.md @@ -78,11 +78,11 @@ With this upgrade guide, there are a few notes to consider: Unless otherwise noted due to extenuating requirements, upgrades from one patch release of Rook to another are as simple as updating the common resources and the image of the Rook operator. For -example, when Rook v1.13.7 is released, the process of updating from v1.13.0 is as simple as running +example, when Rook v1.13.8 is released, the process of updating from v1.13.0 is as simple as running the following: ```console -git clone --single-branch --depth=1 --branch v1.13.7 https://github.com/rook/rook.git +git clone --single-branch --depth=1 --branch v1.13.8 https://github.com/rook/rook.git cd rook/deploy/examples ``` @@ -94,7 +94,7 @@ Then, apply the latest changes from v1.13, and update the Rook Operator image. ```console kubectl apply -f common.yaml -f crds.yaml -kubectl -n rook-ceph set image deploy/rook-ceph-operator rook-ceph-operator=rook/ceph:v1.13.7 +kubectl -n rook-ceph set image deploy/rook-ceph-operator rook-ceph-operator=rook/ceph:v1.13.8 ``` As exemplified above, it is a good practice to update Rook common resources from the example @@ -129,7 +129,7 @@ In order to successfully upgrade a Rook cluster, the following prerequisites mus ## Rook Operator Upgrade The examples given in this guide upgrade a live Rook cluster running `v1.12.11` to -the version `v1.13.7`. This upgrade should work from any official patch release of Rook v1.12 to any +the version `v1.13.8`. This upgrade should work from any official patch release of Rook v1.12 to any official patch release of v1.13. Let's get started! @@ -156,7 +156,7 @@ by the Operator. Also update the Custom Resource Definitions (CRDs). Get the latest common resources manifests that contain the latest changes. ```console -git clone --single-branch --depth=1 --branch v1.13.7 https://github.com/rook/rook.git +git clone --single-branch --depth=1 --branch v1.13.8 https://github.com/rook/rook.git cd rook/deploy/examples ``` @@ -195,7 +195,7 @@ The largest portion of the upgrade is triggered when the operator's image is upd When the operator is updated, it will proceed to update all of the Ceph daemons. ```console -kubectl -n $ROOK_OPERATOR_NAMESPACE set image deploy/rook-ceph-operator rook-ceph-operator=rook/ceph:v1.13.7 +kubectl -n $ROOK_OPERATOR_NAMESPACE set image deploy/rook-ceph-operator rook-ceph-operator=rook/ceph:v1.13.8 ``` ### **3. Update Ceph CSI** @@ -225,16 +225,16 @@ watch --exec kubectl -n $ROOK_CLUSTER_NAMESPACE get deployments -l rook_cluster= ``` As an example, this cluster is midway through updating the OSDs. When all deployments report `1/1/1` -availability and `rook-version=v1.13.7`, the Ceph cluster's core components are fully updated. +availability and `rook-version=v1.13.8`, the Ceph cluster's core components are fully updated. ```console Every 2.0s: kubectl -n rook-ceph get deployment -o j... -rook-ceph-mgr-a req/upd/avl: 1/1/1 rook-version=v1.13.7 -rook-ceph-mon-a req/upd/avl: 1/1/1 rook-version=v1.13.7 -rook-ceph-mon-b req/upd/avl: 1/1/1 rook-version=v1.13.7 -rook-ceph-mon-c req/upd/avl: 1/1/1 rook-version=v1.13.7 -rook-ceph-osd-0 req/upd/avl: 1// rook-version=v1.13.7 +rook-ceph-mgr-a req/upd/avl: 1/1/1 rook-version=v1.13.8 +rook-ceph-mon-a req/upd/avl: 1/1/1 rook-version=v1.13.8 +rook-ceph-mon-b req/upd/avl: 1/1/1 rook-version=v1.13.8 +rook-ceph-mon-c req/upd/avl: 1/1/1 rook-version=v1.13.8 +rook-ceph-osd-0 req/upd/avl: 1// rook-version=v1.13.8 rook-ceph-osd-1 req/upd/avl: 1/1/1 rook-version=v1.12.11 rook-ceph-osd-2 req/upd/avl: 1/1/1 rook-version=v1.12.11 ``` @@ -246,13 +246,13 @@ An easy check to see if the upgrade is totally finished is to check that there i # kubectl -n $ROOK_CLUSTER_NAMESPACE get deployment -l rook_cluster=$ROOK_CLUSTER_NAMESPACE -o jsonpath='{range .items[*]}{"rook-version="}{.metadata.labels.rook-version}{"\n"}{end}' | sort | uniq This cluster is not yet finished: rook-version=v1.12.11 - rook-version=v1.13.7 + rook-version=v1.13.8 This cluster is finished: - rook-version=v1.13.7 + rook-version=v1.13.8 ``` ### **5. Verify the updated cluster** -At this point, the Rook operator should be running version `rook/ceph:v1.13.7`. +At this point, the Rook operator should be running version `rook/ceph:v1.13.8`. Verify the CephCluster health using the [health verification doc](health-verification.md). diff --git a/deploy/charts/rook-ceph/values.yaml b/deploy/charts/rook-ceph/values.yaml index 452926c9f2ae..ab6e95d40ce0 100644 --- a/deploy/charts/rook-ceph/values.yaml +++ b/deploy/charts/rook-ceph/values.yaml @@ -7,7 +7,7 @@ image: repository: rook/ceph # -- Image tag # @default -- `master` - tag: v1.13.7 + tag: v1.13.8 # -- Image pull policy pullPolicy: IfNotPresent diff --git a/deploy/examples/direct-mount.yaml b/deploy/examples/direct-mount.yaml index 57e0e54179cf..fadd0d7cbae0 100644 --- a/deploy/examples/direct-mount.yaml +++ b/deploy/examples/direct-mount.yaml @@ -18,7 +18,7 @@ spec: dnsPolicy: ClusterFirstWithHostNet containers: - name: rook-direct-mount - image: rook/ceph:v1.13.7 + image: rook/ceph:v1.13.8 command: ["/bin/bash"] args: ["-m", "-c", "/usr/local/bin/toolbox.sh"] imagePullPolicy: IfNotPresent diff --git a/deploy/examples/images.txt b/deploy/examples/images.txt index 1fbb01bc6528..513623783296 100644 --- a/deploy/examples/images.txt +++ b/deploy/examples/images.txt @@ -8,4 +8,4 @@ registry.k8s.io/sig-storage/csi-provisioner:v4.0.0 registry.k8s.io/sig-storage/csi-resizer:v1.10.0 registry.k8s.io/sig-storage/csi-snapshotter:v7.0.1 - rook/ceph:v1.13.7 + rook/ceph:v1.13.8 diff --git a/deploy/examples/multus-validation.yaml b/deploy/examples/multus-validation.yaml index 8baa7e9f10cb..ad42672baa94 100644 --- a/deploy/examples/multus-validation.yaml +++ b/deploy/examples/multus-validation.yaml @@ -101,7 +101,7 @@ spec: serviceAccountName: rook-ceph-multus-validation containers: - name: multus-validation - image: rook/ceph:v1.13.7 + image: rook/ceph:v1.13.8 command: ["rook"] args: - "multus" diff --git a/deploy/examples/operator-openshift.yaml b/deploy/examples/operator-openshift.yaml index 624dd1e515ef..97a675319264 100644 --- a/deploy/examples/operator-openshift.yaml +++ b/deploy/examples/operator-openshift.yaml @@ -663,7 +663,7 @@ spec: serviceAccountName: rook-ceph-system containers: - name: rook-ceph-operator - image: rook/ceph:v1.13.7 + image: rook/ceph:v1.13.8 args: ["ceph", "operator"] securityContext: runAsNonRoot: true diff --git a/deploy/examples/operator.yaml b/deploy/examples/operator.yaml index a46c0623e51c..dbc69e1b7039 100644 --- a/deploy/examples/operator.yaml +++ b/deploy/examples/operator.yaml @@ -588,7 +588,7 @@ spec: serviceAccountName: rook-ceph-system containers: - name: rook-ceph-operator - image: rook/ceph:v1.13.7 + image: rook/ceph:v1.13.8 args: ["ceph", "operator"] securityContext: runAsNonRoot: true diff --git a/deploy/examples/osd-purge.yaml b/deploy/examples/osd-purge.yaml index 59473d402e27..450078f7cad9 100644 --- a/deploy/examples/osd-purge.yaml +++ b/deploy/examples/osd-purge.yaml @@ -28,7 +28,7 @@ spec: serviceAccountName: rook-ceph-purge-osd containers: - name: osd-removal - image: rook/ceph:v1.13.7 + image: rook/ceph:v1.13.8 # TODO: Insert the OSD ID in the last parameter that is to be removed # The OSD IDs are a comma-separated list. For example: "0" or "0,2". # If you want to preserve the OSD PVCs, set `--preserve-pvc true`. diff --git a/deploy/examples/toolbox-job.yaml b/deploy/examples/toolbox-job.yaml index b28072952a8d..eed4d75dfd5a 100644 --- a/deploy/examples/toolbox-job.yaml +++ b/deploy/examples/toolbox-job.yaml @@ -10,7 +10,7 @@ spec: spec: initContainers: - name: config-init - image: rook/ceph:v1.13.7 + image: rook/ceph:v1.13.8 command: ["/usr/local/bin/toolbox.sh"] args: ["--skip-watch"] imagePullPolicy: IfNotPresent @@ -29,7 +29,7 @@ spec: mountPath: /var/lib/rook-ceph-mon containers: - name: script - image: rook/ceph:v1.13.7 + image: rook/ceph:v1.13.8 volumeMounts: - mountPath: /etc/ceph name: ceph-config diff --git a/deploy/examples/toolbox-operator-image.yaml b/deploy/examples/toolbox-operator-image.yaml index 62c40596fb9f..98c5e1aa19d9 100644 --- a/deploy/examples/toolbox-operator-image.yaml +++ b/deploy/examples/toolbox-operator-image.yaml @@ -24,7 +24,7 @@ spec: dnsPolicy: ClusterFirstWithHostNet containers: - name: rook-ceph-tools-operator-image - image: rook/ceph:v1.13.7 + image: rook/ceph:v1.13.8 command: - /bin/bash - -c