From 9678f8cfaf59dd401c64b903646ddc67685fff5e Mon Sep 17 00:00:00 2001 From: Francois Ferrand Date: Tue, 20 Aug 2024 19:38:49 +0200 Subject: [PATCH 01/16] Add artesca-root-ca-issuer cluster issuer Issue: ZENKO-4876 --- .../end2end/install-kind-dependencies.sh | 17 +++++++++-------- 1 file changed, 9 insertions(+), 8 deletions(-) diff --git a/.github/scripts/end2end/install-kind-dependencies.sh b/.github/scripts/end2end/install-kind-dependencies.sh index 461fef235..9a7b1c4bc 100755 --- a/.github/scripts/end2end/install-kind-dependencies.sh +++ b/.github/scripts/end2end/install-kind-dependencies.sh @@ -58,14 +58,15 @@ kubectl rollout status -n ingress-nginx deployment/ingress-nginx-controller --ti # cert-manager kubectl apply --validate=false -f https://github.com/jetstack/cert-manager/releases/download/${CERT_MANAGER_VERSION}/cert-manager.yaml --wait -# kubectl apply --validate=false -f - < Date: Fri, 23 Aug 2024 15:06:01 +0200 Subject: [PATCH 02/16] Test that federated metrics are retrieved Issue: ZENKO-4876 --- tests/ctst/features/pra.feature | 3 ++- tests/ctst/steps/pra.ts | 22 ++++++++++++++++++++-- 2 files changed, 22 insertions(+), 3 deletions(-) diff --git a/tests/ctst/features/pra.feature b/tests/ctst/features/pra.feature index f0aede1e3..0199b137b 100644 --- a/tests/ctst/features/pra.feature +++ b/tests/ctst/features/pra.feature @@ -18,7 +18,8 @@ Feature: PRA operations Given a DR installed Then the DR source should be in phase "Running" And the DR sink should be in phase "Running" - Then the kafka DR volume exists + And the kafka DR volume exists + And prometheus should scrap federated metrics from DR sink # Check that objects are transitioned in the DR site Given access keys for the replicated account diff --git a/tests/ctst/steps/pra.ts b/tests/ctst/steps/pra.ts index f8cc5a172..5fe0739a0 100644 --- a/tests/ctst/steps/pra.ts +++ b/tests/ctst/steps/pra.ts @@ -15,6 +15,7 @@ import { } from 'steps/utils/utils'; import { CacheHelper, Constants, Identity, IdentityEnum, SuperAdmin, Utils } from 'cli-testing'; import { safeJsonParse } from 'common/utils'; +import { PrometheusDriver } from 'prometheus-query'; import assert from 'assert'; import { EntityType } from 'world/Zenko'; @@ -328,8 +329,25 @@ Then('the kafka DR volume exists', { timeout: volumeTimeout + 2000 }, async func assert(volumeParsed.result!['volume phase'] === 'Bound'); }); +Then('prometheus should scrap federated metrics from DR sink', { timeout: 70000 }, async function (this: Zenko) { + const prom = new PrometheusDriver({ + endpoint: `http://${this.parameters.PrometheusService}:9090`, + baseURL: '/api/v1', + }); + + for (;;) { + const t = Date.now(); + const metrics = await prom.series('{drSinkInstance="end2end-pra-sink"}', t - 60 * 1000, t); + if (metrics.length > 0) { + break; + } + + await Utils.sleep(1000); + } +}); + const failoverTimeout = 360000; -When ('I request the failover state for the DR', { timeout: failoverTimeout + 2000 }, async function (this: Zenko) { +When('I request the failover state for the DR', { timeout: failoverTimeout + 2000 }, async function (this: Zenko) { await this.zenkoDrCtl?.failover({ sinkZenkoDrNamespace: 'default', sinkZenkoDrInstance: 'end2end-pra-sink', @@ -339,7 +357,7 @@ When ('I request the failover state for the DR', { timeout: failoverTimeout + 20 }); const failbackTimeout = 360000; -When ('I resume operations for the DR', { timeout: failbackTimeout + 2000 }, async function (this: Zenko) { +When('I resume operations for the DR', { timeout: failbackTimeout + 2000 }, async function (this: Zenko) { await this.zenkoDrCtl?.failback({ sinkZenkoDrNamespace: 'default', sinkZenkoDrInstance: 'end2end-pra-sink', From 8970422853ea246855e6e75879fbeb11c4cd68a8 Mon Sep 17 00:00:00 2001 From: Francois Ferrand Date: Thu, 26 Sep 2024 15:01:07 +0200 Subject: [PATCH 03/16] Set federated prometheus name This is needed to avoid conflict with internal prometheus service, since we are running in the same cluster. Issue: ZENKO-4876 --- .github/scripts/end2end/patch-coredns.sh | 1 + tests/ctst/steps/pra.ts | 3 +-- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/.github/scripts/end2end/patch-coredns.sh b/.github/scripts/end2end/patch-coredns.sh index 8a3c3b640..8785733f3 100755 --- a/.github/scripts/end2end/patch-coredns.sh +++ b/.github/scripts/end2end/patch-coredns.sh @@ -31,6 +31,7 @@ corefile=" rewrite name exact s3.dr.zenko.local ingress-nginx-controller.ingress-nginx.svc.cluster.local rewrite name exact sts.dr.zenko.local ingress-nginx-controller.ingress-nginx.svc.cluster.local rewrite name exact iam.dr.zenko.local ingress-nginx-controller.ingress-nginx.svc.cluster.local + rewrite name exact prom.dr.zenko.local ingress-nginx-controller.ingress-nginx.svc.cluster.local rewrite name exact shell-ui.dr.zenko.local ingress-nginx-controller.ingress-nginx.svc.cluster.local rewrite name exact website.mywebsite.com ingress-nginx-controller.ingress-nginx.svc.cluster.local kubernetes cluster.local in-addr.arpa ip6.arpa { diff --git a/tests/ctst/steps/pra.ts b/tests/ctst/steps/pra.ts index 5fe0739a0..6f35d1942 100644 --- a/tests/ctst/steps/pra.ts +++ b/tests/ctst/steps/pra.ts @@ -71,9 +71,8 @@ async function installPRA(world: Zenko, sinkS3Endpoint = 'http://s3.zenko.local' sourceS3Endpoint: 'http://s3.zenko.local', sinkS3Endpoint, prometheusService: world.parameters.PrometheusService, - // prometheusHostname: 'prom.dr.zenko.local', // could be any name, cert will be auto-generated + prometheusHostname: 'prom.dr.zenko.local', prometheusExternalIpsDiscovery: true, - prometheusDisableTls: true, forceRotateServiceCredentials: (CacheHelper.savedAcrossTests[Zenko.PRA_INSTALL_COUNT_KEY] as number) > 0, ...kafkaExternalIpOption, timeout, From 9826e4c5ef59939a3d323b85e6a5463479088139 Mon Sep 17 00:00:00 2001 From: Francois Ferrand Date: Wed, 23 Oct 2024 10:43:08 +0200 Subject: [PATCH 04/16] Fix prometheus config ServiceMonitor were not scrapped. Issue: ZENKO-4876 --- .github/scripts/end2end/configs/prometheus.yaml | 6 ++++++ 1 file changed, 6 insertions(+) diff --git a/.github/scripts/end2end/configs/prometheus.yaml b/.github/scripts/end2end/configs/prometheus.yaml index e885d6325..6a2273117 100644 --- a/.github/scripts/end2end/configs/prometheus.yaml +++ b/.github/scripts/end2end/configs/prometheus.yaml @@ -36,13 +36,19 @@ spec: evaluationInterval: 30s logFormat: logfmt logLevel: info + serviceMonitorNamespaceSelector: {} + serviceMonitorSelector: + matchLabels: + metalk8s.scality.com/monitor: "" podMonitorNamespaceSelector: {} podMonitorSelector: matchLabels: metalk8s.scality.com/monitor: "" + probeNamespaceSelector: {} probeSelector: matchLabels: metalk8s.scality.com/monitor: "" + ruleNamespaceSelector: {} ruleSelector: matchLabels: metalk8s.scality.com/monitor: "" From f9d647fb4846889c6ad19b38d7a738cd7cd94cf0 Mon Sep 17 00:00:00 2001 From: Francois Ferrand Date: Wed, 23 Oct 2024 11:46:48 +0200 Subject: [PATCH 05/16] Fix zkop install to support sha1 When an SHA1 is used which is not the tip of the branch, it is not available after a `git clone` (which fetches only commits reachable from current branches). This causes issues if a build is started while the component branch with the commit has been rewritten (rebase...) To overcome this, the safe way is to explicitely fetch the commit: which works for any commit (branch, tag, sha1). Issue: ZENKO-4876 --- .github/actions/deploy/action.yaml | 5 +++-- 1 file changed, 3 insertions(+), 2 deletions(-) diff --git a/.github/actions/deploy/action.yaml b/.github/actions/deploy/action.yaml index 1fc7708f0..5c4a76cfc 100644 --- a/.github/actions/deploy/action.yaml +++ b/.github/actions/deploy/action.yaml @@ -64,9 +64,10 @@ runs: docker pull ${OPERATOR_IMAGE_NAME}:${OPERATOR_IMAGE_TAG} kind load docker-image ${OPERATOR_IMAGE_NAME}:${OPERATOR_IMAGE_TAG} cd ./.github/scripts/end2end - git clone https://${GIT_ACCESS_TOKEN}@github.com/scality/zenko-operator.git operator + git init operator cd operator - git checkout ${OPERATOR_IMAGE_TAG} + git fetch --depth 1 --no-tags https://${GIT_ACCESS_TOKEN}@github.com/scality/zenko-operator.git ${OPERATOR_IMAGE_TAG} + git checkout FETCH_HEAD tilt ci env: OPERATOR_IMAGE_TAG: ${{ inputs.zkop_tag }} From 33171e8e5b10fd206f4d388469c6c2ceb7f757e4 Mon Sep 17 00:00:00 2001 From: Francois Ferrand Date: Wed, 23 Oct 2024 19:44:08 +0200 Subject: [PATCH 06/16] Move kafka dump to folder Issue: ZENKO-4876 --- .github/actions/archive-artifacts/action.yaml | 13 ++++++++----- 1 file changed, 8 insertions(+), 5 deletions(-) diff --git a/.github/actions/archive-artifacts/action.yaml b/.github/actions/archive-artifacts/action.yaml index 6d02f751d..747e3b147 100644 --- a/.github/actions/archive-artifacts/action.yaml +++ b/.github/actions/archive-artifacts/action.yaml @@ -74,26 +74,29 @@ runs: set -exu KAFKA=$(kubectl get pods -n ${NAMESPACE} -lkafka_cr=${ZENKO_NAME}-base-queue -o jsonpath='{.items[0].metadata.name}') + KAFKA_PATH="/tmp/artifacts/data/${STAGE}/kafka" + + mkdir -p ${KAFKA_PATH} kubectl exec -in ${NAMESPACE} ${KAFKA} -c kafka -- \ env KAFKA_OPTS= kafka-topics.sh --bootstrap-server :9092 --list \ - > /tmp/artifacts/data/${STAGE}/kafka-topics.log + > ${KAFKA_PATH}/kafka-topics.log kubectl exec -in ${NAMESPACE} ${KAFKA} -c kafka -- \ env KAFKA_OPTS= kafka-consumer-groups.sh --bootstrap-server :9092 --list \ - > /tmp/artifacts/data/${STAGE}/kafka-consumer-groups.log + > ${KAFKA_PATH}/kafka-consumer-groups.log kubectl exec -in ${NAMESPACE} ${KAFKA} -c kafka -- \ env KAFKA_OPTS= kafka-consumer-groups.sh --bootstrap-server :9092 --describe --all-groups \ - > /tmp/artifacts/data/${STAGE}/kafka-offsets.log + > ${KAFKA_PATH}/kafka-offsets.log KAFKA_SERVICE=$(kubectl get services -n ${NAMESPACE} -lkafka_cr=${ZENKO_NAME}-base-queue -o jsonpath='{.items[0].metadata.name}') kubectl run -n ${NAMESPACE} kcat --image=edenhill/kcat:1.7.1 --restart=Never --command -- sleep 300 kubectl wait -n ${NAMESPACE} pod kcat --for=condition=ready - cat /tmp/artifacts/data/${STAGE}/kafka-topics.log | grep -v '^__' | xargs -P 15 -I {} \ + cat ${KAFKA_PATH}/kafka-topics.log | grep -v '^__' | xargs -P 15 -I {} \ sh -c "kubectl exec -i -n ${NAMESPACE} kcat -- \ kcat -L -b ${KAFKA_SERVICE} -t {} -C -o beginning -e -q -J \ - > /tmp/artifacts/data/${STAGE}/kafka-messages-{}.log" + > ${KAFKA_PATH}/kafka-messages-{}.log" env: STAGE: ${{ inputs.stage }} NAMESPACE: ${{ inputs.zenko-namespace }} From a4cb9ae624a59f0c44031ffb32343843ee3366b4 Mon Sep 17 00:00:00 2001 From: Francois Ferrand Date: Wed, 23 Oct 2024 19:41:44 +0200 Subject: [PATCH 07/16] Merge alerts workflow No reason to keep separate, and it helps avoid a github bug/limitation when creating a check: and the Issue: ZENKO-4876 --- .github/workflows/alerts.yaml | 9 +++++---- .github/workflows/end2end.yaml | 15 ++++++++++----- 2 files changed, 15 insertions(+), 9 deletions(-) diff --git a/.github/workflows/alerts.yaml b/.github/workflows/alerts.yaml index 540caa655..73f818c6f 100644 --- a/.github/workflows/alerts.yaml +++ b/.github/workflows/alerts.yaml @@ -1,10 +1,11 @@ name: Test alerts on: - push: - branches-ignore: - - 'development/**' - - 'q/*' + workflow_call: + secrets: + GIT_ACCESS_TOKEN: + description: 'GitHub token' + required: true jobs: run-alert-tests: diff --git a/.github/workflows/end2end.yaml b/.github/workflows/end2end.yaml index 5dcb1ddcb..dd838f5a4 100644 --- a/.github/workflows/end2end.yaml +++ b/.github/workflows/end2end.yaml @@ -121,6 +121,10 @@ jobs: - name: Verify monitoring dashboard versions run: bash ./.github/scripts/check_versions.sh + check-alerts: + uses: ./.github/workflows/alerts.yaml + secrets: inherit + check-workflows: runs-on: ubuntu-22.04 steps: @@ -379,7 +383,7 @@ jobs: cache-to: type=gha,mode=max,scope=end2end-ctst end2end-http: - needs: [build-kafka, build-test-image, check-dashboard-versions] + needs: [build-kafka, build-test-image] runs-on: - ubuntu - focal @@ -428,7 +432,7 @@ jobs: run: kind delete cluster end2end-pra: - needs: [build-kafka, check-dashboard-versions, lint-and-build-ctst] + needs: [build-kafka, lint-and-build-ctst] runs-on: ubuntu-22.04-16core env: GIT_ACCESS_TOKEN: ${{ secrets.GIT_ACCESS_TOKEN }} @@ -488,7 +492,7 @@ jobs: run: kind delete cluster end2end-https: - needs: [build-kafka, build-test-image, check-dashboard-versions] + needs: [build-kafka, build-test-image] runs-on: - ubuntu - focal @@ -540,7 +544,7 @@ jobs: run: kind delete cluster end2end-sharded: - needs: [build-kafka, build-test-image, check-dashboard-versions] + needs: [build-kafka, build-test-image] runs-on: - ubuntu-22.04-8core # Enable this for Ring-based tests @@ -580,7 +584,7 @@ jobs: run: kind delete cluster ctst-end2end-sharded: - needs: [build-kafka, lint-and-build-ctst, check-dashboard-versions] + needs: [build-kafka, lint-and-build-ctst] runs-on: - ubuntu-22.04-8core steps: @@ -629,6 +633,7 @@ jobs: write-final-status: runs-on: ubuntu-latest needs: + - check-alerts - check-dashboard-versions - check-workflows - build-doc From b88d9623154ec48bc7cd4d0c62678225ae38b4bd Mon Sep 17 00:00:00 2001 From: williamlardier Date: Thu, 14 Nov 2024 15:40:38 +0100 Subject: [PATCH 08/16] Ensure callback is called once in doWhilst with S3 operations The trigger seem to be https://github.com/aws/aws-sdk-js/issues/1678 ... causing the callback to be called twice, which is a problem in these loops. Issue: ZENKO-4925 --- .../node_tests/backbeat/ReplicationUtility.js | 29 +++++++++++-------- 1 file changed, 17 insertions(+), 12 deletions(-) diff --git a/tests/zenko_tests/node_tests/backbeat/ReplicationUtility.js b/tests/zenko_tests/node_tests/backbeat/ReplicationUtility.js index d80bf5e1d..0e0739186 100644 --- a/tests/zenko_tests/node_tests/backbeat/ReplicationUtility.js +++ b/tests/zenko_tests/node_tests/backbeat/ReplicationUtility.js @@ -1,6 +1,7 @@ const assert = require('assert'); const crypto = require('crypto'); const async = require('async'); +const { jsutil } = require('arsenal'); const { scalityS3Client, awsS3Client } = require('../s3SDK'); @@ -596,8 +597,9 @@ class ReplicationUtility { Key: key, VersionId: versionId, }, (err, data) => { + const cbOnce = jsutil.once(callback); if (err) { - return callback(err); + return cbOnce(err); } status = data.ReplicationStatus; assert.notStrictEqual( @@ -606,9 +608,9 @@ class ReplicationUtility { `Unexpected CRR failure occurred: ${JSON.stringify(data)}`, ); if (status === 'PENDING' || status === 'PROCESSING') { - return setTimeout(callback, 2000); + return setTimeout(cbOnce, 2000); } - return callback(); + return cbOnce(); }), () => (status === 'PENDING' || status === 'PROCESSING'), cb, @@ -622,14 +624,15 @@ class ReplicationUtility { const expectedCode = client === 'azure' ? 'BlobNotFound' : 'NoSuchKey'; return async.doWhilst( callback => this[method](bucketName, key, err => { + const cbOnce = jsutil.once(callback); if (err && err.code !== expectedCode) { - return callback(err); + return cbOnce(err); } objectExists = err === null; if (!objectExists) { - return callback(); + return cbOnce(); } - return setTimeout(callback, 2000); + return setTimeout(cbOnce, 2000); }), () => objectExists, cb, @@ -644,8 +647,9 @@ class ReplicationUtility { Bucket: bucketName, Key: key, }, (err, data) => { + const cbOnce = jsutil.once(callback); if (err) { - return callback(err); + return cbOnce(err); } const statuses = []; // We cannot rely on the global status for one-to-many, so check @@ -657,9 +661,9 @@ class ReplicationUtility { }); shouldContinue = statuses.includes('PENDING'); if (shouldContinue) { - return setTimeout(callback, 2000); + return setTimeout(cbOnce, 2000); } - return callback(); + return cbOnce(); }), () => shouldContinue, cb, @@ -674,14 +678,15 @@ class ReplicationUtility { Bucket: bucketName, Key: key, }, (err, data) => { + const cbOnce = jsutil.once(callback); if (err) { - return callback(err); + return cbOnce(err); } shouldContinue = data.ReplicationStatus === 'FAILED'; if (shouldContinue) { - return setTimeout(callback, 2000); + return setTimeout(cbOnce, 2000); } - return callback(); + return cbOnce(); }), () => shouldContinue, cb, From 88384bc75727d92cbf730a1021b06fe26e8cb9ea Mon Sep 17 00:00:00 2001 From: williamlardier Date: Fri, 15 Nov 2024 09:32:22 +0100 Subject: [PATCH 09/16] Improve variable in doWhilst Issue: ZENKO-4925 --- .../node_tests/backbeat/ReplicationUtility.js | 9 ++++----- 1 file changed, 4 insertions(+), 5 deletions(-) diff --git a/tests/zenko_tests/node_tests/backbeat/ReplicationUtility.js b/tests/zenko_tests/node_tests/backbeat/ReplicationUtility.js index 0e0739186..3e5c8abaa 100644 --- a/tests/zenko_tests/node_tests/backbeat/ReplicationUtility.js +++ b/tests/zenko_tests/node_tests/backbeat/ReplicationUtility.js @@ -590,7 +590,6 @@ class ReplicationUtility { // Continue getting head object while the status is PENDING or PROCESSING. waitUntilReplicated(bucketName, key, versionId, cb) { - let status; return async.doWhilst( callback => this.s3.headObject({ Bucket: bucketName, @@ -601,18 +600,18 @@ class ReplicationUtility { if (err) { return cbOnce(err); } - status = data.ReplicationStatus; + const status = data.ReplicationStatus; assert.notStrictEqual( status, 'FAILED', `Unexpected CRR failure occurred: ${JSON.stringify(data)}`, ); if (status === 'PENDING' || status === 'PROCESSING') { - return setTimeout(cbOnce, 2000); + return setTimeout(() => cbOnce(null, status), 2000); } - return cbOnce(); + return cbOnce(null, status); }), - () => (status === 'PENDING' || status === 'PROCESSING'), + status => (status === 'PENDING' || status === 'PROCESSING'), cb, ); } From 47efee63950fc1ff6691b78858bfa580b605477f Mon Sep 17 00:00:00 2001 From: Francois Ferrand Date: Tue, 12 Nov 2024 15:28:27 +0100 Subject: [PATCH 10/16] Bump backbeat 8.6.51 Issue: ZENKO-4926 --- solution/deps.yaml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/solution/deps.yaml b/solution/deps.yaml index 2a7f50bc8..d41f4b7d9 100644 --- a/solution/deps.yaml +++ b/solution/deps.yaml @@ -6,7 +6,7 @@ backbeat: dashboard: backbeat/backbeat-dashboards image: backbeat policy: backbeat/backbeat-policies - tag: 8.6.49 + tag: 8.6.51 envsubst: BACKBEAT_TAG busybox: image: busybox From 4a877c7454738401a438105ba0cf27b958b46439 Mon Sep 17 00:00:00 2001 From: Francois Ferrand Date: Tue, 12 Nov 2024 15:28:44 +0100 Subject: [PATCH 11/16] Bump zenko-operator 1.6.4 Issue: ZENKO-4926 --- solution/deps.yaml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/solution/deps.yaml b/solution/deps.yaml index d41f4b7d9..55ec83504 100644 --- a/solution/deps.yaml +++ b/solution/deps.yaml @@ -136,7 +136,7 @@ vault: zenko-operator: sourceRegistry: ghcr.io/scality image: zenko-operator - tag: v1.6.3 + tag: v1.6.4 envsubst: ZENKO_OPERATOR_TAG zenko-ui: sourceRegistry: ghcr.io/scality From d15032f781a9c77671502f70468dfa0c790d5919 Mon Sep 17 00:00:00 2001 From: Francois Ferrand Date: Mon, 18 Nov 2024 16:33:28 +0100 Subject: [PATCH 12/16] Bump zenko-operator 1.6.5 Issue: ZENKO-4919 --- solution/deps.yaml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/solution/deps.yaml b/solution/deps.yaml index 55ec83504..88e6f1e85 100644 --- a/solution/deps.yaml +++ b/solution/deps.yaml @@ -136,7 +136,7 @@ vault: zenko-operator: sourceRegistry: ghcr.io/scality image: zenko-operator - tag: v1.6.4 + tag: v1.6.5 envsubst: ZENKO_OPERATOR_TAG zenko-ui: sourceRegistry: ghcr.io/scality From 044896093569e1e7ad24663a77dfc25c9586f7ab Mon Sep 17 00:00:00 2001 From: Francois Ferrand Date: Mon, 18 Nov 2024 16:33:46 +0100 Subject: [PATCH 13/16] Bump cloudserver 8.8.36 Issue: ZENKO-4919 --- solution/deps.yaml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/solution/deps.yaml b/solution/deps.yaml index 88e6f1e85..f6e15203f 100644 --- a/solution/deps.yaml +++ b/solution/deps.yaml @@ -16,7 +16,7 @@ cloudserver: sourceRegistry: ghcr.io/scality dashboard: cloudserver/cloudserver-dashboards image: cloudserver - tag: 8.8.35 + tag: 8.8.36 envsubst: CLOUDSERVER_TAG drctl: sourceRegistry: ghcr.io/scality From d202d0b386a8bfe3e03783e47ce7f967ebcbbc4e Mon Sep 17 00:00:00 2001 From: Francois Ferrand Date: Mon, 18 Nov 2024 16:35:06 +0100 Subject: [PATCH 14/16] Bump sorbet 1.1.13 Issue: ZENKO-4919 --- solution/deps.yaml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/solution/deps.yaml b/solution/deps.yaml index f6e15203f..e470beaa5 100644 --- a/solution/deps.yaml +++ b/solution/deps.yaml @@ -113,7 +113,7 @@ sorbet: policy: sorbet/sorbet-policies dashboard: sorbet/sorbet-dashboards image: sorbet - tag: v1.1.12 + tag: v1.1.13 envsubst: SORBET_TAG stern: # tail any pod logs with pattern matchin tag: 1.30.0 From 4679d665e719e1e27d40dd7af6c54059d13cd1a6 Mon Sep 17 00:00:00 2001 From: Francois Ferrand Date: Mon, 18 Nov 2024 19:35:44 +0100 Subject: [PATCH 15/16] Fix flaky cluster issuer setup We need to retry `kubectl apply`, esp as mutating web hooks may not be ready, c.f. [1]. [1] https://github.com/cert-manager/cert-manager/issues/2908 Issue: ZENKO-4919 --- .github/scripts/end2end/install-kind-dependencies.sh | 8 +++++++- 1 file changed, 7 insertions(+), 1 deletion(-) diff --git a/.github/scripts/end2end/install-kind-dependencies.sh b/.github/scripts/end2end/install-kind-dependencies.sh index 9a7b1c4bc..ed12de80e 100755 --- a/.github/scripts/end2end/install-kind-dependencies.sh +++ b/.github/scripts/end2end/install-kind-dependencies.sh @@ -59,7 +59,9 @@ kubectl rollout status -n ingress-nginx deployment/ingress-nginx-controller --ti # cert-manager kubectl apply --validate=false -f https://github.com/jetstack/cert-manager/releases/download/${CERT_MANAGER_VERSION}/cert-manager.yaml --wait kubectl rollout status -n cert-manager deployment/cert-manager-webhook --timeout=10m -kubectl apply -f - < Date: Mon, 18 Nov 2024 16:38:27 +0100 Subject: [PATCH 16/16] Release Zenko 2.6.70 Issue: ZENKO-4919 --- VERSION | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/VERSION b/VERSION index d97820210..57c2ca6ba 100644 --- a/VERSION +++ b/VERSION @@ -1,4 +1,4 @@ -VERSION="2.6.69" +VERSION="2.6.70" VERSION_SUFFIX=