ArgoCD ECR credentials updater is failing due to bitnami/kubectl:1.23 image not found, causing authentication issues for pulling ECR images.
FAILING COMPONENTS:
- β
argocd-ecr-credentials-updater- CronJob usingbitnami/kubectl:1.23(CRITICAL)- Error:
failed to pull and unpack image "docker.io/bitnami/kubectl:1.23": not found - Impact: ECR authentication broken, causing cascade failures
- Error:
WORKING COMPONENTS:
- β
pod-monitor- Already updated tobitnamisecure/kubectl:latest
OTHER BITNAMI IMAGES IN USE:
docker.io/bitnami/redis:6.2.6-debian-10-r120(did-auth-proxy, iam-did-auth-proxy)docker.io/bitnami/redis:7.2.4-debian-12-r13(gp4btc-paypal)docker.io/bitnami/postgresql:14.4.0-debian-11-r9(gp4btc, tgc)docker.io/bitnami/postgresql:17.0.0-debian-12-r11(gpdive)docker.io/bitnami/postgresql:16.1.0-debian-11-r26(ika)docker.io/bitnami/nginx:1.29.1-debian-12-r0(ewx staking-dapp)docker.io/bitnami/nginx:1.21.6-debian-10-r93(iam-did-auth-proxy)bitnami/sealed-secrets-controller:v0.18.0(kube-system)
FAILING COMPONENTS:
- β
argocd-ecr-credentials-updater- Same issue as dev
OTHER BITNAMI IMAGES:
docker.io/bitnami/postgresql:14.4.0-debian-11-r9(gp4btc)docker.io/bitnami/redis:6.2.6-debian-10-r120(iam-did-auth-proxy)docker.io/bitnami/nginx:1.21.6-debian-10-r93(iam-did-auth-proxy)bitnami/sealed-secrets-controller:v0.18.0(kube-system)
- β
docker.io/bitnami/sealed-secrets-controller:v0.24.0(newer version, likely working)
Repository: energywebfoundation/argocd-ecr-credentials-updater
File: templates/cronJob.yaml
Current:
image: bitnami/kubectl:1.23Fix:
image: bitnamisecure/kubectl:1.23Deployment Commands:
# Update both clusters immediately
kubectl --context=dev-cluster patch cronjob argocd-ecr-credentials-updater -n argocd --type='merge' -p='{"spec":{"jobTemplate":{"spec":{"template":{"spec":{"containers":[{"name":"kubectl","image":"bitnamisecure/kubectl:1.23"}]}}}}}}'
kubectl --context=staging-k8s-misc patch cronjob argocd-ecr-credentials-updater -n argocd --type='merge' -p='{"spec":{"jobTemplate":{"spec":{"template":{"spec":{"containers":[{"name":"kubectl","image":"bitnamisecure/kubectl:1.23"}]}}}}}}'A. energywebfoundation/argocd-ecr-credentials-updater
- File:
templates/cronJob.yamlline ~23 - Change:
bitnami/kubectl:1.23βbitnamisecure/kubectl:1.23
B. energywebfoundation/erpc-helm
- File:
values.yamllines 23-26 - Current:
kubectl: image: repository: bitnami/kubectl pullPolicy: IfNotPresent tag: "1.21"
- Fix:
kubectl: image: repository: bitnamisecure/kubectl pullPolicy: IfNotPresent tag: "1.21"
C. energywebfoundation/operating-envelope-dle-node
- File:
helm/sandbox/substrate-node-0/values.yamllines 55-58 - Current:
kubectl: image: repository: bitnami/kubectl tag: latest
- Fix:
kubectl: image: repository: bitnamisecure/kubectl tag: latest
These images are currently working but should be monitored for future migration:
- PostgreSQL images (multiple versions in use)
- Redis images (multiple versions in use)
- Nginx images (multiple versions in use)
- Sealed Secrets Controller (consider upgrading to newer version)
- Bitnami Registry Migration: Bitnami announced migration from
bitnami/tobitnamisecure/namespace - Pod Monitor Fix: Already applied -
pod-monitorcronjob updated tobitnamisecure/kubectl:latest - Cascade Effect: ECR credentials updater failure prevents pulling other ECR images, causing scheduling issues
After applying fixes:
# Check ArgoCD ECR updater status
kubectl --context=dev-cluster get pods -n argocd -l job-name -o wide
# Verify ECR credentials are being updated
kubectl --context=dev-cluster get secret aws-ecr-creds -n argocd -o yaml
# Check for image pull errors
kubectl --context=dev-cluster get events --all-namespaces | grep -i "failed to pull\|ImagePull"- Immediate (< 30 minutes): Apply kubectl patches to both clusters
- Same day: Create PRs for repository updates
- This week: Monitor other bitnami images for deprecation notices
- Ongoing: Set up monitoring for registry migration announcements
Current Dev Cluster Nodes: 22 nodes (all in default group)
- No issues found with node availability
- The mentioned
EWX_marketplace_EKS-2024041809412703080000001cworker group is not visible - Current scheduling working fine on default node group
# IMMEDIATE FIX - Run these now:
# Dev cluster
kubectl --context=dev-cluster patch cronjob argocd-ecr-credentials-updater -n argocd --type='merge' -p='{"spec":{"jobTemplate":{"spec":{"template":{"spec":{"containers":[{"name":"kubectl","image":"bitnamisecure/kubectl:1.23"}]}}}}}}'
# Staging cluster
kubectl --context=staging-k8s-misc patch cronjob argocd-ecr-credentials-updater -n argocd --type='merge' -p='{"spec":{"jobTemplate":{"spec":{"template":{"spec":{"containers":[{"name":"kubectl","image":"bitnamisecure/kubectl:1.23"}]}}}}}}'
# Verify the fix
kubectl --context=dev-cluster get cronjob argocd-ecr-credentials-updater -n argocd -o yaml | grep image:
kubectl --context=staging-k8s-misc get cronjob argocd-ecr-credentials-updater -n argocd -o yaml | grep image:Status: