Replies: 13 comments 5 replies
-
@cjroebuck We usually see the |
Beta Was this translation helpful? Give feedback.
-
@ihcsim No I haven't done anything custom regarding TLS. Here's output of the linkerd check --proxy:
It doesn't complete and gets stuck. Also the logs from linkerd-prometheus that you asked in the slack, just an endless stream of these access denied errors:
|
Beta Was this translation helpful? Give feedback.
-
Can you try reinstalling Linkerd? This includes both the control and data planes. Instructions on uninstalling Linkerd can be found here. |
Beta Was this translation helpful? Give feedback.
-
I have this same problem on my system. I am running this on minikube and have run this several times and it is reproducible with the same error. The linkerd-proxy fails with AccessDenied error. The logs show that it tries several ports before it gives up. I have attached the logs for linkerd-proxy, linkerd-Init and the result of linkerd check in this gist https://gist.github.com/seizadi/88417cf9fb6313ed40e23d8222d10873 |
Beta Was this translation helpful? Give feedback.
-
@seizadi There are two things I will check. First, try restarting the |
Beta Was this translation helpful? Give feedback.
-
I am running minikube with 3 nodes and 6GB of memory: minikube start --cpus=3 I am using argo rollout CRD which manages a ReplicaSet which creates the Pod which has I tried to restart the pod from argo but the ReplicaSet was in a bad state and it did not restart: ❯ k argo rollouts restart rollouts-demo
rollout 'rollouts-demo' restarts in 0s
❯ k get rs
NAME DESIRED CURRENT READY AGE
rollouts-demo-868f9df8cd 1 1 0 18m Argo Rollout log: INFO[2020-07-27T14:06:37-07:00] cannot restart pods as not all ReplicasSets are fully available Reconciler=PodRestarter ReplicaSet=rollouts-demo-868f9df8cd namespace=test rollout=rollouts-demo I deleted the ReplicaSet attached to the argo rollout, which forced the ReplicaSet to be recreated and still have same problem with linkerd-proxy: ❯ k delete rs rollouts-demo-868f9df8cd
replicaset.apps "rollouts-demo-868f9df8cd" deleted
❯ k get rs
NAME DESIRED CURRENT READY AGE
rollouts-demo-868f9df8cd 1 1 0 23s
❯ k get pods
NAME READY STATUS RESTARTS AGE
rollouts-demo-868f9df8cd-wr6xb 1/2 Running 0 33s
❯ k describe pod rollouts-demo-868f9df8cd-wr6xb
Name: rollouts-demo-868f9df8cd-wr6xb
Namespace: test
Priority: 0
Node: minikube/192.168.64.36
Start Time: Mon, 27 Jul 2020 14:11:59 -0700
Labels: app=rollouts-demo
linkerd.io/control-plane-ns=linkerd
linkerd.io/workload-ns=test
rollouts-pod-template-hash=868f9df8cd
Annotations: linkerd.io/created-by: linkerd/proxy-injector stable-2.8.0
linkerd.io/identity-mode: default
linkerd.io/inject: enabled
linkerd.io/proxy-version: stable-2.8.0
Status: Running
IP: 172.17.0.17
IPs:
IP: 172.17.0.17
Controlled By: ReplicaSet/rollouts-demo-868f9df8cd
....
Normal Created 45s kubelet, minikube Created container linkerd-proxy
Normal Started 45s kubelet, minikube Started container linkerd-proxy
Warning Unhealthy 8s (x4 over 38s) kubelet, minikube Readiness probe failed: HTTP probe failed with statuscode: 503 |
Beta Was this translation helpful? Give feedback.
-
@seizadi The Linkerd commands in your Makefile looks right to me. Not sure what's going on. Can you try this on a new instance of Minikube? fwiw, I'm don't see the |
Beta Was this translation helpful? Give feedback.
-
Every-time you run 'make cluster' it deletes the old minikube instance and creates a new one and I have reproduced this several times, you can see that action here: https://github.com/seizadi/argo/blob/master/examples/argorollout/Makefile#L7 You don't need to specify the memory minikube adjusts memory as a function of core you add, see log here: You should be able to recreate the problem on your minikube by forking the repo and at the top level: cd examples/argorollout
make rollout
make test The make rollout target will create minikube, install linkerd and rollout operator. The make test target will create the test application with the 500 error for the linkerd proxy. |
Beta Was this translation helpful? Give feedback.
-
Ok, looks like latest image now has the necessary changes.... ❯ cd examples/argorollout
❯ make rollout
minikube stop; minikube delete;
✋ Stopping "minikube" in hyperkit ...
🛑 Node "minikube" stopped.
🔥 Deleting "minikube" in hyperkit ...
💀 Removed all traces of the "minikube" cluster.
minikube start --cpus=3
😄 minikube v1.11.0 on Darwin 10.15.5
✨ Automatically selected the hyperkit driver. Other choices: docker, virtualbox
👍 Starting control plane node minikube in cluster minikube
🔥 Creating hyperkit VM (CPUs=3, Memory=6000MB, Disk=20000MB) ...
🐳 Preparing Kubernetes v1.18.3 on Docker 19.03.8 ...
🔎 Verifying Kubernetes components...
🌟 Enabled addons: default-storageclass, storage-provisioner
🏄 Done! kubectl is now configured to use "minikube"
minikube addons enable ingress
🌟 The 'ingress' addon is enabled
minikube addons enable metrics-server
🌟 The 'metrics-server' addon is enabled
Built minikube cluster
linkerd check --pre # validate that Linkerd can be installed
kubernetes-api
--------------
√ can initialize the client
√ can query the Kubernetes API
......
linkerd-grafana
---------------
√ grafana add-on service account exists
√ grafana add-on config map exists
√ grafana pod is running
Status check results are √
Done with deploy Linkerd
kubectl create namespace argo-rollouts
namespace/argo-rollouts created
......
service/argo-rollouts-metrics created
deployment.apps/argo-rollouts created
Done with deploy Argo Rollout
❯ make test
kubectl apply -f examples/smi/namespace.yaml
namespace/test created
kubectl apply -f examples/smi/rollout.yaml
rollout.argoproj.io/rollouts-demo created
kubectl apply -f examples/smi/services.yaml
service/rollouts-demo-canary created
service/rollouts-demo-stable created
kubectl apply -f examples/smi/ingress.yaml
ingress.networking.k8s.io/rollouts-demo-stable created Now you have a namespace test and the problem with linkerd-proxy kubectl -n test describe pods
....
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal Scheduled 17m default-scheduler Successfully assigned test/rollouts-demo-868f9df8cd-brmqt to minikube
Normal Pulled 17m kubelet, minikube Container image "gcr.io/linkerd-io/proxy-init:v1.3.3" already present on machine
Normal Created 17m kubelet, minikube Created container linkerd-init
Normal Started 17m kubelet, minikube Started container linkerd-init
Normal Pulling 17m kubelet, minikube Pulling image "argoproj/rollouts-demo:blue"
Normal Pulled 17m kubelet, minikube Successfully pulled image "argoproj/rollouts-demo:blue"
Normal Created 17m kubelet, minikube Created container rollouts-demo
Normal Started 17m kubelet, minikube Started container rollouts-demo
Normal Pulled 17m kubelet, minikube Container image "gcr.io/linkerd-io/proxy:stable-2.8.0" already present on machine
Normal Created 17m kubelet, minikube Created container linkerd-proxy
Normal Started 17m kubelet, minikube Started container linkerd-proxy
Warning Unhealthy 2m54s (x90 over 17m) kubelet, minikube Readiness probe failed: HTTP probe failed with statuscode: 503 |
Beta Was this translation helpful? Give feedback.
-
can anyone help me with this please. |
Beta Was this translation helpful? Give feedback.
-
I got the same error, and it turned out an ancient version of linkerd-proxy sidecar was running in the pod, so I ran linkerd inject on the deployments in the namespace, and that solved it in my cluster :) |
Beta Was this translation helpful? Give feedback.
-
Had this issue because of Removing this and falling back to the actual dnsPolicy default ( |
Beta Was this translation helpful? Give feedback.
-
We saw this same problem, although it was in a seldom used environment and our short log rotation caused me to lose most logs other than the ongoing So we first tried restarting these 3 deployments: So ultimately, we just did a rolling restart of all the linkerd deployments which fixed it: for deploy1 in $(kubectl get deploy -n linkerd -oname); do
echo ${deploy1}
kubectl rollout restart ${deploy1} -n linkerd
done |
Beta Was this translation helpful? Give feedback.
-
Bug Report
What is the issue?
My container starts up ok and passes live and readiness tests. When I port-forward to the linkerd-proxy container it is returning a 503 on the /ready endpoint. This causes the pod to not be ready.
How can it be reproduced?
All I've done is manually inject linkerd using annotations, skipping redis and mongodb ports.
Logs, error output, etc
Logs from the linkerd-proxy container:
The AccessDenied log is continuous every few milliseconds. The thing running at 10.244.157.32 is linkerd-prometheus pod.
linkerd check
outputEnvironment
Beta Was this translation helpful? Give feedback.
All reactions