Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Issue cluster - component - konnectivity #1113

Closed
wheestermans31 opened this issue Sep 7, 2021 · 29 comments
Closed

Issue cluster - component - konnectivity #1113

wheestermans31 opened this issue Sep 7, 2021 · 29 comments
Labels
bug Something isn't working

Comments

@wheestermans31
Copy link

I have a up-and-running cluster with three controllers and two workers.

sh-4.2# /usr/local/bin/k0s kubectl get node
NAME         STATUS   ROLES    AGE   VERSION
lxdocapa22   Ready    <none>   75m   v1.22.1+k0s
lxdocapa23   Ready    <none>   71m   v1.22.1+k0s
sh-4.2# /usr/local/bin/k0s kubectl -n kube-system get pod -o wide
NAME                              READY   STATUS    RESTARTS   AGE   IP               NODE         NOMINATED NODE   READINESS GATES
coredns-5ccbdcc4c4-8d9fz          1/1     Running   0          45m   10.244.1.25      lxdocapa23   <none>           <none>
coredns-5ccbdcc4c4-trgvz          1/1     Running   0          45m   10.244.0.41      lxdocapa22   <none>           <none>
konnectivity-agent-8zbw4          1/1     Running   0          41m   10.244.0.47      lxdocapa22   <none>           <none>
konnectivity-agent-lfdq7          1/1     Running   0          41m   10.244.1.32      lxdocapa23   <none>           <none>
kube-proxy-76fwz                  1/1     Running   0          45m   10.100.114.101   lxdocapa23   <none>           <none>
kube-proxy-7m8mw                  1/1     Running   0          45m   10.100.114.100   lxdocapa22   <none>           <none>
kube-router-7jt6l                 1/1     Running   0          44m   10.100.114.101   lxdocapa23   <none>           <none>
kube-router-kzgfz                 1/1     Running   0          44m   10.100.114.100   lxdocapa22   <none>           <none>
metrics-server-6bd95db5f4-vckcb   1/1     Running   0          45m   10.244.0.42      lxdocapa22   <none>           <none>
sh-4.2# 

The kubectl is not giving me any issue, from that point of view it looks fine.

But the kubectl command is fast on one of the controllers, and very slow on the two others. probably fast on the leader of the cluster.

The only point I see are the following logging related to the konnectivity component.

Logs on node that is fast:

Sep  7 17:08:12 lxdocapa20 k0s: time="2021-09-07 17:08:12" level=info msg="E0907 17:08:12.677469   18582 server.go:382] \"Stream read from frontend failure\" err=\"rpc error: code = Canceled desc = context canceled\"" component=konnectivity
Sep  7 17:08:12 lxdocapa20 k0s: time="2021-09-07 17:08:12" level=info msg="E0907 17:08:12.678119   18582 server.go:736] \"could not get frontend client\" err=\"can't find connID 79 in the frontends[a8980db4-e38e-47be-a189-c73aef053331]\" connectionID=79" component=konnectivity

But on the ones that are slow I see the following logs:

Sep  7 17:05:46 lxdocapa17 k0s: time="2021-09-07 17:05:46" level=info msg="E0907 17:05:46.132937   27648 server.go:382] \"Stream read from frontend failure\" err=\"rpc error: code = Canceled desc = context canceled\"" component=konnectivity
Sep  7 17:05:46 lxdocapa17 k0s: time="2021-09-07 17:05:46" level=info msg="E0907 17:05:46.132950   27648 server.go:382] \"Stream read from frontend failure\" err=\"rpc error: code = Canceled desc = context canceled\"" component=konnectivity
Sep  7 17:05:46 lxdocapa17 k0s: time="2021-09-07 17:05:46" level=info msg="E0907 17:05:46.132995   27648 server.go:382] \"Stream read from frontend failure\" err=\"rpc error: code = Canceled desc = context canceled\"" component=konnectivity
Sep  7 17:05:46 lxdocapa17 k0s: time="2021-09-07 17:05:46" level=info msg="E0907 17:05:46.132998   27648 server.go:382] \"Stream read from frontend failure\" err=\"rpc error: code = Canceled desc = context canceled\"" component=konnectivity
Sep  7 17:05:46 lxdocapa17 k0s: time="2021-09-07 17:05:46" level=info msg="E0907 17:05:46.133012   27648 server.go:382] \"Stream read from frontend failure\" err=\"rpc error: code = Canceled desc = context canceled\"" component=konnectivity
Sep  7 17:05:51 lxdocapa17 k0s: time="2021-09-07 17:05:51" level=info msg="I0907 17:05:51.133034   27623 trace.go:205] Trace[641514455]: \"Proxy via grpc protocol over uds\" address:10.97.148.61:443 (07-Sep-2021 17:05:46.131) (total time: 5001ms):" component=kube-apiserver
Sep  7 17:05:51 lxdocapa17 k0s: time="2021-09-07 17:05:51" level=info msg="Trace[641514455]: [5.001040281s] [5.001040281s] END" component=kube-apiserver
Sep  7 17:05:51 lxdocapa17 k0s: time="2021-09-07 17:05:51" level=info msg="I0907 17:05:51.133057   27623 trace.go:205] Trace[787933724]: \"Proxy via grpc protocol over uds\" address:10.97.148.61:443 (07-Sep-2021 17:05:46.131) (total time: 5001ms):" component=kube-apiserver
Sep  7 17:05:51 lxdocapa17 k0s: time="2021-09-07 17:05:51" level=info msg="Trace[787933724]: [5.001144092s] [5.001144092s] END" component=kube-apiserver
Sep  7 17:05:51 lxdocapa17 k0s: time="2021-09-07 17:05:51" level=info msg="I0907 17:05:51.133097   27623 trace.go:205] Trace[679539168]: \"Proxy via grpc protocol over uds\" address:10.97.148.61:443 (07-Sep-2021 17:05:46.131) (total time: 5001ms):" component=kube-apiserver
Sep  7 17:05:51 lxdocapa17 k0s: time="2021-09-07 17:05:51" level=info msg="Trace[679539168]: [5.001097475s] [5.001097475s] END" component=kube-apiserver
Sep  7 17:05:51 lxdocapa17 k0s: time="2021-09-07 17:05:51" level=info msg="I0907 17:05:51.133124   27623 trace.go:205] Trace[497685480]: \"Proxy via grpc protocol over uds\" address:10.97.148.61:443 (07-Sep-2021 17:05:46.131) (total time: 5001ms):" component=kube-apiserver
Sep  7 17:05:51 lxdocapa17 k0s: time="2021-09-07 17:05:51" level=info msg="Trace[497685480]: [5.001187233s] [5.001187233s] END" component=kube-apiserver
Sep  7 17:05:51 lxdocapa17 k0s: time="2021-09-07 17:05:51" level=info msg="I0907 17:05:51.133176   27623 trace.go:205] Trace[1558194979]: \"Proxy via grpc protocol over uds\" address:10.97.148.61:443 (07-Sep-2021 17:05:46.132) (total time: 5001ms):" component=kube-apiserver
Sep  7 17:05:51 lxdocapa17 k0s: time="2021-09-07 17:05:51" level=info msg="Trace[1558194979]: [5.001150444s] [5.001150444s] END" component=kube-apiserver
Sep  7 17:05:51 lxdocapa17 k0s: time="2021-09-07 17:05:51" level=info msg="E0907 17:05:51.138133   27623 available_controller.go:524] v1beta1.metrics.k8s.io failed with: Operation cannot be fulfilled on apiservices.apiregistration.k8s.io \"v1beta1.metrics.k8s.io\": the object has been modified; please apply your changes to the latest version and try again" component=kube-apiserver
Sep  7 17:05:51 lxdocapa17 k0s: time="2021-09-07 17:05:51" level=info msg="E0907 17:05:51.139243   27648 server.go:413] \"Failed to get a backend\" err=\"No backend available\"" component=konnectivity
Sep  7 17:05:51 lxdocapa17 k0s: time="2021-09-07 17:05:51" level=info msg="E0907 17:05:51.139271   27648 server.go:413] \"Failed to get a backend\" err=\"No backend available\"" component=konnectivity
Sep  7 17:05:51 lxdocapa17 k0s: time="2021-09-07 17:05:51" level=info msg="E0907 17:05:51.139245   27648 server.go:413] \"Failed to get a backend\" err=\"No backend available\"" component=konnectivity
Sep  7 17:05:51 lxdocapa17 k0s: time="2021-09-07 17:05:51" level=info msg="E0907 17:05:51.139252   27648 server.go:413] \"Failed to get a backend\" err=\"No backend available\"" component=konnectivity
Sep  7 17:05:51 lxdocapa17 k0s: time="2021-09-07 17:05:51" level=info msg="E0907 17:05:51.139252   27648 server.go:413] \"Failed to get a backend\" err=\"No backend available\"" component=konnectivity
Sep  7 17:05:51 lxdocapa17 k0s: time="2021-09-07 17:05:51" level=info msg="I0907 17:05:51.139497   27623 client.go:117] DialResp not recognized; dropped" component=kube-apiserver
Sep  7 17:05:51 lxdocapa17 k0s: time="2021-09-07 17:05:51" level=info msg="I0907 17:05:51.139498   27623 client.go:117] DialResp not recognized; dropped" component=kube-apiserver
Sep  7 17:05:51 lxdocapa17 k0s: time="2021-09-07 17:05:51" level=info msg="I0907 17:05:51.139559   27623 client.go:117] DialResp not recognized; dropped" component=kube-apiserver
Sep  7 17:05:51 lxdocapa17 k0s: time="2021-09-07 17:05:51" level=info msg="I0907 17:05:51.139567   27623 client.go:117] DialResp not recognized; dropped" component=kube-apiserver
Sep  7 17:05:51 lxdocapa17 k0s: time="2021-09-07 17:05:51" level=info msg="I0907 17:05:51.139494   27623 client.go:117] DialResp not recognized; dropped" component=kube-apiserver
Sep  7 17:05:51 lxdocapa17 k0s: time="2021-09-07 17:05:51" level=info msg="E0907 17:05:51.139631   27648 server.go:382] \"Stream read from frontend failure\" err=\"rpc error: code = Canceled desc = context canceled\"" component=konnectivity
Sep  7 17:05:51 lxdocapa17 k0s: time="2021-09-07 17:05:51" level=info msg="E0907 17:05:51.139631   27648 server.go:382] \"Stream read from frontend failure\" err=\"rpc error: code = Canceled desc = context canceled\"" component=konnectivity
Sep  7 17:05:51 lxdocapa17 k0s: time="2021-09-07 17:05:51" level=info msg="E0907 17:05:51.139655   27648 server.go:382] \"Stream read from frontend failure\" err=\"rpc error: code = Canceled desc = context canceled\"" component=konnectivity
Sep  7 17:05:51 lxdocapa17 k0s: time="2021-09-07 17:05:51" level=info msg="E0907 17:05:51.139668   27648 server.go:382] \"Stream read from frontend failure\" err=\"rpc error: code = Canceled desc = context canceled\"" component=konnectivity
Sep  7 17:05:51 lxdocapa17 k0s: time="2021-09-07 17:05:51" level=info msg="E0907 17:05:51.139681   27648 server.go:382] \"Stream read from frontend failure\" err=\"rpc error: code = Canceled desc = context canceled\"" component=konnectivity
Sep  7 17:05:52 lxdocapa17 k0s: time="2021-09-07 17:05:52" level=info msg="E0907 17:05:52.997037   27648 server.go:413] \"Failed to get a backend\" err=\"No backend available\"" component=konnectivity
Sep  7 17:05:52 lxdocapa17 k0s: time="2021-09-07 17:05:52" level=info msg="I0907 17:05:52.997255   27623 client.go:117] DialResp not recognized; dropped" component=kube-apiserver
Sep  7 17:05:52 lxdocapa17 k0s: time="2021-09-07 17:05:52" level=info msg="E0907 17:05:52.997373   27648 server.go:382] \"Stream read from frontend failure\" err=\"rpc error: code = Canceled desc = context canceled\"" component=konnectivity
Sep  7 17:05:54 lxdocapa17 k0s: time="2021-09-07 17:05:54" level=info msg="current cfg matches existing, not gonna do anything" component=kubeproxy
Sep  7 17:05:54 lxdocapa17 k0s: time="2021-09-07 17:05:54" level=info msg="current cfg matches existing, not gonna do anything" component=coredns
Sep  7 17:05:56 lxdocapa17 k0s: time="2021-09-07 17:05:56" level=info msg="I0907 17:05:56.139050   27623 trace.go:205] Trace[701055894]: \"Proxy via grpc protocol over uds\" address:10.97.148.61:443 (07-Sep-2021 17:05:51.138) (total time: 5000ms):" component=kube-apiserver
Sep  7 17:05:56 lxdocapa17 k0s: time="2021-09-07 17:05:56" level=info msg="Trace[701055894]: [5.000532546s] [5.000532546s] END" component=kube-apiserver
Sep  7 17:05:56 lxdocapa17 k0s: time="2021-09-07 17:05:56" level=info msg="I0907 17:05:56.139050   27623 trace.go:205] Trace[1382020849]: \"Proxy via grpc protocol over uds\" address:10.97.148.61:443 (07-Sep-2021 17:05:51.138) (total time: 5000ms):" component=kube-apiserver
Sep  7 17:05:56 lxdocapa17 k0s: time="2021-09-07 17:05:56" level=info msg="Trace[1382020849]: [5.00075802s] [5.00075802s] END" component=kube-apiserver
Sep  7 17:05:56 lxdocapa17 k0s: time="2021-09-07 17:05:56" level=info msg="I0907 17:05:56.139073   27623 trace.go:205] Trace[1290110452]: \"Proxy via grpc protocol over uds\" address:10.97.148.61:443 (07-Sep-2021 17:05:51.138) (total time: 5000ms):" component=kube-apiserver
Sep  7 17:05:56 lxdocapa17 k0s: time="2021-09-07 17:05:56" level=info msg="Trace[1290110452]: [5.000770225s] [5.000770225s] END" component=kube-apiserver
Sep  7 17:05:56 lxdocapa17 k0s: time="2021-09-07 17:05:56" level=info msg="I0907 17:05:56.139096   27623 trace.go:205] Trace[1456862550]: \"Proxy via grpc protocol over uds\" address:10.97.148.61:443 (07-Sep-2021 17:05:51.138) (total time: 5000ms):" component=kube-apiserver
Sep  7 17:05:56 lxdocapa17 k0s: time="2021-09-07 17:05:56" level=info msg="Trace[1456862550]: [5.000653482s] [5.000653482s] END" component=kube-apiserver
Sep  7 17:05:56 lxdocapa17 k0s: time="2021-09-07 17:05:56" level=info msg="I0907 17:05:56.139119   27623 trace.go:205] Trace[77502346]: \"Proxy via grpc protocol over uds\" address:10.97.148.61:443 (07-Sep-2021 17:05:51.138) (total time: 5000ms):" component=kube-apiserver
Sep  7 17:05:56 lxdocapa17 k0s: time="2021-09-07 17:05:56" level=info msg="Trace[77502346]: [5.000738134s] [5.000738134s] END" component=kube-apiserver
Sep  7 17:05:56 lxdocapa17 k0s: time="2021-09-07 17:05:56" level=info msg="E0907 17:05:56.146144   27623 available_controller.go:524] v1beta1.metrics.k8s.io failed with: Operation cannot be fulfilled on apiservices.apiregistration.k8s.io \"v1beta1.metrics.k8s.io\": the object has been modified; please apply your changes to the latest version and try again" component=kube-apiserver
Sep  7 17:05:56 lxdocapa17 k0s: time="2021-09-07 17:05:56" level=info msg="E0907 17:05:56.147087   27648 server.go:413] \"Failed to get a backend\" err=\"No backend available\"" component=konnectivity
Sep  7 17:05:56 lxdocapa17 k0s: time="2021-09-07 17:05:56" level=info msg="E0907 17:05:56.147106   27648 server.go:413] \"Failed to get a backend\" err=\"No backend available\"" component=konnectivity
Sep  7 17:05:56 lxdocapa17 k0s: time="2021-09-07 17:05:56" level=info msg="E0907 17:05:56.147110   27648 server.go:413] \"Failed to get a backend\" err=\"No backend available\"" component=konnectivity
Sep  7 17:05:56 lxdocapa17 k0s: time="2021-09-07 17:05:56" level=info msg="E0907 17:05:56.147107   27648 server.go:413] \"Failed to get a backend\" err=\"No backend available\"" component=konnectivity
Sep  7 17:05:56 lxdocapa17 k0s: time="2021-09-07 17:05:56" level=info msg="E0907 17:05:56.147087   27648 server.go:413] \"Failed to get a backend\" err=\"No backend available\"" component=konnectivity
Sep  7 17:05:56 lxdocapa17 k0s: time="2021-09-07 17:05:56" level=info msg="I0907 17:05:56.147250   27623 client.go:117] DialResp not recognized; dropped" component=kube-apiserver
Sep  7 17:05:56 lxdocapa17 k0s: time="2021-09-07 17:05:56" level=info msg="I0907 17:05:56.147255   27623 client.go:117] DialResp not recognized; dropped" component=kube-apiserver
Sep  7 17:05:56 lxdocapa17 k0s: time="2021-09-07 17:05:56" level=info msg="I0907 17:05:56.147287   27623 client.go:117] DialResp not recognized; dropped" component=kube-apiserver
Sep  7 17:05:56 lxdocapa17 k0s: time="2021-09-07 17:05:56" level=info msg="I0907 17:05:56.147300   27623 client.go:117] DialResp not recognized; dropped" component=kube-apiserver
Sep  7 17:05:56 lxdocapa17 k0s: time="2021-09-07 17:05:56" level=info msg="I0907 17:05:56.147257   27623 client.go:117] DialResp not recognized; dropped" component=kube-apiserver
Sep  7 17:05:56 lxdocapa17 k0s: time="2021-09-07 17:05:56" level=info msg="E0907 17:05:56.147338   27648 server.go:382] \"Stream read from frontend failure\" err=\"rpc error: code = Canceled desc = context canceled\"" component=konnectivity
Sep  7 17:05:56 lxdocapa17 k0s: time="2021-09-07 17:05:56" level=info msg="E0907 17:05:56.147337   27648 server.go:382] \"Stream read from frontend failure\" err=\"rpc error: code = Canceled desc = context canceled\"" component=konnectivity
Sep  7 17:05:56 lxdocapa17 k0s: time="2021-09-07 17:05:56" level=info msg="E0907 17:05:56.147388   27648 server.go:382] \"Stream read from frontend failure\" err=\"rpc error: code = Canceled desc = context canceled\"" component=konnectivity
Sep  7 17:05:56 lxdocapa17 k0s: time="2021-09-07 17:05:56" level=info msg="E0907 17:05:56.147395   27648 server.go:382] \"Stream read from frontend failure\" err=\"rpc error: code = Canceled desc = context canceled\"" component=konnectivity
Sep  7 17:05:56 lxdocapa17 k0s: time="2021-09-07 17:05:56" level=info msg="E0907 17:05:56.147431   27648 server.go:382] \"Stream read from frontend failure\" err=\"rpc error: code = Canceled desc = context canceled\"" component=konnectivity
Sep  7 17:05:57 lxdocapa17 k0s: time="2021-09-07 17:05:57" level=info msg="E0907 17:05:57.549191   27648 server.go:413] \"Failed to get a backend\" err=\"No backend available\"" component=konnectivity
Sep  7 17:05:57 lxdocapa17 k0s: time="2021-09-07 17:05:57" level=info msg="I0907 17:05:57.549484   27623 client.go:117] DialResp not recognized; dropped" component=kube-apiserver
Sep  7 17:05:57 lxdocapa17 k0s: time="2021-09-07 17:05:57" level=info msg="E0907 17:05:57.549664   27648 server.go:382] \"Stream read from frontend failure\" err=\"rpc error: code = Canceled desc = context canceled\"" component=konnectivity
Sep  7 17:05:57 lxdocapa17 k0s: time="2021-09-07 17:05:57" level=info msg="I0907 17:05:57.996948   27623 trace.go:205] Trace[1232079086]: \"Proxy via grpc protocol over uds\" address:10.97.148.61:443 (07-Sep-2021 17:05:52.996) (total time: 5000ms):" component=kube-apiserver
Sep  7 17:05:57 lxdocapa17 k0s: time="2021-09-07 17:05:57" level=info msg="Trace[1232079086]: [5.000754682s] [5.000754682s] END" component=kube-apiserver

Probably related...

@wheestermans31 wheestermans31 added the bug Something isn't working label Sep 7, 2021
@jnummelin
Copy link
Member

Do you have and LB setup in front of the controllers? As stated in docs HA controlplane needs LB so that e.g. konnectivity-agents can establish proper connections with ALL controllers.

@wheestermans31
Copy link
Author

wheestermans31 commented Sep 8, 2021

It was not clear that this was a requirement, requesting now for load-balancer, and then change the configuration of the controllers.

I will update the config file used during the installation process

spec:
  api:
    externalAddress: <load balancer public ip address>
    sans:
    - <load balancer public ip address>

I have executed the following installation for the cluster setup:

Controller 1: /usr/local/bin/k0s default-config > /opt/docker/common/docker-configuration/k0s/k0s-cluster-config.yaml
Controller 2: /usr/local/bin/k0s install controller --token-file /opt/docker/common/docker-configuration/k0s/join-token-controller
Controller 3: /usr/local/bin/k0s install controller --token-file /opt/docker/common/docker-configuration/k0s/join-token-controller

I will update the /opt/docker/common/docker-configuration/k0s/k0s-cluster-config.yaml file witht the load-balancer reference.

Then I need to stop/start the controllers, this is clear for the controller 1 where the config file was used, but what about the controller2/3?

@jnummelin
Copy link
Member

Currently you need to copy over the same config bits to all controllers so they get booted up the same way. And naturally a restart of all is needed if the config changes.

We're building a more "robust" way to handle the config as a kubernetes object but that work is still ongoing (#1008)

@jnummelin
Copy link
Member

The requirement for LB basically stems from the way konnectivity is used to "tunnel" the api --> worker communications via konnectivity-agents. As we see from the logs, each API instance is trying to connect to the workers and pods (e.g. the metrics stuff) thus each API servers needs to have a konnectivity-agent connected so that it can actually reach the worker services.

When there's no LB and no spec.api.externalAddress specified, the agents will connect only with one API instance. Thus APIs 2..n will have no connectivity to the workers.

@wheestermans31
Copy link
Author

wheestermans31 commented Sep 8, 2021

HA F5 LB in place, but I see

Sep  8 17:53:47 lxdocapa21 k0s: time="2021-09-08 17:53:47" level=info msg="E0908 17:53:47.921739    3704 server.go:382] \"Stream read from frontend failure\" err=\"rpc error: code = Canceled desc = context canceled\"" component=konnectivity
Sep  8 17:53:47 lxdocapa21 k0s: time="2021-09-08 17:53:47" level=info msg="E0908 17:53:47.922505    3704 server.go:736] \"could not get frontend client\" err=\"can't find connID 260 in the frontends[dd6f386b-3f13-4a4d-bd4f-30d45b29d258]\" connectionID=260" component=konnectivity
Sep  8 17:53:47 lxdocapa21 k0s: time="2021-09-08 17:53:47" level=info msg="E0908 17:53:47.991983    3704 server.go:382] \"Stream read from frontend failure\" err=\"rpc error: code = Canceled desc = context canceled\"" component=konnectivity
Sep  8 17:53:47 lxdocapa21 k0s: time="2021-09-08 17:53:47" level=info msg="E0908 17:53:47.992474    3704 server.go:736] \"could not get frontend client\" err=\"can't find connID 262 in the frontends[dd6f386b-3f13-4a4d-bd4f-30d45b29d258]\" connectionID=262" component=konnectivity

@wheestermans31
Copy link
Author

We continue to have these errors, I have checked some scenarios but still no solution for this.

@wheestermans31
Copy link
Author

Linked to this issue I see that kubectl commands from two of the three controllers is slow.

@wheestermans31
Copy link
Author

This seems to be related to a running metrics pod. When not running I see that "/usr/local/bin/k0s kubectl -n kube-system get deploy" on all my controllers is fast.

At the same time I don't see those errors:

Sep  9 18:01:42 lxdocapa17 k0s: time="2021-09-09 18:01:42" level=info msg="E0909 18:01:42.208063   13043 server.go:413] \"Failed to get a backend\" err=\"No backend available\"" componentkonnectivity
Sep  9 18:01:42 lxdocapa17 k0s: time="2021-09-09 18:01:42" level=info msg="E0909 18:01:42.208085   13043 server.go:413] \"Failed to get a backend\" err=\"No backend available\"" componentkonnectivity
Sep  9 18:01:42 lxdocapa17 k0s: time="2021-09-09 18:01:42" level=info msg="E0909 18:01:42.208104   13043 server.go:413] \"Failed to get a backend\" err=\"No backend available\"" componentkonnectivity
Sep  9 18:01:42 lxdocapa17 k0s: time="2021-09-09 18:01:42" level=info msg="E0909 18:01:42.208108   13043 server.go:413] \"Failed to get a backend\" err=\"No backend available\"" componentkonnectivity
Sep  9 18:01:42 lxdocapa17 k0s: time="2021-09-09 18:01:42" level=info msg="E0909 18:01:42.208254   13043 server.go:382] \"Stream read from frontend failure\" err=\"rpc error: code = Canceled desc = context canceled\"" component=konnectivity
Sep  9 18:01:42 lxdocapa17 k0s: time="2021-09-09 18:01:42" level=info msg="E0909 18:01:42.208423   13043 server.go:382] \"Stream read from frontend failure\" err=\"rpc error: code = Canceled desc = context canceled\"" component=konnectivity
Sep  9 18:01:42 lxdocapa17 k0s: time="2021-09-09 18:01:42" level=info msg="E0909 18:01:42.208484   13043 server.go:382] \"Stream read from frontend failure\" err=\"rpc error: code = Canceled desc = context canceled\"" component=konnectivity
Sep  9 18:01:42 lxdocapa17 k0s: time="2021-09-09 18:01:42" level=info msg="E0909 18:01:42.208489   13043 server.go:382] \"Stream read from frontend failure\" err=\"rpc error: code = Canceled desc = context canceled\"" component=konnectivity
Sep  9 18:01:42 lxdocapa17 k0s: time="2021-09-09 18:01:42" level=info msg="E0909 18:01:42.208512   13043 server.go:382] \"Stream read from frontend failure\" err=\"rpc error: code = Canceled desc = context canceled\"" component=konnectivity

Once the metrics is started, the above errors are back, and the kubectl command on two of the three is very slow. I think the one that is fast is the leader.

@wheestermans31
Copy link
Author

wheestermans31 commented Sep 10, 2021

Logs metrics pod

sh-4.2# /usr/local/bin/k0s kubectl -n kube-system logs -f metrics-server-6bd95db5f4-b252p  
I0910 05:39:08.043193       1 serving.go:341] Generated self-signed cert (/tmp/apiserver.crt, /tmp/apiserver.key)
I0910 05:39:08.607465       1 configmap_cafile_content.go:202] Starting client-ca::kube-system::extension-apiserver-authentication::client-ca-file
I0910 05:39:08.607488       1 configmap_cafile_content.go:202] Starting client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file
I0910 05:39:08.607507       1 shared_informer.go:240] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file
I0910 05:39:08.607496       1 shared_informer.go:240] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file
I0910 05:39:08.607470       1 requestheader_controller.go:169] Starting RequestHeaderAuthRequestController
I0910 05:39:08.607562       1 shared_informer.go:240] Waiting for caches to sync for RequestHeaderAuthRequestController
I0910 05:39:08.607714       1 dynamic_serving_content.go:130] Starting serving-cert::/tmp/apiserver.crt::/tmp/apiserver.key
I0910 05:39:08.608189       1 secure_serving.go:197] Serving securely on 0.0.0.0:443
I0910 05:39:08.608279       1 tlsconfig.go:240] Starting DynamicServingCertificateController
I0910 05:39:08.708005       1 shared_informer.go:247] Caches are synced for RequestHeaderAuthRequestController 
I0910 05:39:08.708065       1 shared_informer.go:247] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file 
I0910 05:39:08.708064       1 shared_informer.go:247] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::client-ca-file 

@wheestermans31
Copy link
Author

wheestermans31 commented Sep 23, 2021

I restarted the activities on K0s. Anybody a clue what the issue may be?

@jnummelin
Copy link
Member

Are you able to access metrics through kubectl top pod,node?

What do you see in k0s kc get apiservices v1beta1.metrics.k8s.io -o yaml, is the status reporting that API has been able to properly connect to the metrics? Maybe check the same on all controllers

What we have seen is that metrics being "not ready" is slowing down the API discovery (the process where kubectl/other clients discover which APIs are available on the kube apiserver) as it reports some 5xx responses for the discovery. Maybe check the kubectl command in verbose mode (maybe with -v=7) if it reveals where the time is spent.

@wheestermans31
Copy link
Author

/usr/local/bin/k0s kubectl get apiservices v1beta1.metrics.k8s.io -o yaml
apiVersion: apiregistration.k8s.io/v1
kind: APIService
metadata:
annotations:
k0s.k0sproject.io/last-applied-configuration: |
{"apiVersion":"apiregistration.k8s.io/v1","kind":"APIService","metadata":{"labels":{"k8s-app":"metrics-server"},"name":"v1beta1.metrics.k8s.io"},"spec":{"group":"metrics.k8s.io","groupPriorityMinimum":100,"insecureSkipTLSVerify":true,"service":{"name":"metrics-server","namespace":"kube-system"},"version":"v1beta1","versionPriority":100}}
k0s.k0sproject.io/stack-checksum: f46335047740a30cd79c90be43ee087f
creationTimestamp: "2021-09-07T14:08:02Z"
labels:
k0s.k0sproject.io/stack: metricserver
k8s-app: metrics-server
name: v1beta1.metrics.k8s.io
resourceVersion: "3299679"
uid: b35ac743-896f-4f40-bd1a-e16c78294824
spec:
group: metrics.k8s.io
groupPriorityMinimum: 100
insecureSkipTLSVerify: true
service:
name: metrics-server
namespace: kube-system
port: 443
version: v1beta1
versionPriority: 100
status:
conditions:

  • lastTransitionTime: "2021-09-27T14:35:25Z"
    message: all checks passed
    reason: Passed
    status: "True"
    type: Available

Same on all controllers. Out of the three controllers there is always one fast where the other two replies, but much slower.

@wheestermans31
Copy link
Author

On the ones that are slow I see

Sep 27 16:44:16 lxdocapa20 k0s: time="2021-09-27 16:44:16" level=info msg="E0927 16:44:16.778876 10295 server.go:413] "Failed to get a backend" err="No backend available"" component=konnectivity
Sep 27 16:44:16 lxdocapa20 k0s: time="2021-09-27 16:44:16" level=info msg="E0927 16:44:16.778914 10295 server.go:413] "Failed to get a backend" err="No backend available"" component=konnectivity
Sep 27 16:44:16 lxdocapa20 k0s: time="2021-09-27 16:44:16" level=info msg="E0927 16:44:16.778940 10295 server.go:413] "Failed to get a backend" err="No backend available"" component=konnectivity
Sep 27 16:44:16 lxdocapa20 k0s: time="2021-09-27 16:44:16" level=info msg="E0927 16:44:16.778885 10295 server.go:413] "Failed to get a backend" err="No backend available"" component=konnectivity
Sep 27 16:44:16 lxdocapa20 k0s: time="2021-09-27 16:44:16" level=info msg="E0927 16:44:16.778960 10295 server.go:413] "Failed to get a backend" err="No backend available"" component=konnectivity

I don't see this on the node that replies fast to kubectl commands.

@wheestermans31
Copy link
Author

Logs metrics pod doesn't say much:

sh-4.2# /usr/local/bin/k0s kubectl -n kube-system logs -f metrics-server-6bd95db5f4-snvmw
I0927 14:34:11.800514 1 serving.go:341] Generated self-signed cert (/tmp/apiserver.crt, /tmp/apiserver.key)
I0927 14:34:12.417074 1 requestheader_controller.go:169] Starting RequestHeaderAuthRequestController
I0927 14:34:12.417095 1 configmap_cafile_content.go:202] Starting client-ca::kube-system::extension-apiserver-authentication::client-ca-file
I0927 14:34:12.417100 1 shared_informer.go:240] Waiting for caches to sync for RequestHeaderAuthRequestController
I0927 14:34:12.417108 1 shared_informer.go:240] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file
I0927 14:34:12.417120 1 configmap_cafile_content.go:202] Starting client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file
I0927 14:34:12.417151 1 shared_informer.go:240] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file
I0927 14:34:12.417450 1 dynamic_serving_content.go:130] Starting serving-cert::/tmp/apiserver.crt::/tmp/apiserver.key
I0927 14:34:12.418002 1 secure_serving.go:197] Serving securely on 0.0.0.0:443
I0927 14:34:12.418092 1 tlsconfig.go:240] Starting DynamicServingCertificateController
I0927 14:34:12.517725 1 shared_informer.go:247] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::client-ca-file
I0927 14:34:12.517783 1 shared_informer.go:247] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file
I0927 14:34:12.517834 1 shared_informer.go:247] Caches are synced for RequestHeaderAuthRequestController

Neither inside the /var/log/messages, on host where the metrics pod is running, there is nothing special.

From time to time I see inside the konnectivity pods, seems to be connection to the metrics server:

E0927 14:54:20.745230 1 client.go:515] "conn write failure" err="write tcp 10.244.1.161:54652->10.97.148.61:443: use of closed network connection" connectionID=44253
E0927 14:54:20.745252 1 client.go:486] "connection read failure" err="read tcp 10.244.1.161:54652->10.97.148.61:443: use of closed network connection"

/usr/local/bin/k0s kubectl -n kube-system get svc
NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE
kube-dns ClusterIP 10.96.0.10 53/UDP,53/TCP,9153/TCP 20d
metrics-server ClusterIP 10.97.148.61 443/TCP 20d

@wheestermans31
Copy link
Author

wheestermans31 commented Sep 27, 2021

Can this be related, #596?

But when I check

sh-4.2# while true; do lsof -p 12838 |wc -l && date ; sleep 5; done <<< API-server
169
Mon 27 Sep 17:08:33 CEST 2021
169
Mon 27 Sep 17:08:38 CEST 2021
169
Mon 27 Sep 17:08:43 CEST 2021
169
Mon 27 Sep 17:08:48 CEST 2021
169
Mon 27 Sep 17:08:53 CEST 2021

sh-4.2# while true; do lsof -p 119702 |wc -l && date ; sleep 5; done <<<< Konnectivity server
2420
Mon 27 Sep 17:11:01 CEST 2021
2420
Mon 27 Sep 17:11:06 CEST 2021
2420
Mon 27 Sep 17:11:11 CEST 2021
2420
Mon 27 Sep 17:11:17 CEST 2021
2420
Mon 27 Sep 17:11:22 CEST 2021

@wheestermans31
Copy link
Author

wheestermans31 commented Sep 27, 2021

Checked lsof on all controllers for /var/lib/k0s/bin/konnectivity-server

Controller 1

sh-4.2# while true; do lsof -p 119702 |wc -l && date ; sleep 5; done
2420
Mon 27 Sep 17:14:40 CEST 2021
2420
Mon 27 Sep 17:14:45 CEST 2021
2420
Mon 27 Sep 17:14:50 CEST 2021
2420
Mon 27 Sep 17:14:56 CEST 2021
2420

Controller 2

sh-4.2# while true; do lsof -p 10295 |wc -l && date ; sleep 5; done
54
Mon 27 Sep 17:15:47 CEST 2021
54
Mon 27 Sep 17:15:52 CEST 2021
54
Mon 27 Sep 17:15:57 CEST 2021
54

Controller 3

sh-4.2# while true; do lsof -p 11413 |wc -l && date ; sleep 5; done
112
Mon 27 Sep 17:17:40 CEST 2021
112
Mon 27 Sep 17:17:45 CEST 2021

@wheestermans31
Copy link
Author

The lsof for the apiserver is better

Controller 1

sh-4.2# while true; do lsof -p 12838 |wc -l && date ; sleep 5; done
169
Mon 27 Sep 17:19:52 CEST 2021
169
Mon 27 Sep 17:19:57 CEST 2021
169

Controller 2

sh-4.2# while true; do lsof -p 10250 |wc -l && date ; sleep 5; done
143
Mon 27 Sep 17:19:49 CEST 2021
143
Mon 27 Sep 17:19:54 CEST 2021
143

Controller 3

sh-4.2# while true; do lsof -p 11387 |wc -l && date ; sleep 5; done
196
Mon 27 Sep 17:19:46 CEST 2021
197
Mon 27 Sep 17:19:51 CEST 2021
196

@jnummelin
Copy link
Member

Can this be related, #596?

It does sound slightly similar at least. But in that issue the root cause was identified and fixed.

Could you also check that the konnektivity-agent pods are properly configured for the externalAddress. You should see args like this:

--proxy-server-host=1.2.3.4

where 1.2.3.4 is the externalAddress you've configured for the cluster.

@wheestermans31
Copy link
Author

When looking at the konnectivity agents

Args:
  --logtostderr=true
  --ca-cert=/var/run/secrets/kubernetes.io/serviceaccount/ca.crt
  **--proxy-server-host=10.100.114.99
  --proxy-server-port=8132**
  --service-account-token-path=/var/run/secrets/tokens/konnectivity-agent-token

10.100.114.99 --> this is the IP where the agent runs

This is what we use to access the cluster:

https://k8s-k0s-test.toyota-europe.com:6443

@jnummelin
Copy link
Member

10.100.114.99 --> this is the IP where the agent runs

What do you mean by agent, the konnectivity-agent? If you've configured the controllers with api.externalAddress pointing to the LB you have in front of controllers this api.externalAddress should be used for konnectivity-agents as the address to connect to.

What config you have on the konnectivity-server processes on controllers? You can check that e.g. with cat /proc/$(pidof konnectivity-server)/cmdline

@wheestermans31
Copy link
Author

I have changed the '--proxy-server-host=10.100.114.99' to '--proxy-server-host=k8s-k0s-test.toyota-europe.com', and this solved the issue, now all three managers are executing kubectl commands the same.

For high available solution I did the following changes, see external address.

apiVersion: k0s.k0sproject.io/v1beta1
kind: Cluster
metadata:
name: k0s
spec:
api:
address: 10.100.114.94
port: 6443
k0sApiPort: 9443
externalAddress: k8s-k0s-test.toyota-europe.com
sans:
- 10.100.114.94
- 10.100.8.112
- 10.100.13.110
- 172.17.0.1
- k8s-k0s-test.toyota-europe.com
storage:
type: etcd
etcd:
peerAddress: 10.100.114.94

But seems I had to update the daemonset konnectivity-agent too:

...
spec:
containers:
- args:
- --logtostderr=true
- --ca-cert=/var/run/secrets/kubernetes.io/serviceaccount/ca.crt
- --proxy-server-host=k8s-k0s-test.toyota-europe.com
- --proxy-server-port=8132
- --service-account-token-path=/var/run/secrets/tokens/konnectivity-agent-token
...

@jnummelin
Copy link
Member

hmm, k0s should detect the change in the externalAddress and go and re-configure konnectivity daemonset. apparently this did not happen properly in your setup for some reason. need to dive into why...

@jnummelin
Copy link
Member

I haven't been able to reproduce a case where the externalAddress change would not get reflected into the konnectivity daemonset. Maybe a silly question, but when you changed the config, did you change it to same address on ALL controllers and restarted the controllers?

@wheestermans31
Copy link
Author

I just did a complete reinstall of K0s cluster

apiVersion: k0s.k0sproject.io/v1beta1
kind: Cluster
metadata:
  name: k0s
spec:
  api:
    address: 10.100.114.94
    port: 6443
    k0sApiPort: 9443
    externalAddress: k8s-k0s-test.toyota-europe.com
    sans:
    - 10.100.114.94
#    - 10.100.80.171
#    - 10.100.13.119    
#    - 10.100.8.112
#    - 10.100.13.110
#    - 172.17.0.1
    - k8s-k0s-test.toyota-europe.com
  storage:
    type: etcd
    etcd:
      peerAddress: 10.100.114.94
  network:
    podCIDR: 10.244.0.0/16
    serviceCIDR: 10.96.0.0/12
    provider: calico
    calico: 
      mode: vxlan
      vxlanPort: 4789
      vxlanVNI: 4096
      mtu: 0
      wireguard: false
      flexVolumeDriverPath: /usr/libexec/k0s/kubelet-plugins/volume/exec/nodeagent~uds
      withWindowsNodes: false
      overlay: Always
    kuberouter:
      mtu: 0
      peerRouterIPs: ""
      peerRouterASNs: ""
      autoMTU: true
    kubeProxy:
      disabled: false
      mode: iptables
  podSecurityPolicy:
    defaultPolicy: 00-k0s-privileged
  telemetry:
    enabled: true
  installConfig:
    users:
      etcdUser: etcd
      kineUser: kube-apiserver
      konnectivityUser: konnectivity-server
      kubeAPIserverUser: kube-apiserver
      kubeSchedulerUser: kube-scheduler
  images:
    konnectivity:
      image: us.gcr.io/k8s-artifacts-prod/kas-network-proxy/proxy-agent
      version: v0.0.24
    metricsserver:
      image: gcr.io/k8s-staging-metrics-server/metrics-server
      version: v0.5.0
    kubeproxy:
      image: k8s.gcr.io/kube-proxy
      version: v1.22.1
    coredns:
      image: docker.io/coredns/coredns
      version: 1.7.0
    calico:
      cni:
        image: docker.io/calico/cni
        version: v3.18.1
      node:
        image: docker.io/calico/node
        version: v3.18.1
      kubecontrollers:
        image: docker.io/calico/kube-controllers
        version: v3.18.1
    kuberouter:
      cni:
        image: docker.io/cloudnativelabs/kube-router
        version: v1.2.1
      cniInstaller:
        image: quay.io/k0sproject/cni-node
        version: 0.1.0
    default_pull_policy: IfNotPresent
  konnectivity:
    agentPort: 8132
    adminPort: 8133

But inside the konnectivity daemonset I still see:

...
spec:
  revisionHistoryLimit: 10
  selector:
    matchLabels:
      k8s-app: konnectivity-agent
  template:
    metadata:
      creationTimestamp: null
      labels:
        k8s-app: konnectivity-agent
    spec:
      containers:
      - args:
        - --logtostderr=true
        - --ca-cert=/var/run/secrets/kubernetes.io/serviceaccount/ca.crt
        **- --proxy-server-host=10.100.114.99**
        - --proxy-server-port=8132
        - --service-account-token-path=/var/run/secrets/tokens/konnectivity-agent-token
...

So, still the IP address and not the external address?

@mikhail-sakhnov
Copy link
Contributor

@wheestermans31 I am not able to reproduce this.
I tried using your config (the toyota domain name was linked to 127.0.0.1 by overriding it throug /etc/hosts and etcd peer address was dropped in favor of default mode)

After that the new cluster has following definition for konnectivity agent (see below)

Let's check, how do you start the k0s binary and which version do you use?
Also, could you share the logs from controller process, starting from the bootstrap?

I have a feeling that your set up due to some reason ignores the k0s.yaml, is it possible that you pass the wrong value for -c command argument?

kind: DaemonSet
metadata:
  annotations:
    deprecated.daemonset.template.generation: "1"
    k0s.k0sproject.io/last-applied-configuration: |
      {"apiVersion":"apps/v1","kind":"DaemonSet","metadata":{"labels":{"addonmanager.kubernetes.io/mode":"Reconcile","k8s-app":"konnectivity-agent"},"name":"konnectivity-agent","namespace":"kube-system"},"spec":{"selector":{"matchLabels":{"k8s-app":"konnectivity-agent"}},"template":{"metadata":{"labels":{"k8s-app":"konnectivity-agent"}},"spec":{"containers":[{"args":["--logtostderr=true","--ca-cert=/var/run/secrets/kubernetes.io/serviceaccount/ca.crt","--proxy-server-host=k8s-k0s-test.toyota-europe.com","--proxy-server-port=8132","--service-account-token-path=/var/run/secrets/tokens/konnectivity-agent-token"],"command":["/proxy-agent"],"image":"us.gcr.io/k8s-artifacts-prod/kas-network-proxy/proxy-agent:v0.0.24","imagePullPolicy":"IfNotPresent","livenessProbe":{"httpGet":{"path":"/healthz","port":8093},"initialDelaySeconds":15,"timeoutSeconds":15},"name":"konnectivity-agent","volumeMounts":[{"mountPath":"/var/run/secrets/tokens","name":"konnectivity-agent-token"}]}],"nodeSelector":{"kubernetes.io/os":"linux"},"priorityClassName":"system-cluster-critical","serviceAccountName":"konnectivity-agent","tolerations":[{"key":"CriticalAddonsOnly","operator":"Exists"}],"volumes":[{"name":"konnectivity-agent-token","projected":{"sources":[{"serviceAccountToken":{"audience":"system:konnectivity-server","path":"konnectivity-agent-token"}}]}}]}}}}
    k0s.k0sproject.io/stack-checksum: 4a88f382105eb1fd93f19f3dc42e338a
  creationTimestamp: "2021-10-20T10:32:23Z"
  generation: 1
  labels:
    addonmanager.kubernetes.io/mode: Reconcile
    k0s.k0sproject.io/stack: konnectivity
    k8s-app: konnectivity-agent
  name: konnectivity-agent
  namespace: kube-system
  resourceVersion: "603"
  uid: 0cf0f2d5-ad17-41a1-b29b-13cf6b55acdf
spec:
  revisionHistoryLimit: 10
  selector:
    matchLabels:
      k8s-app: konnectivity-agent
  template:
    metadata:
      creationTimestamp: null
      labels:
        k8s-app: konnectivity-agent
    spec:
      containers:
      - args:
        - --logtostderr=true
        - --ca-cert=/var/run/secrets/kubernetes.io/serviceaccount/ca.crt
        - --proxy-server-host=k8s-k0s-test.toyota-europe.com
        - --proxy-server-port=8132
        - --service-account-token-path=/var/run/secrets/tokens/konnectivity-agent-token
        command:
        - /proxy-agent
        image: us.gcr.io/k8s-artifacts-prod/kas-network-proxy/proxy-agent:v0.0.24
        imagePullPolicy: IfNotPresent
        livenessProbe:
          failureThreshold: 3
          httpGet:
            path: /healthz
            port: 8093
            scheme: HTTP
          initialDelaySeconds: 15
          periodSeconds: 10
          successThreshold: 1
          timeoutSeconds: 15
        name: konnectivity-agent
        resources: {}
        terminationMessagePath: /dev/termination-log
        terminationMessagePolicy: File
        volumeMounts:
        - mountPath: /var/run/secrets/tokens
          name: konnectivity-agent-token
      dnsPolicy: ClusterFirst
      nodeSelector:
        kubernetes.io/os: linux
      priorityClassName: system-cluster-critical
      restartPolicy: Always
      schedulerName: default-scheduler
      securityContext: {}
      serviceAccount: konnectivity-agent
      serviceAccountName: konnectivity-agent
      terminationGracePeriodSeconds: 30
      tolerations:
      - key: CriticalAddonsOnly
        operator: Exists
      volumes:
      - name: konnectivity-agent-token
        projected:
          defaultMode: 420
          sources:
          - serviceAccountToken:
              audience: system:konnectivity-server
              expirationSeconds: 3600
              path: konnectivity-agent-token
  updateStrategy:
    rollingUpdate:
      maxSurge: 0
      maxUnavailable: 1
    type: RollingUpdate
status:
  currentNumberScheduled: 1
  desiredNumberScheduled: 1
  numberMisscheduled: 0
  numberReady: 0
  numberUnavailable: 1
  observedGeneration: 1
  updatedNumberScheduled: 1```

@sacesare
Copy link

sacesare commented Oct 29, 2021

Same problem with latest v1.22.2+k0s.2 (single controller+worker host) , every 5 minutes in k0s.log:

time="2021-10-29 18:22:26" level=info msg="E1029 18:22:26.904002   30064 server.go:382] \"Stream read from frontend failure\" err=\"rpc error: code = Canceled desc = context canceled\"" component=konnectivity

time="2021-10-29 18:27:27" level=info msg="E1029 18:27:27.265707   30064 server.go:382] \"Stream read from frontend failure\" err=\"rpc error: code = Canceled desc = context canceled\"" component=konnectivity

@mikhail-sakhnov
Copy link
Contributor

@sacesare I am not sure if it is related to the issue we discussed here (which I think originally was caused by different configs for controller nodes)

The log you shown looks fishy though my gut feelings say that it is related to the connection timeout between agent and proxy and generally it's a noise.

@sacesare
Copy link

sacesare commented Oct 29, 2021

@Soider maybe ... but why it's qualified as "error" and no mention of timeout? And ... it's timeout on single node cluster (localhost) ? @wheestermans31 log contains very similar lines.

@jnummelin
Copy link
Member

Based on "out of Github" discussions this original issue was solved. So closing.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

No branches or pull requests

4 participants