Skip to content

Ingress controller crash randomly with exit code 111 #695

@mdecalf

Description

@mdecalf

Hello,
I run haproxy ingress controller ad Daemonset (4 replicas) on rke2.
I upgrade from helm release 1.39.4 (v1.11) which working well to 1.4X (I tried all v3.X) and the pod are crashing randomly and terminate with exit code 111.
I tried to run it with hostNetwork and without to see, same problem.
Tried also to run it as root and non root same problem.

The only thing that I can found in the logs except the access log and health check is

xxx:24088 [30/Jan/2025:09:52:21.788] https/v4: SSL handshake failure
xxx:14690 [30/Jan/2025:09:52:21.788] https/v4: SSL handshake failure
xxx:39793 [30/Jan/2025:09:52:21.788] https/v4: SSL handshake failure
xxx:42447 [30/Jan/2025:09:52:21.788] https/v4: SSL handshake failure
xxx:35800 [30/Jan/2025:09:52:21.788] https/v4: SSL handshake failure
xxx:39367 [30/Jan/2025:09:52:21.789] https/v4: SSL handshake failure
xxx:10664 [30/Jan/2025:09:52:21.789] https/v4: SSL handshake failure
xxx:40857 [30/Jan/2025:09:52:21.789] https/v4: SSL handshake failure
xxx:52486 [30/Jan/2025:09:52:21.789] https/v4: SSL handshake failure
xxx:58673 [30/Jan/2025:09:52:21.789] https/v4: SSL handshake failure
xxx:33828 [30/Jan/2025:09:52:21.789] https/v4: SSL handshake failure
xxx:38355 [30/Jan/2025:09:52:21.789] https/v4: SSL handshake failure
xxx:32023 [30/Jan/2025:09:52:21.789] https/v4: SSL handshake failure
xxx:59652 [30/Jan/2025:09:52:21.789] https/v4: SSL handshake failure
xxx:29675 [30/Jan/2025:09:52:21.789] https/v4: SSL handshake failure

and the termination log

s6-rc: info: service legacy-services: stopping
s6-rc: info: service legacy-services successfully stopped
s6-rc: info: service ingress-controller: stopping
Ingress Controller exited with code 0, restarting...
s6-rc: info: service ingress-controller successfully stopped
s6-rc: info: service haproxy: stopping
[WARNING]  (69) : Exiting Master process...
[ALERT]    (69) : Current worker (213) exited with code 143 (Terminated)
[WARNING]  (69) : All workers exited. Exiting... (0)
s6-rc: info: service haproxy successfully stopped
s6-rc: info: service legacy-cont-init: stopping
s6-rc: info: service sigusr1: stopping
s6-rc: info: service aux-cfg: stopping
s6-rc: info: service sigusr1 successfully stopped
s6-rc: info: service aux-cfg successfully stopped
s6-rc: info: service legacy-cont-init successfully stopped
s6-rc: info: service fix-attrs: stopping
s6-rc: info: service fix-attrs successfully stopped
s6-rc: info: service s6rc-oneshot-runner: stopping
s6-rc: info: service s6rc-oneshot-runner successfully stopped

The parameters of the chart :

controller:
  kind: DaemonSet
  nodeSelector:
    role.xxx.com/haproxy: "true"
  extraArgs:
    - --healthz-bind-port=1042
    - --stats-bind-port=1024
    - --localpeer-port=10000
    - --prometheus
  containerPort:
    http: 80
    https: 443
    stat: 1024
    healthz: 1042
    prometheus: 6060
  ingressClass: haproxy
  ingressClassResource:
    enabled: false
    name: haproxy
    default: false
  defaultTLSSecret:
    enabled: false
  service:
    enabled: true
    type: ClusterIP
    ports:
      http: 8080
      https: 8443
      stat: 1024
      healthz: 1042
    prometheus: 6060
  daemonset:
    useHostPort: true
    hostPorts:
      http: 8080
      https: 8443
      stat: 1024
      healthz: 1042
    prometheus: 6060
  logging:
    level: warning
    traffic:
      address: stdout
      format: raw
      facility: daemon
  startupProbe:
    periodSeconds: 10
    timeoutSeconds: 5

Describe of the pod :

    State:          Running
      Started:      Thu, 30 Jan 2025 10:52:30 +0100
    Last State:     Terminated
      Reason:       Error
      Exit Code:    111
      Started:      Wed, 29 Jan 2025 18:13:01 +0100
      Finished:     Thu, 30 Jan 2025 10:52:29 +0100
    Ready:          True
    Restart Count:  10
    Requests:
      cpu:      250m
      memory:   400Mi
    Liveness:   http-get http://:1042/healthz delay=0s timeout=1s period=10s #success=1 #failure=3
    Readiness:  http-get http://:1042/healthz delay=0s timeout=1s period=10s #success=1 #failure=3
    Startup:    http-get http://:1042/healthz delay=0s timeout=5s period=10s #success=1 #failure=20

The only specific thing that we have are snippet to have backup on backend

haproxy.org/backend-config-snippet: |
        option allbackups

Do you have an idea about what could be the issue?

Metadata

Metadata

Assignees

No one assigned

    Labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions