Skip to content

* HTTP 1.0, assume close after body < HTTP/1.0 503 Service Unavailable #2526

@aditivw

Description

@aditivw

System Info

We have deployed 'ghcr.io/huggingface/text-generation-inference:2.2.0' using kubernetes and the application logs say that the server is now sucessfully connected

2024-09-17T09:21:04.948326Z INFO text_generation_router: router/src/main.rs:357: Using config Some(Llama)
2024-09-17T09:21:04.948342Z WARN text_generation_router: router/src/main.rs:384: Invalid hostname, defaulting to 0.0.0.0
2024-09-17T09:21:05.378802Z INFO text_generation_router::server: router/src/server.rs:1572: Warming up model
2024-09-17T09:21:26.517495Z INFO text_generation_launcher: Cuda Graphs are enabled for sizes [32, 16, 8, 4, 2, 1]
2024-09-17T09:21:28.936537Z INFO text_generation_router::server: router/src/server.rs:1599: Using scheduler V3
2024-09-17T09:21:28.936563Z INFO text_generation_router::server: router/src/server.rs:1651: Setting max batch total tokens to 37568
2024-09-17T09:21:28.997788Z INFO text_generation_router::server: router/src/server.rs:1889: Connected

But when we hit the route/endpoint we are unable to consume the LLM.
We have checked the route and service configurations thoroughly and they seem to be correct.

when we hit the route the error message is:

  • IPv6: (none)
  • IPv4: 11.241.137.145
  • Trying 11.241.137.145:443...
  • Connected to *********************************.apps.ocp4.dlocp.prd.eu.bp.aws.cloud.vwgroup.com (11.241.137.145) port 443
  • schannel: disabled automatic use of client certificate
  • ALPN: curl offers http/1.1
  • ALPN: server did not agree on a protocol. Uses default.
  • using HTTP/1.x

GET / HTTP/1.1
Host: *******************************.apps.ocp4.dlocp.prd.eu.bp.aws.cloud.vwgroup.com
User-Agent: curl/8.8.0
Accept: /

  • Request completely sent off

  • schannel: remote party requests renegotiation

  • schannel: renegotiating SSL/TLS connection

  • schannel: SSL/TLS connection renegotiated

  • schannel: remote party requests renegotiation

  • schannel: renegotiating SSL/TLS connection

  • schannel: SSL/TLS connection renegotiated

  • schannel: failed to decrypt data, need more data

  • schannel: server close notification received (close_notify)

  • HTTP 1.0, assume close after body
    < HTTP/1.0 503 Service Unavailable

  • schannel: server indicated shutdown in a prior call

  • Closing connection

  • schannel: shutting down SSL/TLS connection with ***************************.apps.ocp4.dlocp.prd.eu.bp.aws.cloud.vwgroup.com port 443

Information

  • Docker
  • The CLI directly

Tasks

  • An officially supported command
  • My own modifications

Reproduction

  1. We followed https://github.com/rh-aiservices-bu/llm-on-openshift these steps for the deployment of TGI
  2. curl http://(your-host)

Expected behavior

Establish a connection to the LLM endpoint

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions