Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Support gRPC health checks #21493

Closed
therc opened this issue Feb 18, 2016 · 26 comments
Closed

Support gRPC health checks #21493

therc opened this issue Feb 18, 2016 · 26 comments
Labels
kind/feature Categorizes issue or PR as related to a new feature. lifecycle/frozen Indicates that an issue or PR should not be auto-closed due to staleness. sig/network Categorizes an issue or PR as relevant to SIG Network. sig/node Categorizes an issue or PR as relevant to SIG Node.

Comments

@therc
Copy link
Member

therc commented Feb 18, 2016

We already have google.golang.org/grpc/health/grpc_health_v1alpha/... in Godeps. My understanding is that all gRPC implementations should be interoperable for health checks and, if they're not, it's a bug.

Add a new health check type grpc with an optional service name. Per the official documentation, an empty name triggers a check for generic health of the whole server, no matter how many gRPC services it understands. The backend is expected to implement service /grpc.health.v1alpha.Health's Check() method, in addition, of course, to whatever other services it supports to do real work. Sending health checks through other random protobuf service definitions is NOT in scope.

I can work on this if nobody else steps up. This issue is really about getting feedback. I assume the only contentious point is whether we want to tie ourselves to the v1alpha service name.

@thockin
Copy link
Member

thockin commented Feb 19, 2016

I don't have any real objection to this, assuming we'll eventually have
grpc linked in anyway, but it's a slippery slope precedent. Do we add HC
support for other RPC protocols? On what principle do we say yes or no?

Can we find a way to make this more modular?

On Thu, Feb 18, 2016 at 11:25 AM, Rudi C notifications@github.com wrote:

We already have google.golang.org/grpc/health/grpc_health_v1alpha/... in
Godeps. My understanding is that all gRPC implementations should be
interoperable for health checks and, if they're not, it's a bug.

Add a new health check type grpc with an optional service name. Per the
official documentation
https://github.com/grpc/grpc/blob/master/doc/health-checking.md, an
empty name triggers a check for generic health of the whole server, no
matter how many gRPC services it understands. The backend is expected to
implement service /grpc.health.v1alpha.Health's Check() method, in
addition, of course, to whatever other services it supports to do real
work. Sending health checks through other random protobuf service
definitions is NOT in scope.

I can work on this if nobody else steps up. The issue is really about
getting feedback. I assume the only contentious point is whether we want to
tie ourselves to the v1alpha service name.


Reply to this email directly or view it on GitHub
#21493.

@therc
Copy link
Member Author

therc commented Feb 19, 2016

One extreme way to make it more modular is having external plugins (some grpc_ping binary, etc.), but then it's not very different from an exec health check. I'm not sure there's anything halfway that would be actually practical to use. I'm assuming that whatever modularity there might be between HTTP/exec checks is already there. As to yes/no criterion, in this case it wouldn't add extra dependencies that don't already exist.

@thockin
Copy link
Member

thockin commented Feb 19, 2016

We actually have a sidecar container that makes exec into http, dodging a
docker bug (sigh). I actually like this pattern - the core of Kube stays
simple, with just a few built-in options, and anyone can extend it by
building a new sidecar container. It also means we have precedent for
saying "no" to niche feature requests, which is becoming more important as
we mature the product.

So that said, why not build a trivial grpc client program that pokes grpc
healthcheck and exposes the result as an HTTP get? That way anyone can use
it, fix it, tweak it, extend it, without putting me in between them and
their goals :)
On Feb 19, 2016 5:46 AM, "Rudi C" notifications@github.com wrote:

One extreme way to make it more modular is having external plugins (some
grpc_ping binary, etc.), but then it's not very different an exec health
check. I'm not sure there's anything halfway that would be actually
practical to use. I'm assuming that whatever modularity there might be
between HTTP/exec checks is already there. As to yes/no criterion, in this
case it wouldn't add extra dependencies that don't already exist.


Reply to this email directly or view it on GitHub
#21493 (comment)
.

@therc
Copy link
Member Author

therc commented Feb 19, 2016

The sidecar container (I just contributed a patch to make it stop spamming logs...) is nifty, but also one more ~6MB thing you need to set up and monitor, since it's an additional POF. And another layer when debugging issues. Services like Datadog charge you by the container, not the pod. I'm not counting the extra connections, cycles and latency. :-) And if you don't let your nodes fetch random stuff from the network, it's one more image to audit and mirror/rebuild. I just want to serve ~5 health checks per minute...

In terms of user experience, it drives me nuts when I can't run kubectl logs on a kube-dns pod without also specifying the container; the sidecar spreads that virus further.

That said, I can write the client (either returning an exit code or serving HTTP), but would rather spare people the ugliness in the long run.

@thockin
Copy link
Member

thockin commented Feb 20, 2016

The sidecar container (I just contributed a patch to make it stop spamming logs...) is nifty, but also one more ~6MB thing you need to set up and monitor, since it's an additional POF. And another layer when debugging issues. Services like Datadog charge you by the container, not the pod. I'm not counting the extra connections, cycles and latency. :-) And if you don't let your nodes fetch random stuff from the network, it's one more image to audit and mirror/rebuild. I just want to serve ~5 health checks per minute...

In terms of user experience, it drives me nuts when I can't run kubectl logs on a kube-dns pod without also specifying the container; the sidecar spreads that virus further.

That said, I can write the client (either returning an exit code or serving HTTP), but would rather spare people the ugliness in the long run.

I hear your concerns, but at the same time, it is untenable in general to bundle everything people need into the core system. We could call gRPC a special case or MAYBE we could make health-checks into plugins like we do with volumes, but those come with their own problems. And it still puts the Kubernetes API and release process in between a user and their goals.

Side-car containers have the wonderful property of being entirely in the user's control and even better - they exist TODAY. Pods exist for a reason, and the reason is, first and foremost, composability. We can debate the need for "monitoring" sidecars, but yeah, stats are good. If DataDog gets enough signal that multi-container pods are an important abstraction, I am confident we can get a conversation going about how to evolve their cost models. @aronchick

@dchen1107 @smarterclayton @bgrant0607 @erictune for consideration. Clearly a gRPC healtcheck is not going to make the 1.2 release, so we have a couple months at least to debate this.

@bgrant0607
Copy link
Member

We should not build in a grpc health check. Use exec or http.

@smarterclayton
Copy link
Contributor

I think a gap that we'll want to sort out is the pattern by which side car
node containers can dynamically participate in the kubelet lifecycle.
Storage volumes, network plugins, lifecycle checks, pod lifecycle
interceptors (like the proposed container commit + push work) all seem to
want to play a part in the lifecycle of the node - some are obvious call
outs, some are not. Reactive callouts are a bit easier though since they
can watch the same state as the master.

If we rephrased this as "given an annotation on a pod, allow it to be
gracefully shutdown and restarted if the condition the annotation describes
is no longer possible", it doesn't seem too hard to sketch a path that
addresses the efficiency concerns. A small go binary is going to be using
a few megs per node, which is easier to amortize. A daemon set can keep it
running. The binary can signal the process through docker to terminate.
What is missing is conveying that information back to the master, but a
restart will at least be detected.

On Sat, Feb 20, 2016 at 8:42 PM, Brian Grant notifications@github.com
wrote:

We should not build in a grpc health check. Use exec or http.


Reply to this email directly or view it on GitHub
#21493 (comment)
.

@aronchick
Copy link
Contributor

@therc So, just want to capture your issues:

  • You don't want to have to monitor the sidecar (if there was an automated way to do this would that help)?
  • Third parties charge per container (effectively doubling cost)
  • You don't want to have to build a separate image for tracking/auditing/etc.

What if, rather that building a sidecar, you had a simple client in the main container itself that responded via exec or http?

@therc
Copy link
Member Author

therc commented Feb 22, 2016

This is definitely not something for 1.2. I really just wanted to spend a week or two in codegen hell. :-)

The main drivers were having services that only talk gRPC and not wanting to add an HTTP handler just for health checks. That calls for two ports; although grpc-go gained support for HTTP serving on the same port last week, I don't think the other languages are as lucky. I was assuming that in the 1.3 timeframe gRPC health checks would be performed somewhere else anyway. I do understand the slippery slope argument. Maybe by the time 2.0 happens, half of users out there will be running gRPC stuff and will ask to reopen this issue.

The second best option is an exec liveness probe, but apparently there's some bug in Docker that requires the clunky exechealthz sidecar. Is that still the case? How do I find out more?

I'll write the simple health check binary and submit it to /contrib.

As for the charges per container, I'm trying to figure if there are any exceptions. In a cluster with just a bunch of base services and a few applications, there are 19 pods and 44 containers. 19 of those 44 are pause containers, of course...

@thockin
Copy link
Member

thockin commented Feb 22, 2016

On Sun, Feb 21, 2016 at 7:58 PM, Rudi C notifications@github.com wrote:

This is definitely not something for 1.2. I really just wanted to spend a week or two in codegen hell. :-)

The main drivers were having services that only talk gRPC and not wanting to add an HTTP handler just for health checks. That calls for two ports; although grpc-go gained support for HTTP serving on the same port last week, I don't think the other languages are as lucky. I was assuming that in the 1.3 timeframe gRPC health checks would be performed somewhere else anyway. I do understand the slippery slope argument. Maybe by the time 2.0 happens, half of users out there will be running gRPC stuff and will ask to reopen this issue.

The second best option is an exec liveness probe, but apparently there's some bug in Docker that requires the clunky exechealthz sidecar. Is that still the case? How do I find out more?

I forget if this is properly fixed in docker. @dchen1107 @bprashanth
@vishh might recall.

I'll write the simple health check binary and submit it to /contrib.

As for the charges per container, I'm trying to figure if there are any exceptions. In a cluster with just a bunch of base services and a few applications, there are 19 pods and 44 containers. 19 of those 44 are pause containers, of course...


Reply to this email directly or view it on GitHub.

@bprashanth
Copy link
Contributor

That bug should be fixed, but the feeling I got from the response to the bug was that docker exec is not something they'd ever meant to be taken seriously in production. That might've changed since, you can infer for yourself: moby/moby#14444. The exechealthz sidecar just gives you control over the docker exec stack.

@thockin
Copy link
Member

thockin commented Feb 22, 2016

Yeah, I would make your side-car export HTTP - Docker gave us exec but they
don't seem to actually want people to use it :(

On Sun, Feb 21, 2016 at 9:01 PM, Prashanth B notifications@github.com
wrote:

That bug should be fixed, but the feeling I got from the response to
the bug was that docker exec is not something they'd ever meant to be taken
seriously in production. That might've changed since, you can infer for
yourself: moby/moby#14444
moby/moby#14444. The exechealthz sidecar
just gives you control over the docker exec stack.


Reply to this email directly or view it on GitHub
#21493 (comment)
.

@smarterclayton
Copy link
Contributor

We'll continue to need exec - if we need an alternate performant implementation that is actually maintained we should document it. Practically, both runc and rocket need an implementation as well.

@thockin
Copy link
Member

thockin commented Jun 29, 2016

We're not implementing gRPC health checks right now, so I am closing this.

@thockin thockin closed this as completed Jun 29, 2016
@irfn
Copy link

irfn commented Dec 24, 2017

@thockin Any chance this can be re-opened.
A plain TCP socket check causes issues such as this. grpc/grpc-go/issues/875

Since kubernetes core does not want to build this into core, how about allowing a way which provides us with a workaround. sending hex data/ asserting predefined hexdata would work fine.
These can be additional params in the existing tcp check. Its already a popular approach with grpc.

Here is the envoyproxy discussion on Implementing this.
envoyproxy/envoy/issues/369
This in turn is very similar to the haproxy tcp-check send binary
haproxy heathchecks

@lalomartins
Copy link

Alternatively, how about either an http2 check, or a flag in the normal http check to tell it to use http2? I can live without checking the actual call results, but pointing to a grpc endpoint with an http check currently gives me “malformed HTTP status code”.

@bgrant0607 bgrant0607 added sig/network Categorizes an issue or PR as relevant to SIG Network. sig/node Categorizes an issue or PR as relevant to SIG Node. and removed team/cluster (deprecated - do not use) labels Feb 9, 2018
@bgrant0607
Copy link
Member

http2 checks seem reasonable, and probably inevitable

@therc
Copy link
Member Author

therc commented Feb 12, 2018

The problem with only adding an HTTP/2 variant, for users of gRPC, is that the latter always returns 200 as the HTTP status in the headers, as long as the server is up. The errors, whether they're raised by the server's own code or from its server-side gRPC library, are reflected in two trailers, grpc-status and grpc-message.
There are thus two options:

  1. add fields so people can inspect trailers
  2. document that gRPC servers should not use health.proto when running under Kubernetes; they should rather return a non-200 HTTP/2 status. gRPC clients are supposed to handle the case when load balancers, proxies and similar break environments that way.

(I'd suggest the original feature request as the third option, but it's not clear that anything has changed since it was first turned down.)

@lalomartins
Copy link

Those are valid points but as long as the server is up tests the most basic case of a health check, no? Maybe my database setup is busted or whatever else… but as long as the process is alive and not overloaded and not in some sort of deadlock, it will respond with 200 (or rather, a no-header response, which I understand is the equivalent of a 200?) to a GET request. That's good enough for my use case.

@therc
Copy link
Member Author

therc commented Feb 12, 2018

@lalomartins that's good enough for a liveness check, but it does not work well for readiness checks. Servers can be up, but unavailable for a number of reasons:

  • they are still in setup phase
  • they are in pushback, because they received too many requests or very large ones
  • they are processing the remaining in-flight requests, after which they will terminate on their own (this is different from when a pod is in terminating state)
  • they are reloading a new configuration
  • they are passive masters, in standby and waiting for the active master to lose its lease

kube-proxy uses the readiness state to determine whether a pod should be in a Service's list of endpoints or not.

A binary health check doesn't convey the (at least) three states. You end up having your process run two servers (gRPC + health check), which is not easy to do in all gRPC-supported languages, especially if you want to share the same port (IIRC, it's doable in Go, but not cleanly in Python).

@bgrant0607
Copy link
Member

Example adaptor: https://github.com/otsimo/grpc-health

@bgrant0607 bgrant0607 reopened this Feb 15, 2018
@bgrant0607 bgrant0607 added the kind/feature Categorizes issue or PR as related to a new feature. label Feb 15, 2018
@bgrant0607
Copy link
Member

cc @kubernetes/sig-node-feature-requests @kubernetes/sig-network-feature-requests

@jtattermusch
Copy link

CC @jtattermusch

@bgrant0607 bgrant0607 added the lifecycle/frozen Indicates that an issue or PR should not be auto-closed due to staleness. label Jun 1, 2018
@thockin
Copy link
Member

thockin commented Aug 20, 2018

Why is this frozen? We're not adding grpc any time soon. HTTP2 should be a different topic entirely.

@thockin thockin closed this as completed Aug 20, 2018
@bhack
Copy link

bhack commented Oct 3, 2018

@ahmetb
Copy link
Member

ahmetb commented Oct 9, 2018

For anyone interested, as @bhack has said we released a tool named grpc-health-probe from grpc-ecosystem that makes it easier to health-check gRPC apps. The only requirement is that you implement the gRPC health protocol in your application (which is pretty trivial and every grpc ships the health library for every language).

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
kind/feature Categorizes issue or PR as related to a new feature. lifecycle/frozen Indicates that an issue or PR should not be auto-closed due to staleness. sig/network Categorizes an issue or PR as relevant to SIG Network. sig/node Categorizes an issue or PR as relevant to SIG Node.
Projects
None yet
Development

No branches or pull requests