Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

JupyterHub spawner: upstream request timeout #261

Closed
jlewi opened this issue Feb 17, 2018 · 6 comments
Closed

JupyterHub spawner: upstream request timeout #261

jlewi opened this issue Feb 17, 2018 · 6 comments
Labels

Comments

@jlewi
Copy link
Contributor

jlewi commented Feb 17, 2018

When I go to JupyterHub and spawn my notebook I get

upstream request timeout

The jupyter pod is actually in the container creating state. Pulling the image take a long time because it is so large.

If I refresh it shows up.

My guess is there is some sort of timeout in JupyterHub and maybe we just need to increase it? Perhaps its a setting in KubeSpawner?

/cc @yuvipanda

@jlewi jlewi added area/jupyter Issues related to Jupyter priority/p1 help wanted labels Feb 27, 2018
@jlewi
Copy link
Contributor Author

jlewi commented Mar 1, 2018

We already set a large timeout of 10 minutes here
https://github.com/kubeflow/kubeflow/blob/master/kubeflow/core/jupyterhub_spawner.py#L81

@lluunn
Copy link
Contributor

lluunn commented Mar 20, 2018

I tried to reproduce:

In my own cluster, create new ksonnet app (ks init) and deploy the kubeflow-core.
After clicking the "Spawn", it took a long time (keep spinning for ~5min).
Then it worked, didn't see "upstream request timeout".

Is this fixed? Is the 5min wait expected?

@jlewi
Copy link
Contributor Author

jlewi commented Mar 20, 2018

The 5 minute timeout is probably do to the image pull timeout.

How were you connecting to the cluster?

I've observed this upstream request timeout when using IAP on GKE. So one possibility is that its a timeout in the GCE loadbalancer, or envoy. If we observe it when using IAP but not when using kubectl port-forward to connect directly to JupyterHub that would indicate the problem is upstream of JupyterHub.

@lluunn
Copy link
Contributor

lluunn commented Mar 21, 2018

Ah, I didn't enable IAP

@ankushagarwal
Copy link
Contributor

I can take a look at this. I think this is coming from envoy as well.

@ankushagarwal
Copy link
Contributor

The timeout specified in envoy config is 10 seconds for jupyterhub https://github.com/kubeflow/kubeflow/blob/master/kubeflow/core/iap.libsonnet#L397

I'll try increasing this to see if it helps

k8s-ci-robot pushed a commit that referenced this issue Mar 23, 2018
Downloading images takes some time - envoy simply
shows upstream request timeout which is a bad user
experience

Run jsonnet autoformat

Fixes #261
yanniszark pushed a commit to arrikto/kubeflow that referenced this issue Nov 1, 2019
* add selector for application controller
statefulset

* fix tests
yanniszark pushed a commit to arrikto/kubeflow that referenced this issue Feb 15, 2021
* Add readinessProbe for vizier-db

Signed-off-by: Koichiro Den <den@valinux.co.jp>

* Fix MYSQL_ROOT_PASSWORD

Fixes: 67e94c7 ("Set MYSQL_ROOT_PASSWORD via Secret (kubeflow#253)")
Signed-off-by: Koichiro Den <den@valinux.co.jp>

* Add simple loop to wait for DB connection successfully opened

Signed-off-by: Koichiro Den <den@valinux.co.jp>
elenzio9 pushed a commit to arrikto/kubeflow that referenced this issue Oct 31, 2022
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Projects
None yet
Development

No branches or pull requests

3 participants