Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Need help deploying simple model with tf-serving and enable Rest API #2292

Closed
fboylu opened this issue Jan 17, 2019 · 6 comments
Closed

Need help deploying simple model with tf-serving and enable Rest API #2292

fboylu opened this issue Jan 17, 2019 · 6 comments

Comments

@fboylu
Copy link

fboylu commented Jan 17, 2019

Hi, I am having issues with the following.

VERSION=v0.4.0
ks pkg install kubeflow/tf-serving@${VERSION}

ks generate tf-serving tf-serving
ks generate tf-serving-service tf-serving-service

ks param set tf-serving-service modelName half_plus_two
ks param set tf-serving modelName half_plus_two
ks param set tf-serving modelPath /models/half_plus_two
ks param set tf-serving numGpus 1
ks param set tf-serving defaultGpuImage fboylu/tf_serving_gpu
ks param set tf-serving serviceType LoadBalancer

ks apply default -c tf-serving-service
ks apply default -c tf-serving

I have my model baked into the container and logs from tf-serving pod confirm it is serving (it works locally anyways).

I see the LoadBalancer external IP but the following times out
curl -d '{"instances": [1.0, 2.0, 5.0]}' -X POST http://EXTERNAL_IP/v1/models/half_plus_two:predict

Can you let me know what is missing/wrong?

thanks

@lluunn
Copy link
Contributor

lluunn commented Jan 17, 2019

tf-serving is the legacy prototype, and we are removing it soon.
tf-serving-service should be used with tf-serving-deployment-gcp or tf-serving-deployment-aws

document:
https://www.kubeflow.org/docs/guides/components/tfserving_new/

@fboylu
Copy link
Author

fboylu commented Jan 17, 2019

I am not usig gcp or aws, I have the model locally in the container. Can I still use tf-serving-deployment-gcp or tf-serving-deployment-aws?

@lluunn
Copy link
Contributor

lluunn commented Jan 17, 2019

should be. The two prototypes just have some credential setup to read from gcp/aws.

@fboylu
Copy link
Author

fboylu commented Jan 17, 2019

Ok, I tried it and seems like it now serves on 8500.

@jlewi
Copy link
Contributor

jlewi commented Jan 19, 2019

@fboylu Is this working now for you? Can we close this issue?

@fboylu
Copy link
Author

fboylu commented Jan 19, 2019

Yes it is working, please close the issue. Many thanks for the help.

@fboylu fboylu closed this as completed Jan 19, 2019
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

3 participants