From 497aef794f6bc767ee198e8a8d652bcac26e660d Mon Sep 17 00:00:00 2001 From: David Eliahu Date: Fri, 20 Nov 2020 19:07:15 -0800 Subject: [PATCH] Add image tag to API config docs --- docs/deployments/batch-api/api-configuration.md | 11 +++++++---- docs/deployments/realtime-api/api-configuration.md | 11 +++++++---- 2 files changed, 14 insertions(+), 8 deletions(-) diff --git a/docs/deployments/batch-api/api-configuration.md b/docs/deployments/batch-api/api-configuration.md index b6f31c1c02..8079ee944c 100644 --- a/docs/deployments/batch-api/api-configuration.md +++ b/docs/deployments/batch-api/api-configuration.md @@ -8,6 +8,7 @@ Reference the section below which corresponds to your Predictor type: [Python](# ## Python Predictor + ```yaml - name: # API name (required) kind: BatchAPI @@ -16,7 +17,7 @@ Reference the section below which corresponds to your Predictor type: [Python](# path: # path to a python file with a PythonPredictor class definition, relative to the Cortex root (required) config: # arbitrary dictionary passed to the constructor of the Predictor (can be overridden by config passed in job submission) (optional) python_path: # path to the root of your Python folder that will be appended to PYTHONPATH (default: folder containing cortex.yaml) - image: # docker image to use for the Predictor (default: quay.io/cortexlabs/python-predictor-cpu or quay.io/cortexlabs/python-predictor-gpu based on compute) + image: # docker image to use for the Predictor (default: quay.io/cortexlabs/python-predictor-cpu:master or quay.io/cortexlabs/python-predictor-gpu:master based on compute) env: # dictionary of environment variables networking: endpoint: # the endpoint for the API (default: ) @@ -32,6 +33,7 @@ See additional documentation for [compute](../compute.md), [networking](../netwo ## TensorFlow Predictor + ```yaml - name: # API name (required) kind: BatchAPI @@ -50,8 +52,8 @@ See additional documentation for [compute](../compute.md), [networking](../netwo batch_interval: # the maximum amount of time to spend waiting for additional requests before running inference on the batch of requests config: # arbitrary dictionary passed to the constructor of the Predictor (can be overridden by config passed in job submission) (optional) python_path: # path to the root of your Python folder that will be appended to PYTHONPATH (default: folder containing cortex.yaml) - image: # docker image to use for the Predictor (default: quay.io/cortexlabs/tensorflow-predictor) - tensorflow_serving_image: # docker image to use for the TensorFlow Serving container (default: quay.io/cortexlabs/tensorflow-serving-gpu or quay.io/cortexlabs/tensorflow-serving-cpu based on compute) + image: # docker image to use for the Predictor (default: quay.io/cortexlabs/tensorflow-predictor:master) + tensorflow_serving_image: # docker image to use for the TensorFlow Serving container (default: quay.io/cortexlabs/tensorflow-serving-gpu:master or quay.io/cortexlabs/tensorflow-serving-cpu:master based on compute) env: # dictionary of environment variables networking: endpoint: # the endpoint for the API (default: ) @@ -67,6 +69,7 @@ See additional documentation for [compute](../compute.md), [networking](../netwo ## ONNX Predictor + ```yaml - name: # API name (required) kind: BatchAPI @@ -81,7 +84,7 @@ See additional documentation for [compute](../compute.md), [networking](../netwo ... config: # arbitrary dictionary passed to the constructor of the Predictor (can be overridden by config passed in job submission) (optional) python_path: # path to the root of your Python folder that will be appended to PYTHONPATH (default: folder containing cortex.yaml) - image: # docker image to use for the Predictor (default: quay.io/cortexlabs/onnx-predictor-gpu or quay.io/cortexlabs/onnx-predictor-cpu based on compute) + image: # docker image to use for the Predictor (default: quay.io/cortexlabs/onnx-predictor-gpu:master or quay.io/cortexlabs/onnx-predictor-cpu:master based on compute) env: # dictionary of environment variables networking: endpoint: # the endpoint for the API (default: ) diff --git a/docs/deployments/realtime-api/api-configuration.md b/docs/deployments/realtime-api/api-configuration.md index 2fe235dec6..51c09f84bc 100644 --- a/docs/deployments/realtime-api/api-configuration.md +++ b/docs/deployments/realtime-api/api-configuration.md @@ -8,6 +8,7 @@ Reference the section below which corresponds to your Predictor type: [Python](# ## Python Predictor + ```yaml - name: # API name (required) kind: RealtimeAPI @@ -27,7 +28,7 @@ Reference the section below which corresponds to your Predictor type: [Python](# threads_per_process: # the number of threads per process (default: 1) config: # arbitrary dictionary passed to the constructor of the Predictor (optional) python_path: # path to the root of your Python folder that will be appended to PYTHONPATH (default: folder containing cortex.yaml) - image: # docker image to use for the Predictor (default: quay.io/cortexlabs/python-predictor-cpu or quay.io/cortexlabs/python-predictor-gpu based on compute) + image: # docker image to use for the Predictor (default: quay.io/cortexlabs/python-predictor-cpu:master or quay.io/cortexlabs/python-predictor-gpu:master based on compute) env: # dictionary of environment variables networking: endpoint: # the endpoint for the API (aws only) (default: ) @@ -63,6 +64,7 @@ See additional documentation for [models](models.md), [parallelism](parallelism. ## TensorFlow Predictor + ```yaml - name: # API name (required) kind: RealtimeAPI @@ -88,8 +90,8 @@ See additional documentation for [models](models.md), [parallelism](parallelism. threads_per_process: # the number of threads per process (default: 1) config: # arbitrary dictionary passed to the constructor of the Predictor (optional) python_path: # path to the root of your Python folder that will be appended to PYTHONPATH (default: folder containing cortex.yaml) - image: # docker image to use for the Predictor (default: quay.io/cortexlabs/tensorflow-predictor) - tensorflow_serving_image: # docker image to use for the TensorFlow Serving container (default: quay.io/cortexlabs/tensorflow-serving-gpu or quay.io/cortexlabs/tensorflow-serving-cpu based on compute) + image: # docker image to use for the Predictor (default: quay.io/cortexlabs/tensorflow-predictor:master) + tensorflow_serving_image: # docker image to use for the TensorFlow Serving container (default: quay.io/cortexlabs/tensorflow-serving-gpu:master or quay.io/cortexlabs/tensorflow-serving-cpu:master based on compute) env: # dictionary of environment variables networking: endpoint: # the endpoint for the API (aws only) (default: ) @@ -125,6 +127,7 @@ See additional documentation for [models](models.md), [parallelism](parallelism. ## ONNX Predictor + ```yaml - name: # API name (required) kind: RealtimeAPI @@ -145,7 +148,7 @@ See additional documentation for [models](models.md), [parallelism](parallelism. threads_per_process: # the number of threads per process (default: 1) config: # arbitrary dictionary passed to the constructor of the Predictor (optional) python_path: # path to the root of your Python folder that will be appended to PYTHONPATH (default: folder containing cortex.yaml) - image: # docker image to use for the Predictor (default: quay.io/cortexlabs/onnx-predictor-gpu or quay.io/cortexlabs/onnx-predictor-cpu based on compute) + image: # docker image to use for the Predictor (default: quay.io/cortexlabs/onnx-predictor-gpu:master or quay.io/cortexlabs/onnx-predictor-cpu:master based on compute) env: # dictionary of environment variables networking: endpoint: # the endpoint for the API (aws only) (default: )