You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/deployments/batch-api/api-configuration.md
+7-4Lines changed: 7 additions & 4 deletions
Original file line number
Diff line number
Diff line change
@@ -8,6 +8,7 @@ Reference the section below which corresponds to your Predictor type: [Python](#
8
8
9
9
## Python Predictor
10
10
11
+
<!-- CORTEX_VERSION_BRANCH_STABLE x2 -->
11
12
```yaml
12
13
- name: <string> # API name (required)
13
14
kind: BatchAPI
@@ -16,7 +17,7 @@ Reference the section below which corresponds to your Predictor type: [Python](#
16
17
path: <string> # path to a python file with a PythonPredictor class definition, relative to the Cortex root (required)
17
18
config: <string: value> # arbitrary dictionary passed to the constructor of the Predictor (can be overridden by config passed in job submission) (optional)
18
19
python_path: <string> # path to the root of your Python folder that will be appended to PYTHONPATH (default: folder containing cortex.yaml)
19
-
image: <string> # docker image to use for the Predictor (default: quay.io/cortexlabs/python-predictor-cpu or quay.io/cortexlabs/python-predictor-gpu based on compute)
20
+
image: <string> # docker image to use for the Predictor (default: quay.io/cortexlabs/python-predictor-cpu:master or quay.io/cortexlabs/python-predictor-gpu:master based on compute)
20
21
env: <string: string> # dictionary of environment variables
21
22
networking:
22
23
endpoint: <string> # the endpoint for the API (default: <api_name>)
@@ -32,6 +33,7 @@ See additional documentation for [compute](../compute.md), [networking](../netwo
32
33
33
34
## TensorFlow Predictor
34
35
36
+
<!-- CORTEX_VERSION_BRANCH_STABLE x3 -->
35
37
```yaml
36
38
- name: <string> # API name (required)
37
39
kind: BatchAPI
@@ -50,8 +52,8 @@ See additional documentation for [compute](../compute.md), [networking](../netwo
50
52
batch_interval: <duration> # the maximum amount of time to spend waiting for additional requests before running inference on the batch of requests
51
53
config: <string: value> # arbitrary dictionary passed to the constructor of the Predictor (can be overridden by config passed in job submission) (optional)
52
54
python_path: <string> # path to the root of your Python folder that will be appended to PYTHONPATH (default: folder containing cortex.yaml)
53
-
image: <string> # docker image to use for the Predictor (default: quay.io/cortexlabs/tensorflow-predictor)
54
-
tensorflow_serving_image: <string> # docker image to use for the TensorFlow Serving container (default: quay.io/cortexlabs/tensorflow-serving-gpu or quay.io/cortexlabs/tensorflow-serving-cpu based on compute)
55
+
image: <string> # docker image to use for the Predictor (default: quay.io/cortexlabs/tensorflow-predictor:master)
56
+
tensorflow_serving_image: <string> # docker image to use for the TensorFlow Serving container (default: quay.io/cortexlabs/tensorflow-serving-gpu:master or quay.io/cortexlabs/tensorflow-serving-cpu:master based on compute)
55
57
env: <string: string> # dictionary of environment variables
56
58
networking:
57
59
endpoint: <string> # the endpoint for the API (default: <api_name>)
@@ -67,6 +69,7 @@ See additional documentation for [compute](../compute.md), [networking](../netwo
67
69
68
70
## ONNX Predictor
69
71
72
+
<!-- CORTEX_VERSION_BRANCH_STABLE x2 -->
70
73
```yaml
71
74
- name: <string> # API name (required)
72
75
kind: BatchAPI
@@ -81,7 +84,7 @@ See additional documentation for [compute](../compute.md), [networking](../netwo
81
84
...
82
85
config: <string: value> # arbitrary dictionary passed to the constructor of the Predictor (can be overridden by config passed in job submission) (optional)
83
86
python_path: <string> # path to the root of your Python folder that will be appended to PYTHONPATH (default: folder containing cortex.yaml)
84
-
image: <string> # docker image to use for the Predictor (default: quay.io/cortexlabs/onnx-predictor-gpu or quay.io/cortexlabs/onnx-predictor-cpu based on compute)
87
+
image: <string> # docker image to use for the Predictor (default: quay.io/cortexlabs/onnx-predictor-gpu:master or quay.io/cortexlabs/onnx-predictor-cpu:master based on compute)
85
88
env: <string: string> # dictionary of environment variables
86
89
networking:
87
90
endpoint: <string> # the endpoint for the API (default: <api_name>)
Copy file name to clipboardExpand all lines: docs/deployments/realtime-api/api-configuration.md
+7-4Lines changed: 7 additions & 4 deletions
Original file line number
Diff line number
Diff line change
@@ -8,6 +8,7 @@ Reference the section below which corresponds to your Predictor type: [Python](#
8
8
9
9
## Python Predictor
10
10
11
+
<!-- CORTEX_VERSION_BRANCH_STABLE x2 -->
11
12
```yaml
12
13
- name: <string> # API name (required)
13
14
kind: RealtimeAPI
@@ -27,7 +28,7 @@ Reference the section below which corresponds to your Predictor type: [Python](#
27
28
threads_per_process: <int> # the number of threads per process (default: 1)
28
29
config: <string: value> # arbitrary dictionary passed to the constructor of the Predictor (optional)
29
30
python_path: <string> # path to the root of your Python folder that will be appended to PYTHONPATH (default: folder containing cortex.yaml)
30
-
image: <string> # docker image to use for the Predictor (default: quay.io/cortexlabs/python-predictor-cpu or quay.io/cortexlabs/python-predictor-gpu based on compute)
31
+
image: <string> # docker image to use for the Predictor (default: quay.io/cortexlabs/python-predictor-cpu:master or quay.io/cortexlabs/python-predictor-gpu:master based on compute)
31
32
env: <string: string> # dictionary of environment variables
32
33
networking:
33
34
endpoint: <string> # the endpoint for the API (aws only) (default: <api_name>)
@@ -63,6 +64,7 @@ See additional documentation for [models](models.md), [parallelism](parallelism.
63
64
64
65
## TensorFlow Predictor
65
66
67
+
<!-- CORTEX_VERSION_BRANCH_STABLE x3 -->
66
68
```yaml
67
69
- name: <string> # API name (required)
68
70
kind: RealtimeAPI
@@ -88,8 +90,8 @@ See additional documentation for [models](models.md), [parallelism](parallelism.
88
90
threads_per_process: <int> # the number of threads per process (default: 1)
89
91
config: <string: value> # arbitrary dictionary passed to the constructor of the Predictor (optional)
90
92
python_path: <string> # path to the root of your Python folder that will be appended to PYTHONPATH (default: folder containing cortex.yaml)
91
-
image: <string> # docker image to use for the Predictor (default: quay.io/cortexlabs/tensorflow-predictor)
92
-
tensorflow_serving_image: <string> # docker image to use for the TensorFlow Serving container (default: quay.io/cortexlabs/tensorflow-serving-gpu or quay.io/cortexlabs/tensorflow-serving-cpu based on compute)
93
+
image: <string> # docker image to use for the Predictor (default: quay.io/cortexlabs/tensorflow-predictor:master)
94
+
tensorflow_serving_image: <string> # docker image to use for the TensorFlow Serving container (default: quay.io/cortexlabs/tensorflow-serving-gpu:master or quay.io/cortexlabs/tensorflow-serving-cpu:master based on compute)
93
95
env: <string: string> # dictionary of environment variables
94
96
networking:
95
97
endpoint: <string> # the endpoint for the API (aws only) (default: <api_name>)
@@ -125,6 +127,7 @@ See additional documentation for [models](models.md), [parallelism](parallelism.
125
127
126
128
## ONNX Predictor
127
129
130
+
<!-- CORTEX_VERSION_BRANCH_STABLE x2 -->
128
131
```yaml
129
132
- name: <string> # API name (required)
130
133
kind: RealtimeAPI
@@ -145,7 +148,7 @@ See additional documentation for [models](models.md), [parallelism](parallelism.
145
148
threads_per_process: <int> # the number of threads per process (default: 1)
146
149
config: <string: value> # arbitrary dictionary passed to the constructor of the Predictor (optional)
147
150
python_path: <string> # path to the root of your Python folder that will be appended to PYTHONPATH (default: folder containing cortex.yaml)
148
-
image: <string> # docker image to use for the Predictor (default: quay.io/cortexlabs/onnx-predictor-gpu or quay.io/cortexlabs/onnx-predictor-cpu based on compute)
151
+
image: <string> # docker image to use for the Predictor (default: quay.io/cortexlabs/onnx-predictor-gpu:master or quay.io/cortexlabs/onnx-predictor-cpu:master based on compute)
149
152
env: <string: string> # dictionary of environment variables
150
153
networking:
151
154
endpoint: <string> # the endpoint for the API (aws only) (default: <api_name>)
0 commit comments