diff --git a/docs/clients/install.md b/docs/clients/install.md index b88b21ecff..c5ac2cde8d 100644 --- a/docs/clients/install.md +++ b/docs/clients/install.md @@ -9,10 +9,10 @@ pip install cortex ``` -To install or upgrade to a specific version (e.g. v0.31.0): +To install or upgrade to a specific version (e.g. v0.31.1): ```bash -pip install cortex==0.31.0 +pip install cortex==0.31.1 ``` To upgrade to the latest version: @@ -25,8 +25,8 @@ pip install --upgrade cortex ```bash -# For example to download CLI version 0.31.0 (Note the "v"): -bash -c "$(curl -sS https://raw.githubusercontent.com/cortexlabs/cortex/v0.31.0/get-cli.sh)" +# For example to download CLI version 0.31.1 (Note the "v"): +bash -c "$(curl -sS https://raw.githubusercontent.com/cortexlabs/cortex/v0.31.1/get-cli.sh)" ``` By default, the Cortex CLI is installed at `/usr/local/bin/cortex`. To install the executable elsewhere, export the `CORTEX_INSTALL_PATH` environment variable to your desired location before running the command above. diff --git a/docs/clusters/aws/install.md b/docs/clusters/aws/install.md index 9c36a0c6a5..5686e13529 100644 --- a/docs/clusters/aws/install.md +++ b/docs/clusters/aws/install.md @@ -101,26 +101,26 @@ The docker images used by the Cortex cluster can also be overridden, although th ```yaml -image_operator: quay.io/cortexlabs/operator:0.31.0 -image_manager: quay.io/cortexlabs/manager:0.31.0 -image_downloader: quay.io/cortexlabs/downloader:0.31.0 -image_request_monitor: quay.io/cortexlabs/request-monitor:0.31.0 -image_cluster_autoscaler: quay.io/cortexlabs/cluster-autoscaler:0.31.0 -image_metrics_server: quay.io/cortexlabs/metrics-server:0.31.0 -image_inferentia: quay.io/cortexlabs/inferentia:0.31.0 -image_neuron_rtd: quay.io/cortexlabs/neuron-rtd:0.31.0 -image_nvidia: quay.io/cortexlabs/nvidia:0.31.0 -image_fluent_bit: quay.io/cortexlabs/fluent-bit:0.31.0 -image_istio_proxy: quay.io/cortexlabs/istio-proxy:0.31.0 -image_istio_pilot: quay.io/cortexlabs/istio-pilot:0.31.0 -image_prometheus: quay.io/cortexlabs/prometheus:0.31.0 -image_prometheus_config_reloader: quay.io/cortexlabs/prometheus-config-reloader:0.31.0 -image_prometheus_operator: quay.io/cortexlabs/prometheus-operator:0.31.0 -image_prometheus_statsd_exporter: quay.io/cortexlabs/prometheus-statsd-exporter:0.31.0 -image_prometheus_dcgm_exporter: quay.io/cortexlabs/prometheus-dcgm-exporter:0.31.0 -image_prometheus_kube_state_metrics: quay.io/cortexlabs/prometheus-kube-state-metrics:0.31.0 -image_prometheus_node_exporter: quay.io/cortexlabs/prometheus-node-exporter:0.31.0 -image_kube_rbac_proxy: quay.io/cortexlabs/kube-rbac-proxy:0.31.0 -image_grafana: quay.io/cortexlabs/grafana:0.31.0 -image_event_exporter: quay.io/cortexlabs/event-exporter:0.31.0 +image_operator: quay.io/cortexlabs/operator:0.31.1 +image_manager: quay.io/cortexlabs/manager:0.31.1 +image_downloader: quay.io/cortexlabs/downloader:0.31.1 +image_request_monitor: quay.io/cortexlabs/request-monitor:0.31.1 +image_cluster_autoscaler: quay.io/cortexlabs/cluster-autoscaler:0.31.1 +image_metrics_server: quay.io/cortexlabs/metrics-server:0.31.1 +image_inferentia: quay.io/cortexlabs/inferentia:0.31.1 +image_neuron_rtd: quay.io/cortexlabs/neuron-rtd:0.31.1 +image_nvidia: quay.io/cortexlabs/nvidia:0.31.1 +image_fluent_bit: quay.io/cortexlabs/fluent-bit:0.31.1 +image_istio_proxy: quay.io/cortexlabs/istio-proxy:0.31.1 +image_istio_pilot: quay.io/cortexlabs/istio-pilot:0.31.1 +image_prometheus: quay.io/cortexlabs/prometheus:0.31.1 +image_prometheus_config_reloader: quay.io/cortexlabs/prometheus-config-reloader:0.31.1 +image_prometheus_operator: quay.io/cortexlabs/prometheus-operator:0.31.1 +image_prometheus_statsd_exporter: quay.io/cortexlabs/prometheus-statsd-exporter:0.31.1 +image_prometheus_dcgm_exporter: quay.io/cortexlabs/prometheus-dcgm-exporter:0.31.1 +image_prometheus_kube_state_metrics: quay.io/cortexlabs/prometheus-kube-state-metrics:0.31.1 +image_prometheus_node_exporter: quay.io/cortexlabs/prometheus-node-exporter:0.31.1 +image_kube_rbac_proxy: quay.io/cortexlabs/kube-rbac-proxy:0.31.1 +image_grafana: quay.io/cortexlabs/grafana:0.31.1 +image_event_exporter: quay.io/cortexlabs/event-exporter:0.31.1 ``` diff --git a/docs/clusters/gcp/install.md b/docs/clusters/gcp/install.md index 8eeabda02a..f2aee94afa 100644 --- a/docs/clusters/gcp/install.md +++ b/docs/clusters/gcp/install.md @@ -68,21 +68,21 @@ The docker images used by the Cortex cluster can also be overridden, although th ```yaml -image_operator: quay.io/cortexlabs/operator:0.31.0 -image_manager: quay.io/cortexlabs/manager:0.31.0 -image_downloader: quay.io/cortexlabs/downloader:0.31.0 -image_request_monitor: quay.io/cortexlabs/request-monitor:0.31.0 -image_istio_proxy: quay.io/cortexlabs/istio-proxy:0.31.0 -image_istio_pilot: quay.io/cortexlabs/istio-pilot:0.31.0 -image_google_pause: quay.io/cortexlabs/google-pause:0.31.0 -image_prometheus: quay.io/cortexlabs/prometheus:0.31.0 -image_prometheus_config_reloader: quay.io/cortexlabs/prometheus-config-reloader:0.31.0 -image_prometheus_operator: quay.io/cortexlabs/prometheus-operator:0.31.0 -image_prometheus_statsd_exporter: quay.io/cortexlabs/prometheus-statsd-exporter:0.31.0 -image_prometheus_dcgm_exporter: quay.io/cortexlabs/prometheus-dcgm-exporter:0.31.0 -image_prometheus_kube_state_metrics: quay.io/cortexlabs/prometheus-kube-state-metrics:0.31.0 -image_prometheus_node_exporter: quay.io/cortexlabs/prometheus-node-exporter:0.31.0 -image_kube_rbac_proxy: quay.io/cortexlabs/kube-rbac-proxy:0.31.0 -image_grafana: quay.io/cortexlabs/grafana:0.31.0 -image_event_exporter: quay.io/cortexlabs/event-exporter:0.31.0 +image_operator: quay.io/cortexlabs/operator:0.31.1 +image_manager: quay.io/cortexlabs/manager:0.31.1 +image_downloader: quay.io/cortexlabs/downloader:0.31.1 +image_request_monitor: quay.io/cortexlabs/request-monitor:0.31.1 +image_istio_proxy: quay.io/cortexlabs/istio-proxy:0.31.1 +image_istio_pilot: quay.io/cortexlabs/istio-pilot:0.31.1 +image_google_pause: quay.io/cortexlabs/google-pause:0.31.1 +image_prometheus: quay.io/cortexlabs/prometheus:0.31.1 +image_prometheus_config_reloader: quay.io/cortexlabs/prometheus-config-reloader:0.31.1 +image_prometheus_operator: quay.io/cortexlabs/prometheus-operator:0.31.1 +image_prometheus_statsd_exporter: quay.io/cortexlabs/prometheus-statsd-exporter:0.31.1 +image_prometheus_dcgm_exporter: quay.io/cortexlabs/prometheus-dcgm-exporter:0.31.1 +image_prometheus_kube_state_metrics: quay.io/cortexlabs/prometheus-kube-state-metrics:0.31.1 +image_prometheus_node_exporter: quay.io/cortexlabs/prometheus-node-exporter:0.31.1 +image_kube_rbac_proxy: quay.io/cortexlabs/kube-rbac-proxy:0.31.1 +image_grafana: quay.io/cortexlabs/grafana:0.31.1 +image_event_exporter: quay.io/cortexlabs/event-exporter:0.31.1 ``` diff --git a/docs/workloads/async/configuration.md b/docs/workloads/async/configuration.md index 636f8a1839..a28164a2e4 100644 --- a/docs/workloads/async/configuration.md +++ b/docs/workloads/async/configuration.md @@ -26,7 +26,7 @@ predictor: shell: # relative path to a shell script for system package installation (default: dependencies.sh) config: # arbitrary dictionary passed to the constructor of the Predictor (optional) python_path: # path to the root of your Python folder that will be appended to PYTHONPATH (default: folder containing cortex.yaml) - image: # docker image to use for the Predictor (default: quay.io/cortexlabs/python-predictor-cpu:0.31.0, quay.io/cortexlabs/python-predictor-gpu:0.31.0-cuda10.2-cudnn8, or quay.io/cortexlabs/python-predictor-inf:0.31.0 based on compute) + image: # docker image to use for the Predictor (default: quay.io/cortexlabs/python-predictor-cpu:0.31.1, quay.io/cortexlabs/python-predictor-gpu:0.31.1-cuda10.2-cudnn8, or quay.io/cortexlabs/python-predictor-inf:0.31.1 based on compute) env: # dictionary of environment variables log_level: # log level that can be "debug", "info", "warning" or "error" (default: "info") shm_size: # size of shared memory (/dev/shm) for sharing data between multiple processes, e.g. 64Mi or 1Gi (default: Null) diff --git a/docs/workloads/batch/configuration.md b/docs/workloads/batch/configuration.md index c8099b5b24..7603ded542 100644 --- a/docs/workloads/batch/configuration.md +++ b/docs/workloads/batch/configuration.md @@ -19,7 +19,7 @@ predictor: path: # path to a python file with a PythonPredictor class definition, relative to the Cortex root (required) config: # arbitrary dictionary passed to the constructor of the Predictor (can be overridden by config passed in job submission) (optional) python_path: # path to the root of your Python folder that will be appended to PYTHONPATH (default: folder containing cortex.yaml) - image: # docker image to use for the Predictor (default: quay.io/cortexlabs/python-predictor-cpu:0.31.0 or quay.io/cortexlabs/python-predictor-gpu:0.31.0-cuda10.2-cudnn8 based on compute) + image: # docker image to use for the Predictor (default: quay.io/cortexlabs/python-predictor-cpu:0.31.1 or quay.io/cortexlabs/python-predictor-gpu:0.31.1-cuda10.2-cudnn8 based on compute) env: # dictionary of environment variables log_level: # log level that can be "debug", "info", "warning" or "error" (default: "info") shm_size: # size of shared memory (/dev/shm) for sharing data between multiple processes, e.g. 64Mi or 1Gi (default: Null) @@ -49,8 +49,8 @@ predictor: batch_interval: # the maximum amount of time to spend waiting for additional requests before running inference on the batch of requests config: # arbitrary dictionary passed to the constructor of the Predictor (can be overridden by config passed in job submission) (optional) python_path: # path to the root of your Python folder that will be appended to PYTHONPATH (default: folder containing cortex.yaml) - image: # docker image to use for the Predictor (default: quay.io/cortexlabs/tensorflow-predictor:0.31.0) - tensorflow_serving_image: # docker image to use for the TensorFlow Serving container (default: quay.io/cortexlabs/tensorflow-serving-cpu:0.31.0 or quay.io/cortexlabs/tensorflow-serving-gpu:0.31.0 based on compute) + image: # docker image to use for the Predictor (default: quay.io/cortexlabs/tensorflow-predictor:0.31.1) + tensorflow_serving_image: # docker image to use for the TensorFlow Serving container (default: quay.io/cortexlabs/tensorflow-serving-cpu:0.31.1 or quay.io/cortexlabs/tensorflow-serving-gpu:0.31.1 based on compute) env: # dictionary of environment variables log_level: # log level that can be "debug", "info", "warning" or "error" (default: "info") shm_size: # size of shared memory (/dev/shm) for sharing data between multiple processes, e.g. 64Mi or 1Gi (default: Null) @@ -75,7 +75,7 @@ predictor: ... config: # arbitrary dictionary passed to the constructor of the Predictor (can be overridden by config passed in job submission) (optional) python_path: # path to the root of your Python folder that will be appended to PYTHONPATH (default: folder containing cortex.yaml) - image: # docker image to use for the Predictor (default: quay.io/cortexlabs/onnx-predictor-cpu:0.31.0 or quay.io/cortexlabs/onnx-predictor-gpu:0.31.0 based on compute) + image: # docker image to use for the Predictor (default: quay.io/cortexlabs/onnx-predictor-cpu:0.31.1 or quay.io/cortexlabs/onnx-predictor-gpu:0.31.1 based on compute) env: # dictionary of environment variables log_level: # log level that can be "debug", "info", "warning" or "error" (default: "info") shm_size: # size of shared memory (/dev/shm) for sharing data between multiple processes, e.g. 64Mi or 1Gi (default: Null) diff --git a/docs/workloads/dependencies/images.md b/docs/workloads/dependencies/images.md index 505013fabe..2a778cc4ab 100644 --- a/docs/workloads/dependencies/images.md +++ b/docs/workloads/dependencies/images.md @@ -11,19 +11,19 @@ mkdir my-api && cd my-api && touch Dockerfile Cortex's base Docker images are listed below. Depending on the Cortex Predictor and compute type specified in your API configuration, choose one of these images to use as the base for your Docker image: -* Python Predictor (CPU): `quay.io/cortexlabs/python-predictor-cpu:0.31.0` +* Python Predictor (CPU): `quay.io/cortexlabs/python-predictor-cpu:0.31.1` * Python Predictor (GPU): choose one of the following: - * `quay.io/cortexlabs/python-predictor-gpu:0.31.0-cuda10.0-cudnn7` - * `quay.io/cortexlabs/python-predictor-gpu:0.31.0-cuda10.1-cudnn7` - * `quay.io/cortexlabs/python-predictor-gpu:0.31.0-cuda10.1-cudnn8` - * `quay.io/cortexlabs/python-predictor-gpu:0.31.0-cuda10.2-cudnn7` - * `quay.io/cortexlabs/python-predictor-gpu:0.31.0-cuda10.2-cudnn8` - * `quay.io/cortexlabs/python-predictor-gpu:0.31.0-cuda11.0-cudnn8` - * `quay.io/cortexlabs/python-predictor-gpu:0.31.0-cuda11.1-cudnn8` -* Python Predictor (Inferentia): `quay.io/cortexlabs/python-predictor-inf:0.31.0` -* TensorFlow Predictor (CPU, GPU, Inferentia): `quay.io/cortexlabs/tensorflow-predictor:0.31.0` -* ONNX Predictor (CPU): `quay.io/cortexlabs/onnx-predictor-cpu:0.31.0` -* ONNX Predictor (GPU): `quay.io/cortexlabs/onnx-predictor-gpu:0.31.0` + * `quay.io/cortexlabs/python-predictor-gpu:0.31.1-cuda10.0-cudnn7` + * `quay.io/cortexlabs/python-predictor-gpu:0.31.1-cuda10.1-cudnn7` + * `quay.io/cortexlabs/python-predictor-gpu:0.31.1-cuda10.1-cudnn8` + * `quay.io/cortexlabs/python-predictor-gpu:0.31.1-cuda10.2-cudnn7` + * `quay.io/cortexlabs/python-predictor-gpu:0.31.1-cuda10.2-cudnn8` + * `quay.io/cortexlabs/python-predictor-gpu:0.31.1-cuda11.0-cudnn8` + * `quay.io/cortexlabs/python-predictor-gpu:0.31.1-cuda11.1-cudnn8` +* Python Predictor (Inferentia): `quay.io/cortexlabs/python-predictor-inf:0.31.1` +* TensorFlow Predictor (CPU, GPU, Inferentia): `quay.io/cortexlabs/tensorflow-predictor:0.31.1` +* ONNX Predictor (CPU): `quay.io/cortexlabs/onnx-predictor-cpu:0.31.1` +* ONNX Predictor (GPU): `quay.io/cortexlabs/onnx-predictor-gpu:0.31.1` The sample `Dockerfile` below inherits from Cortex's Python CPU serving image, and installs 3 packages. `tree` is a system package and `pandas` and `rdkit` are Python packages. @@ -31,7 +31,7 @@ The sample `Dockerfile` below inherits from Cortex's Python CPU serving image, a ```dockerfile # Dockerfile -FROM quay.io/cortexlabs/python-predictor-cpu:0.31.0 +FROM quay.io/cortexlabs/python-predictor-cpu:0.31.1 RUN apt-get update \ && apt-get install -y tree \ @@ -49,7 +49,7 @@ If you need to upgrade the Python Runtime version on your image, you can follow ```Dockerfile # Dockerfile -FROM quay.io/cortexlabs/python-predictor-cpu:0.31.0 +FROM quay.io/cortexlabs/python-predictor-cpu:0.31.1 # upgrade python runtime version RUN conda update -n base -c defaults conda diff --git a/docs/workloads/realtime/configuration.md b/docs/workloads/realtime/configuration.md index 48ad91d2fc..537c4c77d3 100644 --- a/docs/workloads/realtime/configuration.md +++ b/docs/workloads/realtime/configuration.md @@ -39,7 +39,7 @@ predictor: threads_per_process: # the number of threads per process (default: 1) config: # arbitrary dictionary passed to the constructor of the Predictor (optional) python_path: # path to the root of your Python folder that will be appended to PYTHONPATH (default: folder containing cortex.yaml) - image: # docker image to use for the Predictor (default: quay.io/cortexlabs/python-predictor-cpu:0.31.0, quay.io/cortexlabs/python-predictor-gpu:0.31.0-cuda10.2-cudnn8, or quay.io/cortexlabs/python-predictor-inf:0.31.0 based on compute) + image: # docker image to use for the Predictor (default: quay.io/cortexlabs/python-predictor-cpu:0.31.1, quay.io/cortexlabs/python-predictor-gpu:0.31.1-cuda10.2-cudnn8, or quay.io/cortexlabs/python-predictor-inf:0.31.1 based on compute) env: # dictionary of environment variables log_level: # log level that can be "debug", "info", "warning" or "error" (default: "info") shm_size: # size of shared memory (/dev/shm) for sharing data between multiple processes, e.g. 64Mi or 1Gi (default: Null) @@ -74,8 +74,8 @@ predictor: threads_per_process: # the number of threads per process (default: 1) config: # arbitrary dictionary passed to the constructor of the Predictor (optional) python_path: # path to the root of your Python folder that will be appended to PYTHONPATH (default: folder containing cortex.yaml) - image: # docker image to use for the Predictor (default: quay.io/cortexlabs/tensorflow-predictor:0.31.0) - tensorflow_serving_image: # docker image to use for the TensorFlow Serving container (default: quay.io/cortexlabs/tensorflow-serving-cpu:0.31.0, quay.io/cortexlabs/tensorflow-serving-gpu:0.31.0, or quay.io/cortexlabs/tensorflow-serving-inf:0.31.0 based on compute) + image: # docker image to use for the Predictor (default: quay.io/cortexlabs/tensorflow-predictor:0.31.1) + tensorflow_serving_image: # docker image to use for the TensorFlow Serving container (default: quay.io/cortexlabs/tensorflow-serving-cpu:0.31.1, quay.io/cortexlabs/tensorflow-serving-gpu:0.31.1, or quay.io/cortexlabs/tensorflow-serving-inf:0.31.1 based on compute) env: # dictionary of environment variables log_level: # log level that can be "debug", "info", "warning" or "error" (default: "info") shm_size: # size of shared memory (/dev/shm) for sharing data between multiple processes, e.g. 64Mi or 1Gi (default: Null) @@ -105,7 +105,7 @@ predictor: threads_per_process: # the number of threads per process (default: 1) config: # arbitrary dictionary passed to the constructor of the Predictor (optional) python_path: # path to the root of your Python folder that will be appended to PYTHONPATH (default: folder containing cortex.yaml) - image: # docker image to use for the Predictor (default: quay.io/cortexlabs/onnx-predictor-cpu:0.31.0 or quay.io/cortexlabs/onnx-predictor-gpu:0.31.0 based on compute) + image: # docker image to use for the Predictor (default: quay.io/cortexlabs/onnx-predictor-cpu:0.31.1 or quay.io/cortexlabs/onnx-predictor-gpu:0.31.1 based on compute) env: # dictionary of environment variables log_level: # log level that can be "debug", "info", "warning" or "error" (default: "info") shm_size: # size of shared memory (/dev/shm) for sharing data between multiple processes, e.g. 64Mi or 1Gi (default: Null) diff --git a/docs/workloads/task/configuration.md b/docs/workloads/task/configuration.md index 9837124df9..cccc808a00 100644 --- a/docs/workloads/task/configuration.md +++ b/docs/workloads/task/configuration.md @@ -12,7 +12,7 @@ conda: # relative path to conda-packages.txt (default: conda-packages.txt) shell: # relative path to a shell script for system package installation (default: dependencies.sh) python_path: # path to the root of your Python folder that will be appended to PYTHONPATH (default: folder containing cortex.yaml) - image: # docker image to use for the Task (default: quay.io/cortexlabs/python-predictor-cpu:0.31.0, quay.io/cortexlabs/python-predictor-gpu:0.31.0-cuda10.2-cudnn8, or quay.io/cortexlabs/python-predictor-inf:0.31.0 based on compute) + image: # docker image to use for the Task (default: quay.io/cortexlabs/python-predictor-cpu:0.31.1, quay.io/cortexlabs/python-predictor-gpu:0.31.1-cuda10.2-cudnn8, or quay.io/cortexlabs/python-predictor-inf:0.31.1 based on compute) env: # dictionary of environment variables log_level: # log level that can be "debug", "info", "warning" or "error" (default: "info") networking: diff --git a/get-cli.sh b/get-cli.sh index 48c964a014..5b56060ed2 100755 --- a/get-cli.sh +++ b/get-cli.sh @@ -16,7 +16,7 @@ set -e -CORTEX_VERSION_BRANCH_STABLE=0.31.0 +CORTEX_VERSION_BRANCH_STABLE=0.31.1 CORTEX_INSTALL_PATH="${CORTEX_INSTALL_PATH:-/usr/local/bin/cortex}" # replace ~ with the home directory path