diff --git a/README.md b/README.md
index dfbe04c7cd..b764efddda 100644
--- a/README.md
+++ b/README.md
@@ -4,10 +4,6 @@ Cortex is an open source platform for deploying machine learning models as produ
-
-
-[install](https://cortex.dev/install) • [tutorial](https://cortex.dev/iris-classifier) • [docs](https://cortex.dev) • [examples](https://github.com/cortexlabs/cortex/tree/0.13/examples) • [we're hiring](https://angel.co/cortex-labs-inc/jobs) • [email us](mailto:hello@cortex.dev) • [chat with us](https://gitter.im/cortexlabs/cortex)
-
![Demo](https://d1zqebknpdh033.cloudfront.net/demo/gif/v0.13_2.gif)
@@ -33,7 +29,7 @@ Cortex is designed to be self-hosted on any AWS account. You can spin up a clust
```bash
# install the CLI on your machine
-$ bash -c "$(curl -sS https://raw.githubusercontent.com/cortexlabs/cortex/0.13/get-cli.sh)"
+$ bash -c "$(curl -sS https://raw.githubusercontent.com/cortexlabs/cortex/0.14/get-cli.sh)"
# provision infrastructure on AWS and spin up a cluster
$ cortex cluster up
@@ -140,8 +136,8 @@ The CLI sends configuration and code to the cluster every time you run `cortex d
## Examples of Cortex deployments
-* [Sentiment analysis](https://github.com/cortexlabs/cortex/tree/0.13/examples/tensorflow/sentiment-analyzer): deploy a BERT model for sentiment analysis.
-* [Image classification](https://github.com/cortexlabs/cortex/tree/0.13/examples/tensorflow/image-classifier): deploy an Inception model to classify images.
-* [Search completion](https://github.com/cortexlabs/cortex/tree/0.13/examples/pytorch/search-completer): deploy Facebook's RoBERTa model to complete search terms.
-* [Text generation](https://github.com/cortexlabs/cortex/tree/0.13/examples/pytorch/text-generator): deploy Hugging Face's DistilGPT2 model to generate text.
-* [Iris classification](https://github.com/cortexlabs/cortex/tree/0.13/examples/sklearn/iris-classifier): deploy a scikit-learn model to classify iris flowers.
+* [Sentiment analysis](https://github.com/cortexlabs/cortex/tree/0.14/examples/tensorflow/sentiment-analyzer): deploy a BERT model for sentiment analysis.
+* [Image classification](https://github.com/cortexlabs/cortex/tree/0.14/examples/tensorflow/image-classifier): deploy an Inception model to classify images.
+* [Search completion](https://github.com/cortexlabs/cortex/tree/0.14/examples/pytorch/search-completer): deploy Facebook's RoBERTa model to complete search terms.
+* [Text generation](https://github.com/cortexlabs/cortex/tree/0.14/examples/pytorch/text-generator): deploy Hugging Face's DistilGPT2 model to generate text.
+* [Iris classification](https://github.com/cortexlabs/cortex/tree/0.14/examples/sklearn/iris-classifier): deploy a scikit-learn model to classify iris flowers.
diff --git a/build/build-image.sh b/build/build-image.sh
index eae2319d9c..f910e298df 100755
--- a/build/build-image.sh
+++ b/build/build-image.sh
@@ -19,7 +19,7 @@ set -euo pipefail
ROOT="$(cd "$(dirname "${BASH_SOURCE[0]}")"/.. >/dev/null && pwd)"
-CORTEX_VERSION=master
+CORTEX_VERSION=0.14.0
dir=$1
image=$2
diff --git a/build/cli.sh b/build/cli.sh
index 18c5c9e3f4..8aab757c82 100755
--- a/build/cli.sh
+++ b/build/cli.sh
@@ -19,7 +19,7 @@ set -euo pipefail
ROOT="$(cd "$(dirname "${BASH_SOURCE[0]}")"/.. >/dev/null && pwd)"
-CORTEX_VERSION=master
+CORTEX_VERSION=0.14.0
arg1=${1:-""}
upload="false"
diff --git a/build/push-image.sh b/build/push-image.sh
index 206f7de948..a6403e8887 100755
--- a/build/push-image.sh
+++ b/build/push-image.sh
@@ -17,7 +17,7 @@
set -euo pipefail
-CORTEX_VERSION=master
+CORTEX_VERSION=0.14.0
image=$1
diff --git a/docs/cluster-management/config.md b/docs/cluster-management/config.md
index 8662076ebf..45e9f1515a 100644
--- a/docs/cluster-management/config.md
+++ b/docs/cluster-management/config.md
@@ -43,28 +43,28 @@ instance_volume_size: 50
log_group: cortex
# whether to use spot instances in the cluster (default: false)
-# see https://cortex.dev/v/master/cluster-management/spot-instances for additional details on spot configuration
+# see https://cortex.dev/v/0.14/cluster-management/spot-instances for additional details on spot configuration
spot: false
# docker image paths
-image_python_serve: cortexlabs/python-serve:master
-image_python_serve_gpu: cortexlabs/python-serve-gpu:master
-image_tf_serve: cortexlabs/tf-serve:master
-image_tf_serve_gpu: cortexlabs/tf-serve-gpu:master
-image_tf_api: cortexlabs/tf-api:master
-image_onnx_serve: cortexlabs/onnx-serve:master
-image_onnx_serve_gpu: cortexlabs/onnx-serve-gpu:master
-image_operator: cortexlabs/operator:master
-image_manager: cortexlabs/manager:master
-image_downloader: cortexlabs/downloader:master
-image_request_monitor: cortexlabs/request-monitor:master
-image_cluster_autoscaler: cortexlabs/cluster-autoscaler:master
-image_metrics_server: cortexlabs/metrics-server:master
-image_nvidia: cortexlabs/nvidia:master
-image_fluentd: cortexlabs/fluentd:master
-image_statsd: cortexlabs/statsd:master
-image_istio_proxy: cortexlabs/istio-proxy:master
-image_istio_pilot: cortexlabs/istio-pilot:master
-image_istio_citadel: cortexlabs/istio-citadel:master
-image_istio_galley: cortexlabs/istio-galley:master
+image_python_serve: cortexlabs/python-serve:0.14.0
+image_python_serve_gpu: cortexlabs/python-serve-gpu:0.14.0
+image_tf_serve: cortexlabs/tf-serve:0.14.0
+image_tf_serve_gpu: cortexlabs/tf-serve-gpu:0.14.0
+image_tf_api: cortexlabs/tf-api:0.14.0
+image_onnx_serve: cortexlabs/onnx-serve:0.14.0
+image_onnx_serve_gpu: cortexlabs/onnx-serve-gpu:0.14.0
+image_operator: cortexlabs/operator:0.14.0
+image_manager: cortexlabs/manager:0.14.0
+image_downloader: cortexlabs/downloader:0.14.0
+image_request_monitor: cortexlabs/request-monitor:0.14.0
+image_cluster_autoscaler: cortexlabs/cluster-autoscaler:0.14.0
+image_metrics_server: cortexlabs/metrics-server:0.14.0
+image_nvidia: cortexlabs/nvidia:0.14.0
+image_fluentd: cortexlabs/fluentd:0.14.0
+image_statsd: cortexlabs/statsd:0.14.0
+image_istio_proxy: cortexlabs/istio-proxy:0.14.0
+image_istio_pilot: cortexlabs/istio-pilot:0.14.0
+image_istio_citadel: cortexlabs/istio-citadel:0.14.0
+image_istio_galley: cortexlabs/istio-galley:0.14.0
```
diff --git a/docs/cluster-management/install.md b/docs/cluster-management/install.md
index 4d88231e6a..f767c65bb5 100644
--- a/docs/cluster-management/install.md
+++ b/docs/cluster-management/install.md
@@ -12,7 +12,7 @@ See [cluster configuration](config.md) to learn how you can customize your clust
```bash
# install the CLI on your machine
-$ bash -c "$(curl -sS https://raw.githubusercontent.com/cortexlabs/cortex/master/get-cli.sh)"
+$ bash -c "$(curl -sS https://raw.githubusercontent.com/cortexlabs/cortex/0.14/get-cli.sh)"
# provision infrastructure on AWS and spin up a cluster
$ cortex cluster up
@@ -38,7 +38,7 @@ your cluster is ready!
```bash
# clone the Cortex repository
-git clone -b master https://github.com/cortexlabs/cortex.git
+git clone -b 0.14 https://github.com/cortexlabs/cortex.git
# navigate to the TensorFlow iris classification example
cd cortex/examples/tensorflow/iris-classifier
diff --git a/docs/cluster-management/update.md b/docs/cluster-management/update.md
index fe20eb9821..401da895c5 100644
--- a/docs/cluster-management/update.md
+++ b/docs/cluster-management/update.md
@@ -22,7 +22,7 @@ cortex cluster update
cortex cluster down
# update your CLI
-bash -c "$(curl -sS https://raw.githubusercontent.com/cortexlabs/cortex/master/get-cli.sh)"
+bash -c "$(curl -sS https://raw.githubusercontent.com/cortexlabs/cortex/0.14/get-cli.sh)"
# confirm version
cortex version
diff --git a/docs/deployments/onnx.md b/docs/deployments/onnx.md
index 8bda1fb35c..b4dda224b1 100644
--- a/docs/deployments/onnx.md
+++ b/docs/deployments/onnx.md
@@ -67,7 +67,7 @@ You can log information about each request by adding a `?debug=true` parameter t
An ONNX Predictor is a Python class that describes how to serve your ONNX model to make predictions.
-Cortex provides an `onnx_client` and a config object to initialize your implementation of the ONNX Predictor class. The `onnx_client` is an instance of [ONNXClient](https://github.com/cortexlabs/cortex/tree/master/pkg/workloads/cortex/lib/client/onnx.py) that manages an ONNX Runtime session and helps make predictions using your model. Once your implementation of the ONNX Predictor class has been initialized, the replica is available to serve requests. Upon receiving a request, your implementation's `predict()` function is called with the JSON payload and is responsible for returning a prediction or batch of predictions. Your `predict()` function should call `onnx_client.predict()` to make an inference against your exported ONNX model. Preprocessing of the JSON payload and postprocessing of predictions can be implemented in your `predict()` function as well.
+Cortex provides an `onnx_client` and a config object to initialize your implementation of the ONNX Predictor class. The `onnx_client` is an instance of [ONNXClient](https://github.com/cortexlabs/cortex/tree/0.14/pkg/workloads/cortex/lib/client/onnx.py) that manages an ONNX Runtime session and helps make predictions using your model. Once your implementation of the ONNX Predictor class has been initialized, the replica is available to serve requests. Upon receiving a request, your implementation's `predict()` function is called with the JSON payload and is responsible for returning a prediction or batch of predictions. Your `predict()` function should call `onnx_client.predict()` to make an inference against your exported ONNX model. Preprocessing of the JSON payload and postprocessing of predictions can be implemented in your `predict()` function as well.
## Implementation
@@ -133,6 +133,6 @@ requests==2.22.0
```
-The pre-installed system packages are listed in the [onnx-serve Dockerfile](https://github.com/cortexlabs/cortex/tree/master/images/onnx-serve/Dockerfile) (for CPU) or the [onnx-serve-gpu Dockerfile](https://github.com/cortexlabs/cortex/tree/master/images/onnx-serve-gpu/Dockerfile) (for GPU).
+The pre-installed system packages are listed in the [onnx-serve Dockerfile](https://github.com/cortexlabs/cortex/tree/0.14/images/onnx-serve/Dockerfile) (for CPU) or the [onnx-serve-gpu Dockerfile](https://github.com/cortexlabs/cortex/tree/0.14/images/onnx-serve-gpu/Dockerfile) (for GPU).
If your application requires additional dependencies, you can [install additional Python packages](../dependency-management/python-packages.md) or [install additional system packages](../dependency-management/system-packages.md).
diff --git a/docs/deployments/python.md b/docs/deployments/python.md
index 55daa795ca..30a2e7fe83 100644
--- a/docs/deployments/python.md
+++ b/docs/deployments/python.md
@@ -171,6 +171,6 @@ xgboost==0.90
```
-The pre-installed system packages are listed in the [python-serve Dockerfile](https://github.com/cortexlabs/cortex/tree/master/images/python-serve/Dockerfile) (for CPU) or the [python-serve-gpu Dockerfile](https://github.com/cortexlabs/cortex/tree/master/images/python-serve-gpu/Dockerfile) (for GPU).
+The pre-installed system packages are listed in the [python-serve Dockerfile](https://github.com/cortexlabs/cortex/tree/0.14/images/python-serve/Dockerfile) (for CPU) or the [python-serve-gpu Dockerfile](https://github.com/cortexlabs/cortex/tree/0.14/images/python-serve-gpu/Dockerfile) (for GPU).
If your application requires additional dependencies, you can [install additional Python packages](../dependency-management/python-packages.md) or [install additional system packages](../dependency-management/system-packages.md).
diff --git a/docs/deployments/tensorflow.md b/docs/deployments/tensorflow.md
index f8b57ede43..7687ba26d9 100644
--- a/docs/deployments/tensorflow.md
+++ b/docs/deployments/tensorflow.md
@@ -68,7 +68,7 @@ You can log information about each request by adding a `?debug=true` parameter t
A TensorFlow Predictor is a Python class that describes how to serve your TensorFlow model to make predictions.
-Cortex provides a `tensorflow_client` and a config object to initialize your implementation of the TensorFlow Predictor class. The `tensorflow_client` is an instance of [TensorFlowClient](https://github.com/cortexlabs/cortex/tree/master/pkg/workloads/cortex/lib/client/tensorflow.py) that manages a connection to a TensorFlow Serving container via gRPC to make predictions using your model. Once your implementation of the TensorFlow Predictor class has been initialized, the replica is available to serve requests. Upon receiving a request, your implementation's `predict()` function is called with the JSON payload and is responsible for returning a prediction or batch of predictions. Your `predict()` function should call `tensorflow_client.predict()` to make an inference against your exported TensorFlow model. Preprocessing of the JSON payload and postprocessing of predictions can be implemented in your `predict()` function as well.
+Cortex provides a `tensorflow_client` and a config object to initialize your implementation of the TensorFlow Predictor class. The `tensorflow_client` is an instance of [TensorFlowClient](https://github.com/cortexlabs/cortex/tree/0.14/pkg/workloads/cortex/lib/client/tensorflow.py) that manages a connection to a TensorFlow Serving container via gRPC to make predictions using your model. Once your implementation of the TensorFlow Predictor class has been initialized, the replica is available to serve requests. Upon receiving a request, your implementation's `predict()` function is called with the JSON payload and is responsible for returning a prediction or batch of predictions. Your `predict()` function should call `tensorflow_client.predict()` to make an inference against your exported TensorFlow model. Preprocessing of the JSON payload and postprocessing of predictions can be implemented in your `predict()` function as well.
## Implementation
@@ -128,6 +128,6 @@ tensorflow==2.1.0
```
-The pre-installed system packages are listed in the [tf-api Dockerfile](https://github.com/cortexlabs/cortex/tree/master/images/tf-api/Dockerfile).
+The pre-installed system packages are listed in the [tf-api Dockerfile](https://github.com/cortexlabs/cortex/tree/0.14/images/tf-api/Dockerfile).
If your application requires additional dependencies, you can [install additional Python packages](../dependency-management/python-packages.md) or [install additional system packages](../dependency-management/system-packages.md).
diff --git a/docs/packaging-models/tensorflow.md b/docs/packaging-models/tensorflow.md
index 692b2e8ba3..8c01c9ced8 100644
--- a/docs/packaging-models/tensorflow.md
+++ b/docs/packaging-models/tensorflow.md
@@ -1,7 +1,7 @@
# TensorFlow
-Export your trained model and upload the export directory, or a checkpoint directory containing the export directory (which is usually the case if you used `estimator.train_and_evaluate`). An example is shown below (here is the [complete example](https://github.com/cortexlabs/cortex/blob/master/examples/tensorflow/sentiment-analyzer)):
+Export your trained model and upload the export directory, or a checkpoint directory containing the export directory (which is usually the case if you used `estimator.train_and_evaluate`). An example is shown below (here is the [complete example](https://github.com/cortexlabs/cortex/blob/0.14/examples/tensorflow/sentiment-analyzer)):
```python
import tensorflow as tf
diff --git a/docs/summary.md b/docs/summary.md
index 081f90e68a..7b42f67e8d 100644
--- a/docs/summary.md
+++ b/docs/summary.md
@@ -4,7 +4,7 @@
* [Install](cluster-management/install.md)
* [Tutorial](../examples/sklearn/iris-classifier/README.md)
* [GitHub](https://github.com/cortexlabs/cortex)
-* [Examples](https://github.com/cortexlabs/cortex/tree/master/examples)
+* [Examples](https://github.com/cortexlabs/cortex/tree/0.14/examples)
* [Chat with us](https://gitter.im/cortexlabs/cortex)
* [Email us](mailto:hello@cortex.dev)
* [We're hiring](https://angel.co/cortex-labs-inc/jobs)
diff --git a/examples/tensorflow/image-classifier/inception.ipynb b/examples/tensorflow/image-classifier/inception.ipynb
index bc6a2c0de2..5dd7036d00 100644
--- a/examples/tensorflow/image-classifier/inception.ipynb
+++ b/examples/tensorflow/image-classifier/inception.ipynb
@@ -204,7 +204,7 @@
},
"source": [
"\n",
- "That's it! See the [example on GitHub](https://github.com/cortexlabs/cortex/tree/master/examples/tensorflow/image-classifier) for how to deploy the model as an API."
+ "That's it! See the [example on GitHub](https://github.com/cortexlabs/cortex/tree/0.14/examples/tensorflow/image-classifier) for how to deploy the model as an API."
]
}
]
diff --git a/examples/tensorflow/iris-classifier/tensorflow.ipynb b/examples/tensorflow/iris-classifier/tensorflow.ipynb
index 5b53b29a7c..61f1608dbc 100644
--- a/examples/tensorflow/iris-classifier/tensorflow.ipynb
+++ b/examples/tensorflow/iris-classifier/tensorflow.ipynb
@@ -289,7 +289,7 @@
},
"source": [
"\n",
- "That's it! See the [example on GitHub](https://github.com/cortexlabs/cortex/tree/master/examples/tensorflow/iris-classifier) for how to deploy the model as an API."
+ "That's it! See the [example on GitHub](https://github.com/cortexlabs/cortex/tree/0.14/examples/tensorflow/iris-classifier) for how to deploy the model as an API."
]
}
]
diff --git a/examples/tensorflow/sentiment-analyzer/bert.ipynb b/examples/tensorflow/sentiment-analyzer/bert.ipynb
index 74f33fdcfa..968cac7fda 100644
--- a/examples/tensorflow/sentiment-analyzer/bert.ipynb
+++ b/examples/tensorflow/sentiment-analyzer/bert.ipynb
@@ -1000,7 +1000,7 @@
},
"source": [
"\n",
- "That's it! See the [example on GitHub](https://github.com/cortexlabs/cortex/tree/master/examples/tensorflow/sentiment-analyzer) for how to deploy the model as an API."
+ "That's it! See the [example on GitHub](https://github.com/cortexlabs/cortex/tree/0.14/examples/tensorflow/sentiment-analyzer) for how to deploy the model as an API."
]
}
]
diff --git a/examples/tensorflow/text-generator/gpt-2.ipynb b/examples/tensorflow/text-generator/gpt-2.ipynb
index 9d96a93975..d1316760f2 100644
--- a/examples/tensorflow/text-generator/gpt-2.ipynb
+++ b/examples/tensorflow/text-generator/gpt-2.ipynb
@@ -346,7 +346,7 @@
},
"source": [
"\n",
- "We also need to upload `vocab.bpe` and `encoder.json`, so that the [encoder](https://github.com/cortexlabs/cortex/blob/master/examples/tensorflow/text-generator/encoder.py) in the [Predictor](https://github.com/cortexlabs/cortex/blob/master/examples/tensorflow/text-generator/predictor.py) can encode the input text before making a request to the model."
+ "We also need to upload `vocab.bpe` and `encoder.json`, so that the [encoder](https://github.com/cortexlabs/cortex/blob/0.14/examples/tensorflow/text-generator/encoder.py) in the [Predictor](https://github.com/cortexlabs/cortex/blob/0.14/examples/tensorflow/text-generator/predictor.py) can encode the input text before making a request to the model."
]
},
{
@@ -376,7 +376,7 @@
},
"source": [
"\n",
- "That's it! See the [example on GitHub](https://github.com/cortexlabs/cortex/tree/master/examples/tensorflow/text-generator) for how to deploy the model as an API."
+ "That's it! See the [example on GitHub](https://github.com/cortexlabs/cortex/tree/0.14/examples/tensorflow/text-generator) for how to deploy the model as an API."
]
}
]
diff --git a/examples/xgboost/iris-classifier/xgboost.ipynb b/examples/xgboost/iris-classifier/xgboost.ipynb
index cd2de8ef00..72b70707c3 100644
--- a/examples/xgboost/iris-classifier/xgboost.ipynb
+++ b/examples/xgboost/iris-classifier/xgboost.ipynb
@@ -237,7 +237,7 @@
},
"source": [
"\n",
- "That's it! See the [example](https://github.com/cortexlabs/cortex/tree/master/examples/xgboost/iris-classifier) for how to deploy the model as an API."
+ "That's it! See the [example](https://github.com/cortexlabs/cortex/tree/0.14/examples/xgboost/iris-classifier) for how to deploy the model as an API."
]
}
]
diff --git a/get-cli.sh b/get-cli.sh
index f5dcb502ed..a65971acec 100755
--- a/get-cli.sh
+++ b/get-cli.sh
@@ -16,7 +16,7 @@
set -e
-CORTEX_VERSION_BRANCH_STABLE=master
+CORTEX_VERSION_BRANCH_STABLE=0.14.0
case "$OSTYPE" in
darwin*) parsed_os="darwin" ;;
diff --git a/manager/install.sh b/manager/install.sh
index 975fa5dc6e..7933e4096c 100755
--- a/manager/install.sh
+++ b/manager/install.sh
@@ -16,13 +16,12 @@
set -e
-CORTEX_VERSION=master
+CORTEX_VERSION=0.14.0
EKSCTL_TIMEOUT=45m
arg1="$1"
function ensure_eks() {
- # Cluster statuses: https://github.com/aws/aws-sdk-go/blob/master/service/eks/api.go#L2785
set +e
cluster_info=$(eksctl get cluster --name=$CORTEX_CLUSTER_NAME --region=$CORTEX_REGION -o json)
cluster_info_exit_code=$?
diff --git a/pkg/consts/consts.go b/pkg/consts/consts.go
index 2f0786833c..c8e453dad7 100644
--- a/pkg/consts/consts.go
+++ b/pkg/consts/consts.go
@@ -17,8 +17,8 @@ limitations under the License.
package consts
var (
- CortexVersion = "master" // CORTEX_VERSION
- CortexVersionMinor = "master" // CORTEX_VERSION_MINOR
+ CortexVersion = "0.14.0" // CORTEX_VERSION
+ CortexVersionMinor = "0.14" // CORTEX_VERSION_MINOR
MaxClassesPerTrackerRequest = 20 // cloudwatch.GeMetricData can get up to 100 metrics per request, avoid multiple requests and have room for other stats
)
diff --git a/pkg/workloads/cortex/client/cortex/client.py b/pkg/workloads/cortex/client/cortex/client.py
index fe2348a47f..1968582b2b 100644
--- a/pkg/workloads/cortex/client/cortex/client.py
+++ b/pkg/workloads/cortex/client/cortex/client.py
@@ -44,7 +44,7 @@ def __init__(self, aws_access_key_id, aws_secret_access_key, operator_url):
self.aws_access_key_id = aws_access_key_id
self.aws_secret_access_key = aws_secret_access_key
self.headers = {
- "CortexAPIVersion": "master", # CORTEX_VERSION
+ "CortexAPIVersion": "0.14.0", # CORTEX_VERSION
"Authorization": "CortexAWS {}|{}".format(
self.aws_access_key_id, self.aws_secret_access_key
),
diff --git a/pkg/workloads/cortex/client/setup.py b/pkg/workloads/cortex/client/setup.py
index 2ae1f40f07..994ccadda7 100644
--- a/pkg/workloads/cortex/client/setup.py
+++ b/pkg/workloads/cortex/client/setup.py
@@ -16,7 +16,7 @@
setup(
name="cortex",
- version="master", # CORTEX_VERSION
+ version="0.14.0", # CORTEX_VERSION
description="",
author="Cortex Labs",
author_email="dev@cortexlabs.com",
diff --git a/pkg/workloads/cortex/consts.py b/pkg/workloads/cortex/consts.py
index 32deb4d541..d3fef85e7d 100644
--- a/pkg/workloads/cortex/consts.py
+++ b/pkg/workloads/cortex/consts.py
@@ -12,4 +12,4 @@
# See the License for the specific language governing permissions and
# limitations under the License.
-CORTEX_VERSION = "master"
+CORTEX_VERSION = "0.14.0"