Skip to content

wenyang/cortex

 
 

Repository files navigation


Build machine learning APIs

Cortex makes deploying, scaling, and managing machine learning systems in production simple. We believe that developers in any organization should be able to add natural language processing, computer vision, and other machine learning capabilities to their applications without having to worry about infrastructure.

installdocumentationexampleswe're hiringchat with us


Key features

Deploy

  • Run Cortex locally or as a production cluster on your AWS account.
  • Deploy TensorFlow, PyTorch, scikit-learn, and other models as web APIs.
  • Define preprocessing and postprocessing steps in Python.

Manage

  • Update APIs with no downtime.
  • Stream logs from your APIs to your CLI.
  • Monitor API performance and track predictions.

Scale

  • Automatically scale APIs to handle production traffic.
  • Reduce your cloud infrastructure spend with spot instances.
  • Maximize resource utilization by deploying multiple models per API.

How it works

Implement your predictor in predictor.py, configure your deployment in cortex.yaml, and run cortex deploy.

Here's how to deploy GPT-2 as a scalable text generation API:

Demo


Get started

Install

bash -c "$(curl -sS https://raw.githubusercontent.com/cortexlabs/cortex/0.18/get-cli.sh)"

See our installation guide, then deploy one of our examples or bring your own models to build custom APIs.

Learn more

Check out our docs and join our community.

About

Build machine learning APIs

Resources

License

Code of conduct

Stars

Watchers

Forks

Packages

No packages published

Languages

  • Go 88.3%
  • Python 6.6%
  • Shell 3.8%
  • Other 1.3%