Skip to content

Releases: pytorch/serve

TorchServe v0.6.1 Release Notes

14 Nov 20:15
Compare
Choose a tag to compare

This is the release of TorchServe v0.6.1.

New Features

New Examples

Dependency Upgrades

Improvements

Build and CI

Documentation

Deprecations

Platform Support

Ubuntu 16.04, Ubuntu 18.04, MacOS 10.14+, Windows 10 Pro, Windows Server 2019, Windows subsystem for Linux (Windows Server 2019, WSLv1, Ubuntu 18.0.4). TorchServe now requires Python 3.8 and above, and JDK17.

GPU Support

Torch 1.11+ Cuda 10.2, 11.3, 11.6
Torch 1.9.0 + Cuda 11.1
Torch 1.8.1 + Cuda 9.2

TorchServe v0.6.0 Release Notes

16 May 20:02
Compare
Choose a tag to compare

This is the release of TorchServe v0.6.0.

New Features

  • Support PyTorch 1.11 and Cuda 11.3 - Added support for PyTorch 1.11 and Cuda 11.3.
  • Universal Auto Benchmark and Dashboard Tool - Added one command line tool for model analyzer to get benchmark report(sample) and dashboard on any device.
  • HuggingFace model parallelism integration - Added example for HuggingFace model parallelism integration.

Build and CI

  • Added nightly benchmark dashboard - Added nightly benchmark dashboard.
  • Migrated CI, nightly binary and docker build to github workflow - Added CI, docker migration.
  • Fixed gpu regression test buildspec.yaml - Added fixing for gpu regression test buildspec.yaml.

Documentation

Deprecations

  • Deprecated old benchmark/automated directory in favor of new Github Action based workflow

Improvements

  • Fixed workflow threads cleanup - Added fixing to clean workflow inference threadpool.
  • Fixed empty model url - Added fixing for empty model url in model archiver.
  • Fixed load model failure - Added support for loading a model directory.
  • HuggingFace text generation example - Added text generation example.
  • Updated metrics json and qlog format log - Added support for metrics json and qlog format log in log4j2.
  • Added cpu, gpu and memory usage - Added cpu, gpu and memory usage in benchmark-ab.py report.
  • Added exception for torch < 1.8.1 - Added exception to notify torch < 1.8.1.
  • Replaced hard code in install_dependencies.py - Added sys.executable in install_dependencies.py.
  • Added default envelope for workflow - Added default envelope in model manager for workflow.
  • Fixed multiple docker build errors - Fixed /home/venv write permission, typo in docker and added common requirements in docker.
  • Fixed snapshot test - Added fixing for snapshot test.
  • Updated model_zoo.md - Added dog breed, mmf and BERT in model zoo.
  • Added nvgpu in common requirements - Added nvgpu in common dependencies.
  • Fixed Inference API ping response - Fixed typo in Inference API ping response.

Platform Support

Ubuntu 16.04, Ubuntu 18.04, MacOS 10.14+, Windows 10 Pro, Windows Server 2019, Windows subsystem for Linux (Windows Server 2019, WSLv1, Ubuntu 18.0.4). TorchServe now requires Python 3.8 and above.

GPU Support

Torch 1.11+ Cuda 10.2, 11.3
Torch 1.9.0 + Cuda 11.1
Torch 1.8.1 + Cuda 9.2

TorchServe v0.5.3 Release Notes

01 Mar 23:52
Compare
Choose a tag to compare

This is the release of TorchServe v0.5.3.

New Features

  • KServe V2 support - Added support for KServe V2 protocol.
  • Model customized metadata support - Extended managementAPI to support customized metadata from handler.

Improvements

  • Upgraded log4j2 version to 2.17.1 - Added log4j upgrade to address CVE-2021-44832.
  • Upgraded pillow to 9.0.0, python support upgraded to py3.8/py3.9 - Added docker, install dependency upgrade.
  • GPU utilization and GPU memory usage metrics support - Added support for GPU utilization and GPU memory usage metrics in benchmarks.
  • Workflow benchmark support - Added support for workflow benchmark.
  • benchmark-ab.py warmup support - Added support for warmup in benchmark-ab.py.
  • Multiple inputs for a model inference example - Added example to support multiple inputs for a model inference.
  • Documentation refactor - Improved documention.
  • Added API auto-discovery - Added support for API auto-discovery.
  • Nightly build support - Added support for Github action nightly build pip install torchserve-nightly

Platform Support

Ubuntu 16.04, Ubuntu 18.04, MacOS 10.14+, Windows 10 Pro, Windows Server 2019, Windows subsystem for Linux (Windows Server 2019, WSLv1, Ubuntu 18.0.4). TorchServe now requires Python 3.8 and above.

GPU Support

Torch 1.10+ Cuda 10.2, 11.3
Torch 1.9.0 + Cuda 11.1
Torch 1.8.1 + Cuda 9.2

Planned Improvements

TorchServe v0.5.2 Release Notes

29 Dec 21:45
Compare
Choose a tag to compare

This is a hotfix release of Log4j issue.

Log4j Fixing

TorchServe v0.5.1 Release Notes

16 Dec 21:21
Compare
Choose a tag to compare

This is a hotfix release of Log4j issue.

Log4j Fixing

New Features

TorchServe v0.5.0 Release Notes

18 Nov 19:18
Compare
Choose a tag to compare

This is the release of TorchServe v0.5.0.

New Features

  • PyTorch 1.10.0 support - TorchServe is now certified working with torch 1.10.0 torchvision 0.11.1, torchtext 0.11.0 and torchaudio 0.10.0
  • Kubernetes HPA support - Added support for Kubernetes HPA.
  • Faster transformer example - Added example for Faster transformer for optimized transformer model inference.
  • (experimental) torchprep support - Added experimental CLI tool to prepare Pytorch models for efficient inference.
  • Custom metrics example - Added example for custom metrics with mtail metrics exporter and Prometheus.
  • Reactjs example for Image Classifier - Added example for Reactjs Image Classifier.

Improvements

  • Batching inference exception support - Optimized batching to fix a concurrent modification exception that was occurring with batch inference.
  • k8s cluster creation support upgrade - Updated Kubernetes cluster creation scripts for v1.17 support.
  • Nvidia devices visibility support - Added support for NVIDIA devices visibility.
  • Large image support - Added support for PIL.Image.MAX_IMAGE_PIXELS.
  • Custom HTTP status support - Added support to return custom http status from a model handler.
  • TS_CONFIG_FILE env var support - Added support for setting TS_CONFIG_FILE as env var.
  • Frontend build optimization - Optimized frontend to reduce build times by 3.7x.
  • Warmup in benchmark - Added support for warmup in benchmark scripts.

Platform Support

Ubuntu 16.04, Ubuntu 18.04, MacOS 10.14+, Windows 10 Pro, Windows Server 2019, Windows subsystem for Linux (Windows Server 2019, WSLv1, Ubuntu 18.0.4)

GPU Support

Torch 1.10+ Cuda 10.2, 11.3
Torch 1.9.0 + Cuda 11.1
Torch 1.8.1 + Cuda 9.2

TorchServe v0.4.2 Release Notes

02 Aug 21:31
Compare
Choose a tag to compare

TorchServe v0.4.1 Release Notes

22 Jul 17:38
Compare
Choose a tag to compare

This is the release of TorchServe v0.4.1.

New Features

  • PyTorch 1.9.0 support - TorchServe is now certified working with torch 1.9.0 torchvision 0.10.0, torchtext 0.10.0 and torchaudio 0.9.0
  • Model configuration support - Added support for model performance tuning on SageMaker via model configuration in config.properties.
  • Serialize config snapshots to DynamoDB - Added support for serializing config snapshots to DDB.
  • Prometheus metrics plugin support - Added support for Prometheus metrics plugin.
  • Kubeflow Pipelines support - Added support for Kubeflow pipelines and Google Vertex AI Manages pipelines, see examples here
  • KFServing docker support - Added production docker for KFServing.
  • Python 3.9 support - TorchServe is now certified working with Python 3.9.

Improvements

  • HF BERT models multiple GPU support - Added multi-gpu support for HuggingFace BERT models.
  • Error log for customer python package installation - Added support to log error of customer python package installation.
  • Workflow documentation optimization - Optimized workflow documentation.

Tooling improvements

  • Mar file automation integration - Integrated mar file generation automation into pytest and postman test.
  • Benchmark automation for AWS neuron support - Added support for AWS neuron benchmark automation.
  • Staging binary build support - Added support for staging binary build.

Platform Support

Ubuntu 16.04, Ubuntu 18.04, MacOS 10.14+, Windows 10 Pro, Windows Server 2019, Windows subsystem for Linux (Windows Server 2019, WSLv1, Ubuntu 18.0.4)

GPU Support

Torch 1.9.0 + Cuda 10.2, 11.1
Torch 1.8.1 + Cuda 9.2, 10.1

TorchServe v0.4.0 Release Notes

22 May 00:09
d1be158
Compare
Choose a tag to compare

This is the release of TorchServe v0.4.0.

New Features

  • Workflow support - Added support for sequential and parallel ensemble models with Language Translation and Computer Vision classification examples.
  • S3 Model Store SSE support - Added support for S3 server side model encryption via KMS.
  • MMF-activity-recognition model example - Added example MMF-activity-recognition model
  • PyTorch 1.8.1 support - TorchServe is now certified working with torch 1.8.1, torchvision 0.9.1, torchtext 0.9.1, and torchaudio 0.8.1

Improvements

Community Contributions

Bug Fixes

Platform Support

Ubuntu 16.04, Ubuntu 18.04, MacOS 10.14+, Windows 10 Pro, Windows Server 2019, Windows subsystem for Linux (Windows Server 2019, WSLv1, Ubuntu 18.0.4)

GPU Support

Cuda 10.1, 10.2, 11.1

TorchServe v0.3.1 Release Notes (Beta)

15 Mar 22:31
Compare
Choose a tag to compare

Patch release. Fixes Model Archiver to Recursively copy all artifacts

  • Make --serialized-file an Optional Argument #994
  • Recursively copy all files during archive #814