Skip to content

Releases: microsoft/onnxruntime-extensions

v0.11.0

01 Jun 00:00
Compare
Choose a tag to compare

What's changed

  • Created Java packaging pipeline and published to Maven repository.
  • Added support for conversion of Huggingface FastTokenizer into ONNX custom operator.
  • Unified the SentencePiece tokenizer with other Byte Pair Encoding (BPE) based tokenizers.
  • Fixed Whisper large model pre-processing bug.
  • Enabled eager execution for custom operator and refactored the header file structure.

Contributions

Contributors to ONNX Runtime Extensions include members across teams at Microsoft, along with our community members: @sayanshaw24 @wenbingl @skottmckay @natke @hariharans29 @jslhcl @snnn @kazssym @YUNQIUGUO @souptc @yihonglyu

v0.10.1

25 Apr 17:53
Compare
Choose a tag to compare

Support Python 3.12 version in PYPI release package.

v0.10.0

06 Feb 19:16
b9b3ebe
Compare
Choose a tag to compare

What's changed

  • Modified gen_processing_model tokenizer model to output int64, unifying output datatype of all tokenizers.
  • Implemented support for post-processing of YOLO v8 within the Python extensions package.
  • Introduced 'fairseq' flag to enhance compatibility with certain Hugging Face tokenizers.
  • Incorporated 'added_token' attribute into the BPE tokenizer to improve CodeGen tokenizer functionality.
  • Enhanced the SentencePiece tokenizer by integrating token indices into the output.
  • Added support for the custom operator implemented with CUDA kernels, including two example operators.
  • Added more tests on the Hugging Face tokenizer and fixed identified bugs.

Contributions

Contributors to ONNX Runtime Extensions include members across teams at Microsoft, along with our community members: @wenbingl @sayanshaw24 @skottmckay @mszhanyi @edgchen1 @YUNQIUGUO @RandySheriffH @samwebster @hyoshioka0128 @baijumeswani @dizcza @Craigacp @jslhcl

v0.9.0

21 Sep 18:28
Compare
Choose a tag to compare

What's Changed

  • New Python API gen_processing_models to export ONNX data processing model from Huggingface Tokenizers such as LLaMA , CLIP, XLM-Roberta, Falcon, BERT, etc.
  • New TrieTokenizer operator for RWKV-like LLM models, and other tokenizer operator enhancements.
  • New operators for Azure EP compatibility: AzureAudioToText, AzureTextToText, AzureTritonInvoker for Python and NuGet packages.
  • Processing operators have been migrated to the new Lite Custom Op API
  • New operator of string strip
  • Using the latest Ort header instead of minimum compatible headers
  • Support offset mapping in most tokenizers like BERT, CLIP, Roberta and etc.
  • Remove the deprecating std::codecvt_utf8 from code base
  • Document are uploaded to https://onnxruntime.ai/docs/extensions/

Contributions
Contributors to ONNX Runtime Extensions include members across teams at Microsoft, along with our community members: @aidanryan-msft @RandySheriffH @edgchen1 @kunal-vaishnavi @sayanshaw24 @skottmckay @snnn @VishalX @wenbingl @wejoncy

v0.8.0

26 May 23:44
Compare
Choose a tag to compare

New Changes:

  1. NuGet package for the .NET platform. This package offers comprehensive platform support, including Windows, Linux, MacOS, Android, and iOS. Both x64 and arm64 architectures are supported, where applicable.
  2. Support for pre-processing and post-processing of the Whisper model, inclusive of Audio and Tokenizer decoding operators.
  3. Extends support for pre-processing and post-processing of object-detection models, including a new DrawBoundingBoxes operator. Pre/post processing tools can add non-max-suppression to the model to select the best bounding boxes, and scale those to the original image. See the end-to-end example in tutorials/yolo_e2e.py.
  4. Introduces the Audio Domain, complemented with AudioCodec and optimized STFT Operators, enhancing audio processing capabilities.
  5. Enabled optional input/output support for some operators such as GPT2Tokenizer, ClipTokenizer, and RobertaTokenizer.
  6. Refined the implementation of offset mapping for BBPE-style tokenizers for more operators and efficiency improvement.
  7. Other bug and security fixes.

Contributions

Contributors to ONNX Runtime Extensions include members across teams at Microsoft, along with our community members: @edgchen1 @kunal-vaishnavi @sayanshaw24 @skottmckay @snnn @VishalX @wenbingl @wejoncy

Full Changelog: v0.7.0...v0.8.0

v0.7.0

02 Mar 00:32
Compare
Choose a tag to compare

General
1. New custom operators: RobertaTokenizer, ClipTokenizer, EncodeImage, DecodeImage
2. ORT custom operator C++ stub generation tool
3. Operator implementation and documentation improved.
4. Python (3.7 - 3.10) and ORT (1.10 above) compatible.

Mobile
1. Android package: Maven
2. iOS package: CocoaPods
3. PrePostProcessor tool for mobile model
4. Super-resolution model pre- / post- processing end-to-end examples

Contributors to this release include members across teams at Microsoft, along with our community members: @edgchen1 @skottmckay @shaahji @sayanshaw24 @snnn @wenbingl @natke @YUNQIUGUO @guschmue @JamieMagee @adrianlizarraga @wejoncy @matheusgomes28

v0.5.0

04 Aug 23:04
134f882
Compare
Choose a tag to compare

This is a C++ source code package only release. Python and other packages will be expected in the next release.

What's Changed

  1. Support OpenCV core and imgproc modules in Custom Op implementation.
  2. Code security compliance fixings.
  3. Some other improvements.

Thanks for the Contributors from: @joburkho @shaahji @TruscaPetre @Sanster @natke @hombreola @snnn @leqiao-1 @wenbingl

v0.4.2

07 Oct 18:27
Compare
Choose a tag to compare

Fix the issue on StringRegexSplitWithOffsets operator.

v0.4.1

04 Oct 23:46
Compare
Choose a tag to compare

Correct memory indexing issue.

v0.4.0

28 Sep 08:48
Compare
Choose a tag to compare

This release improved the process and the features required for ONNXRuntime Mobile and ONNXRuntime Web,
Specifically

  1. Support Android/Web-Assembly build
  2. Support the custom build for the selected operators
  3. Support non-exception C++ build
  4. Add mobile-optimized operators for string manipulation and tokenization
  5. The GPT-2 beam search tool
  6. Some bug fixings