Skip to content

Releases: explosion/curated-transformers

v2.0.1: Fix Python 3.12.3 compatibility

17 Apr 17:14
491b408
Compare
Choose a tag to compare

🔴 Bug fixes

  • Fix Python 3.12.3 activation lookup error (#375).

v1.3.2: Fix Python 3.12.3 compatibility

17 Apr 17:10
3e6180f
Compare
Choose a tag to compare

🔴 Bug fixes

  • Fix Python 3.12.3 activation lookup error (#377).

v2.0.0 (Superposition)

16 Apr 13:17
8debb21
Compare
Choose a tag to compare

✨ New features and improvements

  • Register models using catalogue to support external models in Auto{Decoder,Encoder,CausalLM} (#351, #352).
  • Add support for loading parameters in-place (#370).
  • Support for ELECTRA models (#358).
  • Add support for write/upload operations with HFHubRepository (#354).
  • Add support for converting Curated Transformer configs to HF-compatible configs (#333).

🔴 Bug fixes

  • Support PyTorch 2.2 (#360).

⚠️ Backwards incompatibilities

  • Support for TorchScript tracing is removed (#361).
  • The qkv_split argument is now mandatory for AttentionHeads, AttentionHeads.uniform, AttentionHeads.multi_query, and AttentionHeads.key_value_broadcast (#374).
  • All FromHFHub mixins are renamed to FromHF (#374).
  • FromHF.convert_hf_state_dict is removed in favor of FromHF.state_dict_from_hf (#374).

👥 Contributors

@danieldk, @honnibal, @ines, @KennethEnevoldsen, @shadeMe

v1.3.1 (Venusian 2)

12 Feb 18:51
b192987
Compare
Choose a tag to compare

🔴 Bug fixes

  • Ensure that parameters are leaf nodes when loading a model (#364).
  • Set the Torch upper bound to <2.1.0 (#363).

Note: we have set the Torch upper bound to <2.1.0 because later versions made some incompatible changes. Newer versions of Torch will be supported by Curated Transformers 2.0.0.

v1.3.0 (Venusian 1)

02 Oct 14:26
cd53833
Compare
Choose a tag to compare

✨ New features and improvements

  • Add support for model repositories other than Hugging Face Hub (#331).
  • Add support for fsspec filesystems as a repository type (#327, #331).
  • Add support for NVTX Ranges (#320).
  • Add a config property to models to query their configuration (#328).

🔴 Bug fixes

  • Fix a potential loading issue that may arise when a model's dtype is not set in the Hugging Face configuration (#330).

🏛️ Feature: Model repositories

The new (experimental) repository API adds support for loading models from repositories other than Hugging Face Hub. You can also easily add your own repository types by implementing the Repository interface. Using a repository is as easy as calling the new from_repo method that is provided by all models and tokenizers:

from curated_transformers.models import AutoDecoder

decoder = AutoDecoder.from_repo(MyRepository("mpt-7b-my-qa"))

Curated Transformers comes with two repository classes out-of-the-box:

👥 Contributors

@danieldk, @honnibal, @ines, @shadeMe

v1.2.0 (Hypertension)

12 Sep 11:46
Compare
Choose a tag to compare

✨ New features and improvements

  • Add support for Safetensor checkpoints (#310).
  • Add from_hf_hub_to_cache method to FromHFHub mixins. This method downloads a model from Hugging Face hub to the local cache without loading it (#303).

🔴 Bug fixes

  • MPT: Honor no_bias config option in layer norms (#321).
  • Fix a typing issue in MPTGenerator (#317).

👥 Contributors

@danieldk, @honnibal, @ines, @mayankjobanputra, @shadeMe

v1.1.0 (Tetrachromacy)

10 Aug 12:22
845decc
Compare
Choose a tag to compare

✨ New features and improvements

  • Add decoder and causal LM for MosaicML MPT (#294).
  • Add generator wrapper for MosaicML MPT (#296).

🔴 Bug fixes

  • Do not initialize embeddings when loading a model (#295).

👥 Contributors

@danieldk, @honnibal, @ines, @shadeMe

v1.0.0 (Beginner's Luck)

03 Aug 13:47
2ca4b0c
Compare
Choose a tag to compare

Three weeks on the heels of our tech preview we are excited to announce first stable release of Curated Transformers! 🎉 From this release onwards, we provide a stable API following semantic versioning guidelines. Of course, this release is also packed with new features.

✨ New features and improvements since version 0.9.0

👥 Contributors

@danieldk, @honnibal, @ines, @shadeMe, @svlandeg

v0.9.1: Falcon compatibility fix

14 Jul 07:48
Compare
Choose a tag to compare

🔴 Bug fixes

  • Support old and new Falcon model type and configuration (#243).

👥 Contributors

@danieldk

v0.9.0 (Robot Stop)

13 Jul 13:14
Compare
Choose a tag to compare

We are very happy to announce this major new release of Curated Transformers! 🎉

Curated Transformers started as a small transformer library for spaCy pipelines. Over the last two months we made it a pure PyTorch library that is completely independent of spaCy and Thinc. We also added support for popular LLM models, generation, 8-bit/4-bit quantization, and many other features:

  • Curated Transformers is now a pure PyTorch library.
  • Support for popular LLMs such as Falcon, LLaMA, and Dolly v2.
  • Greedy generation and generation with sampling.
  • 8-bit and 4-bit quantization of models through bitsandbytes.
  • Flash attention and other optimizations through PyTorch Scaled Dot Product Attention.
  • Efficient model loading without unneeded allocations and initialization through the Torch meta devices.
  • Support for modern tokenizer.json tokenizers.
  • Load models from Hugging Face Hub without requiring the transformers package.
  • Extensive API documentation and examples.

Curated Transformers can be used in spaCy using the spacy-curated-transformers package.

👥 Contributors

@danieldk, @honnibal, @ines, @shadeMe