Skip to content

Releases: nebuly-ai/nebuly

ChatLLaMA 0.0.4

27 Mar 13:47
Compare
Choose a tag to compare
ChatLLaMA 0.0.4 Pre-release
Pre-release

Release Notes - ChatLLaMA 0.0.4

Major release of ChatLLaMA adding support to efficient training using LoRA.

New Features

  • HF-based models can now be trained on both actor and RLHF using LoRA. This will help users in reducing the memory needed for training.

ChatLLaMA 0.0.3

27 Mar 08:16
9efe338
Compare
Choose a tag to compare
ChatLLaMA 0.0.3 Pre-release
Pre-release

Release Notes - ChatLLaMA 0.0.3

Major release of ChatLLaMA solving multiple bugs and expanding the support to distrubuted training.

New Features

  • Training produces a log file containing also the training stats
  • Add Template during the dataset creation
  • Changed default training parameters to InstructGPT paper ones
  • Implemented Cosine Scheduler for LR
  • Add Conversation Logs during RLHF
  • Improved management of checkpoints
  • Add support for Accelerate backend for distributed training
  • Defer import of LLaMA modules, avoiding the LLaMA dependecy while another model is used

Bug Fix

  • Fix bug causing crashes during training due to long sequences
  • Fix multiple bugs on deepspeed MultiGPU training
  • Fix bugs on RLHF training
  • Fix bug with different versions of LangChain during synthetic data generation

New Contributors

v0.9.0

21 Mar 13:07
Compare
Choose a tag to compare
v0.9.0 Pre-release
Pre-release

Nebullvm 0.9.0 Release Notes

Major release for Nebullvm, adding support to diffusion model optimization.

New features

  • Add support for diffusers UNet

Bug fixed

  • Fix CI pipelines triggers

Speedster 0.3.0 Release Notes

Major release adding support to the diffusers library.

New features

  • Add support to diffusers library. Speedster can now optimize diffusion models with a single line of code.
  • Update readme for HF models.

Bug fixed

  • Fix import error in Google Colab
  • Fix few typos in the docs and update the benchmarks

New Contributors

v0.8.1

15 Feb 13:18
Compare
Choose a tag to compare
v0.8.1 Pre-release
Pre-release

nebullvm 0.8.1 Release Notes

This is a minor release fixing multiple bugs.

New features

  • Changed the Auto-Installer API
  • Added support to onnxruntime TensorrtExecutionProvider

Bug fixed

  • Fixed bug in torchscript casting integers to fp16.
  • Optimized the memory usage for inference learners
  • Now the tensorrt workspace size is dynamically computed according to the free memory available on the gpu device
  • Fixed a bug in openvino when using static quantization

speedster 0.2.1 Release Notes

This is a minor release fixing multiple bugs.

New Features

  • Added support for device selection when working multi-gpu environment
  • Added support to input data with inconsistent batch size
  • Implemented benchmark functions also for TensorFlow and ONNX

Bug Fixed

  • Optimized the gpu memory usage during speedster optimization

Contributors

v0.8.0

23 Jan 23:20
20ef9d7
Compare
Choose a tag to compare
v0.8.0 Pre-release
Pre-release

nebullvm 0.8.0 Release Notes

This is a major release fixing multiple bugs and implementing two new functions for loading and saving the models.

New Features

  • Implements two new functions for loading and saving inference learners.

Bug fixed

  • Fixed bug for ONNXRuntime models being loaded on the wrong device after the optimization.
  • Fixes TensorRT behaviour when using dynamic shape
  • Improves significantly the performance of TensorRT with ONNX interface
  • Limits the gpu memory used by tensorflow, to avoid memory issues during tensorflow models optimization
  • Fixes some issues with unit tests, and adds additional controls to ensure that an optimized model with dynamic shape works properly.
  • removes setuptools from tensorrt installation, it's not needed anymore by the newer version and it was causing issues with the azure pipelines.

speedster 0.2.0 Release Notes

Major release for Speedster adding the load_model and save_model functions.

New Features

  • Improved the logs
  • Save and load model functions can be imported directly from speedster
  • Brand new documentation moved directly in the GitHub repository and built using mkdocs

Bug Fixed

  • Fixed version compatibility issue on Colab
  • adds missing port forwarding in the docker run command inside the notebooks readme.

Contributors

v0.7.3

12 Jan 23:51
Compare
Choose a tag to compare
v0.7.3 Pre-release
Pre-release

nebullvm 0.7.3 Release Notes

This is a minor release modifying the metric_drop_ths behaviour.

New Features

  • metric_drop_ths by default set to 0.001
  • Half precision is activated for each value of metric_drop_ths>0
  • Int8 quantization is activated just for values of metric_drop_ths>0.01

Speedster 0.1.2 Release Notes

This is a minor release editing the default value for metric_drop_ths and modifying the summarisation table at the end of the optimization.

New Features

  • The summary at the end of the optimization now shows in a clear way the improvements.
  • Warning with tips on how to improve the optimization result is now shown also for speedups lower than 2x.

v0.7.2

12 Jan 08:44
Compare
Choose a tag to compare
v0.7.2 Pre-release
Pre-release

nebullvm 0.7.2 Release Notes

This is a minor release editing the logs style.

New Features

  • Warning logs are now less intrusive.

Speedster 0.1.1 Release Notes

This is a minor release adding tips and improving the logs.

New Features

  • Suggestions are provided to the user on how to increase model performance when optimization does not provide the required speedup.
  • Improved the quality of the logs.

v0.7.1

10 Jan 18:27
7e207a6
Compare
Choose a tag to compare
v0.7.1 Pre-release
Pre-release

nebullvm 0.7.1 Release Notes

This is a minor release fixing few bugs for supporting YOLOv8 models.

Bug fixed

  • Fixed bugs with half-precision in PyTorch for supporting YOLOv8 models.

v0.7.0

09 Jan 17:02
9362fc7
Compare
Choose a tag to compare
v0.7.0 Pre-release
Pre-release

nebullvm 0.7.0 Release Notes

This is a major release simplifying the backends installation and improving the UX.

New Features

  • The auto-installer interface has been simplified and clarified.

Bug fixed

  • Fixed problem with shell not found when installing TensorRT on specific Linux systems.

speedster 0.1.0 Release Notes

Major release for Speedster, now supporting TF backend for HuggingFace transformers.

New Features

  • Add support for HuggingFace models having a TensorFlow backends.
  • Improved the logs style.
  • Optimization results are now summarised into a nice and informative table.

Contributors

v0.6.0

01 Jan 19:15
5fb48f6
Compare
Choose a tag to compare
v0.6.0 Pre-release
Pre-release

nebullvm 0.6.0 Release Notes

This release of Nebullvm modifies the structure of the library. Nebullvm library is transformed into a framework which can be used for building Apps for AI optimization. The end-to-end optimization tool is moved into speedster, an App built using nebullvm as underlying framework.

New Features

  • The API for model optimization has been moved in app/accelerate/speedster.
  • The optimize_model function in nebullvm has been deprecated. It will be removed in the next major release.
  • Added support for windows platforms.

speedster 0.0.1 Release Notes

Speedster is the new library replacing the previous nebullvm API optimizing DL models in a single line of code. Speedster keeps the same interface as the previous nebullvm API.

New Features

  • Model optimization API moved into speedster. Now it is possible to import the optimize_model function running from speedster import optimize_model.

Contributors