All notable changes to this project will be documented in this file.
The format is based on Keep a Changelog, and this project adheres to Semantic Versioning.
- Makefile to upload package to Melior's Pypi server
- Drone config to upload package to Melior's Pypi server automatically.
- Setup to include upload, test, dev and extra requires.
- Pre-commit hooks added.
- Drone config to run just one instance.
- Even when
save_model_every_epoch
the las model was not saved.
os.mkdir
replaced toos.makedirs
to create all the intermediate directories
shutil.rmtree
error not handleddelete_worst_models
was somehow deleted during old merge conflict
Makefile
added withformatter
,linter
andtest
comands.requirements-dev.txt
andrequirementes.txt
with new dependecies.sentence-transformers
wrapper to easaly extract embeddings.- Drone corectly configured.
- Some simple tests.
- Renamed library to
melior-transformers
- Almost all files are formatted, linted and the imports are sorted.
metric_criteria
andsave_n_best_epochs
parameters added. These allow to save the best model.- Renamed library to
meliorTransformers
README
and versioning file update- Added
.drone.yml
placeholder
0.18.2 - 2020-01-15
- Added option to turn off model saving at the end of every epoch with
save_model_every_epoch
.
- Fixed bug with missing
tensorboard_folder
key in certain situations.
- Moved
args
items common to all classes to one place (config/global_args.py
) for maintainability.
0.18.1 - 2020-01-15
- Fixed bug with missing
regression
key when using MultiLabelClassification.
0.18.0 - 2020-01-15
- Sentence pair tasks are now supported.
- Regression tasks are now supported.
use_cached_eval_features
toargs
. Evaluation during training will now use cached features by default. Set toFalse
if features should be reprocessed.
- Checkpoints saved at the end of an epoch now follow the `checkpoint-{global_step}-epoch-{epoch_number} format.
0.17.1 - 2020-01-14
- Fixed
wandb_kwargs
key missing inargs
bug.
0.17.0 - 2020-01-14
- Added new model XLM-RoBERTa. Can now be used with
ClassificationModel
andNERModel
.
0.16.6 - 2020-01-13
- Added evaluation scores from end-of-epoch evaluation to
training_progress_scores.csv
.
- Typos in
README.md
.
0.16.5 - 2020-01-09
- Reverted missed logging commands to print statements.
0.16.4 - 2020-01-09
- Removed logging import.
0.16.3 - 2020-01-09
- Reverted to using print instead of logging as logging seems to be causing issues.
0.16.2 - 2020-01-08
- Changed print statements to logging.
0.16.1 - 2020-01-07
- Added
wandb_kwargs
toargs
which can be used to specify keyword arguments towandb.init()
method.
0.16.0 - 2020-01-07
- Added support for training visualization using the W&B framework.
- Added
save_eval_checkpoints
attribute toargs
which controls whether or not a model checkpoint will be saved with every evaluation.
0.15.7 - 2020-01-05
- Added
**kwargs
for different accuracy measures during multilabel training.
0.15.6 - 2020-01-05
- Added
train_loss
totraining_progress_scores.csv
(which contains the evaluation results of all checkpoints) in the output directory.
0.15.5 - 2020-01-05
- Using
evaluate_during_training
now generatestraining_progress_scores.csv
(which contains the evaluation results of all checkpoints) in the output directory.
sliding_window
removed in order to have direct compatibility with Transformers models.metric_criteria
andsave_n_best_epochs
parameters added. These allow to save the best model.
0.15.4 - 2019-12-31
- Fixed bug in
QuestonAnsweringModel
when usingevaluate_during_training
.
0.15.3 - 2019-12-31
- Fixed bug in MultiLabelClassificationModel due to
tensorboard_dir
being missing in parameter dictionary.
- Renamed
tensorboard_folder
totensorboard_dir
for consistency.
0.15.2 - 2019-12-28
- Added
tensorboard_folder
to parameter dictionary which can be used to specify the directory in which the tensorboard files will be stored.
0.15.1 - 2019-12-27
- Added
**kwargs
to support different accuracy measures at training time.
0.15.0 - 2019-12-24
- Added
evaluate_during_training_steps
parameter that specifies when evaluation should be performed during training.
- A model checkpoint will be created for each evaluation during training and the evaluation results will be saved along with the model.
0.14.0 - 2019-12-24
- Added option to specify a GPU to be used when multiple GPUs are available. E.g.:
cuda_device=1
- Added
do_lower_case
argument for uncased models.
- Fixed possible bug with output directory not being created before evaluation is run when using
evaluate_during_training
.
0.13.4 - 2019-12-21
- Fixed bug with when using
eval_during_training
with QuestionAnswering model.
0.13.3 - 2019-12-21
- Fixed bug with loading Multilabel classification models.
- Fixed formatting in README.md.
0.13.2 - 2019-12-20
- Fixed formatting in README.md.
0.13.1 - 2019-12-20
- Bug in Multilabel Classification due to missing entries in default args dict.
0.13.0 - 2019-12-19
- Sliding window feature for Binary and Multiclass Classification tasks.
0.12.0 - 2019-12-19
- Minimal examples have been added to the
examples
directory as Python scripts.
- Readme updated to include the addition of examples.
0.11.2 - 2019-12-18
- Evaluation during training fixed for multilabel classification.
0.11.1 - 2019-12-18
- Broken multiprocessing support for NER tasks fixed.
0.11.0 - 2019-12-15
- CamemBERT can now be used with NERModel.
- Readme changed to include CamemBERT for NER.
0.10.8 - 2019-12-15
- DistilBERT can now be used with NERModel.
- Readme changed to include DistilBERT for NER.
0.10.7 - 2019-12-15
- This CHANGELOG file to hopefully serve as an evolving example of a standardized open source project CHANGELOG.