-
Notifications
You must be signed in to change notification settings - Fork 0
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Update with forked from #1
Update with forked from #1
Commits on Feb 23, 2023
-
Fix 2 quicktour file doctest (#21742)
* Update expect output values - as Hub repo. files are updated * Update expect output values - as librosa is from 0.9.2 to 0.10.0 on CI docker * fix * update one more --------- Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 36a6a1a - Browse repository at this point
Copy the full SHA 36a6a1aView commit details -
[
GPTNeo
] Fix gradient checkpointing bug (#21733)* fix bug * forward contrib credits from discussions * change logic --------- Co-authored-by: edbeeching <edbeeching@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 78a93d1 - Browse repository at this point
Copy the full SHA 78a93d1View commit details -
Configuration menu - View commit details
-
Copy full SHA for 1d4b797 - Browse repository at this point
Copy the full SHA 1d4b797View commit details -
Configuration menu - View commit details
-
Copy full SHA for aa3787c - Browse repository at this point
Copy the full SHA aa3787cView commit details -
Added Type Hints for modeling_tf_encoder_decoder.py (#21673)
* Ran Black formatting * Added imports and reformatted * Update src/transformers/models/encoder_decoder/modeling_tf_encoder_decoder.py --------- Co-authored-by: Matt <Rocketknight1@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 0ffa22f - Browse repository at this point
Copy the full SHA 0ffa22fView commit details -
Auto api Value Error addition to Troubleshoot (#21708)
* troubleshooting guide: added an error description for missing auto-mapping * minor polishing * changed the example * Apply suggestions from code review Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/en/troubleshooting.mdx Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com> --------- Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com> Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 04d90ac - Browse repository at this point
Copy the full SHA 04d90acView commit details -
[deepspeed tests] fix issues introduced by #21700 (#21769)
* [deepspeed tests] fix issues introduced by #21700 * fix * fix
Configuration menu - View commit details
-
Copy full SHA for 6330626 - Browse repository at this point
Copy the full SHA 6330626View commit details
Commits on Feb 24, 2023
-
* Removed useless check for backend * fix style check for graphormer * Reverted change and corrected requires_backend for cython * code qual
Configuration menu - View commit details
-
Copy full SHA for 4446b6b - Browse repository at this point
Copy the full SHA 4446b6bView commit details -
fix: Change is_last chunk calc and add conditional break in chunk_iter (
#21612) * fix: Change is_last chunk calc and add conditional break * format fix * account for 0 and full stride_rights, add comment * add new test * make style * update slow whisper asr test timestamps * use nested_simplify on output and round timestamp to hundreths place
Connor Henderson committedFeb 24, 2023 Configuration menu - View commit details
-
Copy full SHA for 279008a - Browse repository at this point
Copy the full SHA 279008aView commit details -
[Flax] adding support for batch norm layers (#21581)
* [flax] adding support for batch norm layers * fixing bugs related to pt+flax integration * cleanup, batchnorm support in sharded pt to flax * support for batchnorm tests in pt+flax integration * simplifying checking batch norm layer
Configuration menu - View commit details
-
Copy full SHA for f7ca656 - Browse repository at this point
Copy the full SHA f7ca656View commit details -
[Examples] Generalise run audio classification for log-mel models (#2…
…1756) * [Examples] Generalise run audio classification for log-mel models * batch feature extractor * make style
Configuration menu - View commit details
-
Copy full SHA for 1348924 - Browse repository at this point
Copy the full SHA 1348924View commit details -
Different behavior in DistilBERT when using "inputs_embeds" (#21752)
* Different behavior in DistilBERT when using "inputs_embeds" Fixes #21089 * fix failing test
Configuration menu - View commit details
-
Copy full SHA for 14f3320 - Browse repository at this point
Copy the full SHA 14f3320View commit details -
Configuration menu - View commit details
-
Copy full SHA for 75bd49f - Browse repository at this point
Copy the full SHA 75bd49fView commit details -
[Whisper] Add SpecAugment (#21298)
* Return and rescale attention_mask * Add SpecAugment to Whisper modeling * Fix test * Update docstring * Add SpecAug related parameters to model config * Add the _mask_input_features function to doc * Fix quality * Apply suggestions from code review Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com> * Remove dev comments * Add test * Resolve conflict * feat: mask {feature, time} prob fast tests * Apply suggestions from code review Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com> --------- Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com> Co-authored-by: sanchit-gandhi <sanchit@huggingface.co> Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for c8545d2 - Browse repository at this point
Copy the full SHA c8545d2View commit details -
* fix history * input_features instead of input ids for TFWhisport doctest * use translate intead of transcribe
Configuration menu - View commit details
-
Copy full SHA for 087436c - Browse repository at this point
Copy the full SHA 087436cView commit details -
Generate - update cookie cutters to not initialize cache with trainin…
…g and gradient checkpointing (#21759)
Configuration menu - View commit details
-
Copy full SHA for 440f397 - Browse repository at this point
Copy the full SHA 440f397View commit details -
[time series] updated expected values for integration test. (#21762)
* updated expected * prediction_length fix * prediction_length default value * default prediction_length 24 * revert back prediction_length default * move prediction_length test
Configuration menu - View commit details
-
Copy full SHA for ba0e370 - Browse repository at this point
Copy the full SHA ba0e370View commit details -
[GPT2, ProphetNet] Fix gradient checkpointing bug (#21772)
* fix gradient checkpointing bug * fix gradient checkpointing bug * ran make fix-copies * fixed bug * fixed bug
Configuration menu - View commit details
-
Copy full SHA for 59c1d5b - Browse repository at this point
Copy the full SHA 59c1d5bView commit details -
Configuration menu - View commit details
-
Copy full SHA for 3dae0d7 - Browse repository at this point
Copy the full SHA 3dae0d7View commit details
Commits on Feb 25, 2023
-
Fix resume_from_checkpoint for deepspeed (#21735)
* Fix resume_from_checkpoint for deepspeed Fix resume_from_checkpoint for deepspeed, by ensuring that the deepspeed engine is the one to load the checkpoint. * Empty commit to trigger CI * Removed deepspeed skipping Removed deepspeed skipping inside the _load_from_checkpoint function, as it is obsolete * another adjustment * Trigger CI * trigger circleci * style --------- Co-authored-by: ydshieh <ydshieh@users.noreply.github.com> Co-authored-by: Stas Bekman <stas@stason.org>
Configuration menu - View commit details
-
Copy full SHA for 9ddf4f4 - Browse repository at this point
Copy the full SHA 9ddf4f4View commit details
Commits on Feb 27, 2023
-
[examples/summarization] deal with
max_length
andnum_beams
(#21740)* Override the decoding parameters of Seq2SeqTrainer * Fix quality * Fix max_length parameter * Fix quality * Remove redundant parameter max_length * Separate the preprocess of train and validation to use different max_target_length
Configuration menu - View commit details
-
Copy full SHA for 3c0ce60 - Browse repository at this point
Copy the full SHA 3c0ce60View commit details -
Fix type in gpt2 config docstring (#21782)
Fix docstring gpt2 config
Configuration menu - View commit details
-
Copy full SHA for a369836 - Browse repository at this point
Copy the full SHA a369836View commit details -
Fix en documentation typos (#21799)
* fix wrong url * typos in english documentation
Configuration menu - View commit details
-
Copy full SHA for ba2a5f1 - Browse repository at this point
Copy the full SHA ba2a5f1View commit details -
[FX tracer] Make
concrete_args
from outside available (#21775)make concrete_args from outside available
Configuration menu - View commit details
-
Copy full SHA for 2ea1ef9 - Browse repository at this point
Copy the full SHA 2ea1ef9View commit details -
[Pipeline] Add zero shot audio classificatoin pipeline (#21600)
* add pipeline * update init * add zero shot to init * update inits and correct checkpoints * update base to support input features * add tests * Update src/transformers/pipelines/zero_shot_audio_classification.py Co-authored-by: Younes Belkada <49240599+younesbelkada@users.noreply.github.com> * Update src/transformers/pipelines/zero_shot_audio_classification.py Co-authored-by: Younes Belkada <49240599+younesbelkada@users.noreply.github.com> * update pieline code * use tiny checkpoint * nits and expected value with tiny model * style * last nit on tests values * fix styling * fix collate fn that was casting t float * update --------- Co-authored-by: Younes Belkada <49240599+younesbelkada@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for cc44e72 - Browse repository at this point
Copy the full SHA cc44e72View commit details -
[torch] remove deprecated uint8 in favor of bool (#21384)
* uint8 -> bool * fix copies * style * update test modeling commen when checking attention buffers * style * use logical not on random mask instead of subtraction with 1 * remove torch uint8 * quality * remove modified modeling utils * Update based on review Co-authored-by: sgugger <sylvain.gugger@gmail.com> --------- Co-authored-by: sgugger <sylvain.gugger@gmail.com>
Configuration menu - View commit details
-
Copy full SHA for c51dc4f - Browse repository at this point
Copy the full SHA c51dc4fView commit details -
[
tests
] addaccelerate
marker (#21743)* add `accelerate` marker * add to docs * Update docs/source/en/testing.mdx
Configuration menu - View commit details
-
Copy full SHA for 831f314 - Browse repository at this point
Copy the full SHA 831f314View commit details -
Fix PyTorch Perceiver
PerceiverFourierPositionEncoding
with fp16 (#……21787) * fix perceiver fp16 * hopefully fix tests
Configuration menu - View commit details
-
Copy full SHA for ebf84f0 - Browse repository at this point
Copy the full SHA ebf84f0View commit details -
Fix nn.init.trunc_normal_ call on torch.float16 data (#21789)
fix nn.init.trunc_normal_ call on half data
Configuration menu - View commit details
-
Copy full SHA for 0c7f93f - Browse repository at this point
Copy the full SHA 0c7f93fView commit details -
Fix gradient checkpointing bug in gptneox (#21815)
* Fix gradient checkpointing bug in gptneox * Remove use_cache block
Configuration menu - View commit details
-
Copy full SHA for 7811bf7 - Browse repository at this point
Copy the full SHA 7811bf7View commit details -
Configuration menu - View commit details
-
Copy full SHA for 92dfceb - Browse repository at this point
Copy the full SHA 92dfcebView commit details -
Fix quality with
ruff==0.0.253
(#21828)fix quality with ruff 0.0.253 Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for f95f60c - Browse repository at this point
Copy the full SHA f95f60cView commit details -
introduce
logger.warning_once
and use it for grad checkpointing code (#21804) * logger.warning_once * style
Configuration menu - View commit details
-
Copy full SHA for c7f3abc - Browse repository at this point
Copy the full SHA c7f3abcView commit details
Commits on Feb 28, 2023
-
Rename
MobileViTModelTest
toTFMobileViTModelTest
(#21825)Let's give TF a bit more love ❤️ 🙏 Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for a9dd124 - Browse repository at this point
Copy the full SHA a9dd124View commit details -
Fix gradient checkpointing bug BioGpt (#21844)
Co-authored-by: saswatmeher <saswatmeher@cse.iitb.ac.in>
Configuration menu - View commit details
-
Copy full SHA for 50644cf - Browse repository at this point
Copy the full SHA 50644cfView commit details -
Configuration menu - View commit details
-
Copy full SHA for 50db741 - Browse repository at this point
Copy the full SHA 50db741View commit details -
Fix gradient checkpointing bug in git (#21818)
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for e07a3d9 - Browse repository at this point
Copy the full SHA e07a3d9View commit details -
Fix gradient checkpointing imagegpt (#21816)
* Fix gradient checkpointing bug in gptneox * Fix gradient checkpointing bug in modeling_imagegpt.py * Revert gpt neox changes --------- Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 4fe744f - Browse repository at this point
Copy the full SHA 4fe744fView commit details -
Fix tf random token masking probability in data collator (#21834)
* fix tf random mask tokens probability * fix tf random mask tokens probability in collator for langauge modelling
Configuration menu - View commit details
-
Copy full SHA for 2d506ea - Browse repository at this point
Copy the full SHA 2d506eaView commit details -
[
T5
] Fix torchquant issue (#21843)* fix torchquant issue * add tests
Configuration menu - View commit details
-
Copy full SHA for ae9230a - Browse repository at this point
Copy the full SHA ae9230aView commit details -
[
Blip2
] AddBlip2Model
(#21817)* add v1 * add `Blip2Model` - add relevant functions - add tests - add on automapping * fix docs * fix doctest
Configuration menu - View commit details
-
Copy full SHA for b8de7e4 - Browse repository at this point
Copy the full SHA b8de7e4View commit details -
Fix the issue of blip model returning loss even when the label is not…
… provided. (#21811) * Fix the issue of blip model returning loss even when the label is not provoided * Fix ruff failure * Incorporate PR feedbacks * Incorporate PR feedbacks * Incorporate PR feedbacks * Incorporate PR feedbacks
Configuration menu - View commit details
-
Copy full SHA for eec7604 - Browse repository at this point
Copy the full SHA eec7604View commit details -
[GPTJ] Fix gradient checkpointing bug (#21794)
* If applied, this commit fixes generate bug in gptj * Remove extra same code block * formatting and test fix * Conflict fix and declaration error fix --------- Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 31fa2b6 - Browse repository at this point
Copy the full SHA 31fa2b6View commit details -
Add: task guide for zero shot object detection (#21829)
* zero shot object detection part 1 * added batch prediction section * added image guided object detection section * make style * added the task guide to the TOC * minor polishing * Apply suggestions from code review Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com> Co-authored-by: Alara Dirik <8944735+alaradirik@users.noreply.github.com> * added embedded owlvit demo * Apply suggestions from code review Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * minor fix * make style --------- Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com> Co-authored-by: Alara Dirik <8944735+alaradirik@users.noreply.github.com> Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 6ca8445 - Browse repository at this point
Copy the full SHA 6ca8445View commit details -
Make Slack CI reporting stronger (#21823)
* Use token * Avoid failure * better error * Fix * fix style --------- Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for aab895c - Browse repository at this point
Copy the full SHA aab895cView commit details -
[
Blip2
] Fix Blip-2 multi gpu (#21707)* fix blip multi gpu * fix * final changes * adapt suggestions * fix failing slow test * forward contrib credits from testing and suggestions * reformat --------- Co-authored-by: akkikiki <akkikiki@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 7f4f8b9 - Browse repository at this point
Copy the full SHA 7f4f8b9View commit details -
Add loss for BridgeTowerForMaskedLM and BridgeTowerForImageAndTextRet…
…rieval (#21684) * Add loss for BridgeTowerForMaskedLM and BridgeTowerForImageAndTextRetrieval * minor fix return_dict * implement test for loss computation --------- Co-authored-by: Tiep Le <97980157+tileintel@users.noreply.github.com> Co-authored-by: Tiep Le <tiep.le@intel.com>
Configuration menu - View commit details
-
Copy full SHA for 4cb5ffa - Browse repository at this point
Copy the full SHA 4cb5ffaView commit details -
🔥Rework pipeline testing by removing
PipelineTestCaseMeta
🚀 (#21516)* Add PipelineTesterMixin * remove class PipelineTestCaseMeta * move validate_test_components * Add for ViT * Add to SPECIAL_MODULE_TO_TEST_MAP * style and quality * Add feature-extraction * update * raise instead of skip * add tiny_model_summary.json * more explicit * skip tasks not in mapping * add availability check * Add Copyright * A way to diable irrelevant tests * update with main * remove disable_irrelevant_tests * skip tests * better skip message * better skip message * Add all pipeline task tests * revert * Import PipelineTesterMixin * subclass test classes with PipelineTesterMixin * Add pipieline_model_mapping * Fix import after adding pipieline_model_mapping * Fix style and quality after adding pipieline_model_mapping * Fix one more import after adding pipieline_model_mapping * Fix style and quality after adding pipieline_model_mapping * Fix test issues * Fix import requirements * Fix mapping for MobileViTModelTest * Update * Better skip message * pipieline_model_mapping could not be None * Remove some PipelineTesterMixin * Fix typo * revert tests_fetcher.py * update * rename * revert * Remove PipelineTestCaseMeta from ZeroShotAudioClassificationPipelineTests * style and quality * test fetcher for all pipeline/model tests --------- Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 871c31a - Browse repository at this point
Copy the full SHA 871c31aView commit details -
Improve TF weight loading, especially PT crossloading (#21792)
* First commit for the improved PT-TF weight loading * Remove workarounds from TFEncoderDecoder tests * Allow a custom weight renaming function in from_pretrained and use that to clean up EncoderDecoder * make fixup * First attempt at visionencoderdecoder * Disable tensorfloat32 in tests to get consistent outputs * Quick fix to tf_vision_encoder_decoder tests * make fixup * Update Blenderbot tests * Remove unused arg in modeling_tf_opt * load_tf_sharded_weights had strict=True! This meant transfer learning was impossible, so I'm setting it to False. * Support prefixes when loading sharded TF checkpoints * make fixup * Add test to load sharded models with a weight prefix * Fix sharded weight loading test * Add a test for transfer from a sharded checkpoint * make fixup * Add test to check that crossloading from PT with a prefix works * Refactor from_pretrained in the encoderdecoder classes * Refactor from_pretrained in the encoderdecoder classes * missmatched -> mismatched * Explicitly check for None * No comments showing my very impressive and attractive knowledge of Py3.9+ * Disable TF32 across all TF tests
Configuration menu - View commit details
-
Copy full SHA for acfb714 - Browse repository at this point
Copy the full SHA acfb714View commit details -
Fix flaky test for log level (#21776)
* Fix flaky test for log level * Fix other flaky test
Configuration menu - View commit details
-
Copy full SHA for b29e2dc - Browse repository at this point
Copy the full SHA b29e2dcView commit details
Commits on Mar 1, 2023
-
prepare for "__floordiv__ is deprecated and its behavior will change …
…in a future version of pytorch" (#20211) * rounding_mode = "floor" instead of // to prevent behavioral change * add other TODO * use `torch_int_div` from pytrch_utils * same for tests * fix copies * style * use relative imports when needed * Co-authored-by: sgugger <sylvain.gugger@gmail.com>
Configuration menu - View commit details
-
Copy full SHA for 44e3e3f - Browse repository at this point
Copy the full SHA 44e3e3fView commit details -
[ConvBert] Fix #21523 (#21849)
* fix reshaping Fixes #21523 * add test * styling * last fixes * Update src/transformers/models/convbert/modeling_convbert.py * code quallity
Configuration menu - View commit details
-
Copy full SHA for b599b19 - Browse repository at this point
Copy the full SHA b599b19View commit details -
Configuration menu - View commit details
-
Copy full SHA for 5e6cd51 - Browse repository at this point
Copy the full SHA 5e6cd51View commit details -
Fix gradient checkpointing bug Bart (#21866)
Co-authored-by: saswatmeher <saswatmeher@cse.iitb.ac.in>
Configuration menu - View commit details
-
Copy full SHA for 72e9ca7 - Browse repository at this point
Copy the full SHA 72e9ca7View commit details -
[deepspeed] check whether model is NLP one instead of counting on inp…
…ut type (#21800) * trying to figure out whether model is NLP * drop my changes and apply easier fix * trying to handle all int input types * fix logic --------- Co-authored-by: Stas Bekman <stas@stason.org>
Configuration menu - View commit details
-
Copy full SHA for f71873c - Browse repository at this point
Copy the full SHA f71873cView commit details -
Change the way tensor is reshaped in BartAttention (from .view to .re…
…shape) (#21860) * Change the .view call to .reshape * Change the .view call to .reshape to all the copies from bart attention * Fix copies and style * Fix copies and style * Fix copies and style * Fix copies and style * Fix copies and style * Revert unneccessary changes * Revert unneccessary changes * Revert unneccessary changes * Revert unneccessary changes
Configuration menu - View commit details
-
Copy full SHA for ebd5258 - Browse repository at this point
Copy the full SHA ebd5258View commit details -
Italian translation of community.mdx (#21871)
Italian translation of community.mdx gh-17459
Configuration menu - View commit details
-
Copy full SHA for 619d831 - Browse repository at this point
Copy the full SHA 619d831View commit details -
Configuration menu - View commit details
-
Copy full SHA for 72787c5 - Browse repository at this point
Copy the full SHA 72787c5View commit details -
Removed BLIP mention from the troubleshooting guide (#21872)
removed BLIP mention from the troubleshooting guide
Configuration menu - View commit details
-
Copy full SHA for 9c1d598 - Browse repository at this point
Copy the full SHA 9c1d598View commit details -
update FSDP and add XLA-FSDP documentation (#21812)
* update FSDP and add XLA-FSDP documentation * resolving comments * minor update * fix xla-fsdp docs
Configuration menu - View commit details
-
Copy full SHA for 571dd69 - Browse repository at this point
Copy the full SHA 571dd69View commit details -
Configuration menu - View commit details
-
Copy full SHA for 3eba1dd - Browse repository at this point
Copy the full SHA 3eba1ddView commit details -
Add an utility file to get information from test files (#21856)
* Add an utility file to get information from test files --------- Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 53735d7 - Browse repository at this point
Copy the full SHA 53735d7View commit details -
Add check for different embedding types in examples (#21881)
* Add check for different embedding types in examples * Correctly update summarization example
Configuration menu - View commit details
-
Copy full SHA for 1d3a1cc - Browse repository at this point
Copy the full SHA 1d3a1ccView commit details -
Make loading of pretrained gpt2 faster by avoiding initialization of …
…Conv1D's weights (#21879) apply normal_ after assigning weight as nn.Parameter to avoid unnecessary initialization computation
Configuration menu - View commit details
-
Copy full SHA for 45e1109 - Browse repository at this point
Copy the full SHA 45e1109View commit details -
Add TFVisionTextDualEncoder (#21873)
* Temporary commit to stash everything so far * Temporary commit to stash everything so far * stash commit * Refactor from_pretrained * Fix final test, make fixup * Update dummies * Add model to TEST_FILES_WITH_NO_COMMON_TESTS * Update src/transformers/models/vision_text_dual_encoder/modeling_tf_vision_text_dual_encoder.py Co-authored-by: Joao Gante <joaofranciscocardosogante@gmail.com> * Update src/transformers/models/vision_text_dual_encoder/modeling_tf_vision_text_dual_encoder.py Co-authored-by: Joao Gante <joaofranciscocardosogante@gmail.com> * Update src/transformers/models/vision_text_dual_encoder/modeling_tf_vision_text_dual_encoder.py Co-authored-by: Joao Gante <joaofranciscocardosogante@gmail.com> * Update src/transformers/models/vision_text_dual_encoder/modeling_tf_vision_text_dual_encoder.py Co-authored-by: Joao Gante <joaofranciscocardosogante@gmail.com> * Add TFVisionTextDualEncoder to utils/documentation_tests.txt * make fixup --------- Co-authored-by: Joao Gante <joaofranciscocardosogante@gmail.com>
Configuration menu - View commit details
-
Copy full SHA for f7c618e - Browse repository at this point
Copy the full SHA f7c618eView commit details -
Add ALIGN to transformers (#21741)
Adds the ALIGN model to transformers. ALIGN is introduced in "Scaling Up Visual and Vision-Language Representation Learning With Noisy Text Supervision" by Chao Jia, Yinfei Yang, Ye Xia, Yi-Ting Chen, Zarana Parekh, Hieu Pham, Quoc V. Le, Yunhsuan Sung, Zhen Li, Tom Duerig.
Configuration menu - View commit details
-
Copy full SHA for 269b054 - Browse repository at this point
Copy the full SHA 269b054View commit details -
Fix Gradient checkpointing bug BigBird (#21882)
Co-authored-by: saswatmeher <saswatmeher@cse.iitb.ac.in>
Configuration menu - View commit details
-
Copy full SHA for 4edfd2d - Browse repository at this point
Copy the full SHA 4edfd2dView commit details -
* force on the same device * fix tests --------- Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 36ee128 - Browse repository at this point
Copy the full SHA 36ee128View commit details -
Fix
test_load_default_pipelines_pt
forClapModel
(#21886)* fix tests --------- Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 89359e4 - Browse repository at this point
Copy the full SHA 89359e4View commit details
Commits on Mar 2, 2023
-
Configuration menu - View commit details
-
Copy full SHA for 43299c6 - Browse repository at this point
Copy the full SHA 43299c6View commit details -
[Refactor] Relative imports wherever we can (#21880)
* initial commit * update * second batch * style * fix imports * fix relative import on pipeline
Configuration menu - View commit details
-
Copy full SHA for 633e5e8 - Browse repository at this point
Copy the full SHA 633e5e8View commit details -
add correct revision after model was overwritten
Configuration menu - View commit details
-
Copy full SHA for c256bc6 - Browse repository at this point
Copy the full SHA c256bc6View commit details -
Use PyAV instead of Decord in examples (#21572)
* Use PyAV instead of Decord * Get frame indices * Fix number of frames * Update src/transformers/models/videomae/image_processing_videomae.py * Fix up * Fix copies * Update timesformer doctests * Update docstrings
Configuration menu - View commit details
-
Copy full SHA for 3412f59 - Browse repository at this point
Copy the full SHA 3412f59View commit details -
Add
inputs_embeds
functionality when generating with BioGPT (#21889)* initial commit to add inputs_embeds to generation * formatting
Configuration menu - View commit details
-
Copy full SHA for edbb37f - Browse repository at this point
Copy the full SHA edbb37fView commit details -
[T5 doc] Fix confusing documentation about
d_kv
(#21896)* Confusing documentation in T5 * Fix onfusing documentation in T5 configuration file
Configuration menu - View commit details
-
Copy full SHA for b48c7f7 - Browse repository at this point
Copy the full SHA b48c7f7View commit details -
[Whisper] Add rescaling function with
do_normalize
(#21263)* add `zero_mean_unit_var_norm` function * normalize before MEL computation * fixup * add simple test * quality * Update tests/models/whisper/test_feature_extraction_whisper.py Co-authored-by: Sanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com> * fixup * use attention masks if padding was applied * Update based on review Co-authored-by: bofeng huang <bofenghuang7@gmail.com> --------- Co-authored-by: Sanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com> Co-authored-by: bofeng huang <bofenghuang7@gmail.com>
Configuration menu - View commit details
-
Copy full SHA for c87654d - Browse repository at this point
Copy the full SHA c87654dView commit details -
Configuration menu - View commit details
-
Copy full SHA for 648d0de - Browse repository at this point
Copy the full SHA 648d0deView commit details -
[GPT-J] add deprecation warning (#21869)
* add deprecation warning * remove pos ids from args docstirng * fix failing test
Configuration menu - View commit details
-
Copy full SHA for fb76994 - Browse repository at this point
Copy the full SHA fb76994View commit details -
Configuration menu - View commit details
-
Copy full SHA for b6f47b5 - Browse repository at this point
Copy the full SHA b6f47b5View commit details -
Fix gradient checkpointing bug LED (#21840)
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 7e6dd66 - Browse repository at this point
Copy the full SHA 7e6dd66View commit details -
Fix gradient checkpointing bug M2M 100 (#21841)
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for b405b62 - Browse repository at this point
Copy the full SHA b405b62View commit details -
Fix gradient checkpointing bug marian (#21842)
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for d9e28d9 - Browse repository at this point
Copy the full SHA d9e28d9View commit details -
Mark pipeline tests to skip them easily (#21887)
* Mark pipeline tests to skip them easily * Mark the mixin as pipeline test * Update src/transformers/testing_utils.py Co-authored-by: Yih-Dar <2521628+ydshieh@users.noreply.github.com> --------- Co-authored-by: Yih-Dar <2521628+ydshieh@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 50a8ed3 - Browse repository at this point
Copy the full SHA 50a8ed3View commit details -
Clean up auto mapping names (#21903)
* add new test * fix after new test --------- Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 99ba36e - Browse repository at this point
Copy the full SHA 99ba36eView commit details -
Prophetnet batch dimension inversion fix (#21870)
* decoder forward pass is working * no model has forward pass returning attentions * decoder ngram changed to not mix batch size * current basic forward pass returns identical result * passed test_model attentions * passed test_encoder_decoder_model_generate * passed test_headmasking * removed old block * removed comments bug/fixme * removed bug comments * applied styling * applied fix-copies * applied ngram forward comments * corrected dimension notation * applied styling and comment fixes * changed asserts for raise ValueError * changed question gen test * updated hidden_states integration test * applied styling
Configuration menu - View commit details
-
Copy full SHA for 6bf8853 - Browse repository at this point
Copy the full SHA 6bf8853View commit details -
Make schedulers picklable by making lr_lambda fns global (#21768)
* Make schedulers picklable by making lr_lambda fns global * add unused _get_constant_schedule_lr_lambda arg * remove unneeded _get_constant_schedule_lr_lamda * add test * make style * rebase, remove torch dep, put lambda back * repo-consistency and style
Connor Henderson committedMar 2, 2023 Configuration menu - View commit details
-
Copy full SHA for 8e5a1b2 - Browse repository at this point
Copy the full SHA 8e5a1b2View commit details -
Refactor whisper asr pipeline to include language too. (#21427)
* [WIP] whisper refacto to support language output. * Handling merges. * A bit more cleanup and comments. * Many improvements. Lots of details everywhere. * Cleanup old code and tests. * Handle lone timestamp tokens (just recover when something bad happens). * Adding return_language example. * No ffmpeg. * Hmm. * Some corrections. * Both fast and slow. * New black. * Update src/transformers/models/whisper/tokenization_whisper.py Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com> * Update src/transformers/models/whisper/tokenization_whisper.py Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com> * Remove print. * Undoing tests modifications. * Smaller test modifications. * Rename. * Remove maxDiff. --------- Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 1325459 - Browse repository at this point
Copy the full SHA 1325459View commit details -
Add Blip and Blip2 for pipeline tests (#21904)
* fix * add to tests * style and quality * add missing --------- Co-authored-by: NielsRogge <NielsRogge@users.noreply.github.com> Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for e6de918 - Browse repository at this point
Copy the full SHA e6de918View commit details -
Temporarily skip 3 tests in
BridgeTowerModelTest
(#21908)skip for now Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 88e5c51 - Browse repository at this point
Copy the full SHA 88e5c51View commit details -
Faster zero shot image (#21897)
* Make ZeroShotImageClassificationPipeline faster The pipeline makes separate calls to model for each candidate label. This commit combines all labels into one call. Original code takes more that 60 seconds to process one image and 1000 candidate labels. Updated code takes less than 2 seconds. * implement batching * code formatting * Creating an even faster zero-shot-image-classifiction. Unfortunately super tailored towards CLIP. Co-Authored-By: Yessen Kanapin <yessen@deepinfra.com> * Quality. * Cleanup. * Order different on the CI it seems. * Cleanup. * Quality. --------- Co-authored-by: Yessen Kanapin <yessen@deepinfra.com>
Configuration menu - View commit details
-
Copy full SHA for b2a41d2 - Browse repository at this point
Copy the full SHA b2a41d2View commit details -
[time series] Add Time series inputs tests (#21846)
* intial test of inputs * added test for generation * remove asserts * fixed test * Update tests/models/time_series_transformer/test_modeling_time_series_transformer.py Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com> --------- Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for db979f7 - Browse repository at this point
Copy the full SHA db979f7View commit details -
Avoid modeling tests run in pipeline CI jobs (#21911)
* rework is_pipeline_test * bring back 3 tests --------- Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 9f5bfe1 - Browse repository at this point
Copy the full SHA 9f5bfe1View commit details
Commits on Mar 3, 2023
-
Configuration menu - View commit details
-
Copy full SHA for 37e0974 - Browse repository at this point
Copy the full SHA 37e0974View commit details -
faster forward following what is done for images (#21906)
* faster forward following what is done for images * add missing licence
Configuration menu - View commit details
-
Copy full SHA for dcec327 - Browse repository at this point
Copy the full SHA dcec327View commit details -
Configuration menu - View commit details
-
Copy full SHA for e407b5a - Browse repository at this point
Copy the full SHA e407b5aView commit details -
Configuration menu - View commit details
-
Copy full SHA for 99a6234 - Browse repository at this point
Copy the full SHA 99a6234View commit details -
Configuration menu - View commit details
-
Copy full SHA for c82bd37 - Browse repository at this point
Copy the full SHA c82bd37View commit details -
Update
model_split_percents
forWhisperModelTest
(#21922)Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for fa9d2ad - Browse repository at this point
Copy the full SHA fa9d2adView commit details -
Use large VM for
repo_utils_job
(#21928)upgrade to large VM Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for b05e0be - Browse repository at this point
Copy the full SHA b05e0beView commit details -
Cleanup more auto mapping names (#21909)
* fix auto 2 * fix auto 2 * fix task guide issue * fix --------- Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 02a77fa - Browse repository at this point
Copy the full SHA 02a77faView commit details -
feat: filter try/except when looking at custom code (#21914)
* feat: filter try/except * Update src/transformers/dynamic_module_utils.py Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com> --------- Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for c5a1ff9 - Browse repository at this point
Copy the full SHA c5a1ff9View commit details -
Fix
AlignModelTest
tests (#21923)* fix * fix --------- Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for d4306da - Browse repository at this point
Copy the full SHA d4306daView commit details -
Avoid failure in
check_repo.py
due to missing backends (#21930)* Update utils/check_repo.py Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update utils/check_repo.py Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com> --------- Co-authored-by: ydshieh <ydshieh@users.noreply.github.com> Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 8c40ba7 - Browse repository at this point
Copy the full SHA 8c40ba7View commit details -
Fix wrong documentation about DataCollator padding defaults (#21919)
* Fix wrong documentation about DataCollator padding defaults * Fix styling
Configuration menu - View commit details
-
Copy full SHA for 956ae62 - Browse repository at this point
Copy the full SHA 956ae62View commit details -
[Flan-UL2] Add-flan-ul2 (#21929)
* add doc and readme * add model docs * update toctree and fix copies * update * update doc file * fix * add FLAN-UL2 to configuration mapping * fixup * Apply suggestions from code review * more clarification --------- Co-authored-by: younesbelakda <younesbelkada@gmail.com> Co-authored-by: Younes Belkada <49240599+younesbelkada@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 82aac00 - Browse repository at this point
Copy the full SHA 82aac00View commit details -
Configuration menu - View commit details
-
Copy full SHA for c5fe06c - Browse repository at this point
Copy the full SHA c5fe06cView commit details -
[CLAP] Support batched inputs for CLAP. Fixes pipeline issues (#21931)
* fix pipeline * fix feature_extraction clap * you can now batch the `is_longer` attribute * add tests * fixup * add expected scores * comment on is_longert
Configuration menu - View commit details
-
Copy full SHA for 718e9d7 - Browse repository at this point
Copy the full SHA 718e9d7View commit details -
Configuration menu - View commit details
-
Copy full SHA for 003a7cc - Browse repository at this point
Copy the full SHA 003a7ccView commit details
Commits on Mar 4, 2023
-
Configuration menu - View commit details
-
Copy full SHA for f932ee6 - Browse repository at this point
Copy the full SHA f932ee6View commit details -
Configuration menu - View commit details
-
Copy full SHA for f12c74f - Browse repository at this point
Copy the full SHA f12c74fView commit details -
Configuration menu - View commit details
-
Copy full SHA for 6386eb9 - Browse repository at this point
Copy the full SHA 6386eb9View commit details -
Configuration menu - View commit details
-
Copy full SHA for 6feb39b - Browse repository at this point
Copy the full SHA 6feb39bView commit details -
Fixed gradient_checkpointing/use_cache bug in blenderbot (#21833)
* Fixed gradient_checkpointing/use_cache bug in blenderbot * Update modeling_blenderbot.py * Added back if statement * Formatted using black
Configuration menu - View commit details
-
Copy full SHA for 699a229 - Browse repository at this point
Copy the full SHA 699a229View commit details
Commits on Mar 6, 2023
-
Update expected values in
XLMProphetNetModelIntegrationTest
(#21957)update values Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for fcf8134 - Browse repository at this point
Copy the full SHA fcf8134View commit details -
* fix `get_proposal_pos_embed` * fix order * style * zero shot simplify test * add approximate values for zero shot audio classification
Configuration menu - View commit details
-
Copy full SHA for bc33fbf - Browse repository at this point
Copy the full SHA bc33fbfView commit details -
Disable DDP for neuron (#21953)
Disable DDp for neuron Co-authored-by: EC2 Default User <ec2-user@ip-172-31-42-72.us-west-2.compute.internal>
Configuration menu - View commit details
-
Copy full SHA for 0bb1729 - Browse repository at this point
Copy the full SHA 0bb1729View commit details -
Co-authored-by: saswatmeher <saswatmeher@cse.iitb.ac.in>
Configuration menu - View commit details
-
Copy full SHA for 934d0b8 - Browse repository at this point
Copy the full SHA 934d0b8View commit details -
[Generate] Fix gradient_checkpointing and use_cache bug for BLOOM (#2…
…1956) Step 1 - Change use_cache fix
Configuration menu - View commit details
-
Copy full SHA for f3c75f8 - Browse repository at this point
Copy the full SHA f3c75f8View commit details -
Add missing parameter definition in layoutlm config (#21960)
Four parameters in `LayoutLM` config were missing definitions, Added their definition (copied from BertConfig).
Configuration menu - View commit details
-
Copy full SHA for 64d95c4 - Browse repository at this point
Copy the full SHA 64d95c4View commit details -
Use larger atol in
torch.allclose
for some tests (#21966)Use larger atol Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 9474abd - Browse repository at this point
Copy the full SHA 9474abdView commit details -
Add TF contrastive image text finetuning example (#21939)
* Initial commit * stash commit * Add model checkpointing and pushing * Fix model name inference * Update README * Update README * Remove a couple of Torch references * Update copyright date * make fixup * Update PushToHubCallback args! * Remove the torch summary * Add strategy.scope
Configuration menu - View commit details
-
Copy full SHA for 5d8efc7 - Browse repository at this point
Copy the full SHA 5d8efc7View commit details -
Update expected values for
test_xglm_sample
(#21975)update expected values for xglm Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for f2a2616 - Browse repository at this point
Copy the full SHA f2a2616View commit details -
Configuration menu - View commit details
-
Copy full SHA for 4f84ded - Browse repository at this point
Copy the full SHA 4f84dedView commit details -
Configuration menu - View commit details
-
Copy full SHA for 451263b - Browse repository at this point
Copy the full SHA 451263bView commit details -
Configuration menu - View commit details
-
Copy full SHA for 4a545d1 - Browse repository at this point
Copy the full SHA 4a545d1View commit details -
Configuration menu - View commit details
-
Copy full SHA for de496ef - Browse repository at this point
Copy the full SHA de496efView commit details -
Configuration menu - View commit details
-
Copy full SHA for 0ce5236 - Browse repository at this point
Copy the full SHA 0ce5236View commit details -
docs: improve clarity for language modeling (#21952)
* docs: improve clarity for clm/mlm * docs: remove incorrect explanation * docs: remove incorrect explanation --------- Co-authored-by: pdhall99 <pdhall99>
Configuration menu - View commit details
-
Copy full SHA for 31e3c6c - Browse repository at this point
Copy the full SHA 31e3c6cView commit details
Commits on Mar 7, 2023
-
* update expected values for jukebox * update expected values for jukebox * update expected values for jukebox * update expected values for jukebox * update expected values for jukebox --------- Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 5b28b78 - Browse repository at this point
Copy the full SHA 5b28b78View commit details -
Add check before int casting for PIL conversion (#21969)
* Add check before int casting for PIL conversion * Line length * Tidier logic
Configuration menu - View commit details
-
Copy full SHA for 4063fd9 - Browse repository at this point
Copy the full SHA 4063fd9View commit details -
Fix MinNewTokensLengthLogitsProcessor when used with a list of eos to…
…kens (#21959) * Fix MinNewTokensLengthLogitsProcessor when used with a list of eos tokens * fix docs * Empty commit * formatting
Configuration menu - View commit details
-
Copy full SHA for eec46b4 - Browse repository at this point
Copy the full SHA eec46b4View commit details -
[DETR, YOLOS] Fix device bug (#21974)
* Fix integration test * Add test * Add test
Configuration menu - View commit details
-
Copy full SHA for 95408e9 - Browse repository at this point
Copy the full SHA 95408e9View commit details -
Configuration menu - View commit details
-
Copy full SHA for 10bcbca - Browse repository at this point
Copy the full SHA 10bcbcaView commit details -
Update
notification_service.py
(#21992)* better check * better check --------- Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 99c5c60 - Browse repository at this point
Copy the full SHA 99c5c60View commit details -
Skip
test_multi_gpu_data_parallel_forward
for some model tests (#21991Configuration menu - View commit details
-
Copy full SHA for 9402788 - Browse repository at this point
Copy the full SHA 9402788View commit details -
[Whisper] Add model for audio classification (#21754)
* [Whisper] Add model for audio classification * make fix-copies * add to docs * add docstring * empty returns * add code example * switch to fleurs * stick everything on one line
Configuration menu - View commit details
-
Copy full SHA for 7c39318 - Browse repository at this point
Copy the full SHA 7c39318View commit details -
Stop requiring Torch for our TF examples! (#21997)
* Stop requiring Torch for our TF examples! * Slight tweak to logging in the example itself
Configuration menu - View commit details
-
Copy full SHA for d128f2f - Browse repository at this point
Copy the full SHA d128f2fView commit details -
[TF] Fix creating a PR while pushing in TF framework (#21968)
* add create pr arg * style * add test * ficup * update test * last nit fix typo * add `is_pt_tf_cross_test` marker for the tsts
Configuration menu - View commit details
-
Copy full SHA for 2156662 - Browse repository at this point
Copy the full SHA 2156662View commit details -
[DETR and friends] Remove is_timm_available (#21814)
* First draft * Fix to_dict * Improve conversion script * Update config * Remove timm dependency * Fix dummies * Fix typo, add integration test * Upload 101 model as well * Remove timm dummies * Fix style --------- Co-authored-by: Niels Rogge <nielsrogge@Nielss-MacBook-Pro.local>
Configuration menu - View commit details
-
Copy full SHA for dde718e - Browse repository at this point
Copy the full SHA dde718eView commit details -
[Time-Series] informer model (#21099)
* added informer to gitignore * added informer to gitignore * WIP informer2020 * added checking that instantiate works * added config using gluonTS by kashif * WIP config * adding informeConfig. need to remove FeatureEmbedder * done InformerConfig, but need to change the names * Done informer model init. working on enc-dec * added things to address, after reading again enc-dec in the paper * done modeling - checking initialization work * added informer to gitignore * WIP informer2020 * added checking that instantiate works * added config using gluonTS by kashif * WIP config * adding informeConfig. need to remove FeatureEmbedder * done InformerConfig, but need to change the names * Done informer model init. working on enc-dec * added things to address, after reading again enc-dec in the paper * done modeling - checking initialization work * moved enc-dec init to InformerEncoder/Decoder init * added 'init_std' to config, now model init works! * WIP conversion script, and added code sources * WIP conversion script: loading original informer pth works * WIP conversion script: change defaults in the config * WIP conversion script: supporting Informer input embedding * WIP conversion script: added parameters for the informer embed * WIP conversion script: change dim_feedforward=2048 * WIP conversion script: remove unused args for loading checkpoint * just cleaning up * DataEmbedding removed, after thinking with Kashif * working on forward pass * WIP forward pass: trying to establish working batch for forward pass * cleaning and finalizing * adding HF names and docs * init after cleaning works * WIP in tests * added docs for the informer specific args * fix style * undo change * cleaning informer, now need to work only enc-dec * initial enc-dec classes * added encoder and decoder * added todo * add todos for conv_layers * added decoder docs from vanilla * added encoder docs from vanilla * remove encoder decoder from the original informer * removed AttentionLayer from the original paper * removed TriangularCausalMask, same as decoder_attention_mask * initial sparse attention * use conv_layers * fixed test_config test * fix parenthesis when itearting zip(layers, conv_layers) * error found in prob attention, added sizes as comments * fix sizes * added proposal for q_reduce indexing, and remove unused * WIP ProbMask, and changed factor=2 for testing * remove unused libs for this PR for creating the env * fix checking the attn_weights.size() after bmm * Q_reduce: changed from torch.gather to simple slicing * WIP calculate final attn_output * finish adding v_aggregated, attn_output ready * changed tgt_len to u in attention_mask, need to fix the size error * comment attention_mask for encoder, and fix if cond for v_agg * added ProbMask support (wip), removed old original code * finished ProbMask 😃 * Revert "remove unused libs for this PR for creating the env" This reverts commit 11a081e. * fixes * make style * fix initial tests * fix more tests * dry * make style * remove unused files * style * added integration tests * fix num_static_real_features * fix header * remove unused function * fix example * fix docs * Update src/transformers/models/informer/configuration_informer.py Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com> * Update src/transformers/models/informer/modeling_informer.py Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com> * Update src/transformers/models/informer/configuration_informer.py Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com> * Update src/transformers/models/informer/configuration_informer.py Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com> * Update src/transformers/models/informer/configuration_informer.py Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com> * Update src/transformers/models/informer/configuration_informer.py Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com> * fixes for reviewer * use prediction_length from model * fix style * fixed informer.mdx * added to index * updated readme * undo * make fix-copies * typo * fix copy * added Informer to toctree * in order * fixed comments * remove unneeded new lines in docs * make static real and cat optional * fix use of distil conv layers * fixed integration test * added checkpoint for convlayer * make fix-copies * updated from time series model * make fix-copies * copy decoder * fix unit tests * updated scaling config * fix integration tests * IGNORE_NON_TESTED * IGNORE_NON_AUTO_CONFIGURED * IGNORE_NON_AUTO_CONFIGURED * updated check configs * fix formatting * undo change from time series * prediction_length should not be None * aliign with the blog: prettify ProbSparse and change attention_factor to sampling_factor * make style * make fix-copies * niels CR: update contributed by * niels CR: update configuration_informer.py Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com> * niels CR: update kashif -> huggingface Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com> * niels CR: `sampling_factor` only relevant when `attention_type`=prob * make style * fixed U_part: added multiplication by `L_Q` * fixed bug: remove `is not None` from `if config.distil` * fixed test: `decoder_seq_length` to `encoder_seq_length` in cross_attentions check * fix integration tests * updated model hub * do not shift as in training * undo * fix make-copies * make fix-copies * added `if prediction_length is None` * changed `ProbSparseAttention` to `InformerProbSparseAttention` * changed `V_sum` -> `v_mean_dim_time` * changed `ConvLayer` to `InformerConvLayer` and fixed `super()` * TimeSeriesTansformer->Informer in decoder's Copied from * more descriptive in ProbSparse * make style * fix coped from * Revert "added `if prediction_length is None`" This reverts commit b4cbddf. * fixed indent * use InformerSinusoidalPositionalEmbedding * make fix-style * fix from #21860 * fix name * make fix-copies * use time series utils * fix dec num_heads * docstring * added time series util doc * _import_structure * formatting * changes from review * make style * fix docs * fix doc * removed NegativeLogLikelihood --------- Co-authored-by: Kashif Rasul <kashif.rasul@gmail.com> Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 8abe493 - Browse repository at this point
Copy the full SHA 8abe493View commit details -
Update tiny model creation script and some others files (#22006)
* Update 1 * Update 2 * Update 3 * Update 4 * Update 5 * Update 6 * Update 7 * Update 8 * Update 9 * Update 10 --------- Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for b338414 - Browse repository at this point
Copy the full SHA b338414View commit details
Commits on Mar 8, 2023
-
Generate - add 1 to cur_len to make up the new beam length (#21993)
* add 1 to cur_len to make up the new beam length cur_len is 1 token shorter comparing to the length of the sequence whose best_sum_logprobs is the numerator. * cur_len+=1 before check if beam hyp is done * format code * reformat with black --------- Co-authored-by: Chiming <chiming@biomap.com>
Configuration menu - View commit details
-
Copy full SHA for c1f8559 - Browse repository at this point
Copy the full SHA c1f8559View commit details -
VideoMAE doctest - use valid dummy pixel values (#22022)
Use valid dummy pixel values
Configuration menu - View commit details
-
Copy full SHA for 4130e70 - Browse repository at this point
Copy the full SHA 4130e70View commit details -
Configuration menu - View commit details
-
Copy full SHA for bbd9499 - Browse repository at this point
Copy the full SHA bbd9499View commit details -
Update
AudioClassificationPipelineTests::test_small_model_pt
for PT…… 2.0.0 (#22023) fix Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for dfe9a31 - Browse repository at this point
Copy the full SHA dfe9a31View commit details -
[
bnb
] Fix bnb error message (#22026)* fix error message * make style
Configuration menu - View commit details
-
Copy full SHA for edea08a - Browse repository at this point
Copy the full SHA edea08aView commit details -
[WIP] Add BridgeTowerForContrastiveLearning (#21964)
* Add BridgeTower for ITC * Fix review feedback * Rename BridgeTowerForITC, cleanup * Fix style and quality * implement tests --------- Co-authored-by: Tiep Le <97980157+tileintel@users.noreply.github.com> Co-authored-by: Tiep Le <tiep.le@intel.com>
Configuration menu - View commit details
-
Copy full SHA for de81adf - Browse repository at this point
Copy the full SHA de81adfView commit details -
Configuration menu - View commit details
-
Copy full SHA for a5392ee - Browse repository at this point
Copy the full SHA a5392eeView commit details -
Add tokenize_kwargs parameter definition in the FeatureExtractionPipe…
…line (#22031) add tokenize_kwargs doc in the FeatureExtractionPipeline
Configuration menu - View commit details
-
Copy full SHA for b427b26 - Browse repository at this point
Copy the full SHA b427b26View commit details -
[examples/speech-recognition] Add SpecAugment to run_speech_recogniti…
…on_seq2seq.py (#21942) * Add specaugment to run_speech_recognition_seq2seq.py * Remove useless argument: text_column * Fix quality * Update return_attention_mask condition * Update specaugment arguments only for whisper models * Remove SpecAugment arguments from ModelArguments, only leave default values for simplicity * Apply suggestions from code review Co-authored-by: Sanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com> * Update apply_spec_augment only for whisper models * Apply suggestions from code review Co-authored-by: Sanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com> * Rename return_attention_mask to forward_attention_mask to avoid confusion with wav2vec2 models --------- Co-authored-by: Sanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 6192549 - Browse repository at this point
Copy the full SHA 6192549View commit details -
fixes the gradient checkpointing of whisper (#22019)
* fixing * Update modeling_whisper.py * Update modeling_whisper.py * Update src/transformers/models/whisper/modeling_whisper.py --------- Co-authored-by: Joao Gante <joaofranciscocardosogante@gmail.com>
Configuration menu - View commit details
-
Copy full SHA for 9983950 - Browse repository at this point
Copy the full SHA 9983950View commit details -
Avoid
text_config_dict
andvision_config_dict
being saved for CLI……P-like models (#22035) * Avoid text_config_dict and vision_config_dict being saved * for other CLIP-like models --------- Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for bcc8d30 - Browse repository at this point
Copy the full SHA bcc8d30View commit details -
Mark all
BridgeTower
tests slow for now (#22039)* slow me --------- Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 1cbac68 - Browse repository at this point
Copy the full SHA 1cbac68View commit details -
Bug fix: token classification pipeline while passing offset_mapping (#…
…22034) fix slow tokenizers with passing offset_mapping
Configuration menu - View commit details
-
Copy full SHA for 3ec8171 - Browse repository at this point
Copy the full SHA 3ec8171View commit details
Commits on Mar 9, 2023
-
* Fix typos and add code examples, resources
Configuration menu - View commit details
-
Copy full SHA for 2055d73 - Browse repository at this point
Copy the full SHA 2055d73View commit details -
[21737][T5]: Fix gradient checkpoint bug (#22036)
* [21737][T5]: Fix gradient checkpoint bug * [21737][T5]: Fix gradient checkpoint bug * [21737][T5]: Fix gradient checkpoint bug * Update src/transformers/models/mt5/modeling_mt5.py * Update src/transformers/models/t5/modeling_t5.py --------- Co-authored-by: njindal <njindal@adobe.com> Co-authored-by: Joao Gante <joaofranciscocardosogante@gmail.com>
Configuration menu - View commit details
-
Copy full SHA for 1a77a1a - Browse repository at this point
Copy the full SHA 1a77a1aView commit details -
Docs Improvement - In ZSH, not using ' ' around pip install fails, fi…
…x it (#22045) In ZSH, not using ' ' around pip install fails Running ``` pip install transformers[torch] ``` in the default ZSH terminal will fail with the error `zsh: no matches found: transformers[torch]` The solution is to wrap the installation path in ' ' like ``` pip install 'transformers[torch]' ``` Relevant StackOverflow: https://stackoverflow.com/questions/30539798/zsh-no-matches-found-requestssecurity
Configuration menu - View commit details
-
Copy full SHA for 81cd655 - Browse repository at this point
Copy the full SHA 81cd655View commit details -
Configuration menu - View commit details
-
Copy full SHA for 6847743 - Browse repository at this point
Copy the full SHA 6847743View commit details -
Remove set_access_token usage + fail tests if FutureWarning (#22051)
* Remove set_access_token usage + fail tests if FutureWarning * do not fail on FutureWarning in CI --------- Co-authored-by: testbot <lucainp@hf.co>
Configuration menu - View commit details
-
Copy full SHA for 923110b - Browse repository at this point
Copy the full SHA 923110bView commit details -
Show the number of
huggingface_hub
warnings in CI report (#22054)* show hfh warnings --------- Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 90a7c95 - Browse repository at this point
Copy the full SHA 90a7c95View commit details -
Return analysis for hyperparameter_search with Ray backend (#22040)
* return analysis for hyperparameter_search with ray backend * Revert "return analysis for hyperparameter_search with ray backend" This reverts commit cd51790. * add run_summary attribute to BestRun and return analysis for ray backend * fix typo * add doc for run_summary for ray backend
Configuration menu - View commit details
-
Copy full SHA for 04bfac8 - Browse repository at this point
Copy the full SHA 04bfac8View commit details -
pt-to-tf model architecture override (#22055)
* Add an argument to pt-to-tf to allow overriding the model class * make fixup * Minor fix to error message * Remove unused extra conversion from the script
Configuration menu - View commit details
-
Copy full SHA for fdf8409 - Browse repository at this point
Copy the full SHA fdf8409View commit details -
rm $ symbol from code block from contributing.md (#22057)
rm $ symbol from code block Removed the $ symbol from the code block to make copy-pasting easier.
Configuration menu - View commit details
-
Copy full SHA for d0c19b3 - Browse repository at this point
Copy the full SHA d0c19b3View commit details -
[deepspeed] offload + non-cpuadam optimizer exception (#22043)
* [deepspeed] offload + non-cpuadam optimizer exception * flip * revert min version
Configuration menu - View commit details
-
Copy full SHA for ec24132 - Browse repository at this point
Copy the full SHA ec24132View commit details -
Edit the docstring of
image_processing_donut
to match code (#22033)* Edit the docstring of `image_processing_donut` to match code * improve style * more style improvement after installing quality
Configuration menu - View commit details
-
Copy full SHA for 8434cb8 - Browse repository at this point
Copy the full SHA 8434cb8View commit details -
Skip 3 tests for
WhisperEncoderModelTest
(#22060)* skip 3 tests --------- Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for ab81d31 - Browse repository at this point
Copy the full SHA ab81d31View commit details -
Add setters by type of args to TrainingArguments (#21570)
* Add setters by type of args to TrainingArguments * Define more setters
Configuration menu - View commit details
-
Copy full SHA for 7a2b915 - Browse repository at this point
Copy the full SHA 7a2b915View commit details -
Update tiny model creation script (#22058)
Update the script Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 6d9031f - Browse repository at this point
Copy the full SHA 6d9031fView commit details -
Fix case when using --gradient_accumulation_steps with DDP disabled. (#…
…22007) Co-authored-by: EC2 Default User <ec2-user@ip-172-31-42-72.us-west-2.compute.internal>
Configuration menu - View commit details
-
Copy full SHA for 1a5fc30 - Browse repository at this point
Copy the full SHA 1a5fc30View commit details -
Add a progress bar for the total download of shards (#22062)
* Add a progress bar for the total download of shards * Check for no cache at all * Fix check
Configuration menu - View commit details
-
Copy full SHA for a9bd5df - Browse repository at this point
Copy the full SHA a9bd5dfView commit details
Commits on Mar 10, 2023
-
Fix gradient checkpointing bug in Speech2Text (#22079)
* Fix gradient checkpointing bug in Speech2Text * Update modeling_speech_to_text.py * Update modeling_speech_to_text_2.py
Configuration menu - View commit details
-
Copy full SHA for b927335 - Browse repository at this point
Copy the full SHA b927335View commit details -
Configuration menu - View commit details
-
Copy full SHA for eee195b - Browse repository at this point
Copy the full SHA eee195bView commit details -
[GPT2] Propose fix for #21080 (#21853)
* Make sure position ids are masked * test that padded input produce the same results * fix failing tests * fixup * fix batch test
Configuration menu - View commit details
-
Copy full SHA for a3fef89 - Browse repository at this point
Copy the full SHA a3fef89View commit details -
Fix small typo in flan-ul2.mdx (#22068)
* Update flan-ul2.mdx * Update flan-ul2.mdx
Configuration menu - View commit details
-
Copy full SHA for ade26bf - Browse repository at this point
Copy the full SHA ade26bfView commit details -
Configuration menu - View commit details
-
Copy full SHA for 7014fc3 - Browse repository at this point
Copy the full SHA 7014fc3View commit details -
Fix gradient checkpointing bug in Speecht5 (#22080)
* Fix gradient checkpointing bug in Speecht5 * Update modeling_speech_to_text.py * Update src/transformers/models/speech_to_text/modeling_speech_to_text.py * Fix change errors --------- Co-authored-by: Joao Gante <joaofranciscocardosogante@gmail.com>
Configuration menu - View commit details
-
Copy full SHA for 419d979 - Browse repository at this point
Copy the full SHA 419d979View commit details -
Configuration menu - View commit details
-
Copy full SHA for a70da86 - Browse repository at this point
Copy the full SHA a70da86View commit details -
Configuration menu - View commit details
-
Copy full SHA for 2f4cdd9 - Browse repository at this point
Copy the full SHA 2f4cdd9View commit details -
Configuration menu - View commit details
-
Copy full SHA for bdec276 - Browse repository at this point
Copy the full SHA bdec276View commit details -
Fix imports of TF MobileViT (#22065)
* Fix imports of TF MobileViT * Fix copies
Configuration menu - View commit details
-
Copy full SHA for 499770c - Browse repository at this point
Copy the full SHA 499770cView commit details -
Configuration menu - View commit details
-
Copy full SHA for 2f32066 - Browse repository at this point
Copy the full SHA 2f32066View commit details
Commits on Mar 11, 2023
-
[Whisper] Remove embed_tokens from encoder docstring (#21996)
* [Whisper] Remove embed_tokens from encoder docstring * new line to retrigger CI * remove new line
Configuration menu - View commit details
-
Copy full SHA for b90fbc7 - Browse repository at this point
Copy the full SHA b90fbc7View commit details
Commits on Mar 13, 2023
-
Add AutoModelForZeroShotImageClassification (#22087)
Adds AutoModelForZeroShotImageClassification to transformers
Configuration menu - View commit details
-
Copy full SHA for 32e3466 - Browse repository at this point
Copy the full SHA 32e3466View commit details -
add new model of MGP-STR (#21418)
* add new model of MGP-STR * fix the check failings * remove torch and numpy from mgp_tokenization * remove unused import from modeling_mgp_str * add test_processing_mgp_str * rm test_processing_mgp_str.py * add test_processing_mgp_str * add test_processing_mgp_str * add test_processing_mgp_str * rm test_processing_mgp_str and add softmax outs to model * rm test_processing_mgp_str and add softmax outs to model * rewrite the code of mgp-str according to PR suggestions * rewrite the code of mgp-str according to PR suggestions * add new model of MGP-STR * fix the check failings * remove torch and numpy from mgp_tokenization * remove unused import from modeling_mgp_str * add test_processing_mgp_str * rm test_processing_mgp_str.py * add test_processing_mgp_str * add test_processing_mgp_str * add test_processing_mgp_str * rm test_processing_mgp_str and add softmax outs to model * rewrite the code of mgp-str according to PR suggestions * rewrite the code of mgp-str according to PR suggestions * remove representation_size from MGPSTRConfig * reformat configuration_mgp_str.py * format test_processor_mgp_str.py * add test for tokenizer and complete model/processer test and model file * rm Unnecessary tupple in modeling_mgp_str * reduce hidden_size/layers/label_size in test_model * add integration tests and change MGPSTR to Mgpstr * add test for logit values * reformat test model file --------- Co-authored-by: yue kun <yuekun.wp@alibaba-inc.com>
Configuration menu - View commit details
-
Copy full SHA for 102b5ff - Browse repository at this point
Copy the full SHA 102b5ffView commit details -
Configuration menu - View commit details
-
Copy full SHA for 0c88376 - Browse repository at this point
Copy the full SHA 0c88376View commit details -
Configuration menu - View commit details
-
Copy full SHA for d0876a0 - Browse repository at this point
Copy the full SHA d0876a0View commit details -
Configuration menu - View commit details
-
Copy full SHA for 4c14c1f - Browse repository at this point
Copy the full SHA 4c14c1fView commit details -
Configuration menu - View commit details
-
Copy full SHA for 0768c5e - Browse repository at this point
Copy the full SHA 0768c5eView commit details -
Added big_models.mdx italian translation #17600 (#22115)
* updated toctree * italian translation big_model.mdx * italian translation big_models
Configuration menu - View commit details
-
Copy full SHA for dd3a058 - Browse repository at this point
Copy the full SHA dd3a058View commit details -
Configuration menu - View commit details
-
Copy full SHA for 6652e7d - Browse repository at this point
Copy the full SHA 6652e7dView commit details -
Configuration menu - View commit details
-
Copy full SHA for c1db6a3 - Browse repository at this point
Copy the full SHA c1db6a3View commit details -
Configuration menu - View commit details
-
Copy full SHA for ef74e7e - Browse repository at this point
Copy the full SHA ef74e7eView commit details -
Fix gradient checkpointing bug in trocr (#22126)
* Fix gradient checkpointing bug in trocr * Fix format * Update src/transformers/models/trocr/modeling_trocr.py Co-authored-by: Younes Belkada <49240599+younesbelkada@users.noreply.github.com> --------- Co-authored-by: Younes Belkada <49240599+younesbelkada@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for e61081e - Browse repository at this point
Copy the full SHA e61081eView commit details -
Zero-shot image classification task guide (#22132)
* WIP * WIP * manual inference example * make style * Apply suggestions from code review Co-authored-by: Alara Dirik <8944735+alaradirik@users.noreply.github.com> --------- Co-authored-by: Alara Dirik <8944735+alaradirik@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 8def252 - Browse repository at this point
Copy the full SHA 8def252View commit details -
Configuration menu - View commit details
-
Copy full SHA for 6cb5132 - Browse repository at this point
Copy the full SHA 6cb5132View commit details -
Adding Type Hints to TF_Pegasus model (#21941)
* Adding Type Hints to TF_Pegasus model * Updated some parameters per maintainer comments
Configuration menu - View commit details
-
Copy full SHA for a096eac - Browse repository at this point
Copy the full SHA a096eacView commit details -
Add a new script to check model testers' config (#22063)
* Add script --------- Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 54ee56b - Browse repository at this point
Copy the full SHA 54ee56bView commit details -
Update configuration_align.py (projected_dim=640) (#22139)
Update configuration_align.py updated projected_dim=640 from 512 in arguments of AlignConfig
Configuration menu - View commit details
-
Copy full SHA for 9879723 - Browse repository at this point
Copy the full SHA 9879723View commit details -
[
Whiper
] addget_input_embeddings
to `WhisperForAudioClassificati……on` (#22133) * add `get_input_embeddings` to `WhisperForAudioClassification` * add common tests * fix another common test * Update tests/models/whisper/test_modeling_whisper.py Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com> * fix style --------- Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for d979cf6 - Browse repository at this point
Copy the full SHA d979cf6View commit details -
Trainer: let generate pick its inputs (#22108)
* Let generate pick its inputs * fix squad seq2seq example
Configuration menu - View commit details
-
Copy full SHA for e16cbe8 - Browse repository at this point
Copy the full SHA e16cbe8View commit details -
Configuration menu - View commit details
-
Copy full SHA for 1c801d6 - Browse repository at this point
Copy the full SHA 1c801d6View commit details -
[trainer] fix bug in grad accum with multiple epochs (#22098)
* [trainer] fix bug in grad accum * comment out debug * fix one-off * rename counter
Configuration menu - View commit details
-
Copy full SHA for 5b85add - Browse repository at this point
Copy the full SHA 5b85addView commit details -
[deepspeed docs] Activation Checkpointing (#22099)
* [deepspeed docs] Activation Checkpointing * Apply suggestions from code review Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update deepspeed.mdx --------- Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 618697e - Browse repository at this point
Copy the full SHA 618697eView commit details -
Remove backend check for torch.compile (#22140)
* Remove backend enforcment for torch.compile * Update error * Update src/transformers/training_args.py Co-authored-by: Stas Bekman <stas00@users.noreply.github.com> * Apply suggestions from code review Co-authored-by: Stas Bekman <stas00@users.noreply.github.com> * Style --------- Co-authored-by: Stas Bekman <stas00@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 3a35937 - Browse repository at this point
Copy the full SHA 3a35937View commit details -
[Safetensors] Add explicit flag to from pretrained (#22083)
* [Safetensors] Add explicit flag to from pretrained * add test * remove @ * Apply suggestions from code review Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com> --------- Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for f780557 - Browse repository at this point
Copy the full SHA f780557View commit details -
Prepare daily CI for torch 2.0.0 (#22135)
Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for ba9e019 - Browse repository at this point
Copy the full SHA ba9e019View commit details -
docs: New terms and updates to glossary (#21982)
* Updated glossary with new terms, added abbreviations for certain terms and merged autoencoding models, autoregressive models and causal language modeling into encoder and decoder models * Update docs/source/en/glossary.mdx Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/en/glossary.mdx Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/en/glossary.mdx Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/en/glossary.mdx Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/en/glossary.mdx Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/en/glossary.mdx Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/en/glossary.mdx Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/en/glossary.mdx Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/en/glossary.mdx Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/en/glossary.mdx Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/en/glossary.mdx Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/en/glossary.mdx Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com> * Added link to 'Pipeline for inference' tutorial * Trigger CI * Update docs/source/en/glossary.mdx Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update docs/source/en/glossary.mdx Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com> * Added entry for self supervised learning, added deleted entries + fixed broken links * Update docs/source/en/glossary.mdx Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com> --------- Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com> Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 101a6cd - Browse repository at this point
Copy the full SHA 101a6cdView commit details
Commits on Mar 14, 2023
-
[🛠️] Fix-whisper-breaking-changes (#21965)
* temp fix * temporary fix * update * fix tests * fixup * update based on reveiew Co-authored-by: Sanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com> * update to fix tests * update docstring --------- Co-authored-by: Sanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 2beabd2 - Browse repository at this point
Copy the full SHA 2beabd2View commit details -
Move
is_pipeline_test_to_skip
to specific model test classes (#21999)* Move `is_pipeline_test_to_skip` to specific model test classes --------- Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 6c2ad00 - Browse repository at this point
Copy the full SHA 6c2ad00View commit details -
* Add ConvNeXt V2 to transformers * TF model is separated from the PR to fix issues
Configuration menu - View commit details
-
Copy full SHA for cdddfbf - Browse repository at this point
Copy the full SHA cdddfbfView commit details -
Update 2 doctest expected values for torch 2.0.0 (#22148)
update values Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for ff88703 - Browse repository at this point
Copy the full SHA ff88703View commit details -
Translation Italian: perf_train_cpu and perf_train_cpu_many (#22151)
* added translated files added perf_train_cpu and perf_train_cpu_many * updated toctree
Configuration menu - View commit details
-
Copy full SHA for 7f5ad6c - Browse repository at this point
Copy the full SHA 7f5ad6cView commit details -
Fix big model inference for T5 models in float16 (#22095)
* Fix big model inference for T5 models in float16 * Apply suggestions from code review Co-authored-by: Younes Belkada <49240599+younesbelkada@users.noreply.github.com> * Style * Trigger CI with latest release --------- Co-authored-by: Younes Belkada <49240599+younesbelkada@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for b45192e - Browse repository at this point
Copy the full SHA b45192eView commit details -
Create MaskedImageCompletionOutput and fix ViT docs (#22152)
* create MaskedImageCompletionOutput * fix bugs * fix bugs
Configuration menu - View commit details
-
Copy full SHA for 3b22bfb - Browse repository at this point
Copy the full SHA 3b22bfbView commit details -
to_pil - don't rescale if int and in range 0-255 (#22158)
* Don't rescale if in and in range 0-255 * Raise value error if int values too large * Update tests/test_image_transforms.py * Update tests/test_image_transforms.py
Configuration menu - View commit details
-
Copy full SHA for c6318c3 - Browse repository at this point
Copy the full SHA c6318c3View commit details -
[trainer] add
--optim adamw_torch_fused
for pt-2.0+ (#22144)* [trainer] add --optim adamw_torch_fused * change optim default * deal with non-torch * revert default change; prep; add fp16/amp assert * typo * typo
Configuration menu - View commit details
-
Copy full SHA for 085bf5c - Browse repository at this point
Copy the full SHA 085bf5cView commit details -
Revert "Enforce same behavior as PyTorch 2.0 for older versions" (#22163
Configuration menu - View commit details
-
Copy full SHA for c52c528 - Browse repository at this point
Copy the full SHA c52c528View commit details -
Configuration menu - View commit details
-
Copy full SHA for ebdb185 - Browse repository at this point
Copy the full SHA ebdb185View commit details -
Configuration menu - View commit details
-
Copy full SHA for b7036f4 - Browse repository at this point
Copy the full SHA b7036f4View commit details -
Configuration menu - View commit details
-
Copy full SHA for f732975 - Browse repository at this point
Copy the full SHA f732975View commit details
Commits on Mar 15, 2023
-
Fix: unfinished_sequences with correct device (#22184)
Fix: unfinished_sequences with correct device The original code was causing errors when running torch.jit.trace due to the tensor options being incorrect. I fixed this by using torch.ones to create a tensor with the correct device and dtype. This should resolve the issue with running torch.jit.trace.
Configuration menu - View commit details
-
Copy full SHA for 7b0e2cf - Browse repository at this point
Copy the full SHA 7b0e2cfView commit details -
Revert 22152 MaskedImageCompletionOutput changes (#22187)
Revert changes
Configuration menu - View commit details
-
Copy full SHA for 7376814 - Browse repository at this point
Copy the full SHA 7376814View commit details -
Regression pipeline device (#22190)
* Fix regression in pipeline when device=-1 is passed * Add regression test
Configuration menu - View commit details
-
Copy full SHA for 42ad693 - Browse repository at this point
Copy the full SHA 42ad693View commit details -
Update BridgeTowerForContrastiveLearning (#22145)
* Use return_loss for BridgeTowerForContrastiveLearning, add example * fix tests * Update example in BridgeTowerForContrastiveLearning * Update test_modeling_bridgetower.py * update model output format * minor update * Update src/transformers/models/bridgetower/modeling_bridgetower.py * make style --------- Co-authored-by: Tiep Le <97980157+tileintel@users.noreply.github.com> Co-authored-by: Tiep Le <tiep.le@intel.com> Co-authored-by: Yih-Dar <2521628+ydshieh@users.noreply.github.com> Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 16121ba - Browse repository at this point
Copy the full SHA 16121baView commit details -
t5 remove data dependency (#22097)
* t5 remove data dependency * make style * make fix-copies --------- Co-authored-by: Prathik Rao <prathikrao@microsoft.com>
Configuration menu - View commit details
-
Copy full SHA for 7c4999e - Browse repository at this point
Copy the full SHA 7c4999eView commit details
Commits on Mar 16, 2023
-
* Deal with torch-tensorrt --------- Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 1c4a9ac - Browse repository at this point
Copy the full SHA 1c4a9acView commit details -
Configuration menu - View commit details
-
Copy full SHA for 1485bd9 - Browse repository at this point
Copy the full SHA 1485bd9View commit details -
Update expected values in
MgpstrModelIntegrationTest
(#22195)Update values Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 52a57f7 - Browse repository at this point
Copy the full SHA 52a57f7View commit details -
Italian Translation of migration.mdx (#22183)
* Tranlstion Italian: migration * Update migration.mdx minor fixes * Update _toctree.yml * Delete migration.mdx * Add italian translation of migration.mdx * Update of migration.mdx translation and toctree
Configuration menu - View commit details
-
Copy full SHA for 09922da - Browse repository at this point
Copy the full SHA 09922daView commit details -
* LLaMA * sharding and docs * tweak * black * inits * ruff * LLAMA_PRETRAINED_CONFIG_ARCHIVE_MAP * init * no checkpoint * docs * ruff * type_vocab_size * tokenizer fixes * tokenizer fixes * Update tokenization_llama.py * Update tokenization_llama.py * Update configuration_llama.py * Update modeling_llama.py * tokenizer add_bos by default * licenses * remove decoder * norms and mlp * rope overhaul * tweaks * black * mention OPT implementation * off-by-one naming * typo * fix * tokenization fix and slicing bug * padding config * cleanup * black * update tests * undo typo * fix vocab caching logic * ruff * docbuilder * attn fix from BlackSamorez * initial feedback * typo * docs * llama case * llama case * load checkpoint docs * comment about tokenizer * tokenizer defaults * clear past_key_values if use_cache=False * last tweaks * last tweaks * last tweaks * last tweaks --------- Co-authored-by: Stella Biderman <stellabiderman@gmail.com>
Configuration menu - View commit details
-
Copy full SHA for 0041be5 - Browse repository at this point
Copy the full SHA 0041be5View commit details -
* LLaMA * sharding and docs * tweak * black * inits * ruff * LLAMA_PRETRAINED_CONFIG_ARCHIVE_MAP * init * no checkpoint * docs * ruff * type_vocab_size * tokenizer fixes * tokenizer fixes * Update tokenization_llama.py * Update tokenization_llama.py * Update configuration_llama.py * Update modeling_llama.py * tokenizer add_bos by default * licenses * remove decoder * norms and mlp * rope overhaul * tweaks * black * mention OPT implementation * off-by-one naming * typo * fix * tokenization fix and slicing bug * padding config * cleanup * black * update tests * undo typo * fix vocab caching logic * ruff * docbuilder * attn fix from BlackSamorez * initial feedback * typo * docs * llama case * llama case * load checkpoint docs * comment about tokenizer * tokenizer defaults * clear past_key_values if use_cache=False * last tweaks * last tweaks * last tweaks * last tweaks --------- Co-authored-by: Stella Biderman <stellabiderman@gmail.com>
Configuration menu - View commit details
-
Copy full SHA for 464d420 - Browse repository at this point
Copy the full SHA 464d420View commit details -
Update tiny model creation script (#22202)
* Update UNCONVERTIBLE_MODEL_ARCHITECTURES * Deal with 2 model tester classes in single test file * Deal with 2 model tester classes in single test file * Deal with 2 model tester classes in single test file * make style and quality --------- Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 4c5c0af - Browse repository at this point
Copy the full SHA 4c5c0afView commit details -
Temporarily fix ONNX model exporting error (#21830)
* Temporarily fix https://github.com/microsoft/onnx-converters-private/issues/143 * Reduced column width * Fix formatting. * Revert "Temporarily fix https://github.com/microsoft/onnx-converters-private/issues/143" This reverts commit 6e95a10. * Fix export error. * Revert "Fix formatting." This reverts commit 8310f60. * Propagated changes made in SwinV2 to Swin2SR
Configuration menu - View commit details
-
Copy full SHA for a88a4da - Browse repository at this point
Copy the full SHA a88a4daView commit details -
[
XGLM
] Addaccelerate
support for XGLM (#22207)* add `accelerate` support for XGLM * fix order
Configuration menu - View commit details
-
Copy full SHA for da3ba3a - Browse repository at this point
Copy the full SHA da3ba3aView commit details -
fixes a typo in WhisperFeatureExtractor docs. (#22208)
* fixes a typo * .
Configuration menu - View commit details
-
Copy full SHA for fb366b9 - Browse repository at this point
Copy the full SHA fb366b9View commit details -
* py38 + torch 2 * increment cache versions --------- Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 5110e57 - Browse repository at this point
Copy the full SHA 5110e57View commit details -
Hotfix for natten issue with torch 2.0.0 on CircleCI (#22218)
fix Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 97a3d16 - Browse repository at this point
Copy the full SHA 97a3d16View commit details
Commits on Mar 17, 2023
-
Configuration menu - View commit details
-
Copy full SHA for 33d033d - Browse repository at this point
Copy the full SHA 33d033dView commit details -
fix code example in mgp-str doc (#22219)
Co-authored-by: yue kun <yuekun.wp@alibaba-inc.com>
Configuration menu - View commit details
-
Copy full SHA for af1c864 - Browse repository at this point
Copy the full SHA af1c864View commit details -
Use
dash==2.8.1
for now for daily CI (#22227)Use dash 2.8.1 for now Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for 5321867 - Browse repository at this point
Copy the full SHA 5321867View commit details -
Depth estimation task guide (#22205)
* added doc to toc, auto tip with supported models, mention of task guide in model docs * make style * removed "see also" * minor fix
Configuration menu - View commit details
-
Copy full SHA for 42f8f76 - Browse repository at this point
Copy the full SHA 42f8f76View commit details -
Configuration menu - View commit details
-
Copy full SHA for 0093402 - Browse repository at this point
Copy the full SHA 0093402View commit details -
fix AutoTP in deepspeed could not work for bloom (#22196)
* fix AutoTP in deepspeed could not work for bloom Signed-off-by: Wang, Yi A <yi.a.wang@intel.com> * add a method in BloomModel to build ailib Signed-off-by: Wang, Yi A <yi.a.wang@intel.com> --------- Signed-off-by: Wang, Yi A <yi.a.wang@intel.com>
Configuration menu - View commit details
-
Copy full SHA for 675d2a5 - Browse repository at this point
Copy the full SHA 675d2a5View commit details -
Add LlamaForSequenceClassification (#22209)
* Add LlamaForSequenceClassification * Update src/transformers/models/llama/modeling_llama.py Co-authored-by: Younes Belkada <49240599+younesbelkada@users.noreply.github.com> * Update src/transformers/models/llama/modeling_llama.py Co-authored-by: Younes Belkada <49240599+younesbelkada@users.noreply.github.com> * Add docstring * Add test * Add input embedding getter and setter * Remove dead code --------- Co-authored-by: Younes Belkada <49240599+younesbelkada@users.noreply.github.com>
Configuration menu - View commit details
-
Copy full SHA for f251441 - Browse repository at this point
Copy the full SHA f251441View commit details -
Removed .mdx extension in two links (#22230)
removed .mdx extension
Configuration menu - View commit details
-
Copy full SHA for 314cdf7 - Browse repository at this point
Copy the full SHA 314cdf7View commit details -
fix(docs): fix task guide links in model docs (#22226)
fix(docs): task guide links in model docs
Configuration menu - View commit details
-
Copy full SHA for 074490b - Browse repository at this point
Copy the full SHA 074490bView commit details -
* Add kernel size to NATTEN's QK arguments. The new NATTEN 0.14.5 supports PyTorch 2.0, but also adds an additional argument to the QK operation to allow optional RPBs. This ends up failing NATTEN tests. This commit adds NATTEN back to circleci and adds the arguments to get it working again. * Force NATTEN >= 0.14.5
Configuration menu - View commit details
-
Copy full SHA for 3028b20 - Browse repository at this point
Copy the full SHA 3028b20View commit details -
Configuration menu - View commit details
-
Copy full SHA for bec0756 - Browse repository at this point
Copy the full SHA bec0756View commit details -
Configuration menu - View commit details
-
Copy full SHA for cf601b9 - Browse repository at this point
Copy the full SHA cf601b9View commit details -
[trainer] param count for deepspeed zero3 (#22193)
[trainer] param count for zero3
Configuration menu - View commit details
-
Copy full SHA for 60d51ef - Browse repository at this point
Copy the full SHA 60d51efView commit details