-
Notifications
You must be signed in to change notification settings - Fork 346
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Sync/v4.12.x #255
Sync/v4.12.x #255
Commits on Sep 27, 2021
-
Configuration menu - View commit details
-
Copy full SHA for 11c69b8 - Browse repository at this point
Copy the full SHA 11c69b8View commit details -
Configuration menu - View commit details
-
Copy full SHA for 5e3b4a7 - Browse repository at this point
Copy the full SHA 5e3b4a7View commit details
Commits on Sep 28, 2021
-
Configuration menu - View commit details
-
Copy full SHA for 83d3dc0 - Browse repository at this point
Copy the full SHA 83d3dc0View commit details -
Configuration menu - View commit details
-
Copy full SHA for a21ee1f - Browse repository at this point
Copy the full SHA a21ee1fView commit details
Commits on Sep 29, 2021
-
Configuration menu - View commit details
-
Copy full SHA for aa018a7 - Browse repository at this point
Copy the full SHA aa018a7View commit details -
Keras callback to push to hub each epoch, or after N steps (#13773)
* Keras callback to push to hub each epoch, or after N steps * Reworked the callback to use Repository * Use an Enum for save_strategy * Style pass * Correct type for tokenizer * Update src/transformers/keras_callbacks.py Co-authored-by: Sylvain Gugger <[email protected]> * Update src/transformers/keras_callbacks.py Co-authored-by: Sylvain Gugger <[email protected]> * Update src/transformers/keras_callbacks.py Co-authored-by: Sylvain Gugger <[email protected]> * Update src/transformers/keras_callbacks.py Co-authored-by: Sylvain Gugger <[email protected]> * Update src/transformers/keras_callbacks.py Co-authored-by: Sylvain Gugger <[email protected]> * Update src/transformers/keras_callbacks.py Co-authored-by: Sylvain Gugger <[email protected]> * Adding print message to the final upload * Adding print message to the final upload * Change how we wait for the last process to finish * is_done is a property, not a method, derp * Docstrings and documentation * Style pass * Style edit * Docstring reformat * Docstring rewrite * Replacing print with internal logger Co-authored-by: Sylvain Gugger <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 3a8a801 - Browse repository at this point
Copy the full SHA 3a8a801View commit details -
Configuration menu - View commit details
-
Copy full SHA for a1ea3ad - Browse repository at this point
Copy the full SHA a1ea3adView commit details -
Enable readme link synchronization (#13785)
* Enable readme link synchronization * Style * Reuse regex pattern * Apply suggestions * Update
Configuration menu - View commit details
-
Copy full SHA for 7d84c3a - Browse repository at this point
Copy the full SHA 7d84c3aView commit details -
Fix length of IterableDatasetShard and add test (#13792)
* Fix length of IterableDatasetShard and add test * Add comments
Configuration menu - View commit details
-
Copy full SHA for 63cc5bd - Browse repository at this point
Copy the full SHA 63cc5bdView commit details -
Configuration menu - View commit details
-
Copy full SHA for 2a51b15 - Browse repository at this point
Copy the full SHA 2a51b15View commit details -
Configuration menu - View commit details
-
Copy full SHA for cf4aa35 - Browse repository at this point
Copy the full SHA cf4aa35View commit details -
Configuration menu - View commit details
-
Copy full SHA for 55695df - Browse repository at this point
Copy the full SHA 55695dfView commit details -
[docs/gpt-j] addd instructions for how minimize CPU RAM usage (#13795)
* add a note about tokenizer * add tips to load model is less RAM * fix link * fix more links
Configuration menu - View commit details
-
Copy full SHA for bf6118e - Browse repository at this point
Copy the full SHA bf6118eView commit details -
[examples
run_glue.py
] missing requirementsscipy
,sklearn
(#13……768) * missing requirement * list both
Configuration menu - View commit details
-
Copy full SHA for b90096f - Browse repository at this point
Copy the full SHA b90096fView commit details
Commits on Sep 30, 2021
-
[examples/flax] use Repository API for push_to_hub (#13672)
* use Repository for push_to_hub * update readme * update other flax scripts * update readme * update qa example * fix push_to_hub call * fix typo * fix more typos * update readme * use abosolute path to get repo name * fix glue script
Configuration menu - View commit details
-
Copy full SHA for 7db2a79 - Browse repository at this point
Copy the full SHA 7db2a79View commit details -
Configuration menu - View commit details
-
Copy full SHA for 269c3d1 - Browse repository at this point
Copy the full SHA 269c3d1View commit details -
Configuration menu - View commit details
-
Copy full SHA for 5f25855 - Browse repository at this point
Copy the full SHA 5f25855View commit details -
Configuration menu - View commit details
-
Copy full SHA for e1d1c7c - Browse repository at this point
Copy the full SHA e1d1c7cView commit details -
Add MultiBERTs conversion script (#13077)
* Init multibert checkpoint conversion script * Rename conversion script * Fix MultiBerts Conversion Script * Apply suggestions from code review Co-authored-by: NielsRogge <[email protected]> Co-authored-by: Patrick von Platen <[email protected]> Co-authored-by: NielsRogge <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 9a9805f - Browse repository at this point
Copy the full SHA 9a9805fView commit details -
Configuration menu - View commit details
-
Copy full SHA for 44eb8bd - Browse repository at this point
Copy the full SHA 44eb8bdView commit details -
* update * add to docs and init * make fix-copies
Configuration menu - View commit details
-
Copy full SHA for 41436d3 - Browse repository at this point
Copy the full SHA 41436d3View commit details -
Configuration menu - View commit details
-
Copy full SHA for 8bbb53e - Browse repository at this point
Copy the full SHA 8bbb53eView commit details
Commits on Oct 1, 2021
-
Fix warning situation: UserWarning: max_length is ignored when paddin…
…g=True" (#13829) * Removed wrong warning * Raise a warning when `max_length` is given with wrong `truncation` * Update the error message * Update the warning message Co-authored-by: Sylvain Gugger <[email protected]> Co-authored-by: Sylvain Gugger <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 90f980e - Browse repository at this point
Copy the full SHA 90f980eView commit details -
Configuration menu - View commit details
-
Copy full SHA for c411372 - Browse repository at this point
Copy the full SHA c411372View commit details -
[Examples] Add an official audio classification example (#13722)
* Restore broken merge * Additional args, DDP, remove CommonLanguage * Update examples for V100, add training results * Style * Apply suggestions from code review Co-authored-by: Sylvain Gugger <[email protected]> * Remove custom datasets for simplicity, apply suggestions from code review * Add the attention_mask flag, reorganize README Co-authored-by: Sylvain Gugger <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 4213728 - Browse repository at this point
Copy the full SHA 4213728View commit details -
Bart: check if decoder_inputs_embeds is set (#13800)
In BartForConditionalGeneration.forward, if labels are provided, decoder_input_ids are set to the labels shifted to the right. This is problematic: if decoder_inputs_embeds is also set, the call to self.model, which eventually gets to BartDecoder.forward, will raise an error. The fix is quite simple, similar to what is there already in BartModel.forward. Mainly, we should not compute decoder_input_ids if decoder_inputs_embeds is provided. Co-authored-by: Silviu Vlad Oprea <[email protected]>
Silviu Oprea and Silviu Vlad Oprea authoredOct 1, 2021 Configuration menu - View commit details
-
Copy full SHA for 707f7eb - Browse repository at this point
Copy the full SHA 707f7ebView commit details -
Configuration menu - View commit details
-
Copy full SHA for bcc3f7b - Browse repository at this point
Copy the full SHA bcc3f7bView commit details
Commits on Oct 4, 2021
-
Configuration menu - View commit details
-
Copy full SHA for de94835 - Browse repository at this point
Copy the full SHA de94835View commit details -
Configuration menu - View commit details
-
Copy full SHA for 955fd4f - Browse repository at this point
Copy the full SHA 955fd4fView commit details -
Add Mistral GPT-2 Stability Tweaks (#13573)
* Add layer-wise scaling * Add reorder & upcasting argument * Add OpenAI GPT-2 weight initialization scheme * start `layer_idx` count at zero for consistency * disentangle attn and reordered and upscaled attn function * rename `scale_attn_by_layer` to `scale_attn_by_layer_id` * make autocast from amp compatible with pytorch<1.6 * fix docstring * style fixes * Add fixes from PR feedback, style tweaks * Fix doc whitespace * Reformat * First pass scale_attn_by_layer_idx and reorder_and_upcast_attn tests * Rename scale_attn_by_layer_idx, add tip * Remove extra newline * add test for weight initialization * update code format * add assert check weights are fp32 * remove assert * Fix incorrect merge * Fix shape mismatch in baddbmm * Add generation test for Mistral flags Co-authored-by: leandro <[email protected]> Co-authored-by: Keshav Santhanam <[email protected]> Co-authored-by: J38 <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 3a8de58 - Browse repository at this point
Copy the full SHA 3a8de58View commit details -
Fix broken link to distill models in docs (#13848)
* Fix broken link to distill models * Missing symbol * Fix spaces
Configuration menu - View commit details
-
Copy full SHA for 6c08840 - Browse repository at this point
Copy the full SHA 6c08840View commit details -
✨ update image classification example (#13824)
* ✨ update image classification example * 📌 update reqs
Configuration menu - View commit details
-
Copy full SHA for cc0a415 - Browse repository at this point
Copy the full SHA cc0a415View commit details -
Update no_* argument (HfArgumentParser) (#13865)
* update no_* argument Changes the order so that the no_* argument is created after the original argument AND sets the default for this no_* argument to False * import copy * update test * make style * Use kwargs to set default=False * make style
Bram Vanroy authoredOct 4, 2021 Configuration menu - View commit details
-
Copy full SHA for 12b4d66 - Browse repository at this point
Copy the full SHA 12b4d66View commit details
Commits on Oct 5, 2021
-
Configuration menu - View commit details
-
Copy full SHA for 7051b89 - Browse repository at this point
Copy the full SHA 7051b89View commit details -
Configuration menu - View commit details
-
Copy full SHA for 7079a99 - Browse repository at this point
Copy the full SHA 7079a99View commit details -
Fix: save checkpoint after each epoch and push checkpoint to the hub …
…(#13872) Co-authored-by: ydshieh <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for a6ea244 - Browse repository at this point
Copy the full SHA a6ea244View commit details -
Fixing empty prompts for text-generation when BOS exists. (#13859)
* Fixing empty prompts for text-generation when BOS exists. * Fixing odd case with Pegasus. * Fixing Bert is Assertion Error.
Configuration menu - View commit details
-
Copy full SHA for 3a9c0f2 - Browse repository at this point
Copy the full SHA 3a9c0f2View commit details -
Improve error message when loading models from Hub (#13836)
* Improve error message when loading models from Hub * Adjust error message wording
Configuration menu - View commit details
-
Copy full SHA for 46efc58 - Browse repository at this point
Copy the full SHA 46efc58View commit details -
Initial support for symbolic tracing with torch.fx allowing dynamic a…
…xes (#13579) * Symbolic trace dynamic axes support for BERT like models (albert, bert, distilbert, mobilebert, electra, megatron-bert) * Sanity checks before tracing that make sure the model to trace is supported * Adapted to PyTorch 1.9 Co-authored-by: Michael Benayoun <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for d4e4efc - Browse repository at this point
Copy the full SHA d4e4efcView commit details -
Allow dataset to be an optional argument for (Distributed)LengthGroup…
…edSampler (#13820) * Allow dataset to be an optional argument for (Distributed)LengthGroupedSampler * Fix
Configuration menu - View commit details
-
Copy full SHA for 1b74af7 - Browse repository at this point
Copy the full SHA 1b74af7View commit details -
Fixing question-answering with long contexts (#13873)
* Tmp. * Fixing BC for question answering with long context. * Capping model_max_length to avoid tf overflow. * Bad workaround bugged roberta. * Fixing name.
Configuration menu - View commit details
-
Copy full SHA for 0ddadbf - Browse repository at this point
Copy the full SHA 0ddadbfView commit details -
Configuration menu - View commit details
-
Copy full SHA for f099249 - Browse repository at this point
Copy the full SHA f099249View commit details -
Configuration menu - View commit details
-
Copy full SHA for 7af7d7c - Browse repository at this point
Copy the full SHA 7af7d7cView commit details
Commits on Oct 6, 2021
-
Update parallelism.md (#13892)
* Update parallelism.md * Update docs/source/parallelism.md Co-authored-by: Stas Bekman <[email protected]> * Update docs/source/parallelism.md Co-authored-by: Stas Bekman <[email protected]> * Update docs/source/parallelism.md Co-authored-by: Stas Bekman <[email protected]> * Update docs/source/parallelism.md Co-authored-by: Stas Bekman <[email protected]> * Update docs/source/parallelism.md Co-authored-by: Stas Bekman <[email protected]> * Update docs/source/parallelism.md Co-authored-by: Stas Bekman <[email protected]> Co-authored-by: Stas Bekman <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 36fc401 - Browse repository at this point
Copy the full SHA 36fc401View commit details -
Configuration menu - View commit details
-
Copy full SHA for 7d83655 - Browse repository at this point
Copy the full SHA 7d83655View commit details -
Fixing GPU for token-classification in a better way. (#13856)
Co-authored-by: Pierre Snell <[email protected]> Co-authored-by: Pierre Snell <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for e7b16f3 - Browse repository at this point
Copy the full SHA e7b16f3View commit details -
Update FSNER code in examples->research_projects->fsner (#13864)
* Add example use of few-shot named entity recognition model in research_projects folder. * Apply suggestions from code review Co-authored-by: NielsRogge <[email protected]> * Update fsner example README.md. - Change wrong import FSNERTokenizerWrapper to FSNERTokenizerUtils in the example code - Add a link to the model identifier * Update examples/research_projects/fsner/src/fsner/model.py Fix spelling mistake in the default parameter of pretrained model name. Co-authored-by: Stefan Schweter <[email protected]> * Add example use of few-shot named entity recognition model in research_projects folder. * Apply suggestions from code review Co-authored-by: NielsRogge <[email protected]> * Update fsner example README.md. - Change wrong import FSNERTokenizerWrapper to FSNERTokenizerUtils in the example code - Add a link to the model identifier * Update examples/research_projects/fsner/src/fsner/model.py Fix spelling mistake in the default parameter of pretrained model name. Co-authored-by: Stefan Schweter <[email protected]> * Run Checking/fixing examples/flax/language-modeling/run_clm_flax.py examples/flax/question-answering/run_qa.py examples/flax/question-answering/utils_qa.py examples/flax/token-classification/run_flax_ner.py examples/legacy/multiple_choice/utils_multiple_choice.py examples/legacy/seq2seq/seq2seq_trainer.py examples/legacy/token-classification/utils_ner.py examples/pytorch/image-classification/run_image_classification.py examples/pytorch/language-modeling/run_clm.py examples/pytorch/language-modeling/run_clm_no_trainer.py examples/pytorch/language-modeling/run_mlm.py examples/pytorch/language-modeling/run_mlm_no_trainer.py examples/pytorch/language-modeling/run_plm.py examples/pytorch/multiple-choice/run_swag.py examples/pytorch/multiple-choice/run_swag_no_trainer.py examples/pytorch/question-answering/run_qa.py examples/pytorch/question-answering/run_qa_beam_search.py examples/pytorch/question-answering/run_qa_beam_search_no_trainer.py examples/pytorch/question-answering/run_qa_no_trainer.py examples/pytorch/summarization/run_summarization.py examples/pytorch/summarization/run_summarization_no_trainer.py examples/pytorch/test_examples.py examples/pytorch/text-classification/run_glue.py examples/pytorch/text-classification/run_glue_no_trainer.py examples/pytorch/text-classification/run_xnli.py examples/pytorch/token-classification/run_ner.py examples/pytorch/token-classification/run_ner_no_trainer.py examples/pytorch/translation/run_translation.py examples/pytorch/translation/run_translation_no_trainer.py examples/research_projects/adversarial/utils_hans.py examples/research_projects/distillation/grouped_batch_sampler.py examples/research_projects/fsner/setup.py examples/research_projects/fsner/src/fsner/__init__.py examples/research_projects/fsner/src/fsner/model.py examples/research_projects/fsner/src/fsner/tokenizer_utils.py examples/research_projects/jax-projects/big_bird/evaluate.py examples/research_projects/jax-projects/hybrid_clip/run_hybrid_clip.py examples/tensorflow/language-modeling/run_clm.py examples/tensorflow/multiple-choice/run_swag.py examples/tensorflow/question-answering/run_qa.py examples/tensorflow/summarization/run_summarization.py examples/tensorflow/text-classification/run_glue.py examples/tensorflow/translation/run_translation.py src/transformers/__init__.py src/transformers/commands/add_new_model.py src/transformers/configuration_utils.py src/transformers/convert_slow_tokenizer.py src/transformers/data/__init__.py src/transformers/data/data_collator.py src/transformers/data/datasets/glue.py src/transformers/data/datasets/language_modeling.py src/transformers/data/datasets/squad.py src/transformers/deepspeed.py src/transformers/dependency_versions_table.py src/transformers/feature_extraction_sequence_utils.py src/transformers/file_utils.py src/transformers/generation_flax_utils.py src/transformers/generation_logits_process.py src/transformers/generation_tf_utils.py src/transformers/generation_utils.py src/transformers/integrations.py src/transformers/modelcard.py src/transformers/modeling_flax_utils.py src/transformers/modeling_outputs.py src/transformers/modeling_tf_utils.py src/transformers/modeling_utils.py src/transformers/models/__init__.py src/transformers/models/albert/__init__.py src/transformers/models/albert/modeling_albert.py src/transformers/models/albert/modeling_flax_albert.py src/transformers/models/albert/tokenization_albert_fast.py src/transformers/models/auto/__init__.py src/transformers/models/auto/auto_factory.py src/transformers/models/auto/configuration_auto.py src/transformers/models/auto/dynamic.py src/transformers/models/auto/feature_extraction_auto.py src/transformers/models/auto/modeling_auto.py src/transformers/models/auto/modeling_flax_auto.py src/transformers/models/auto/modeling_tf_auto.py src/transformers/models/auto/tokenization_auto.py src/transformers/models/bart/configuration_bart.py src/transformers/models/bart/modeling_bart.py src/transformers/models/bart/modeling_flax_bart.py src/transformers/models/bart/modeling_tf_bart.py src/transformers/models/barthez/tokenization_barthez_fast.py src/transformers/models/beit/__init__.py src/transformers/models/beit/configuration_beit.py src/transformers/models/beit/modeling_beit.py src/transformers/models/beit/modeling_flax_beit.py src/transformers/models/bert/configuration_bert.py src/transformers/models/bert/modeling_bert.py src/transformers/models/bert/modeling_flax_bert.py src/transformers/models/bert_generation/configuration_bert_generation.py src/transformers/models/bert_generation/modeling_bert_generation.py src/transformers/models/big_bird/configuration_big_bird.py src/transformers/models/big_bird/modeling_big_bird.py src/transformers/models/big_bird/modeling_flax_big_bird.py src/transformers/models/big_bird/tokenization_big_bird_fast.py src/transformers/models/bigbird_pegasus/configuration_bigbird_pegasus.py src/transformers/models/bigbird_pegasus/modeling_bigbird_pegasus.py src/transformers/models/blenderbot/configuration_blenderbot.py src/transformers/models/blenderbot/modeling_blenderbot.py src/transformers/models/blenderbot/modeling_tf_blenderbot.py src/transformers/models/blenderbot_small/configuration_blenderbot_small.py src/transformers/models/blenderbot_small/modeling_blenderbot_small.py src/transformers/models/blenderbot_small/modeling_tf_blenderbot_small.py src/transformers/models/byt5/tokenization_byt5.py src/transformers/models/camembert/tokenization_camembert_fast.py src/transformers/models/canine/configuration_canine.py src/transformers/models/canine/modeling_canine.py src/transformers/models/clip/configuration_clip.py src/transformers/models/clip/convert_clip_original_pytorch_to_hf.py src/transformers/models/clip/modeling_clip.py src/transformers/models/clip/modeling_flax_clip.py src/transformers/models/clip/tokenization_clip.py src/transformers/models/convbert/modeling_convbert.py src/transformers/models/ctrl/configuration_ctrl.py src/transformers/models/deberta/modeling_tf_deberta.py src/transformers/models/deberta_v2/__init__.py src/transformers/models/deberta_v2/modeling_deberta_v2.py src/transformers/models/deberta_v2/modeling_tf_deberta_v2.py src/transformers/models/deit/configuration_deit.py src/transformers/models/deit/modeling_deit.py src/transformers/models/detr/configuration_detr.py src/transformers/models/detr/modeling_detr.py src/transformers/models/distilbert/__init__.py src/transformers/models/distilbert/configuration_distilbert.py src/transformers/models/distilbert/modeling_distilbert.py src/transformers/models/distilbert/modeling_flax_distilbert.py src/transformers/models/dpr/configuration_dpr.py src/transformers/models/dpr/modeling_dpr.py src/transformers/models/electra/modeling_electra.py src/transformers/models/electra/modeling_flax_electra.py src/transformers/models/encoder_decoder/__init__.py src/transformers/models/encoder_decoder/modeling_encoder_decoder.py src/transformers/models/encoder_decoder/modeling_flax_encoder_decoder.py src/transformers/models/flaubert/configuration_flaubert.py src/transformers/models/flaubert/modeling_flaubert.py src/transformers/models/fnet/__init__.py src/transformers/models/fnet/configuration_fnet.py src/transformers/models/fnet/convert_fnet_original_flax_checkpoint_to_pytorch.py src/transformers/models/fnet/modeling_fnet.py src/transformers/models/fnet/tokenization_fnet.py src/transformers/models/fnet/tokenization_fnet_fast.py src/transformers/models/fsmt/configuration_fsmt.py src/transformers/models/fsmt/modeling_fsmt.py src/transformers/models/funnel/configuration_funnel.py src/transformers/models/gpt2/__init__.py src/transformers/models/gpt2/configuration_gpt2.py src/transformers/models/gpt2/modeling_flax_gpt2.py src/transformers/models/gpt2/modeling_gpt2.py src/transformers/models/gpt2/modeling_tf_gpt2.py src/transformers/models/gpt_neo/configuration_gpt_neo.py src/transformers/models/gpt_neo/modeling_gpt_neo.py src/transformers/models/gptj/__init__.py src/transformers/models/gptj/configuration_gptj.py src/transformers/models/gptj/modeling_gptj.py src/transformers/models/herbert/tokenization_herbert_fast.py src/transformers/models/hubert/__init__.py src/transformers/models/hubert/configuration_hubert.py src/transformers/models/hubert/convert_hubert_original_s3prl_checkpoint_to_pytorch.py src/transformers/models/hubert/modeling_hubert.py src/transformers/models/hubert/modeling_tf_hubert.py src/transformers/models/ibert/modeling_ibert.py src/transformers/models/layoutlm/__init__.py src/transformers/models/layoutlm/configuration_layoutlm.py src/transformers/models/layoutlm/modeling_layoutlm.py src/transformers/models/layoutlmv2/__init__.py src/transformers/models/layoutlmv2/configuration_layoutlmv2.py src/transformers/models/layoutlmv2/feature_extraction_layoutlmv2.py src/transformers/models/layoutlmv2/modeling_layoutlmv2.py src/transformers/models/layoutlmv2/processing_layoutlmv2.py src/transformers/models/layoutlmv2/tokenization_layoutlmv2.py src/transformers/models/layoutlmv2/tokenization_layoutlmv2_fast.py src/transformers/models/led/configuration_led.py src/transformers/models/led/modeling_led.py src/transformers/models/longformer/modeling_longformer.py src/transformers/models/luke/configuration_luke.py src/transformers/models/luke/modeling_luke.py src/transformers/models/luke/tokenization_luke.py src/transformers/models/lxmert/configuration_lxmert.py src/transformers/models/m2m_100/configuration_m2m_100.py src/transformers/models/m2m_100/modeling_m2m_100.py src/transformers/models/m2m_100/tokenization_m2m_100.py src/transformers/models/marian/configuration_marian.py src/transformers/models/marian/modeling_flax_marian.py src/transformers/models/marian/modeling_marian.py src/transformers/models/marian/modeling_tf_marian.py src/transformers/models/mbart/configuration_mbart.py src/transformers/models/mbart/modeling_flax_mbart.py src/transformers/models/mbart/modeling_mbart.py src/transformers/models/mbart/tokenization_mbart.py src/transformers/models/mbart/tokenization_mbart_fast.py src/transformers/models/mbart50/tokenization_mbart50.py src/transformers/models/mbart50/tokenization_mbart50_fast.py src/transformers/models/megatron_bert/configuration_megatron_bert.py src/transformers/models/megatron_bert/convert_megatron_bert_checkpoint.py src/transformers/models/megatron_bert/modeling_megatron_bert.py src/transformers/models/megatron_gpt2/convert_megatron_gpt2_checkpoint.py src/transformers/models/openai/configuration_openai.py src/transformers/models/pegasus/__init__.py src/transformers/models/pegasus/configuration_pegasus.py src/transformers/models/pegasus/modeling_flax_pegasus.py src/transformers/models/pegasus/modeling_pegasus.py src/transformers/models/pegasus/modeling_tf_pegasus.py src/transformers/models/pegasus/tokenization_pegasus_fast.py src/transformers/models/prophetnet/configuration_prophetnet.py src/transformers/models/prophetnet/modeling_prophetnet.py src/transformers/models/rag/modeling_rag.py src/transformers/models/rag/modeling_tf_rag.py src/transformers/models/reformer/configuration_reformer.py src/transformers/models/reformer/tokenization_reformer_fast.py src/transformers/models/rembert/configuration_rembert.py src/transformers/models/rembert/modeling_rembert.py src/transformers/models/rembert/tokenization_rembert_fast.py src/transformers/models/roberta/modeling_flax_roberta.py src/transformers/models/roberta/modeling_roberta.py src/transformers/models/roberta/modeling_tf_roberta.py src/transformers/models/roformer/configuration_roformer.py src/transformers/models/roformer/modeling_roformer.py src/transformers/models/speech_encoder_decoder/__init__.py src/transformers/models/speech_encoder_decoder/configuration_speech_encoder_decoder.py src/transformers/models/speech_encoder_decoder/convert_speech_to_text_wav2vec2_seq2seq_original_to_pytorch.py src/transformers/models/speech_encoder_decoder/modeling_speech_encoder_decoder.py src/transformers/models/speech_to_text/configuration_speech_to_text.py src/transformers/models/speech_to_text/feature_extraction_speech_to_text.py src/transformers/models/speech_to_text/modeling_speech_to_text.py src/transformers/models/speech_to_text_2/__init__.py src/transformers/models/speech_to_text_2/configuration_speech_to_text_2.py src/transformers/models/speech_to_text_2/modeling_speech_to_text_2.py src/transformers/models/speech_to_text_2/processing_speech_to_text_2.py src/transformers/models/speech_to_text_2/tokenization_speech_to_text_2.py src/transformers/models/splinter/configuration_splinter.py src/transformers/models/splinter/modeling_splinter.py src/transformers/models/t5/configuration_t5.py src/transformers/models/t5/modeling_flax_t5.py src/transformers/models/t5/modeling_t5.py src/transformers/models/t5/modeling_tf_t5.py src/transformers/models/t5/tokenization_t5_fast.py src/transformers/models/tapas/__init__.py src/transformers/models/tapas/configuration_tapas.py src/transformers/models/tapas/convert_tapas_original_tf_checkpoint_to_pytorch.py src/transformers/models/tapas/modeling_tapas.py src/transformers/models/tapas/tokenization_tapas.py src/transformers/models/transfo_xl/configuration_transfo_xl.py src/transformers/models/visual_bert/modeling_visual_bert.py src/transformers/models/vit/configuration_vit.py src/transformers/models/vit/convert_dino_to_pytorch.py src/transformers/models/vit/modeling_flax_vit.py src/transformers/models/vit/modeling_vit.py src/transformers/models/wav2vec2/__init__.py src/transformers/models/wav2vec2/configuration_wav2vec2.py src/transformers/models/wav2vec2/convert_wav2vec2_original_s3prl_checkpoint_to_pytorch.py src/transformers/models/wav2vec2/feature_extraction_wav2vec2.py src/transformers/models/wav2vec2/modeling_flax_wav2vec2.py src/transformers/models/wav2vec2/modeling_wav2vec2.py src/transformers/models/wav2vec2/tokenization_wav2vec2.py src/transformers/models/xlm/configuration_xlm.py src/transformers/models/xlm_roberta/tokenization_xlm_roberta.py src/transformers/models/xlm_roberta/tokenization_xlm_roberta_fast.py src/transformers/models/xlnet/configuration_xlnet.py src/transformers/models/xlnet/tokenization_xlnet_fast.py src/transformers/onnx/convert.py src/transformers/onnx/features.py src/transformers/optimization.py src/transformers/pipelines/__init__.py src/transformers/pipelines/audio_classification.py src/transformers/pipelines/automatic_speech_recognition.py src/transformers/pipelines/base.py src/transformers/pipelines/conversational.py src/transformers/pipelines/feature_extraction.py src/transformers/pipelines/fill_mask.py src/transformers/pipelines/image_classification.py src/transformers/pipelines/object_detection.py src/transformers/pipelines/question_answering.py src/transformers/pipelines/table_question_answering.py src/transformers/pipelines/text2text_generation.py src/transformers/pipelines/text_classification.py src/transformers/pipelines/text_generation.py src/transformers/pipelines/token_classification.py src/transformers/pipelines/zero_shot_classification.py src/transformers/testing_utils.py src/transformers/tokenization_utils.py src/transformers/tokenization_utils_base.py src/transformers/tokenization_utils_fast.py src/transformers/trainer.py src/transformers/trainer_callback.py src/transformers/trainer_pt_utils.py src/transformers/trainer_seq2seq.py src/transformers/trainer_utils.py src/transformers/training_args.py src/transformers/training_args_seq2seq.py src/transformers/utils/dummy_detectron2_objects.py src/transformers/utils/dummy_flax_objects.py src/transformers/utils/dummy_pt_objects.py src/transformers/utils/dummy_tf_objects.py src/transformers/utils/dummy_tokenizers_objects.py src/transformers/utils/dummy_vision_objects.py tests/deepspeed/test_deepspeed.py tests/sagemaker/conftest.py tests/sagemaker/test_multi_node_data_parallel.py tests/test_configuration_auto.py tests/test_configuration_common.py tests/test_data_collator.py tests/test_feature_extraction_auto.py tests/test_feature_extraction_layoutlmv2.py tests/test_feature_extraction_speech_to_text.py tests/test_feature_extraction_wav2vec2.py tests/test_file_utils.py tests/test_modeling_auto.py tests/test_modeling_bart.py tests/test_modeling_beit.py tests/test_modeling_bert.py tests/test_modeling_clip.py tests/test_modeling_common.py tests/test_modeling_convbert.py tests/test_modeling_deit.py tests/test_modeling_distilbert.py tests/test_modeling_encoder_decoder.py tests/test_modeling_flaubert.py tests/test_modeling_flax_albert.py tests/test_modeling_flax_bart.py tests/test_modeling_flax_beit.py tests/test_modeling_flax_distilbert.py tests/test_modeling_flax_encoder_decoder.py tests/test_modeling_flax_gpt2.py tests/test_modeling_flax_gpt_neo.py tests/test_modeling_flax_mt5.py tests/test_modeling_flax_pegasus.py tests/test_modeling_fnet.py tests/test_modeling_gpt2.py tests/test_modeling_gpt_neo.py tests/test_modeling_gptj.py tests/test_modeling_hubert.py tests/test_modeling_layoutlmv2.py tests/test_modeling_pegasus.py tests/test_modeling_rag.py tests/test_modeling_reformer.py tests/test_modeling_speech_encoder_decoder.py tests/test_modeling_speech_to_text.py tests/test_modeling_speech_to_text_2.py tests/test_modeling_tf_auto.py tests/test_modeling_tf_deberta_v2.py tests/test_modeling_tf_hubert.py tests/test_modeling_tf_pytorch.py tests/test_modeling_tf_wav2vec2.py tests/test_modeling_wav2vec2.py tests/test_onnx_v2.py tests/test_pipelines_audio_classification.py tests/test_pipelines_automatic_speech_recognition.py tests/test_pipelines_common.py tests/test_pipelines_conversational.py tests/test_pipelines_feature_extraction.py tests/test_pipelines_fill_mask.py tests/test_pipelines_image_classification.py tests/test_pipelines_object_detection.py tests/test_pipelines_question_answering.py tests/test_pipelines_summarization.py tests/test_pipelines_table_question_answering.py tests/test_pipelines_text2text_generation.py tests/test_pipelines_text_classification.py tests/test_pipelines_text_generation.py tests/test_pipelines_token_classification.py tests/test_pipelines_translation.py tests/test_pipelines_zero_shot.py tests/test_processor_layoutlmv2.py tests/test_processor_wav2vec2.py tests/test_sequence_feature_extraction_common.py tests/test_tokenization_auto.py tests/test_tokenization_byt5.py tests/test_tokenization_canine.py tests/test_tokenization_common.py tests/test_tokenization_fnet.py tests/test_tokenization_layoutlmv2.py tests/test_tokenization_luke.py tests/test_tokenization_mbart.py tests/test_tokenization_mbart50.py tests/test_tokenization_speech_to_text_2.py tests/test_tokenization_t5.py tests/test_tokenization_tapas.py tests/test_tokenization_xlm_roberta.py tests/test_trainer.py tests/test_trainer_distributed.py tests/test_trainer_tpu.py tests/test_utils_check_copies.py utils/check_copies.py utils/check_repo.py utils/notification_service.py utils/release.py utils/tests_fetcher.py python utils/custom_init_isort.py python utils/style_doc.py src/transformers docs/source --max_len 119 running deps_table_update updating src/transformers/dependency_versions_table.py python utils/check_copies.py python utils/check_table.py python utils/check_dummies.py python utils/check_repo.py Checking all models are public. Checking all models are properly tested. Checking all objects are properly documented. Checking all models are in at least one auto class. python utils/check_inits.py python utils/tests_fetcher.py --sanity_check and fix suggested changes. * Run black examples tests src utils isort examples tests src utils Skipped 1 files make autogenerate_code make[1]: Entering directory '/mnt/c/Users/Admin/Desktop/Home/Projects/transformers' running deps_table_update updating src/transformers/dependency_versions_table.py make[1]: Leaving directory '/mnt/c/Users/Admin/Desktop/Home/Projects/transformers' make extra_style_checks make[1]: Entering directory '/mnt/c/Users/Admin/Desktop/Home/Projects/transformers' python utils/custom_init_isort.py python utils/style_doc.py src/transformers docs/source --max_len 119 make[1]: Leaving directory '/mnt/c/Users/Admin/Desktop/Home/Projects/transformers' for reformatting code. * Add installation dependencies for examples/research_projects/fsner. * Add support to pass in variable numbers of examples to FSNER model. * Retrieve start_token_id and end_token_id from tokenizer instead of hardcoding in the FSNER model. * Run black examples tests src utils isort examples tests src utils Skipped 1 files make autogenerate_code make[1]: Entering directory '/home/saif/transformers' running deps_table_update updating src/transformers/dependency_versions_table.py make[1]: Leaving directory '/home/saif/transformers' make extra_style_checks make[1]: Entering directory '/home/saif/transformers' python utils/custom_init_isort.py python utils/style_doc.py src/transformers docs/source --max_len 119 make[1]: Leaving directory '/home/saif/transformers' for FSNER * Update FSNER readme.md with a header image. * Update FSNER readme Co-authored-by: NielsRogge <[email protected]> Co-authored-by: Stefan Schweter <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 155b230 - Browse repository at this point
Copy the full SHA 155b230View commit details -
Configuration menu - View commit details
-
Copy full SHA for 9f58bec - Browse repository at this point
Copy the full SHA 9f58becView commit details -
Configuration menu - View commit details
-
Copy full SHA for 013bdc6 - Browse repository at this point
Copy the full SHA 013bdc6View commit details -
Configuration menu - View commit details
-
Copy full SHA for dac7798 - Browse repository at this point
Copy the full SHA dac7798View commit details -
T5ForConditionalGeneration: enabling using past_key_values and labels…
… in training (#13805) * enabling using past_key_values together with labels when training in T5ForConditionalGeneration * test * Enable past_key_values in T5ForconditionalGeneration while training. * delete comments
Configuration menu - View commit details
-
Copy full SHA for aea7c5b - Browse repository at this point
Copy the full SHA aea7c5bView commit details -
Fix trainer logging_nan_inf_filter in torch_xla mode (#13896)
* Fix logging_nan_inf_filter in torch_xla mode * Update src/transformers/trainer.py Co-authored-by: Sylvain Gugger <[email protected]> * Fix format Co-authored-by: Sylvain Gugger <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 77770ec - Browse repository at this point
Copy the full SHA 77770ecView commit details -
Configuration menu - View commit details
-
Copy full SHA for 8f2c07d - Browse repository at this point
Copy the full SHA 8f2c07dView commit details -
Configuration menu - View commit details
-
Copy full SHA for 5d390e9 - Browse repository at this point
Copy the full SHA 5d390e9View commit details -
Configuration menu - View commit details
-
Copy full SHA for 5be59a3 - Browse repository at this point
Copy the full SHA 5be59a3View commit details
Commits on Oct 7, 2021
-
Configuration menu - View commit details
-
Copy full SHA for 0d309ce - Browse repository at this point
Copy the full SHA 0d309ceView commit details -
Add an example of exporting BartModel + BeamSearch to ONNX module. (#…
…13765) * Add all example files. * Reformat files by black. * Style. * Remove unused imports. Co-authored-by: Morgan Funtowicz <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 279ce5b - Browse repository at this point
Copy the full SHA 279ce5bView commit details -
#12789 Replace assert statements with exceptions (#13909)
* #12789 Replace assert statements with exceptions * fix-copies: made copy changes to utils_qa.py in examples/pytorch/question-answering and examples/tensorflow/question-answering * minor refactor for clarity
Configuration menu - View commit details
-
Copy full SHA for 319beb6 - Browse repository at this point
Copy the full SHA 319beb6View commit details -
Configuration menu - View commit details
-
Copy full SHA for 57420b1 - Browse repository at this point
Copy the full SHA 57420b1View commit details -
[Wav2Vec2] Fix mask_feature_prob (#13921)
* up * overwrite hubert
Configuration menu - View commit details
-
Copy full SHA for 0f5488f - Browse repository at this point
Copy the full SHA 0f5488fView commit details -
Add missing character (#13922)
Mishig Davaadorj authoredOct 7, 2021 Configuration menu - View commit details
-
Copy full SHA for 5f34163 - Browse repository at this point
Copy the full SHA 5f34163View commit details -
Fix incorrect output shapes for TF/PT LED (#13882)
* Fix issues with LED model * Style pass * Bugfixes * correct attentions as well Co-authored-by: Patrick von Platen <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 61cf2ea - Browse repository at this point
Copy the full SHA 61cf2eaView commit details -
[trainer] memory metrics: add memory at the start report (#13915)
* [trainer] memory metrics: add memory at start * fix for no-gpu
Configuration menu - View commit details
-
Copy full SHA for be71ac3 - Browse repository at this point
Copy the full SHA be71ac3View commit details
Commits on Oct 8, 2021
-
Image Segmentation pipeline (#13828)
* Implement img seg pipeline * Update src/transformers/pipelines/image_segmentation.py Co-authored-by: NielsRogge <[email protected]> * Update src/transformers/pipelines/image_segmentation.py Co-authored-by: NielsRogge <[email protected]> * Update output shape with individual masks * Rm dev change * Remove loops in test Co-authored-by: NielsRogge <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 026866d - Browse repository at this point
Copy the full SHA 026866dView commit details -
Adding support for tokens being suffixes or part of each other. (#13918)
* Adding support for tokens being suffixes or part of each other. * Better test name.
Configuration menu - View commit details
-
Copy full SHA for d70919e - Browse repository at this point
Copy the full SHA d70919eView commit details -
Adds
PreTrainedModel.framework
attribute (#13817)* Added `framework` attribute * Update modeling_utils.py * Update modeling_flax_utils.py * Update modeling_tf_utils.py * Update modeling_utils.py * Update modeling_tf_utils.py * Update modeling_tf_utils.py * Update modeling_flax_utils.py * Update modeling_tf_utils.py * Update modeling_utils.py * Update modeling_utils.py * Update modeling_tf_utils.py * Update modeling_flax_utils.py * string -> str * Update modeling_tf_utils.py * string -> str * fixup * make flake happy Co-authored-by: patil-suraj <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for de34481 - Browse repository at this point
Copy the full SHA de34481View commit details -
Configuration menu - View commit details
-
Copy full SHA for 23ee06e - Browse repository at this point
Copy the full SHA 23ee06eView commit details -
Configuration menu - View commit details
-
Copy full SHA for 5a1b5e4 - Browse repository at this point
Copy the full SHA 5a1b5e4View commit details -
[Generation] Fix max_new_tokens (#13919)
* up * Update src/transformers/generation_stopping_criteria.py * finish
Configuration menu - View commit details
-
Copy full SHA for c8b0761 - Browse repository at this point
Copy the full SHA c8b0761View commit details -
Configuration menu - View commit details
-
Copy full SHA for cb911e5 - Browse repository at this point
Copy the full SHA cb911e5View commit details -
Configuration menu - View commit details
-
Copy full SHA for 9e15b51 - Browse repository at this point
Copy the full SHA 9e15b51View commit details -
Configuration menu - View commit details
-
Copy full SHA for 3e21852 - Browse repository at this point
Copy the full SHA 3e21852View commit details -
Configuration menu - View commit details
-
Copy full SHA for 46dfe99 - Browse repository at this point
Copy the full SHA 46dfe99View commit details -
* Update bug-report.md * Update .github/ISSUE_TEMPLATE/bug-report.md Co-authored-by: Suraj Patil <[email protected]> * Update .github/ISSUE_TEMPLATE/bug-report.md Co-authored-by: Suraj Patil <[email protected]> * Update .github/ISSUE_TEMPLATE/bug-report.md Co-authored-by: NielsRogge <[email protected]> * Update .github/ISSUE_TEMPLATE/bug-report.md Co-authored-by: Anton Lozhkov <[email protected]> Co-authored-by: Suraj Patil <[email protected]> Co-authored-by: NielsRogge <[email protected]> Co-authored-by: Anton Lozhkov <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 239bd61 - Browse repository at this point
Copy the full SHA 239bd61View commit details
Commits on Oct 9, 2021
-
fix issue 13904 -attribute does not exist- by change self_.mapping to…
… self._model_mapping (#13942)
Configuration menu - View commit details
-
Copy full SHA for 91758e3 - Browse repository at this point
Copy the full SHA 91758e3View commit details
Commits on Oct 11, 2021
-
Raise ValueError instead of asserts in src/transformers/benchmark/ben…
…chmark.py (#13951) * Raise ValueError exception instead of assert * Remove f unnecessary f-strings * Remove unused f-strings
Configuration menu - View commit details
-
Copy full SHA for 3c0c699 - Browse repository at this point
Copy the full SHA 3c0c699View commit details -
Honor existing attention mask in tokenzier.pad (#13926)
* Honor existing attention mask in tokenzier.pad * Fix initialization of attention mask * Roll the implem on all subclasses * Fix tests
Configuration menu - View commit details
-
Copy full SHA for 4a18337 - Browse repository at this point
Copy the full SHA 4a18337View commit details -
[Gradient checkpoining] Correct disabling
find_unused_parameters
in…… Trainer when gradient checkpointing is enabled (#13961) * up * correct test
Configuration menu - View commit details
-
Copy full SHA for dca6796 - Browse repository at this point
Copy the full SHA dca6796View commit details -
Configuration menu - View commit details
-
Copy full SHA for 6e4c8f6 - Browse repository at this point
Copy the full SHA 6e4c8f6View commit details -
Configuration menu - View commit details
-
Copy full SHA for e1bb2eb - Browse repository at this point
Copy the full SHA e1bb2ebView commit details -
Configuration menu - View commit details
-
Copy full SHA for 708ffff - Browse repository at this point
Copy the full SHA 708ffffView commit details -
Configuration menu - View commit details
-
Copy full SHA for 32634bc - Browse repository at this point
Copy the full SHA 32634bcView commit details -
Configuration menu - View commit details
-
Copy full SHA for 64743d0 - Browse repository at this point
Copy the full SHA 64743d0View commit details -
Replace assert by ValueError of src/transformers/models/electra/model…
…ing_{electra,tf_electra}.py and all other models that had copies (#13955) * Replace all assert by ValueError in src/transformers/models/electra * Reformat with black to pass check_code_quality test * Change some assert to ValueError of modeling_bert & modeling_tf_albert * Change some assert in multiples models * Change multiples models assertion to ValueError in order to validate check_code_style test and models template test. * Black reformat * Change some more asserts in multiples models * Change assert to ValueError in modeling_layoutlm.py to fix copy error in code_style_check * Add proper message to ValueError in modeling_tf_albert.py Co-authored-by: Sylvain Gugger <[email protected]> * Simplify logic in models/bert/modeling_bert.py Co-authored-by: Sylvain Gugger <[email protected]> * Add ValueError message to models/convbert/modeling_tf_convbert.py Co-authored-by: Sylvain Gugger <[email protected]> * Add error message for ValueError to modeling_tf_electra.py Co-authored-by: Sylvain Gugger <[email protected]> * Simplify logic in models/tapas/modeling_tapas.py Co-authored-by: Sylvain Gugger <[email protected]> * Simplify logic in models/electra/modeling_electra.py Co-authored-by: Sylvain Gugger <[email protected]> * Add ValueError message in src/transformers/models/bert/modeling_tf_bert.py Co-authored-by: Sylvain Gugger <[email protected]> * Simplify logic in src/transformers/models/rembert/modeling_rembert.py Co-authored-by: Sylvain Gugger <[email protected]> * Simplify logic in src/transformers/models/albert/modeling_albert.py Co-authored-by: Sylvain Gugger <[email protected]> Co-authored-by: Sylvain Gugger <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 3499728 - Browse repository at this point
Copy the full SHA 3499728View commit details -
[Speech Examples] Add pytorch speech pretraining (#13877)
* adapt wav2vec2 * add example * add files * adapt * remove bogus file * Apply suggestions from code review * adapt files more * upload changes * del old files * up * up * up * up * up * correct gradient checkpoitning * add readme * finish * finish * up * more fixes * up * up * add demo run to readme * up
Configuration menu - View commit details
-
Copy full SHA for d45fc7d - Browse repository at this point
Copy the full SHA d45fc7dView commit details
Commits on Oct 12, 2021
-
Configuration menu - View commit details
-
Copy full SHA for 990de2c - Browse repository at this point
Copy the full SHA 990de2cView commit details -
Configuration menu - View commit details
-
Copy full SHA for 85d69a7 - Browse repository at this point
Copy the full SHA 85d69a7View commit details -
Specify im-seg mask greyscole mode (#13974)
Mishig Davaadorj authoredOct 12, 2021 Configuration menu - View commit details
-
Copy full SHA for 11c043d - Browse repository at this point
Copy the full SHA 11c043dView commit details -
[Wav2Vec2] Make sure tensors are always bool for mask_indices (#13977)
* correct long to bool * up * correct code
Configuration menu - View commit details
-
Copy full SHA for 58bf882 - Browse repository at this point
Copy the full SHA 58bf882View commit details -
Fixing the lecture values by making sure defaults are not changed (#1…
…3976) 384 // 4 < 128 would break `doc_stride`.
Configuration menu - View commit details
-
Copy full SHA for 26b6ef7 - Browse repository at this point
Copy the full SHA 26b6ef7View commit details -
Add TFEncoderDecoderModel + Add cross-attention to some TF models (#1…
…3222) * Add cross attentions to TFGPT2Model * Add TFEncoderDecoderModel * Add TFBaseModelOutputWithPoolingAndCrossAttentions * Add cross attentions to TFBertModel * Fix past or past_key_values argument issue * Fix generation * Fix save and load * Add some checks and comments * Clean the code that deals with past keys/values * Add kwargs to processing_inputs * Add serving_output to TFEncoderDecoderModel * Some cleaning + fix use_cache value issue * Fix tests + add bert2bert/bert2gpt2 tests * Fix more tests * Ignore crossattention.bias when loading GPT2 weights into TFGPT2 * Fix return_dict_in_generate in tf generation * Fix is_token_logit_eos_token bug in tf generation * Finalize the tests after fixing some bugs * Fix another is_token_logit_eos_token bug in tf generation * Add/Update docs * Add TFBertEncoderDecoderModelTest * Clean test script * Add TFEncoderDecoderModel to the library * Add cross attentions to TFRobertaModel * Add TFRobertaEncoderDecoderModelTest * make style * Change the way of position_ids computation * bug fix * Fix copies in tf_albert * Remove some copied from and apply some fix-copies * Remove some copied * Add cross attentions to some other TF models * Remove encoder_hidden_states from TFLayoutLMModel.call for now * Make style * Fix TFRemBertForCausalLM * Revert the change to longformer + Remove copies * Revert the change to albert and convbert + Remove copies * make quality * make style * Add TFRembertEncoderDecoderModelTest * make quality and fix-copies * test TFRobertaForCausalLM * Fixes for failed tests * Fixes for failed tests * fix more tests * Fixes for failed tests * Fix Auto mapping order * Fix TFRemBertEncoder return value * fix tf_rembert * Check copies are OK * Fix missing TFBaseModelOutputWithPastAndCrossAttentions is not defined * Add TFEncoderDecoderModelSaveLoadTests * fix tf weight loading * check the change of use_cache * Revert the change * Add missing test_for_causal_lm for TFRobertaModelTest * Try cleaning past * fix _reorder_cache * Revert some files to original versions * Keep as many copies as possible * Apply suggested changes - Use raise ValueError instead of assert * Move import to top * Fix wrong require_torch * Replace more assert by raise ValueError * Add test_pt_tf_model_equivalence (the test won't pass for now) * add test for loading/saving * finish * finish * Remove test_pt_tf_model_equivalence * Update tf modeling template * Remove pooling, added in the prev. commit, from MainLayer * Update tf modeling test template * Move inputs["use_cache"] = False to modeling_tf_utils.py * Fix torch.Tensor in the comment * fix use_cache * Fix missing use_cache in ElectraConfig * Add a note to from_pretrained * Fix style * Change test_encoder_decoder_save_load_from_encoder_decoder_from_pt * Fix TFMLP (in TFGPT2) activation issue * Fix None past_key_values value in serving_output * Don't call get_encoderdecoder_model in TFEncoderDecoderModelTest.test_configuration_tie until we have a TF checkpoint on Hub * Apply review suggestions - style for cross_attns in serving_output * Apply review suggestions - change assert + docstrings * break the error message to respect the char limit * deprecate the argument past * fix docstring style * Update the encoder-decoder rst file * fix Unknown interpreted text role "method" * fix typo Co-authored-by: ydshieh <[email protected]> Co-authored-by: Patrick von Platen <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 8b240a0 - Browse repository at this point
Copy the full SHA 8b240a0View commit details -
Configuration menu - View commit details
-
Copy full SHA for 61f6426 - Browse repository at this point
Copy the full SHA 61f6426View commit details
Commits on Oct 13, 2021
-
Add TrOCR + VisionEncoderDecoderModel (#13874)
* First draft * Update self-attention of RoBERTa as proposition * Improve conversion script * Add TrOCR decoder-only model * More improvements * Make forward pass with pretrained weights work * More improvements * Some more improvements * More improvements * Make conversion work * Clean up print statements * Add documentation, processor * Add test files * Small improvements * Some more improvements * Make fix-copies, improve docs * Make all vision encoder decoder model tests pass * Make conversion script support other models * Update URL for OCR image * Update conversion script * Fix style & quality * Add support for the large-printed model * Fix some issues * Add print statement for debugging * Add print statements for debugging * Make possible fix for sinusoidal embedding * Further debugging * Potential fix v2 * Add more print statements for debugging * Add more print statements for debugging * Deubg more * Comment out print statements * Make conversion of large printed model possible, address review comments * Make it possible to convert the stage1 checkpoints * Clean up code, apply suggestions from code review * Apply suggestions from code review, use Microsoft models in tests * Rename encoder_hidden_size to cross_attention_hidden_size * Improve docs
Configuration menu - View commit details
-
Copy full SHA for 408b2d2 - Browse repository at this point
Copy the full SHA 408b2d2View commit details
Commits on Oct 14, 2021
-
Remove wrong model_args supplied (#13937)
* Remove wrong model_args of config.from_pretrained * Fix tf & flax
Configuration menu - View commit details
-
Copy full SHA for 51ee20f - Browse repository at this point
Copy the full SHA 51ee20fView commit details -
Configuration menu - View commit details
-
Copy full SHA for 5b6bd4e - Browse repository at this point
Copy the full SHA 5b6bd4eView commit details -
Configuration menu - View commit details
-
Copy full SHA for cc36064 - Browse repository at this point
Copy the full SHA cc36064View commit details -
Configuration menu - View commit details
-
Copy full SHA for a5be954 - Browse repository at this point
Copy the full SHA a5be954View commit details -
Configuration menu - View commit details
-
Copy full SHA for 0ef61d3 - Browse repository at this point
Copy the full SHA 0ef61d3View commit details -
Add strong test for configuration attributes (#14000)
* Add strong test for configuration attributes * Add fake modif to trigger all tests * Add a better fake modif * Ignore is_encoder_decoder * Fix faulty configs * Remove fake modif
Configuration menu - View commit details
-
Copy full SHA for f2002fe - Browse repository at this point
Copy the full SHA f2002feView commit details -
Configuration menu - View commit details
-
Copy full SHA for 7604557 - Browse repository at this point
Copy the full SHA 7604557View commit details -
Configuration menu - View commit details
-
Copy full SHA for 7fb2a8b - Browse repository at this point
Copy the full SHA 7fb2a8bView commit details -
Raise exceptions instead of asserts in src/transformers/models/bart/m…
…odeling_flax_[bart, marian, mbart, pegasus].py (#13939) * Raise exceptions instead of asserts * fix: fixed failing quality check with copies * fix: fixed max line length * rerun github ci, failed to install dependencies
Configuration menu - View commit details
-
Copy full SHA for b65c389 - Browse repository at this point
Copy the full SHA b65c389View commit details -
Scatter dummies + skip pipeline tests (#13996)
* Scatter dummies + skip pipeline tests * Add torch scatter to build docs
Configuration menu - View commit details
-
Copy full SHA for 5b317f7 - Browse repository at this point
Copy the full SHA 5b317f7View commit details
Commits on Oct 15, 2021
-
Fixed horizon_length for PPLM (#13886)
* fixed horizon_length * fixed horizon_length * fix style
Configuration menu - View commit details
-
Copy full SHA for d5b82bb - Browse repository at this point
Copy the full SHA d5b82bbView commit details -
Add the SEW and SEW-D speech models (#13962)
* Working encoder * SEW-D and tests * Further conv fixes * Automodels and conv inits * Update integration tests, add docs * Docs cleanup, resolve todos * Conf fix * Fix docs * Fix tests, apply suggestions * Update src/transformers/models/sew/modeling_sew.py Co-authored-by: Patrick von Platen <[email protected]> * Model conversion and updated no-mask tests * Remove copy of feature_proj * Style * Update src/transformers/models/auto/feature_extraction_auto.py Co-authored-by: Patrick von Platen <[email protected]> * Update src/transformers/models/auto/feature_extraction_auto.py Co-authored-by: Patrick von Platen <[email protected]> * Move orgs Co-authored-by: Patrick von Platen <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for cd3166a - Browse repository at this point
Copy the full SHA cd3166aView commit details -
Configuration menu - View commit details
-
Copy full SHA for 47489a6 - Browse repository at this point
Copy the full SHA 47489a6View commit details -
[Docs] More general docstrings (#14028)
* up * finish * up * up * finish
Configuration menu - View commit details
-
Copy full SHA for f5af873 - Browse repository at this point
Copy the full SHA f5af873View commit details
Commits on Oct 16, 2021
-
Configuration menu - View commit details
-
Copy full SHA for 84ad6af - Browse repository at this point
Copy the full SHA 84ad6afView commit details -
Configuration menu - View commit details
-
Copy full SHA for 968ae57 - Browse repository at this point
Copy the full SHA 968ae57View commit details -
Replace assertions with ValueError exceptions (#14018)
* Replace assertions with ValueError exceptions * Change length check for a more explicit one
Configuration menu - View commit details
-
Copy full SHA for cde0c75 - Browse repository at this point
Copy the full SHA cde0c75View commit details
Commits on Oct 17, 2021
-
[Speech Examples] Add new audio feature (#14027)
* finish * up * finish all * up
Configuration menu - View commit details
-
Copy full SHA for 37c5759 - Browse repository at this point
Copy the full SHA 37c5759View commit details
Commits on Oct 18, 2021
-
Mishig Davaadorj authored
Oct 18, 2021 Configuration menu - View commit details
-
Copy full SHA for 4334095 - Browse repository at this point
Copy the full SHA 4334095View commit details -
[Speech] Move all examples to new audio feature (#14045)
* up * up * up * finish
Configuration menu - View commit details
-
Copy full SHA for bdf31d6 - Browse repository at this point
Copy the full SHA bdf31d6View commit details -
Configuration menu - View commit details
-
Copy full SHA for 82b62fa - Browse repository at this point
Copy the full SHA 82b62faView commit details -
Configuration menu - View commit details
-
Copy full SHA for 7c6cd0a - Browse repository at this point
Copy the full SHA 7c6cd0aView commit details -
Add BARTpho: Pre-trained Sequence-to-Sequence Models for Vietnamese (…
…#13788) * Add the pre-trained BARTpho model * Add the pre-trained BARTpho model * Add the pre-trained BARTpho model * Fix incorrectly sorted and/or formatted imports * Fix incorrectly sorted and/or formatted style * Fix check_dummies * Fix check_dummies * Fix check_dummies * Update docs/source/model_doc/bartpho.rst Co-authored-by: Suraj Patil <[email protected]> * Update src/transformers/models/bartpho/__init__.py Co-authored-by: Suraj Patil <[email protected]> * Update src/transformers/models/bartpho/tokenization_bartpho.py Co-authored-by: Suraj Patil <[email protected]> * Update tests/test_tokenization_bartpho.py Co-authored-by: Suraj Patil <[email protected]> * Update src/transformers/models/bartpho/tokenization_bartpho.py Co-authored-by: Suraj Patil <[email protected]> * Update tests/test_tokenization_bartpho.py Co-authored-by: Suraj Patil <[email protected]> * Update docs/source/model_doc/bartpho.rst Co-authored-by: Sylvain Gugger <[email protected]> * Update docs/source/model_doc/bartpho.rst Co-authored-by: Sylvain Gugger <[email protected]> * Update src/transformers/models/bartpho/__init__.py Co-authored-by: Suraj Patil <[email protected]> * Add the pre-trained BARTpho model * Add Tips section in doc and details of monolingual_vocab_file * Fix conflicts * Add another tip related to monolingual_vocab_file * Readd dependency_versions_table.py * Handle failing checks * Remove test_list.txt * Remove md5sum.saved * Revise Readme.md Co-authored-by: Suraj Patil <[email protected]> Co-authored-by: Sylvain Gugger <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 3d587c5 - Browse repository at this point
Copy the full SHA 3d587c5View commit details -
Add an API to register objects to Auto classes (#13989)
* Add API to register a new object in auto classes * Fix test * Documentation * Add to tokenizers and test * Add cleanup after tests * Be more careful * Move import * Move import * Cleanup in TF test too * Add consistency check * Add documentation * Style * Update docs/source/model_doc/auto.rst Co-authored-by: Lysandre Debut <[email protected]> * Update src/transformers/models/auto/auto_factory.py Co-authored-by: Lysandre Debut <[email protected]> Co-authored-by: Lysandre Debut <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 2c60ff2 - Browse repository at this point
Copy the full SHA 2c60ff2View commit details -
Configuration menu - View commit details
-
Copy full SHA for 2024faf - Browse repository at this point
Copy the full SHA 2024fafView commit details -
[Speech] Refactor Examples (#14040)
* adapt_examples * up * up * up * up * add auto models * finish
Configuration menu - View commit details
-
Copy full SHA for d5ff69f - Browse repository at this point
Copy the full SHA d5ff69fView commit details -
Configuration menu - View commit details
-
Copy full SHA for 7a3147e - Browse repository at this point
Copy the full SHA 7a3147eView commit details -
Configuration menu - View commit details
-
Copy full SHA for 9eda0d1 - Browse repository at this point
Copy the full SHA 9eda0d1View commit details
Commits on Oct 19, 2021
-
Configuration menu - View commit details
-
Copy full SHA for fde4867 - Browse repository at this point
Copy the full SHA fde4867View commit details -
TF Model train and eval step metrics for seq2seq models. (#14009)
* TF Model train and eval step metrics for seq2seq models. When using a model with a seq2seq output compute metrics against logits. * Removing vestigial code Co-authored-by: matt <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 122c2f8 - Browse repository at this point
Copy the full SHA 122c2f8View commit details -
update to_py_obj to support np.number (#14064)
Co-authored-by: 眸浩 <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 3892d09 - Browse repository at this point
Copy the full SHA 3892d09View commit details
Commits on Oct 20, 2021
-
Configuration menu - View commit details
-
Copy full SHA for 3fefa29 - Browse repository at this point
Copy the full SHA 3fefa29View commit details -
[Examples] Use Audio feature in speech classification (#14052)
* Update SEW integration test tolerance * Update audio classification * Update test * Remove torchaudio * Add dataset revision * Hub branch naming * Revert dataset revisions * Update datasets
Configuration menu - View commit details
-
Copy full SHA for dbaf492 - Browse repository at this point
Copy the full SHA dbaf492View commit details -
Configuration menu - View commit details
-
Copy full SHA for 0bc2e54 - Browse repository at this point
Copy the full SHA 0bc2e54View commit details -
Configuration menu - View commit details
-
Copy full SHA for 53dc39d - Browse repository at this point
Copy the full SHA 53dc39dView commit details -
replace assert with exception in src/transformers/utils/model_pararal…
…lel_utils.py (#14072) * replace assert with exception in src/transformers/utils/model_parallel_utils.py * fix some code style * fix typo Co-authored-by: skpig <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for a43d935 - Browse repository at this point
Copy the full SHA a43d935View commit details -
Fix missing autocast() in Trainer.prediction_step() (#14075)
Co-authored-by: jonas <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 0106826 - Browse repository at this point
Copy the full SHA 0106826View commit details -
Fix assert in src/transformers/data/datasets/language_modeling.py (#1…
…4077) * replace assertion with ValueError * fix code style Co-authored-by: skpig <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 31560f6 - Browse repository at this point
Copy the full SHA 31560f6View commit details -
Configuration menu - View commit details
-
Copy full SHA for f875fb0 - Browse repository at this point
Copy the full SHA f875fb0View commit details -
* add `ContextManagers` for lists of contexts * fix import sorting * add `ContextManagers` tests
Configuration menu - View commit details
-
Copy full SHA for 0270d44 - Browse repository at this point
Copy the full SHA 0270d44View commit details -
Configuration menu - View commit details
-
Copy full SHA for 49155d2 - Browse repository at this point
Copy the full SHA 49155d2View commit details
Commits on Oct 21, 2021
-
Configuration menu - View commit details
-
Copy full SHA for e9d2a63 - Browse repository at this point
Copy the full SHA e9d2a63View commit details -
Change asserts in src/transformers/models/xlnet/ to raise ValueError …
…(#14088) * Change asserts in src/transformers/models/xlnet/ to raise ValueError * Update src/transformers/models/xlnet/modeling_tf_xlnet.py Co-authored-by: Sylvain Gugger <[email protected]> Co-authored-by: Sylvain Gugger <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 9e4ea25 - Browse repository at this point
Copy the full SHA 9e4ea25View commit details -
Replace assertions with ValueError exceptions (#14061)
* Replace assertions with ValueError exceptions * Format error messages as suggested
Configuration menu - View commit details
-
Copy full SHA for 3187228 - Browse repository at this point
Copy the full SHA 3187228View commit details -
Configuration menu - View commit details
-
Copy full SHA for f9c16b0 - Browse repository at this point
Copy the full SHA f9c16b0View commit details -
Configuration menu - View commit details
-
Copy full SHA for 0f50268 - Browse repository at this point
Copy the full SHA 0f50268View commit details -
[Examples] Add audio classification notebooks (#14099)
* Update SEW integration test tolerance * Add audio classification notebooks
Configuration menu - View commit details
-
Copy full SHA for e03544a - Browse repository at this point
Copy the full SHA e03544aView commit details -
Fix ignore_mismatched_sizes (#14085)
* Fix * Style * Name * Fix tests * Style * Remove embed sizes checking * Disable some tests * Fix * Apply suggestion
Configuration menu - View commit details
-
Copy full SHA for 234cfef - Browse repository at this point
Copy the full SHA 234cfefView commit details -
Configuration menu - View commit details
-
Copy full SHA for f00bcea - Browse repository at this point
Copy the full SHA f00bceaView commit details -
Configuration menu - View commit details
-
Copy full SHA for 7af55d3 - Browse repository at this point
Copy the full SHA 7af55d3View commit details -
fix typo in license docstring (#14094)
last line: "# limitations under the License." is missing
Configuration menu - View commit details
-
Copy full SHA for d432a65 - Browse repository at this point
Copy the full SHA d432a65View commit details -
Configuration menu - View commit details
-
Copy full SHA for 7888914 - Browse repository at this point
Copy the full SHA 7888914View commit details -
Configuration menu - View commit details
-
Copy full SHA for ca2ef7d - Browse repository at this point
Copy the full SHA ca2ef7dView commit details
Commits on Oct 22, 2021
-
Configuration menu - View commit details
-
Copy full SHA for 70f186f - Browse repository at this point
Copy the full SHA 70f186fView commit details -
Replace assert statements with exceptions (#13871) (#13901)
* Replace assert statements with exceptions (#13871) * Change f-strings when not needed (flake8) * Replace assert statements with exceptions (#13871) * Change f-strings when not needed (flake8) * Improve error message as suggested by reviewer * Fix identation bug * Fix style errors
Configuration menu - View commit details
-
Copy full SHA for f5a49bf - Browse repository at this point
Copy the full SHA f5a49bfView commit details -
Translate README.md to Korean (#14015)
* Create README_ko.md * Update README.md * Update README_zh-hans.md * Update README_zh-hant.md * Update README_ko.md * Update check_copies.py * Update README_ko.md * typo * match with readme_ko
Configuration menu - View commit details
-
Copy full SHA for 9f53f04 - Browse repository at this point
Copy the full SHA 9f53f04View commit details -
Replace assertions with valueError Exeptions (#14117)
* Replace assertions with valueError Exeptions * Reformatted
Configuration menu - View commit details
-
Copy full SHA for fa4abdb - Browse repository at this point
Copy the full SHA fa4abdbView commit details -
Configuration menu - View commit details
-
Copy full SHA for 16d7b70 - Browse repository at this point
Copy the full SHA 16d7b70View commit details -
Fix assertion in models (#14090)
* replace assertions in src/transformers/models/luke/convert_luke_original_pytorch_checkpoint_to_pytorch.py * replace assertions in src/transformers/models/marian/convert_marian_to_pytorch.py * Update src/transformers/models/luke/convert_luke_original_pytorch_checkpoint_to_pytorch.py Co-authored-by: Sylvain Gugger <[email protected]> * Update src/transformers/models/marian/convert_marian_to_pytorch.py Co-authored-by: Sylvain Gugger <[email protected]> * Update src/transformers/models/marian/convert_marian_to_pytorch.py Co-authored-by: Sylvain Gugger <[email protected]> * Update src/transformers/models/marian/convert_marian_to_pytorch.py Co-authored-by: Sylvain Gugger <[email protected]> * Update src/transformers/models/marian/convert_marian_to_pytorch.py Co-authored-by: Sylvain Gugger <[email protected]> * Update src/transformers/models/marian/convert_marian_to_pytorch.py Co-authored-by: Sylvain Gugger <[email protected]> * Update src/transformers/models/marian/convert_marian_to_pytorch.py Co-authored-by: Sylvain Gugger <[email protected]> * Update src/transformers/models/marian/convert_marian_to_pytorch.py Co-authored-by: Sylvain Gugger <[email protected]> * Update src/transformers/models/marian/convert_marian_to_pytorch.py Co-authored-by: Sylvain Gugger <[email protected]> Co-authored-by: skpig <[email protected]> Co-authored-by: Sylvain Gugger <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for c7ccb2e - Browse repository at this point
Copy the full SHA c7ccb2eView commit details -
Configuration menu - View commit details
-
Copy full SHA for 05a2afc - Browse repository at this point
Copy the full SHA 05a2afcView commit details -
Rename variables with unclear naming (#14122)
* Rename var * Add comments
Configuration menu - View commit details
-
Copy full SHA for 62ccbe0 - Browse repository at this point
Copy the full SHA 62ccbe0View commit details -
Update TP parallel GEMM image (#14112)
* Update TP parallel GEMM image * Delete parallelism-tp-parallel_gemm.png * Update parallelism-tp-parallel_gemm.png
Configuration menu - View commit details
-
Copy full SHA for 95bab53 - Browse repository at this point
Copy the full SHA 95bab53View commit details
Commits on Oct 25, 2021
-
Fix some typos in the docs (#14126)
* Fix some typos in the docs * Fix a styling issue * Fix code quality check error
Configuration menu - View commit details
-
Copy full SHA for 6b83090 - Browse repository at this point
Copy the full SHA 6b83090View commit details -
Supporting Seq2Seq model for question answering task (#13432)
* Add seq2seq example for QnA on SQuAD Dataset. * Changes from review - Fixing styling mistakes. * Added how to example in README, simplified the access to dataset's preprocess function. * Added tests for the seq2seq QA example. * Change dataset column name to fix tests. * Fix test command mistake. * Add missing argument 'ignore_pad_token_for_loss' from DataTrainingArguments. * Add missing argument 'num_beams' from DataTrainingArguments. * Fix processing of output predicted token ids so that tokenizer decode gets appropriate input. Updated assertion conditions on the tests.
Configuration menu - View commit details
-
Copy full SHA for 1b871e0 - Browse repository at this point
Copy the full SHA 1b871e0View commit details -
Configuration menu - View commit details
-
Copy full SHA for 2ac6555 - Browse repository at this point
Copy the full SHA 2ac6555View commit details -
Fix some writing issues in the docs (#14136)
* Fix some writing issues in the docs * Run code quality check
Configuration menu - View commit details
-
Copy full SHA for 3e04a41 - Browse repository at this point
Copy the full SHA 3e04a41View commit details -
BartEnocder add set_input_embeddings (#13960)
* BartEnocder add set_input_embeddings To unify the interface, add set_input_embeddings to BartEncoder. * BartEnocder add get_input_embeddings
Configuration menu - View commit details
-
Copy full SHA for 1967c43 - Browse repository at this point
Copy the full SHA 1967c43View commit details -
Configuration menu - View commit details
-
Copy full SHA for 84b9579 - Browse repository at this point
Copy the full SHA 84b9579View commit details -
Configuration menu - View commit details
-
Copy full SHA for 3e8761a - Browse repository at this point
Copy the full SHA 3e8761aView commit details -
Configuration menu - View commit details
-
Copy full SHA for 1a9381c - Browse repository at this point
Copy the full SHA 1a9381cView commit details -
Configuration menu - View commit details
-
Copy full SHA for c99a283 - Browse repository at this point
Copy the full SHA c99a283View commit details -
Configuration menu - View commit details
-
Copy full SHA for 8560b55 - Browse repository at this point
Copy the full SHA 8560b55View commit details -
Add TF<>PT and Flax<>PT everywhere (#14047)
* up * up * up * up * up * up * up * add clip * fix clip PyTorch * fix clip PyTorch * up * up * up * up * up * up * up
Configuration menu - View commit details
-
Copy full SHA for 0c3174c - Browse repository at this point
Copy the full SHA 0c3174cView commit details
Commits on Oct 26, 2021
-
Add Camembert to models exportable with ONNX (#14059)
Add Camembert to models exportable with ONNX Co-authored-by: Thomas.Chaigneau <[email protected]> Co-authored-by: Michael Benayoun <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 1f60df8 - Browse repository at this point
Copy the full SHA 1f60df8View commit details -
Configuration menu - View commit details
-
Copy full SHA for e248e9b - Browse repository at this point
Copy the full SHA e248e9bView commit details -
Add vision_encoder_decoder to models/__init__.py (#14151)
* Add vision_encoder_decoder * Update _ignore_modules in get_model_modules() Co-authored-by: ydshieh <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 840fc8d - Browse repository at this point
Copy the full SHA 840fc8dView commit details -
[Speech Recognition] - Distributed training: Make sure vocab file rem…
…oval and creation don't interfer (#14161) * up * better
Configuration menu - View commit details
-
Copy full SHA for f5ed19f - Browse repository at this point
Copy the full SHA f5ed19fView commit details -
Include Keras tensor in the allowed types (#14155)
* Include KerasTensor in allowed types - This allows propagating symbolic tensors through TFBert models and layers' call(), which allows converting the subclass models to functional models. * Style pass Co-authored-by: Sergio Valcarcel Macua <[email protected]> Co-authored-by: matt <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 919a964 - Browse repository at this point
Copy the full SHA 919a964View commit details -
[megatron_gpt2] dynamic gelu, add tokenizer, save config (#13928)
* [megatron_gpt2] dynamic gelu, add tokenizer, save config * cleanup * Update src/transformers/models/megatron_gpt2/convert_megatron_gpt2_checkpoint.py Co-authored-by: Sylvain Gugger <[email protected]> * apply suggestions Co-authored-by: Sylvain Gugger <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for bfd8176 - Browse repository at this point
Copy the full SHA bfd8176View commit details -
Configuration menu - View commit details
-
Copy full SHA for 9799f4e - Browse repository at this point
Copy the full SHA 9799f4eView commit details -
Add Unispeech & Unispeech-SAT (#13963)
* unispeech * add copy from * remove hubert copy from * finish for today * add unispeech-sat * adapt more * up * up * up * up * add modeling * add tests * up * up * finish * up * Apply suggestions from code review * up * up * Apply suggestions from code review Co-authored-by: Sylvain Gugger <[email protected]> * up * up Co-authored-by: Sylvain Gugger <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 9f3aa46 - Browse repository at this point
Copy the full SHA 9f3aa46View commit details -
[ONNX] Add symbolic function for XSoftmax op for exporting to ONNX. (…
…#14013) * Add symbolic function for XSoftmax op for exporting to ONNX. * Fix format issues. * Fix a CI issue relative to copies.
Configuration menu - View commit details
-
Copy full SHA for 3f23634 - Browse repository at this point
Copy the full SHA 3f23634View commit details -
Configuration menu - View commit details
-
Copy full SHA for 27c888d - Browse repository at this point
Copy the full SHA 27c888dView commit details -
Configuration menu - View commit details
-
Copy full SHA for 41dad89 - Browse repository at this point
Copy the full SHA 41dad89View commit details -
fix typos in error messages in speech recognition example and modelca…
…rd.py (#14166) * specify the text column name in the error message * pluralize the word fields
Configuration menu - View commit details
-
Copy full SHA for 42bfb83 - Browse repository at this point
Copy the full SHA 42bfb83View commit details -
Replace assertions with ValueError exception (#14142)
Updated masked-language modeling examples in pytorch with convention defined by #12789
Configuration menu - View commit details
-
Copy full SHA for ebd48c6 - Browse repository at this point
Copy the full SHA ebd48c6View commit details -
switch to inference_mode from no_gard (#13667)
* switch to inference_mode from no_gard faster inference * added switch to support older version of pytorch
Configuration menu - View commit details
-
Copy full SHA for 8ddbfe9 - Browse repository at this point
Copy the full SHA 8ddbfe9View commit details
Commits on Oct 27, 2021
-
Configuration menu - View commit details
-
Copy full SHA for 1e53fae - Browse repository at this point
Copy the full SHA 1e53faeView commit details -
* Add SEW CTC models * Update paths * Update paths
Configuration menu - View commit details
-
Copy full SHA for e1dc5af - Browse repository at this point
Copy the full SHA e1dc5afView commit details -
[Gradient checkpointing] Enable for Deberta + DebertaV2 + SEW-D (#14175)
* up * up * finish * up * final changes
Configuration menu - View commit details
-
Copy full SHA for 6200fd7 - Browse repository at this point
Copy the full SHA 6200fd7View commit details -
Configuration menu - View commit details
-
Copy full SHA for 25ceb81 - Browse repository at this point
Copy the full SHA 25ceb81View commit details -
Replace assert of data/data_collator.py by ValueError (#14131)
* Replace assert of data_collator.py by ValueError * Replace assert of data_collator.py by ValueError
Configuration menu - View commit details
-
Copy full SHA for e5b8ffb - Browse repository at this point
Copy the full SHA e5b8ffbView commit details -
* Add conversion * Rename * Add an integration test and remove layer_norm * Remove layer_norm from the converter * wording * Fix imports
Configuration menu - View commit details
-
Copy full SHA for 232822f - Browse repository at this point
Copy the full SHA 232822fView commit details -
[TPU tests] Enable first TPU examples pytorch (#14121)
* up * up * fix * up * Update examples/pytorch/test_xla_examples.py * correct labels * up * up * up * up * up * up
Configuration menu - View commit details
-
Copy full SHA for 01b1466 - Browse repository at this point
Copy the full SHA 01b1466View commit details -
Configuration menu - View commit details
-
Copy full SHA for e118db1 - Browse repository at this point
Copy the full SHA e118db1View commit details
Commits on Oct 28, 2021
-
Configuration menu - View commit details
-
Copy full SHA for 88cd82e - Browse repository at this point
Copy the full SHA 88cd82eView commit details -
[modeling_utils] respect original dtype in _get_resized_lm_head (#14181)
* respect dtype in _get_resized_lm_head * Update src/transformers/modeling_utils.py Co-authored-by: Sylvain Gugger <[email protected]> * consistency Co-authored-by: Sylvain Gugger <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 123cce6 - Browse repository at this point
Copy the full SHA 123cce6View commit details -
* First draft * Make style & quality * Improve conversion script * Add print statement to see actual slice * Make absolute tolerance smaller * Fix image classification models * Add post_process_semantic method * Disable padding * Improve conversion script * Rename to ForSemanticSegmentation, add integration test, remove post_process methods * Improve docs * Fix code quality * Fix feature extractor tests * Fix tests for image classification model * Delete file * Add is_torch_available to feature extractor * Improve documentation of feature extractor methods * Apply suggestions from @sgugger's code review Co-authored-by: Sylvain Gugger <[email protected]> * Apply some more suggestions of code review * Rebase with master * Fix rebase issues * Make sure model only outputs hidden states when the user wants to * Apply suggestions from code review * Add pad method * Support padding of 2d images * Add print statement * Add print statement * Move padding method to SegformerFeatureExtractor * Fix issue * Add casting of segmentation maps * Add test for padding * Add small note about padding Co-authored-by: Sylvain Gugger <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 1dc96a7 - Browse repository at this point
Copy the full SHA 1dc96a7View commit details -
Configuration menu - View commit details
-
Copy full SHA for 78b6a2e - Browse repository at this point
Copy the full SHA 78b6a2eView commit details -
Fix SEW-D implementation differences (#14191)
* Fix SEW-D * Update tests * isort
Configuration menu - View commit details
-
Copy full SHA for 1251072 - Browse repository at this point
Copy the full SHA 1251072View commit details -
Fix EncoderDecoderModel classes to be more like BART and T5 (#14139)
* First draft * Make tuple output more readable * Replace assertions by value errors * Make it possible to predict_with_generate for vision and speech models * Adapt Seq2SeqTrainer to work with VisionEncoderDecoder/SpeechEncoderDecoder * Add deprecation warning * Add copied from statements to vision and speech encoder decoders * Fix failing test * Apply @patrickvonplaten's suggestion * Use reshape instead of view for consistency
Configuration menu - View commit details
-
Copy full SHA for ac12a5a - Browse repository at this point
Copy the full SHA ac12a5aView commit details -
Fix EncoderDecoderModel docs (#14197)
* Fix docs * Apply suggestions from review + fix bug
Configuration menu - View commit details
-
Copy full SHA for 5f3bf65 - Browse repository at this point
Copy the full SHA 5f3bf65View commit details -
Configuration menu - View commit details
-
Copy full SHA for 62bf536 - Browse repository at this point
Copy the full SHA 62bf536View commit details
Commits on Oct 29, 2021
-
* Torch 1.10 * torch scatter for 1.10 * style * Skip tests ok
Configuration menu - View commit details
-
Copy full SHA for 9f3f335 - Browse repository at this point
Copy the full SHA 9f3f335View commit details -
Configuration menu - View commit details
-
Copy full SHA for e0a5154 - Browse repository at this point
Copy the full SHA e0a5154View commit details -
Fixing image segmentation with inference mode. (#14204)
* Fixing image segmentation for inference mode. * Update src/transformers/pipelines/base.py Co-authored-by: Patrick von Platen <[email protected]> Co-authored-by: Patrick von Platen <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for cde7d78 - Browse repository at this point
Copy the full SHA cde7d78View commit details -
Configuration menu - View commit details
-
Copy full SHA for 2191373 - Browse repository at this point
Copy the full SHA 2191373View commit details
Commits on Nov 2, 2021
-
Configuration menu - View commit details
-
Copy full SHA for ac77639 - Browse repository at this point
Copy the full SHA ac77639View commit details -
Configuration menu - View commit details
-
Copy full SHA for 872c4f3 - Browse repository at this point
Copy the full SHA 872c4f3View commit details -
Configuration menu - View commit details
-
Copy full SHA for 9ab10fc - Browse repository at this point
Copy the full SHA 9ab10fcView commit details -
Configuration menu - View commit details
-
Copy full SHA for 294a920 - Browse repository at this point
Copy the full SHA 294a920View commit details -
Configuration menu - View commit details
-
Copy full SHA for 3ea15d2 - Browse repository at this point
Copy the full SHA 3ea15d2View commit details
Commits on Nov 16, 2021
-
Fix of issue #13327: Wrong weight initialization for TF t5 model (#14…
…241) * Fix of issue #13327: Wrong weight initialization for TF t5 model * run black formatter * fix typo * remove my name tag from comments Co-authored-by: Shirron <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for b6b97c3 - Browse repository at this point
Copy the full SHA b6b97c3View commit details -
Configuration menu - View commit details
-
Copy full SHA for c8206b4 - Browse repository at this point
Copy the full SHA c8206b4View commit details -
Configuration menu - View commit details
-
Copy full SHA for 6bf2027 - Browse repository at this point
Copy the full SHA 6bf2027View commit details -
Configuration menu - View commit details
-
Copy full SHA for 341a059 - Browse repository at this point
Copy the full SHA 341a059View commit details -
Experimenting with adding proper get_config() and from_config() metho…
…ds (#14361) * Experimenting with adding proper get_config() and from_config() methods * Adding a test for get/from config * Fix test for get/from config
Configuration menu - View commit details
-
Copy full SHA for e99a231 - Browse repository at this point
Copy the full SHA e99a231View commit details -
[Wav2Vec2] Make sure that gradient checkpointing is only run if neede…
…d (#14407) * [Wav2Vec2] Make sure that gradient checkpointing is only run if needed * make fix-copies
Configuration menu - View commit details
-
Copy full SHA for db242ae - Browse repository at this point
Copy the full SHA db242aeView commit details -
Fix gradient_checkpointing backward compatibility (#14408)
* Fix gradient_checkpointing backward compatibility * Remove needless line * make sure mask prob is big enough and length small enough * Fix tests Co-authored-by: patrickvonplaten <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 6f40723 - Browse repository at this point
Copy the full SHA 6f40723View commit details -
Configuration menu - View commit details
-
Copy full SHA for 527c763 - Browse repository at this point
Copy the full SHA 527c763View commit details
Commits on Nov 17, 2021
-
Revert "Experimenting with adding proper get_config() and from_config…
…() methods (#14361)" This reverts commit e99a231.
Configuration menu - View commit details
-
Copy full SHA for a5211fc - Browse repository at this point
Copy the full SHA a5211fcView commit details -
Configuration menu - View commit details
-
Copy full SHA for ef3cec0 - Browse repository at this point
Copy the full SHA ef3cec0View commit details
Commits on Dec 1, 2021
-
Configuration menu - View commit details
-
Copy full SHA for b3a8de8 - Browse repository at this point
Copy the full SHA b3a8de8View commit details -
Configuration menu - View commit details
-
Copy full SHA for 2c63bc1 - Browse repository at this point
Copy the full SHA 2c63bc1View commit details
Commits on Dec 7, 2021
-
Configuration menu - View commit details
-
Copy full SHA for f83eb7e - Browse repository at this point
Copy the full SHA f83eb7eView commit details
Commits on Dec 8, 2021
-
Configuration menu - View commit details
-
Copy full SHA for 3356b30 - Browse repository at this point
Copy the full SHA 3356b30View commit details