Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Sync/v4.12.x #255

Merged
merged 219 commits into from
Dec 8, 2021
Merged

Sync/v4.12.x #255

merged 219 commits into from
Dec 8, 2021

Commits on Sep 27, 2021

  1. Docs for version v4.11.0

    LysandreJik committed Sep 27, 2021
    Configuration menu
    Copy the full SHA
    11c69b8 View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    5e3b4a7 View commit details
    Browse the repository at this point in the history

Commits on Sep 28, 2021

  1. Configuration menu
    Copy the full SHA
    83d3dc0 View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    a21ee1f View commit details
    Browse the repository at this point in the history

Commits on Sep 29, 2021

  1. up (#13777)

    patrickvonplaten authored Sep 29, 2021
    Configuration menu
    Copy the full SHA
    aa018a7 View commit details
    Browse the repository at this point in the history
  2. Keras callback to push to hub each epoch, or after N steps (#13773)

    * Keras callback to push to hub each epoch, or after N steps
    
    * Reworked the callback to use Repository
    
    * Use an Enum for save_strategy
    
    * Style pass
    
    * Correct type for tokenizer
    
    * Update src/transformers/keras_callbacks.py
    
    Co-authored-by: Sylvain Gugger <[email protected]>
    
    * Update src/transformers/keras_callbacks.py
    
    Co-authored-by: Sylvain Gugger <[email protected]>
    
    * Update src/transformers/keras_callbacks.py
    
    Co-authored-by: Sylvain Gugger <[email protected]>
    
    * Update src/transformers/keras_callbacks.py
    
    Co-authored-by: Sylvain Gugger <[email protected]>
    
    * Update src/transformers/keras_callbacks.py
    
    Co-authored-by: Sylvain Gugger <[email protected]>
    
    * Update src/transformers/keras_callbacks.py
    
    Co-authored-by: Sylvain Gugger <[email protected]>
    
    * Adding print message to the final upload
    
    * Adding print message to the final upload
    
    * Change how we wait for the last process to finish
    
    * is_done is a property, not a method, derp
    
    * Docstrings and documentation
    
    * Style pass
    
    * Style edit
    
    * Docstring reformat
    
    * Docstring rewrite
    
    * Replacing print with internal logger
    
    Co-authored-by: Sylvain Gugger <[email protected]>
    Rocketknight1 and sgugger authored Sep 29, 2021
    Configuration menu
    Copy the full SHA
    3a8a801 View commit details
    Browse the repository at this point in the history
  3. Fix LayoutLM ONNX test error (#13710)

    Fix LayoutLM ONNX test error
    nishprabhu authored Sep 29, 2021
    Configuration menu
    Copy the full SHA
    a1ea3ad View commit details
    Browse the repository at this point in the history
  4. Enable readme link synchronization (#13785)

    * Enable readme link synchronization
    
    * Style
    
    * Reuse regex pattern
    
    * Apply suggestions
    
    * Update
    qqaatw authored Sep 29, 2021
    Configuration menu
    Copy the full SHA
    7d84c3a View commit details
    Browse the repository at this point in the history
  5. Fix length of IterableDatasetShard and add test (#13792)

    * Fix length of IterableDatasetShard and add test
    
    * Add comments
    sgugger authored Sep 29, 2021
    Configuration menu
    Copy the full SHA
    63cc5bd View commit details
    Browse the repository at this point in the history
  6. Configuration menu
    Copy the full SHA
    2a51b15 View commit details
    Browse the repository at this point in the history
  7. Update doc for v4.11.1

    sgugger committed Sep 29, 2021
    Configuration menu
    Copy the full SHA
    cf4aa35 View commit details
    Browse the repository at this point in the history
  8. Configuration menu
    Copy the full SHA
    55695df View commit details
    Browse the repository at this point in the history
  9. [docs/gpt-j] addd instructions for how minimize CPU RAM usage (#13795)

    * add a note about tokenizer
    
    * add  tips to load model is less RAM
    
    * fix link
    
    * fix more links
    patil-suraj authored Sep 29, 2021
    Configuration menu
    Copy the full SHA
    bf6118e View commit details
    Browse the repository at this point in the history
  10. [examples run_glue.py] missing requirements scipy, sklearn (#13…

    …768)
    
    * missing requirement
    
    * list both
    stas00 authored Sep 29, 2021
    Configuration menu
    Copy the full SHA
    b90096f View commit details
    Browse the repository at this point in the history

Commits on Sep 30, 2021

  1. [examples/flax] use Repository API for push_to_hub (#13672)

    * use Repository for push_to_hub
    
    * update readme
    
    * update other flax scripts
    
    * update readme
    
    * update qa example
    
    * fix push_to_hub call
    
    * fix typo
    
    * fix more typos
    
    * update readme
    
    * use abosolute path to get repo name
    
    * fix glue script
    patil-suraj authored Sep 30, 2021
    Configuration menu
    Copy the full SHA
    7db2a79 View commit details
    Browse the repository at this point in the history
  2. Fix gather for TPU (#13813)

    sgugger authored Sep 30, 2021
    Configuration menu
    Copy the full SHA
    269c3d1 View commit details
    Browse the repository at this point in the history
  3. Update doc for v4.11.2

    sgugger committed Sep 30, 2021
    Configuration menu
    Copy the full SHA
    5f25855 View commit details
    Browse the repository at this point in the history
  4. Configuration menu
    Copy the full SHA
    e1d1c7c View commit details
    Browse the repository at this point in the history
  5. Add MultiBERTs conversion script (#13077)

    * Init multibert checkpoint conversion script
    
    * Rename conversion script
    
    * Fix MultiBerts Conversion Script
    
    * Apply suggestions from code review
    
    Co-authored-by: NielsRogge <[email protected]>
    
    Co-authored-by: Patrick von Platen <[email protected]>
    Co-authored-by: NielsRogge <[email protected]>
    3 people authored Sep 30, 2021
    Configuration menu
    Copy the full SHA
    9a9805f View commit details
    Browse the repository at this point in the history
  6. Configuration menu
    Copy the full SHA
    44eb8bd View commit details
    Browse the repository at this point in the history
  7. [DPR] Correct init (#13796)

    * update
    
    * add to docs and init
    
    * make fix-copies
    patrickvonplaten authored Sep 30, 2021
    Configuration menu
    Copy the full SHA
    41436d3 View commit details
    Browse the repository at this point in the history
  8. Configuration menu
    Copy the full SHA
    8bbb53e View commit details
    Browse the repository at this point in the history

Commits on Oct 1, 2021

  1. Fix warning situation: UserWarning: max_length is ignored when paddin…

    …g=True" (#13829)
    
    * Removed wrong warning
    
    * Raise a warning when `max_length` is given with wrong `truncation`
    
    * Update the error message
    
    * Update the warning message
    
    Co-authored-by: Sylvain Gugger <[email protected]>
    
    Co-authored-by: Sylvain Gugger <[email protected]>
    shirayu and sgugger authored Oct 1, 2021
    Configuration menu
    Copy the full SHA
    90f980e View commit details
    Browse the repository at this point in the history
  2. Update CITATION.cff (#13833)

    arfon authored Oct 1, 2021
    Configuration menu
    Copy the full SHA
    c411372 View commit details
    Browse the repository at this point in the history
  3. [Examples] Add an official audio classification example (#13722)

    * Restore broken merge
    
    * Additional args, DDP, remove CommonLanguage
    
    * Update examples for V100, add training results
    
    * Style
    
    * Apply suggestions from code review
    
    Co-authored-by: Sylvain Gugger <[email protected]>
    
    * Remove custom datasets for simplicity, apply suggestions from code review
    
    * Add the attention_mask flag, reorganize README
    
    Co-authored-by: Sylvain Gugger <[email protected]>
    anton-l and sgugger authored Oct 1, 2021
    Configuration menu
    Copy the full SHA
    4213728 View commit details
    Browse the repository at this point in the history
  4. Bart: check if decoder_inputs_embeds is set (#13800)

    In BartForConditionalGeneration.forward, if labels are provided,
       decoder_input_ids are set to the labels shifted to the right.
       This is problematic: if decoder_inputs_embeds is also set,
       the call to self.model, which eventually gets to BartDecoder.forward,
       will raise an error.
       The fix is quite simple, similar to what is there already in
       BartModel.forward. Mainly, we should not
       compute decoder_input_ids if decoder_inputs_embeds is provided.
    
    Co-authored-by: Silviu Vlad Oprea <[email protected]>
    Silviu Oprea and Silviu Vlad Oprea authored Oct 1, 2021
    Configuration menu
    Copy the full SHA
    707f7eb View commit details
    Browse the repository at this point in the history
  5. Configuration menu
    Copy the full SHA
    bcc3f7b View commit details
    Browse the repository at this point in the history

Commits on Oct 4, 2021

  1. Configuration menu
    Copy the full SHA
    de94835 View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    955fd4f View commit details
    Browse the repository at this point in the history
  3. Add Mistral GPT-2 Stability Tweaks (#13573)

    * Add layer-wise scaling
    
    * Add reorder & upcasting argument
    
    * Add OpenAI GPT-2 weight initialization scheme
    
    * start `layer_idx` count at zero for consistency
    
    * disentangle attn and reordered and upscaled attn function
    
    * rename `scale_attn_by_layer` to `scale_attn_by_layer_id`
    
    * make autocast from amp compatible with pytorch<1.6
    
    * fix docstring
    
    * style fixes
    
    * Add fixes from PR feedback, style tweaks
    
    * Fix doc whitespace
    
    * Reformat
    
    * First pass scale_attn_by_layer_idx and reorder_and_upcast_attn tests
    
    * Rename scale_attn_by_layer_idx, add tip
    
    * Remove extra newline
    
    * add test for weight initialization
    
    * update code format
    
    * add assert check weights are fp32
    
    * remove assert
    
    * Fix incorrect merge
    
    * Fix shape mismatch in baddbmm
    
    * Add generation test for Mistral flags
    
    Co-authored-by: leandro <[email protected]>
    Co-authored-by: Keshav Santhanam <[email protected]>
    Co-authored-by: J38 <[email protected]>
    4 people authored Oct 4, 2021
    Configuration menu
    Copy the full SHA
    3a8de58 View commit details
    Browse the repository at this point in the history
  4. Fix broken link to distill models in docs (#13848)

    * Fix broken link to distill models
    
    * Missing symbol
    
    * Fix spaces
    Randl authored Oct 4, 2021
    Configuration menu
    Copy the full SHA
    6c08840 View commit details
    Browse the repository at this point in the history
  5. ✨ update image classification example (#13824)

    * ✨ update image classification example
    
    * 📌 update reqs
    nateraw authored Oct 4, 2021
    Configuration menu
    Copy the full SHA
    cc0a415 View commit details
    Browse the repository at this point in the history
  6. Update no_* argument (HfArgumentParser) (#13865)

    * update no_* argument
    
    Changes the order so that the no_* argument is created after the original argument AND sets the default for this no_* argument to False
    
    * import copy
    
    * update test
    
    * make style
    
    * Use kwargs to set default=False
    
    * make style
    Bram Vanroy authored Oct 4, 2021
    Configuration menu
    Copy the full SHA
    12b4d66 View commit details
    Browse the repository at this point in the history

Commits on Oct 5, 2021

  1. Update Tatoeba conversion (#13757)

    * Update Tatoeba conversion
    Traubert authored Oct 5, 2021
    Configuration menu
    Copy the full SHA
    7051b89 View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    7079a99 View commit details
    Browse the repository at this point in the history
  3. Fix: save checkpoint after each epoch and push checkpoint to the hub …

    …(#13872)
    
    Co-authored-by: ydshieh <[email protected]>
    ydshieh and ydshieh authored Oct 5, 2021
    Configuration menu
    Copy the full SHA
    a6ea244 View commit details
    Browse the repository at this point in the history
  4. Fixing empty prompts for text-generation when BOS exists. (#13859)

    * Fixing empty prompts for text-generation when BOS exists.
    
    * Fixing odd case with Pegasus.
    
    * Fixing Bert is Assertion Error.
    Narsil authored Oct 5, 2021
    Configuration menu
    Copy the full SHA
    3a9c0f2 View commit details
    Browse the repository at this point in the history
  5. Improve error message when loading models from Hub (#13836)

    * Improve error message when loading models from Hub
    
    * Adjust error message wording
    aphedges authored Oct 5, 2021
    Configuration menu
    Copy the full SHA
    46efc58 View commit details
    Browse the repository at this point in the history
  6. Initial support for symbolic tracing with torch.fx allowing dynamic a…

    …xes (#13579)
    
    * Symbolic trace dynamic axes support for BERT like models (albert, bert, distilbert, mobilebert, electra, megatron-bert)
    * Sanity checks before tracing that make sure the model to trace is supported
    * Adapted to PyTorch 1.9
    
    Co-authored-by: Michael Benayoun <[email protected]>
    michaelbenayoun and michaelbenayoun authored Oct 5, 2021
    Configuration menu
    Copy the full SHA
    d4e4efc View commit details
    Browse the repository at this point in the history
  7. Allow dataset to be an optional argument for (Distributed)LengthGroup…

    …edSampler (#13820)
    
    * Allow dataset to be an optional argument for (Distributed)LengthGroupedSampler
    
    * Fix
    ZhaofengWu authored Oct 5, 2021
    Configuration menu
    Copy the full SHA
    1b74af7 View commit details
    Browse the repository at this point in the history
  8. Fixing question-answering with long contexts (#13873)

    * Tmp.
    
    * Fixing BC for question answering with long context.
    
    * Capping model_max_length to avoid tf overflow.
    
    * Bad workaround bugged roberta.
    
    * Fixing name.
    Narsil authored Oct 5, 2021
    Configuration menu
    Copy the full SHA
    0ddadbf View commit details
    Browse the repository at this point in the history
  9. Configuration menu
    Copy the full SHA
    f099249 View commit details
    Browse the repository at this point in the history
  10. Configuration menu
    Copy the full SHA
    7af7d7c View commit details
    Browse the repository at this point in the history

Commits on Oct 6, 2021

  1. Update parallelism.md (#13892)

    * Update parallelism.md
    
    * Update docs/source/parallelism.md
    
    Co-authored-by: Stas Bekman <[email protected]>
    
    * Update docs/source/parallelism.md
    
    Co-authored-by: Stas Bekman <[email protected]>
    
    * Update docs/source/parallelism.md
    
    Co-authored-by: Stas Bekman <[email protected]>
    
    * Update docs/source/parallelism.md
    
    Co-authored-by: Stas Bekman <[email protected]>
    
    * Update docs/source/parallelism.md
    
    Co-authored-by: Stas Bekman <[email protected]>
    
    * Update docs/source/parallelism.md
    
    Co-authored-by: Stas Bekman <[email protected]>
    
    Co-authored-by: Stas Bekman <[email protected]>
    hyunwoongko and stas00 authored Oct 6, 2021
    Configuration menu
    Copy the full SHA
    36fc401 View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    7d83655 View commit details
    Browse the repository at this point in the history
  3. Fixing GPU for token-classification in a better way. (#13856)

    Co-authored-by:  Pierre Snell <[email protected]>
    
    Co-authored-by: Pierre Snell <[email protected]>
    Narsil and ierezell authored Oct 6, 2021
    Configuration menu
    Copy the full SHA
    e7b16f3 View commit details
    Browse the repository at this point in the history
  4. Update FSNER code in examples->research_projects->fsner (#13864)

    * Add example use of few-shot named entity recognition model in research_projects folder.
    
    * Apply suggestions from code review
    
    Co-authored-by: NielsRogge <[email protected]>
    
    * Update fsner example README.md.
    
    - Change wrong import FSNERTokenizerWrapper to FSNERTokenizerUtils in the example code
    - Add a link to the model identifier
    
    * Update examples/research_projects/fsner/src/fsner/model.py
    
    Fix spelling mistake in the default parameter of pretrained model name.
    
    Co-authored-by: Stefan Schweter <[email protected]>
    
    * Add example use of few-shot named entity recognition model in research_projects folder.
    
    * Apply suggestions from code review
    
    Co-authored-by: NielsRogge <[email protected]>
    
    * Update fsner example README.md.
    
    - Change wrong import FSNERTokenizerWrapper to FSNERTokenizerUtils in the example code
    - Add a link to the model identifier
    
    * Update examples/research_projects/fsner/src/fsner/model.py
    
    Fix spelling mistake in the default parameter of pretrained model name.
    
    Co-authored-by: Stefan Schweter <[email protected]>
    
    * Run Checking/fixing examples/flax/language-modeling/run_clm_flax.py examples/flax/question-answering/run_qa.py examples/flax/question-answering/utils_qa.py examples/flax/token-classification/run_flax_ner.py examples/legacy/multiple_choice/utils_multiple_choice.py examples/legacy/seq2seq/seq2seq_trainer.py examples/legacy/token-classification/utils_ner.py examples/pytorch/image-classification/run_image_classification.py examples/pytorch/language-modeling/run_clm.py examples/pytorch/language-modeling/run_clm_no_trainer.py examples/pytorch/language-modeling/run_mlm.py examples/pytorch/language-modeling/run_mlm_no_trainer.py examples/pytorch/language-modeling/run_plm.py examples/pytorch/multiple-choice/run_swag.py examples/pytorch/multiple-choice/run_swag_no_trainer.py examples/pytorch/question-answering/run_qa.py examples/pytorch/question-answering/run_qa_beam_search.py examples/pytorch/question-answering/run_qa_beam_search_no_trainer.py examples/pytorch/question-answering/run_qa_no_trainer.py examples/pytorch/summarization/run_summarization.py examples/pytorch/summarization/run_summarization_no_trainer.py examples/pytorch/test_examples.py examples/pytorch/text-classification/run_glue.py examples/pytorch/text-classification/run_glue_no_trainer.py examples/pytorch/text-classification/run_xnli.py examples/pytorch/token-classification/run_ner.py examples/pytorch/token-classification/run_ner_no_trainer.py examples/pytorch/translation/run_translation.py examples/pytorch/translation/run_translation_no_trainer.py examples/research_projects/adversarial/utils_hans.py examples/research_projects/distillation/grouped_batch_sampler.py examples/research_projects/fsner/setup.py examples/research_projects/fsner/src/fsner/__init__.py examples/research_projects/fsner/src/fsner/model.py examples/research_projects/fsner/src/fsner/tokenizer_utils.py examples/research_projects/jax-projects/big_bird/evaluate.py examples/research_projects/jax-projects/hybrid_clip/run_hybrid_clip.py examples/tensorflow/language-modeling/run_clm.py examples/tensorflow/multiple-choice/run_swag.py examples/tensorflow/question-answering/run_qa.py examples/tensorflow/summarization/run_summarization.py examples/tensorflow/text-classification/run_glue.py examples/tensorflow/translation/run_translation.py src/transformers/__init__.py src/transformers/commands/add_new_model.py src/transformers/configuration_utils.py src/transformers/convert_slow_tokenizer.py src/transformers/data/__init__.py src/transformers/data/data_collator.py src/transformers/data/datasets/glue.py src/transformers/data/datasets/language_modeling.py src/transformers/data/datasets/squad.py src/transformers/deepspeed.py src/transformers/dependency_versions_table.py src/transformers/feature_extraction_sequence_utils.py src/transformers/file_utils.py src/transformers/generation_flax_utils.py src/transformers/generation_logits_process.py src/transformers/generation_tf_utils.py src/transformers/generation_utils.py src/transformers/integrations.py src/transformers/modelcard.py src/transformers/modeling_flax_utils.py src/transformers/modeling_outputs.py src/transformers/modeling_tf_utils.py src/transformers/modeling_utils.py src/transformers/models/__init__.py src/transformers/models/albert/__init__.py src/transformers/models/albert/modeling_albert.py src/transformers/models/albert/modeling_flax_albert.py src/transformers/models/albert/tokenization_albert_fast.py src/transformers/models/auto/__init__.py src/transformers/models/auto/auto_factory.py src/transformers/models/auto/configuration_auto.py src/transformers/models/auto/dynamic.py src/transformers/models/auto/feature_extraction_auto.py src/transformers/models/auto/modeling_auto.py src/transformers/models/auto/modeling_flax_auto.py src/transformers/models/auto/modeling_tf_auto.py src/transformers/models/auto/tokenization_auto.py src/transformers/models/bart/configuration_bart.py src/transformers/models/bart/modeling_bart.py src/transformers/models/bart/modeling_flax_bart.py src/transformers/models/bart/modeling_tf_bart.py src/transformers/models/barthez/tokenization_barthez_fast.py src/transformers/models/beit/__init__.py src/transformers/models/beit/configuration_beit.py src/transformers/models/beit/modeling_beit.py src/transformers/models/beit/modeling_flax_beit.py src/transformers/models/bert/configuration_bert.py src/transformers/models/bert/modeling_bert.py src/transformers/models/bert/modeling_flax_bert.py src/transformers/models/bert_generation/configuration_bert_generation.py src/transformers/models/bert_generation/modeling_bert_generation.py src/transformers/models/big_bird/configuration_big_bird.py src/transformers/models/big_bird/modeling_big_bird.py src/transformers/models/big_bird/modeling_flax_big_bird.py src/transformers/models/big_bird/tokenization_big_bird_fast.py src/transformers/models/bigbird_pegasus/configuration_bigbird_pegasus.py src/transformers/models/bigbird_pegasus/modeling_bigbird_pegasus.py src/transformers/models/blenderbot/configuration_blenderbot.py src/transformers/models/blenderbot/modeling_blenderbot.py src/transformers/models/blenderbot/modeling_tf_blenderbot.py src/transformers/models/blenderbot_small/configuration_blenderbot_small.py src/transformers/models/blenderbot_small/modeling_blenderbot_small.py src/transformers/models/blenderbot_small/modeling_tf_blenderbot_small.py src/transformers/models/byt5/tokenization_byt5.py src/transformers/models/camembert/tokenization_camembert_fast.py src/transformers/models/canine/configuration_canine.py src/transformers/models/canine/modeling_canine.py src/transformers/models/clip/configuration_clip.py src/transformers/models/clip/convert_clip_original_pytorch_to_hf.py src/transformers/models/clip/modeling_clip.py src/transformers/models/clip/modeling_flax_clip.py src/transformers/models/clip/tokenization_clip.py src/transformers/models/convbert/modeling_convbert.py src/transformers/models/ctrl/configuration_ctrl.py src/transformers/models/deberta/modeling_tf_deberta.py src/transformers/models/deberta_v2/__init__.py src/transformers/models/deberta_v2/modeling_deberta_v2.py src/transformers/models/deberta_v2/modeling_tf_deberta_v2.py src/transformers/models/deit/configuration_deit.py src/transformers/models/deit/modeling_deit.py src/transformers/models/detr/configuration_detr.py src/transformers/models/detr/modeling_detr.py src/transformers/models/distilbert/__init__.py src/transformers/models/distilbert/configuration_distilbert.py src/transformers/models/distilbert/modeling_distilbert.py src/transformers/models/distilbert/modeling_flax_distilbert.py src/transformers/models/dpr/configuration_dpr.py src/transformers/models/dpr/modeling_dpr.py src/transformers/models/electra/modeling_electra.py src/transformers/models/electra/modeling_flax_electra.py src/transformers/models/encoder_decoder/__init__.py src/transformers/models/encoder_decoder/modeling_encoder_decoder.py src/transformers/models/encoder_decoder/modeling_flax_encoder_decoder.py src/transformers/models/flaubert/configuration_flaubert.py src/transformers/models/flaubert/modeling_flaubert.py src/transformers/models/fnet/__init__.py src/transformers/models/fnet/configuration_fnet.py src/transformers/models/fnet/convert_fnet_original_flax_checkpoint_to_pytorch.py src/transformers/models/fnet/modeling_fnet.py src/transformers/models/fnet/tokenization_fnet.py src/transformers/models/fnet/tokenization_fnet_fast.py src/transformers/models/fsmt/configuration_fsmt.py src/transformers/models/fsmt/modeling_fsmt.py src/transformers/models/funnel/configuration_funnel.py src/transformers/models/gpt2/__init__.py src/transformers/models/gpt2/configuration_gpt2.py src/transformers/models/gpt2/modeling_flax_gpt2.py src/transformers/models/gpt2/modeling_gpt2.py src/transformers/models/gpt2/modeling_tf_gpt2.py src/transformers/models/gpt_neo/configuration_gpt_neo.py src/transformers/models/gpt_neo/modeling_gpt_neo.py src/transformers/models/gptj/__init__.py src/transformers/models/gptj/configuration_gptj.py src/transformers/models/gptj/modeling_gptj.py src/transformers/models/herbert/tokenization_herbert_fast.py src/transformers/models/hubert/__init__.py src/transformers/models/hubert/configuration_hubert.py src/transformers/models/hubert/convert_hubert_original_s3prl_checkpoint_to_pytorch.py src/transformers/models/hubert/modeling_hubert.py src/transformers/models/hubert/modeling_tf_hubert.py src/transformers/models/ibert/modeling_ibert.py src/transformers/models/layoutlm/__init__.py src/transformers/models/layoutlm/configuration_layoutlm.py src/transformers/models/layoutlm/modeling_layoutlm.py src/transformers/models/layoutlmv2/__init__.py src/transformers/models/layoutlmv2/configuration_layoutlmv2.py src/transformers/models/layoutlmv2/feature_extraction_layoutlmv2.py src/transformers/models/layoutlmv2/modeling_layoutlmv2.py src/transformers/models/layoutlmv2/processing_layoutlmv2.py src/transformers/models/layoutlmv2/tokenization_layoutlmv2.py src/transformers/models/layoutlmv2/tokenization_layoutlmv2_fast.py src/transformers/models/led/configuration_led.py src/transformers/models/led/modeling_led.py src/transformers/models/longformer/modeling_longformer.py src/transformers/models/luke/configuration_luke.py src/transformers/models/luke/modeling_luke.py src/transformers/models/luke/tokenization_luke.py src/transformers/models/lxmert/configuration_lxmert.py src/transformers/models/m2m_100/configuration_m2m_100.py src/transformers/models/m2m_100/modeling_m2m_100.py src/transformers/models/m2m_100/tokenization_m2m_100.py src/transformers/models/marian/configuration_marian.py src/transformers/models/marian/modeling_flax_marian.py src/transformers/models/marian/modeling_marian.py src/transformers/models/marian/modeling_tf_marian.py src/transformers/models/mbart/configuration_mbart.py src/transformers/models/mbart/modeling_flax_mbart.py src/transformers/models/mbart/modeling_mbart.py src/transformers/models/mbart/tokenization_mbart.py src/transformers/models/mbart/tokenization_mbart_fast.py src/transformers/models/mbart50/tokenization_mbart50.py src/transformers/models/mbart50/tokenization_mbart50_fast.py src/transformers/models/megatron_bert/configuration_megatron_bert.py src/transformers/models/megatron_bert/convert_megatron_bert_checkpoint.py src/transformers/models/megatron_bert/modeling_megatron_bert.py src/transformers/models/megatron_gpt2/convert_megatron_gpt2_checkpoint.py src/transformers/models/openai/configuration_openai.py src/transformers/models/pegasus/__init__.py src/transformers/models/pegasus/configuration_pegasus.py src/transformers/models/pegasus/modeling_flax_pegasus.py src/transformers/models/pegasus/modeling_pegasus.py src/transformers/models/pegasus/modeling_tf_pegasus.py src/transformers/models/pegasus/tokenization_pegasus_fast.py src/transformers/models/prophetnet/configuration_prophetnet.py src/transformers/models/prophetnet/modeling_prophetnet.py src/transformers/models/rag/modeling_rag.py src/transformers/models/rag/modeling_tf_rag.py src/transformers/models/reformer/configuration_reformer.py src/transformers/models/reformer/tokenization_reformer_fast.py src/transformers/models/rembert/configuration_rembert.py src/transformers/models/rembert/modeling_rembert.py src/transformers/models/rembert/tokenization_rembert_fast.py src/transformers/models/roberta/modeling_flax_roberta.py src/transformers/models/roberta/modeling_roberta.py src/transformers/models/roberta/modeling_tf_roberta.py src/transformers/models/roformer/configuration_roformer.py src/transformers/models/roformer/modeling_roformer.py src/transformers/models/speech_encoder_decoder/__init__.py src/transformers/models/speech_encoder_decoder/configuration_speech_encoder_decoder.py src/transformers/models/speech_encoder_decoder/convert_speech_to_text_wav2vec2_seq2seq_original_to_pytorch.py src/transformers/models/speech_encoder_decoder/modeling_speech_encoder_decoder.py src/transformers/models/speech_to_text/configuration_speech_to_text.py src/transformers/models/speech_to_text/feature_extraction_speech_to_text.py src/transformers/models/speech_to_text/modeling_speech_to_text.py src/transformers/models/speech_to_text_2/__init__.py src/transformers/models/speech_to_text_2/configuration_speech_to_text_2.py src/transformers/models/speech_to_text_2/modeling_speech_to_text_2.py src/transformers/models/speech_to_text_2/processing_speech_to_text_2.py src/transformers/models/speech_to_text_2/tokenization_speech_to_text_2.py src/transformers/models/splinter/configuration_splinter.py src/transformers/models/splinter/modeling_splinter.py src/transformers/models/t5/configuration_t5.py src/transformers/models/t5/modeling_flax_t5.py src/transformers/models/t5/modeling_t5.py src/transformers/models/t5/modeling_tf_t5.py src/transformers/models/t5/tokenization_t5_fast.py src/transformers/models/tapas/__init__.py src/transformers/models/tapas/configuration_tapas.py src/transformers/models/tapas/convert_tapas_original_tf_checkpoint_to_pytorch.py src/transformers/models/tapas/modeling_tapas.py src/transformers/models/tapas/tokenization_tapas.py src/transformers/models/transfo_xl/configuration_transfo_xl.py src/transformers/models/visual_bert/modeling_visual_bert.py src/transformers/models/vit/configuration_vit.py src/transformers/models/vit/convert_dino_to_pytorch.py src/transformers/models/vit/modeling_flax_vit.py src/transformers/models/vit/modeling_vit.py src/transformers/models/wav2vec2/__init__.py src/transformers/models/wav2vec2/configuration_wav2vec2.py src/transformers/models/wav2vec2/convert_wav2vec2_original_s3prl_checkpoint_to_pytorch.py src/transformers/models/wav2vec2/feature_extraction_wav2vec2.py src/transformers/models/wav2vec2/modeling_flax_wav2vec2.py src/transformers/models/wav2vec2/modeling_wav2vec2.py src/transformers/models/wav2vec2/tokenization_wav2vec2.py src/transformers/models/xlm/configuration_xlm.py src/transformers/models/xlm_roberta/tokenization_xlm_roberta.py src/transformers/models/xlm_roberta/tokenization_xlm_roberta_fast.py src/transformers/models/xlnet/configuration_xlnet.py src/transformers/models/xlnet/tokenization_xlnet_fast.py src/transformers/onnx/convert.py src/transformers/onnx/features.py src/transformers/optimization.py src/transformers/pipelines/__init__.py src/transformers/pipelines/audio_classification.py src/transformers/pipelines/automatic_speech_recognition.py src/transformers/pipelines/base.py src/transformers/pipelines/conversational.py src/transformers/pipelines/feature_extraction.py src/transformers/pipelines/fill_mask.py src/transformers/pipelines/image_classification.py src/transformers/pipelines/object_detection.py src/transformers/pipelines/question_answering.py src/transformers/pipelines/table_question_answering.py src/transformers/pipelines/text2text_generation.py src/transformers/pipelines/text_classification.py src/transformers/pipelines/text_generation.py src/transformers/pipelines/token_classification.py src/transformers/pipelines/zero_shot_classification.py src/transformers/testing_utils.py src/transformers/tokenization_utils.py src/transformers/tokenization_utils_base.py src/transformers/tokenization_utils_fast.py src/transformers/trainer.py src/transformers/trainer_callback.py src/transformers/trainer_pt_utils.py src/transformers/trainer_seq2seq.py src/transformers/trainer_utils.py src/transformers/training_args.py src/transformers/training_args_seq2seq.py src/transformers/utils/dummy_detectron2_objects.py src/transformers/utils/dummy_flax_objects.py src/transformers/utils/dummy_pt_objects.py src/transformers/utils/dummy_tf_objects.py src/transformers/utils/dummy_tokenizers_objects.py src/transformers/utils/dummy_vision_objects.py tests/deepspeed/test_deepspeed.py tests/sagemaker/conftest.py tests/sagemaker/test_multi_node_data_parallel.py tests/test_configuration_auto.py tests/test_configuration_common.py tests/test_data_collator.py tests/test_feature_extraction_auto.py tests/test_feature_extraction_layoutlmv2.py tests/test_feature_extraction_speech_to_text.py tests/test_feature_extraction_wav2vec2.py tests/test_file_utils.py tests/test_modeling_auto.py tests/test_modeling_bart.py tests/test_modeling_beit.py tests/test_modeling_bert.py tests/test_modeling_clip.py tests/test_modeling_common.py tests/test_modeling_convbert.py tests/test_modeling_deit.py tests/test_modeling_distilbert.py tests/test_modeling_encoder_decoder.py tests/test_modeling_flaubert.py tests/test_modeling_flax_albert.py tests/test_modeling_flax_bart.py tests/test_modeling_flax_beit.py tests/test_modeling_flax_distilbert.py tests/test_modeling_flax_encoder_decoder.py tests/test_modeling_flax_gpt2.py tests/test_modeling_flax_gpt_neo.py tests/test_modeling_flax_mt5.py tests/test_modeling_flax_pegasus.py tests/test_modeling_fnet.py tests/test_modeling_gpt2.py tests/test_modeling_gpt_neo.py tests/test_modeling_gptj.py tests/test_modeling_hubert.py tests/test_modeling_layoutlmv2.py tests/test_modeling_pegasus.py tests/test_modeling_rag.py tests/test_modeling_reformer.py tests/test_modeling_speech_encoder_decoder.py tests/test_modeling_speech_to_text.py tests/test_modeling_speech_to_text_2.py tests/test_modeling_tf_auto.py tests/test_modeling_tf_deberta_v2.py tests/test_modeling_tf_hubert.py tests/test_modeling_tf_pytorch.py tests/test_modeling_tf_wav2vec2.py tests/test_modeling_wav2vec2.py tests/test_onnx_v2.py tests/test_pipelines_audio_classification.py tests/test_pipelines_automatic_speech_recognition.py tests/test_pipelines_common.py tests/test_pipelines_conversational.py tests/test_pipelines_feature_extraction.py tests/test_pipelines_fill_mask.py tests/test_pipelines_image_classification.py tests/test_pipelines_object_detection.py tests/test_pipelines_question_answering.py tests/test_pipelines_summarization.py tests/test_pipelines_table_question_answering.py tests/test_pipelines_text2text_generation.py tests/test_pipelines_text_classification.py tests/test_pipelines_text_generation.py tests/test_pipelines_token_classification.py tests/test_pipelines_translation.py tests/test_pipelines_zero_shot.py tests/test_processor_layoutlmv2.py tests/test_processor_wav2vec2.py tests/test_sequence_feature_extraction_common.py tests/test_tokenization_auto.py tests/test_tokenization_byt5.py tests/test_tokenization_canine.py tests/test_tokenization_common.py tests/test_tokenization_fnet.py tests/test_tokenization_layoutlmv2.py tests/test_tokenization_luke.py tests/test_tokenization_mbart.py tests/test_tokenization_mbart50.py tests/test_tokenization_speech_to_text_2.py tests/test_tokenization_t5.py tests/test_tokenization_tapas.py tests/test_tokenization_xlm_roberta.py tests/test_trainer.py tests/test_trainer_distributed.py tests/test_trainer_tpu.py tests/test_utils_check_copies.py utils/check_copies.py utils/check_repo.py utils/notification_service.py utils/release.py utils/tests_fetcher.py
    python utils/custom_init_isort.py
    python utils/style_doc.py src/transformers docs/source --max_len 119
    running deps_table_update
    updating src/transformers/dependency_versions_table.py
    python utils/check_copies.py
    python utils/check_table.py
    python utils/check_dummies.py
    python utils/check_repo.py
    Checking all models are public.
    Checking all models are properly tested.
    Checking all objects are properly documented.
    Checking all models are in at least one auto class.
    python utils/check_inits.py
    python utils/tests_fetcher.py --sanity_check and fix suggested changes.
    
    * Run black examples tests src utils
    isort examples tests src utils
    Skipped 1 files
    make autogenerate_code
    make[1]: Entering directory '/mnt/c/Users/Admin/Desktop/Home/Projects/transformers'
    running deps_table_update
    updating src/transformers/dependency_versions_table.py
    make[1]: Leaving directory '/mnt/c/Users/Admin/Desktop/Home/Projects/transformers'
    make extra_style_checks
    make[1]: Entering directory '/mnt/c/Users/Admin/Desktop/Home/Projects/transformers'
    python utils/custom_init_isort.py
    python utils/style_doc.py src/transformers docs/source --max_len 119
    make[1]: Leaving directory '/mnt/c/Users/Admin/Desktop/Home/Projects/transformers' for reformatting code.
    
    * Add installation dependencies for examples/research_projects/fsner.
    
    * Add support to pass in variable numbers of examples to FSNER model.
    
    * Retrieve start_token_id and end_token_id from tokenizer instead of hardcoding in the FSNER model.
    
    * Run black examples tests src utils
    isort examples tests src utils
    Skipped 1 files
    make autogenerate_code
    make[1]: Entering directory '/home/saif/transformers'
    running deps_table_update
    updating src/transformers/dependency_versions_table.py
    make[1]: Leaving directory '/home/saif/transformers'
    make extra_style_checks
    make[1]: Entering directory '/home/saif/transformers'
    python utils/custom_init_isort.py
    python utils/style_doc.py src/transformers docs/source --max_len 119
    make[1]: Leaving directory '/home/saif/transformers' for FSNER
    
    * Update FSNER readme.md with a header image.
    
    * Update FSNER readme
    
    Co-authored-by: NielsRogge <[email protected]>
    Co-authored-by: Stefan Schweter <[email protected]>
    3 people authored Oct 6, 2021
    Configuration menu
    Copy the full SHA
    155b230 View commit details
    Browse the repository at this point in the history
  5. Configuration menu
    Copy the full SHA
    9f58bec View commit details
    Browse the repository at this point in the history
  6. Fixing Backward compatiblity for zero-shot (#13855)

    Fixes #13846
    Narsil authored Oct 6, 2021
    Configuration menu
    Copy the full SHA
    013bdc6 View commit details
    Browse the repository at this point in the history
  7. Configuration menu
    Copy the full SHA
    dac7798 View commit details
    Browse the repository at this point in the history
  8. T5ForConditionalGeneration: enabling using past_key_values and labels…

    … in training (#13805)
    
    * enabling using past_key_values together with labels when training in T5ForConditionalGeneration
    
    * test
    
    * Enable past_key_values in T5ForconditionalGeneration while training.
    
    * delete comments
    yssjtu authored Oct 6, 2021
    Configuration menu
    Copy the full SHA
    aea7c5b View commit details
    Browse the repository at this point in the history
  9. Fix trainer logging_nan_inf_filter in torch_xla mode (#13896)

    * Fix logging_nan_inf_filter in torch_xla mode
    
    * Update src/transformers/trainer.py
    
    Co-authored-by: Sylvain Gugger <[email protected]>
    
    * Fix format
    
    Co-authored-by: Sylvain Gugger <[email protected]>
    ymwangg and sgugger authored Oct 6, 2021
    Configuration menu
    Copy the full SHA
    77770ec View commit details
    Browse the repository at this point in the history
  10. Configuration menu
    Copy the full SHA
    8f2c07d View commit details
    Browse the repository at this point in the history
  11. Configuration menu
    Copy the full SHA
    5d390e9 View commit details
    Browse the repository at this point in the history
  12. Deploy docs for v4.11.3

    LysandreJik committed Oct 6, 2021
    Configuration menu
    Copy the full SHA
    5be59a3 View commit details
    Browse the repository at this point in the history

Commits on Oct 7, 2021

  1. Configuration menu
    Copy the full SHA
    0d309ce View commit details
    Browse the repository at this point in the history
  2. Add an example of exporting BartModel + BeamSearch to ONNX module. (#…

    …13765)
    
    * Add all example files.
    
    * Reformat files by black.
    
    * Style.
    
    * Remove unused imports.
    
    Co-authored-by: Morgan Funtowicz <[email protected]>
    fatcat-z and mfuntowicz authored Oct 7, 2021
    Configuration menu
    Copy the full SHA
    279ce5b View commit details
    Browse the repository at this point in the history
  3. #12789 Replace assert statements with exceptions (#13909)

    * #12789 Replace assert statements with exceptions
    
    * fix-copies: made copy changes to utils_qa.py in examples/pytorch/question-answering and examples/tensorflow/question-answering
    
    * minor refactor for clarity
    djroxx2000 authored Oct 7, 2021
    Configuration menu
    Copy the full SHA
    319beb6 View commit details
    Browse the repository at this point in the history
  4. Configuration menu
    Copy the full SHA
    57420b1 View commit details
    Browse the repository at this point in the history
  5. [Wav2Vec2] Fix mask_feature_prob (#13921)

    * up
    
    * overwrite hubert
    patrickvonplaten authored Oct 7, 2021
    Configuration menu
    Copy the full SHA
    0f5488f View commit details
    Browse the repository at this point in the history
  6. Add missing character (#13922)

    Mishig Davaadorj authored Oct 7, 2021
    Configuration menu
    Copy the full SHA
    5f34163 View commit details
    Browse the repository at this point in the history
  7. Fix incorrect output shapes for TF/PT LED (#13882)

    * Fix issues with LED model
    
    * Style pass
    
    * Bugfixes
    
    * correct attentions as well
    
    Co-authored-by: Patrick von Platen <[email protected]>
    Rocketknight1 and patrickvonplaten authored Oct 7, 2021
    Configuration menu
    Copy the full SHA
    61cf2ea View commit details
    Browse the repository at this point in the history
  8. [trainer] memory metrics: add memory at the start report (#13915)

    * [trainer] memory metrics: add memory at start
    
    * fix for no-gpu
    stas00 authored Oct 7, 2021
    Configuration menu
    Copy the full SHA
    be71ac3 View commit details
    Browse the repository at this point in the history

Commits on Oct 8, 2021

  1. Image Segmentation pipeline (#13828)

    * Implement img seg pipeline
    
    * Update src/transformers/pipelines/image_segmentation.py
    
    Co-authored-by: NielsRogge <[email protected]>
    
    * Update src/transformers/pipelines/image_segmentation.py
    
    Co-authored-by: NielsRogge <[email protected]>
    
    * Update output shape with individual masks
    
    * Rm dev change
    
    * Remove loops in test
    
    Co-authored-by: NielsRogge <[email protected]>
    Mishig Davaadorj and NielsRogge authored Oct 8, 2021
    Configuration menu
    Copy the full SHA
    026866d View commit details
    Browse the repository at this point in the history
  2. Adding support for tokens being suffixes or part of each other. (#13918)

    * Adding support for tokens being suffixes or part of each other.
    
    * Better test name.
    Narsil authored Oct 8, 2021
    Configuration menu
    Copy the full SHA
    d70919e View commit details
    Browse the repository at this point in the history
  3. Adds PreTrainedModel.framework attribute (#13817)

    * Added `framework` attribute
    
    * Update modeling_utils.py
    
    * Update modeling_flax_utils.py
    
    * Update modeling_tf_utils.py
    
    * Update modeling_utils.py
    
    * Update modeling_tf_utils.py
    
    * Update modeling_tf_utils.py
    
    * Update modeling_flax_utils.py
    
    * Update modeling_tf_utils.py
    
    * Update modeling_utils.py
    
    * Update modeling_utils.py
    
    * Update modeling_tf_utils.py
    
    * Update modeling_flax_utils.py
    
    * string -> str
    
    * Update modeling_tf_utils.py
    
    * string -> str
    
    * fixup
    
    * make flake happy
    
    Co-authored-by: patil-suraj <[email protected]>
    StellaAthena and patil-suraj authored Oct 8, 2021
    Configuration menu
    Copy the full SHA
    de34481 View commit details
    Browse the repository at this point in the history
  4. Configuration menu
    Copy the full SHA
    23ee06e View commit details
    Browse the repository at this point in the history
  5. Configuration menu
    Copy the full SHA
    5a1b5e4 View commit details
    Browse the repository at this point in the history
  6. [Generation] Fix max_new_tokens (#13919)

    * up
    
    * Update src/transformers/generation_stopping_criteria.py
    
    * finish
    patrickvonplaten authored Oct 8, 2021
    Configuration menu
    Copy the full SHA
    c8b0761 View commit details
    Browse the repository at this point in the history
  7. Style

    sgugger committed Oct 8, 2021
    Configuration menu
    Copy the full SHA
    cb911e5 View commit details
    Browse the repository at this point in the history
  8. Move to TF only

    sgugger committed Oct 8, 2021
    Configuration menu
    Copy the full SHA
    9e15b51 View commit details
    Browse the repository at this point in the history
  9. Configuration menu
    Copy the full SHA
    3e21852 View commit details
    Browse the repository at this point in the history
  10. Configuration menu
    Copy the full SHA
    46dfe99 View commit details
    Browse the repository at this point in the history
  11. Update bug-report.md (#13934)

    * Update bug-report.md
    
    * Update .github/ISSUE_TEMPLATE/bug-report.md
    
    Co-authored-by: Suraj Patil <[email protected]>
    
    * Update .github/ISSUE_TEMPLATE/bug-report.md
    
    Co-authored-by: Suraj Patil <[email protected]>
    
    * Update .github/ISSUE_TEMPLATE/bug-report.md
    
    Co-authored-by: NielsRogge <[email protected]>
    
    * Update .github/ISSUE_TEMPLATE/bug-report.md
    
    Co-authored-by: Anton Lozhkov <[email protected]>
    
    Co-authored-by: Suraj Patil <[email protected]>
    Co-authored-by: NielsRogge <[email protected]>
    Co-authored-by: Anton Lozhkov <[email protected]>
    4 people authored Oct 8, 2021
    Configuration menu
    Copy the full SHA
    239bd61 View commit details
    Browse the repository at this point in the history

Commits on Oct 9, 2021

  1. fix issue 13904 -attribute does not exist- by change self_.mapping to…

    … self._model_mapping (#13942)
    oraby8 authored Oct 9, 2021
    Configuration menu
    Copy the full SHA
    91758e3 View commit details
    Browse the repository at this point in the history

Commits on Oct 11, 2021

  1. Raise ValueError instead of asserts in src/transformers/benchmark/ben…

    …chmark.py (#13951)
    
    * Raise ValueError exception instead of assert
    
    * Remove f unnecessary f-strings
    
    * Remove unused f-strings
    AkechiShiro authored Oct 11, 2021
    Configuration menu
    Copy the full SHA
    3c0c699 View commit details
    Browse the repository at this point in the history
  2. Honor existing attention mask in tokenzier.pad (#13926)

    * Honor existing attention mask in tokenzier.pad
    
    * Fix initialization of attention mask
    
    * Roll the implem on all subclasses
    
    * Fix tests
    sgugger authored Oct 11, 2021
    Configuration menu
    Copy the full SHA
    4a18337 View commit details
    Browse the repository at this point in the history
  3. [Gradient checkpoining] Correct disabling find_unused_parameters in…

    … Trainer when gradient checkpointing is enabled (#13961)
    
    * up
    
    * correct test
    patrickvonplaten authored Oct 11, 2021
    Configuration menu
    Copy the full SHA
    dca6796 View commit details
    Browse the repository at this point in the history
  4. Configuration menu
    Copy the full SHA
    6e4c8f6 View commit details
    Browse the repository at this point in the history
  5. Configuration menu
    Copy the full SHA
    e1bb2eb View commit details
    Browse the repository at this point in the history
  6. Configuration menu
    Copy the full SHA
    708ffff View commit details
    Browse the repository at this point in the history
  7. Configuration menu
    Copy the full SHA
    32634bc View commit details
    Browse the repository at this point in the history
  8. Configuration menu
    Copy the full SHA
    64743d0 View commit details
    Browse the repository at this point in the history
  9. Replace assert by ValueError of src/transformers/models/electra/model…

    …ing_{electra,tf_electra}.py and all other models that had copies (#13955)
    
    * Replace all assert by ValueError in src/transformers/models/electra
    
    * Reformat with black to pass check_code_quality test
    
    * Change some assert to ValueError of modeling_bert & modeling_tf_albert
    
    * Change some assert in multiples models
    
    * Change multiples models assertion to ValueError in order to validate
      check_code_style test and models template test.
    
    * Black reformat
    
    * Change some more asserts in multiples models
    
    * Change assert to ValueError in modeling_layoutlm.py to fix copy error in code_style_check
    
    * Add proper message to ValueError in modeling_tf_albert.py
    
    Co-authored-by: Sylvain Gugger <[email protected]>
    
    * Simplify logic in models/bert/modeling_bert.py
    
    Co-authored-by: Sylvain Gugger <[email protected]>
    
    * Add ValueError message to models/convbert/modeling_tf_convbert.py
    
    Co-authored-by: Sylvain Gugger <[email protected]>
    
    * Add error message for ValueError to modeling_tf_electra.py
    
    Co-authored-by: Sylvain Gugger <[email protected]>
    
    * Simplify logic in models/tapas/modeling_tapas.py
    
    Co-authored-by: Sylvain Gugger <[email protected]>
    
    * Simplify logic in models/electra/modeling_electra.py
    
    Co-authored-by: Sylvain Gugger <[email protected]>
    
    * Add ValueError message in src/transformers/models/bert/modeling_tf_bert.py
    
    Co-authored-by: Sylvain Gugger <[email protected]>
    
    * Simplify logic in src/transformers/models/rembert/modeling_rembert.py
    
    Co-authored-by: Sylvain Gugger <[email protected]>
    
    * Simplify logic in src/transformers/models/albert/modeling_albert.py
    
    Co-authored-by: Sylvain Gugger <[email protected]>
    
    Co-authored-by: Sylvain Gugger <[email protected]>
    AkechiShiro and sgugger authored Oct 11, 2021
    Configuration menu
    Copy the full SHA
    3499728 View commit details
    Browse the repository at this point in the history
  10. [Speech Examples] Add pytorch speech pretraining (#13877)

    * adapt wav2vec2
    
    * add example
    
    * add files
    
    * adapt
    
    * remove bogus file
    
    * Apply suggestions from code review
    
    * adapt files more
    
    * upload changes
    
    * del old files
    
    * up
    
    * up
    
    * up
    
    * up
    
    * up
    
    * correct gradient checkpoitning
    
    * add readme
    
    * finish
    
    * finish
    
    * up
    
    * more fixes
    
    * up
    
    * up
    
    * add demo run to readme
    
    * up
    patrickvonplaten authored Oct 11, 2021
    Configuration menu
    Copy the full SHA
    d45fc7d View commit details
    Browse the repository at this point in the history

Commits on Oct 12, 2021

  1. Configuration menu
    Copy the full SHA
    990de2c View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    85d69a7 View commit details
    Browse the repository at this point in the history
  3. Specify im-seg mask greyscole mode (#13974)

    Mishig Davaadorj authored Oct 12, 2021
    Configuration menu
    Copy the full SHA
    11c043d View commit details
    Browse the repository at this point in the history
  4. [Wav2Vec2] Make sure tensors are always bool for mask_indices (#13977)

    * correct long to bool
    
    * up
    
    * correct code
    patrickvonplaten authored Oct 12, 2021
    Configuration menu
    Copy the full SHA
    58bf882 View commit details
    Browse the repository at this point in the history
  5. Fixing the lecture values by making sure defaults are not changed (#1…

    …3976)
    
    384 // 4 < 128 would break `doc_stride`.
    Narsil authored Oct 12, 2021
    Configuration menu
    Copy the full SHA
    26b6ef7 View commit details
    Browse the repository at this point in the history
  6. Add TFEncoderDecoderModel + Add cross-attention to some TF models (#1…

    …3222)
    
    * Add cross attentions to TFGPT2Model
    
    * Add TFEncoderDecoderModel
    
    * Add TFBaseModelOutputWithPoolingAndCrossAttentions
    
    * Add cross attentions to TFBertModel
    
    * Fix past or past_key_values argument issue
    
    * Fix generation
    
    * Fix save and load
    
    * Add some checks and comments
    
    * Clean the code that deals with past keys/values
    
    * Add kwargs to processing_inputs
    
    * Add serving_output to TFEncoderDecoderModel
    
    * Some cleaning + fix use_cache value issue
    
    * Fix tests + add bert2bert/bert2gpt2 tests
    
    * Fix more tests
    
    * Ignore crossattention.bias when loading GPT2 weights into TFGPT2
    
    * Fix return_dict_in_generate in tf generation
    
    * Fix is_token_logit_eos_token bug in tf generation
    
    * Finalize the tests after fixing some bugs
    
    * Fix another is_token_logit_eos_token bug in tf generation
    
    * Add/Update docs
    
    * Add TFBertEncoderDecoderModelTest
    
    * Clean test script
    
    * Add TFEncoderDecoderModel to the library
    
    * Add cross attentions to TFRobertaModel
    
    * Add TFRobertaEncoderDecoderModelTest
    
    * make style
    
    * Change the way of position_ids computation
    
    * bug fix
    
    * Fix copies in tf_albert
    
    * Remove some copied from and apply some fix-copies
    
    * Remove some copied
    
    * Add cross attentions to some other TF models
    
    * Remove encoder_hidden_states from TFLayoutLMModel.call for now
    
    * Make style
    
    * Fix TFRemBertForCausalLM
    
    * Revert the change to longformer + Remove copies
    
    * Revert the change to albert and convbert + Remove copies
    
    * make quality
    
    * make style
    
    * Add TFRembertEncoderDecoderModelTest
    
    * make quality and fix-copies
    
    * test TFRobertaForCausalLM
    
    * Fixes for failed tests
    
    * Fixes for failed tests
    
    * fix more tests
    
    * Fixes for failed tests
    
    * Fix Auto mapping order
    
    * Fix TFRemBertEncoder return value
    
    * fix tf_rembert
    
    * Check copies are OK
    
    * Fix missing TFBaseModelOutputWithPastAndCrossAttentions is not defined
    
    * Add TFEncoderDecoderModelSaveLoadTests
    
    * fix tf weight loading
    
    * check the change of use_cache
    
    * Revert the change
    
    * Add missing test_for_causal_lm for TFRobertaModelTest
    
    * Try cleaning past
    
    * fix _reorder_cache
    
    * Revert some files to original versions
    
    * Keep as many copies as possible
    
    * Apply suggested changes - Use raise ValueError instead of assert
    
    * Move import to top
    
    * Fix wrong require_torch
    
    * Replace more assert by raise ValueError
    
    * Add test_pt_tf_model_equivalence (the test won't pass for now)
    
    * add test for loading/saving
    
    * finish
    
    * finish
    
    * Remove test_pt_tf_model_equivalence
    
    * Update tf modeling template
    
    * Remove pooling, added in the prev. commit, from MainLayer
    
    * Update tf modeling test template
    
    * Move inputs["use_cache"] = False to modeling_tf_utils.py
    
    * Fix torch.Tensor in the comment
    
    * fix use_cache
    
    * Fix missing use_cache in ElectraConfig
    
    * Add a note to from_pretrained
    
    * Fix style
    
    * Change test_encoder_decoder_save_load_from_encoder_decoder_from_pt
    
    * Fix TFMLP (in TFGPT2) activation issue
    
    * Fix None past_key_values value in serving_output
    
    * Don't call get_encoderdecoder_model in TFEncoderDecoderModelTest.test_configuration_tie until we have a TF checkpoint on Hub
    
    * Apply review suggestions - style for cross_attns in serving_output
    
    * Apply review suggestions - change assert + docstrings
    
    * break the error message to respect the char limit
    
    * deprecate the argument past
    
    * fix docstring style
    
    * Update the encoder-decoder rst file
    
    * fix Unknown interpreted text role "method"
    
    * fix typo
    
    Co-authored-by: ydshieh <[email protected]>
    Co-authored-by: Patrick von Platen <[email protected]>
    3 people authored Oct 12, 2021
    Configuration menu
    Copy the full SHA
    8b240a0 View commit details
    Browse the repository at this point in the history
  7. Configuration menu
    Copy the full SHA
    61f6426 View commit details
    Browse the repository at this point in the history

Commits on Oct 13, 2021

  1. Add TrOCR + VisionEncoderDecoderModel (#13874)

    * First draft
    
    * Update self-attention of RoBERTa as proposition
    
    * Improve conversion script
    
    * Add TrOCR decoder-only model
    
    * More improvements
    
    * Make forward pass with pretrained weights work
    
    * More improvements
    
    * Some more improvements
    
    * More improvements
    
    * Make conversion work
    
    * Clean up print statements
    
    * Add documentation, processor
    
    * Add test files
    
    * Small improvements
    
    * Some more improvements
    
    * Make fix-copies, improve docs
    
    * Make all vision encoder decoder model tests pass
    
    * Make conversion script support other models
    
    * Update URL for OCR image
    
    * Update conversion script
    
    * Fix style & quality
    
    * Add support for the large-printed model
    
    * Fix some issues
    
    * Add print statement for debugging
    
    * Add print statements for debugging
    
    * Make possible fix for sinusoidal embedding
    
    * Further debugging
    
    * Potential fix v2
    
    * Add more print statements for debugging
    
    * Add more print statements for debugging
    
    * Deubg more
    
    * Comment out print statements
    
    * Make conversion of large printed model possible, address review comments
    
    * Make it possible to convert the stage1 checkpoints
    
    * Clean up code, apply suggestions from code review
    
    * Apply suggestions from code review, use Microsoft models in tests
    
    * Rename encoder_hidden_size to cross_attention_hidden_size
    
    * Improve docs
    NielsRogge authored Oct 13, 2021
    Configuration menu
    Copy the full SHA
    408b2d2 View commit details
    Browse the repository at this point in the history

Commits on Oct 14, 2021

  1. Remove wrong model_args supplied (#13937)

    * Remove wrong model_args of config.from_pretrained
    
    * Fix tf & flax
    qqaatw authored Oct 14, 2021
    Configuration menu
    Copy the full SHA
    51ee20f View commit details
    Browse the repository at this point in the history
  2. Skip faulty test

    sgugger committed Oct 14, 2021
    Configuration menu
    Copy the full SHA
    5b6bd4e View commit details
    Browse the repository at this point in the history
  3. up (#13988)

    patrickvonplaten authored Oct 14, 2021
    Configuration menu
    Copy the full SHA
    cc36064 View commit details
    Browse the repository at this point in the history
  4. Configuration menu
    Copy the full SHA
    a5be954 View commit details
    Browse the repository at this point in the history
  5. Revert "Skip faulty test"

    This reverts commit 5b6bd4e.
    sgugger committed Oct 14, 2021
    Configuration menu
    Copy the full SHA
    0ef61d3 View commit details
    Browse the repository at this point in the history
  6. Add strong test for configuration attributes (#14000)

    * Add strong test for configuration attributes
    
    * Add fake modif to trigger all tests
    
    * Add a better fake modif
    
    * Ignore is_encoder_decoder
    
    * Fix faulty configs
    
    * Remove fake modif
    sgugger authored Oct 14, 2021
    Configuration menu
    Copy the full SHA
    f2002fe View commit details
    Browse the repository at this point in the history
  7. Configuration menu
    Copy the full SHA
    7604557 View commit details
    Browse the repository at this point in the history
  8. up (#14008)

    patrickvonplaten authored Oct 14, 2021
    Configuration menu
    Copy the full SHA
    7fb2a8b View commit details
    Browse the repository at this point in the history
  9. Raise exceptions instead of asserts in src/transformers/models/bart/m…

    …odeling_flax_[bart, marian, mbart, pegasus].py (#13939)
    
    * Raise exceptions instead of asserts
    
    * fix: fixed failing quality check with copies
    
    * fix: fixed max line length
    
    * rerun github ci, failed to install dependencies
    killazz67 authored Oct 14, 2021
    Configuration menu
    Copy the full SHA
    b65c389 View commit details
    Browse the repository at this point in the history
  10. Scatter dummies + skip pipeline tests (#13996)

    * Scatter dummies + skip pipeline tests
    
    * Add torch scatter to build docs
    LysandreJik authored Oct 14, 2021
    Configuration menu
    Copy the full SHA
    5b317f7 View commit details
    Browse the repository at this point in the history

Commits on Oct 15, 2021

  1. Fixed horizon_length for PPLM (#13886)

    * fixed horizon_length
    
    * fixed horizon_length
    
    * fix style
    jacksukk authored Oct 15, 2021
    Configuration menu
    Copy the full SHA
    d5b82bb View commit details
    Browse the repository at this point in the history
  2. Add the SEW and SEW-D speech models (#13962)

    * Working encoder
    
    * SEW-D and tests
    
    * Further conv fixes
    
    * Automodels and conv inits
    
    * Update integration tests, add docs
    
    * Docs cleanup, resolve todos
    
    * Conf fix
    
    * Fix docs
    
    * Fix tests, apply suggestions
    
    * Update src/transformers/models/sew/modeling_sew.py
    
    Co-authored-by: Patrick von Platen <[email protected]>
    
    * Model conversion and updated no-mask tests
    
    * Remove copy of feature_proj
    
    * Style
    
    * Update src/transformers/models/auto/feature_extraction_auto.py
    
    Co-authored-by: Patrick von Platen <[email protected]>
    
    * Update src/transformers/models/auto/feature_extraction_auto.py
    
    Co-authored-by: Patrick von Platen <[email protected]>
    
    * Move orgs
    
    Co-authored-by: Patrick von Platen <[email protected]>
    anton-l and patrickvonplaten authored Oct 15, 2021
    Configuration menu
    Copy the full SHA
    cd3166a View commit details
    Browse the repository at this point in the history
  3. Configuration menu
    Copy the full SHA
    47489a6 View commit details
    Browse the repository at this point in the history
  4. [Docs] More general docstrings (#14028)

    * up
    
    * finish
    
    * up
    
    * up
    
    * finish
    patrickvonplaten authored Oct 15, 2021
    Configuration menu
    Copy the full SHA
    f5af873 View commit details
    Browse the repository at this point in the history

Commits on Oct 16, 2021

  1. minor fixes (#14026)

    patil-suraj authored Oct 16, 2021
    Configuration menu
    Copy the full SHA
    84ad6af View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    968ae57 View commit details
    Browse the repository at this point in the history
  3. Replace assertions with ValueError exceptions (#14018)

    * Replace assertions with ValueError exceptions
    
    * Change length check for a more explicit one
    ddrm86 authored Oct 16, 2021
    Configuration menu
    Copy the full SHA
    cde0c75 View commit details
    Browse the repository at this point in the history

Commits on Oct 17, 2021

  1. [Speech Examples] Add new audio feature (#14027)

    * finish
    
    * up
    
    * finish all
    
    * up
    patrickvonplaten authored Oct 17, 2021
    Configuration menu
    Copy the full SHA
    37c5759 View commit details
    Browse the repository at this point in the history

Commits on Oct 18, 2021

  1. Fix typo (#14044)

    Mishig Davaadorj authored Oct 18, 2021
    Configuration menu
    Copy the full SHA
    4334095 View commit details
    Browse the repository at this point in the history
  2. [Speech] Move all examples to new audio feature (#14045)

    * up
    
    * up
    
    * up
    
    * finish
    patrickvonplaten authored Oct 18, 2021
    Configuration menu
    Copy the full SHA
    bdf31d6 View commit details
    Browse the repository at this point in the history
  3. Configuration menu
    Copy the full SHA
    82b62fa View commit details
    Browse the repository at this point in the history
  4. up (#14046)

    patrickvonplaten authored Oct 18, 2021
    Configuration menu
    Copy the full SHA
    7c6cd0a View commit details
    Browse the repository at this point in the history
  5. Add BARTpho: Pre-trained Sequence-to-Sequence Models for Vietnamese (…

    …#13788)
    
    * Add the pre-trained BARTpho model
    
    * Add the pre-trained BARTpho model
    
    * Add the pre-trained BARTpho model
    
    * Fix incorrectly sorted and/or formatted imports
    
    * Fix incorrectly sorted and/or formatted style
    
    * Fix check_dummies
    
    * Fix check_dummies
    
    * Fix check_dummies
    
    * Update docs/source/model_doc/bartpho.rst
    
    Co-authored-by: Suraj Patil <[email protected]>
    
    * Update src/transformers/models/bartpho/__init__.py
    
    Co-authored-by: Suraj Patil <[email protected]>
    
    * Update src/transformers/models/bartpho/tokenization_bartpho.py
    
    Co-authored-by: Suraj Patil <[email protected]>
    
    * Update tests/test_tokenization_bartpho.py
    
    Co-authored-by: Suraj Patil <[email protected]>
    
    * Update src/transformers/models/bartpho/tokenization_bartpho.py
    
    Co-authored-by: Suraj Patil <[email protected]>
    
    * Update tests/test_tokenization_bartpho.py
    
    Co-authored-by: Suraj Patil <[email protected]>
    
    * Update docs/source/model_doc/bartpho.rst
    
    Co-authored-by: Sylvain Gugger <[email protected]>
    
    * Update docs/source/model_doc/bartpho.rst
    
    Co-authored-by: Sylvain Gugger <[email protected]>
    
    * Update src/transformers/models/bartpho/__init__.py
    
    Co-authored-by: Suraj Patil <[email protected]>
    
    * Add the pre-trained BARTpho model
    
    * Add Tips section in doc and details of monolingual_vocab_file
    
    * Fix conflicts
    
    * Add another tip related to monolingual_vocab_file
    
    * Readd dependency_versions_table.py
    
    * Handle failing checks
    
    * Remove test_list.txt
    
    * Remove md5sum.saved
    
    * Revise Readme.md
    
    Co-authored-by: Suraj Patil <[email protected]>
    Co-authored-by: Sylvain Gugger <[email protected]>
    3 people authored Oct 18, 2021
    Configuration menu
    Copy the full SHA
    3d587c5 View commit details
    Browse the repository at this point in the history
  6. Add an API to register objects to Auto classes (#13989)

    * Add API to register a new object in auto classes
    
    * Fix test
    
    * Documentation
    
    * Add to tokenizers and test
    
    * Add cleanup after tests
    
    * Be more careful
    
    * Move import
    
    * Move import
    
    * Cleanup in TF test too
    
    * Add consistency check
    
    * Add documentation
    
    * Style
    
    * Update docs/source/model_doc/auto.rst
    
    Co-authored-by: Lysandre Debut <[email protected]>
    
    * Update src/transformers/models/auto/auto_factory.py
    
    Co-authored-by: Lysandre Debut <[email protected]>
    
    Co-authored-by: Lysandre Debut <[email protected]>
    sgugger and LysandreJik authored Oct 18, 2021
    Configuration menu
    Copy the full SHA
    2c60ff2 View commit details
    Browse the repository at this point in the history
  7. Configuration menu
    Copy the full SHA
    2024faf View commit details
    Browse the repository at this point in the history
  8. [Speech] Refactor Examples (#14040)

    * adapt_examples
    
    * up
    
    * up
    
    * up
    
    * up
    
    * add auto models
    
    * finish
    patrickvonplaten authored Oct 18, 2021
    Configuration menu
    Copy the full SHA
    d5ff69f View commit details
    Browse the repository at this point in the history
  9. fix typo (#14049)

    yyy-Apple authored Oct 18, 2021
    Configuration menu
    Copy the full SHA
    7a3147e View commit details
    Browse the repository at this point in the history
  10. Fix typo (#14056)

    ihoromi4 authored Oct 18, 2021
    Configuration menu
    Copy the full SHA
    9eda0d1 View commit details
    Browse the repository at this point in the history

Commits on Oct 19, 2021

  1. Configuration menu
    Copy the full SHA
    fde4867 View commit details
    Browse the repository at this point in the history
  2. TF Model train and eval step metrics for seq2seq models. (#14009)

    * TF Model train and eval step metrics for seq2seq models.
    
    When using a model with a seq2seq output compute metrics against logits.
    
    * Removing vestigial code
    
    Co-authored-by: matt <[email protected]>
    pedro-r-marques and Rocketknight1 authored Oct 19, 2021
    Configuration menu
    Copy the full SHA
    122c2f8 View commit details
    Browse the repository at this point in the history
  3. update to_py_obj to support np.number (#14064)

    Co-authored-by: 眸浩 <[email protected]>
    PrettyMeng and 眸浩 authored Oct 19, 2021
    Configuration menu
    Copy the full SHA
    3892d09 View commit details
    Browse the repository at this point in the history

Commits on Oct 20, 2021

  1. Configuration menu
    Copy the full SHA
    3fefa29 View commit details
    Browse the repository at this point in the history
  2. [Examples] Use Audio feature in speech classification (#14052)

    * Update SEW integration test tolerance
    
    * Update audio classification
    
    * Update test
    
    * Remove torchaudio
    
    * Add dataset revision
    
    * Hub branch naming
    
    * Revert dataset revisions
    
    * Update datasets
    anton-l authored Oct 20, 2021
    Configuration menu
    Copy the full SHA
    dbaf492 View commit details
    Browse the repository at this point in the history
  3. Add ASR colabs (#14067)

    * up
    
    * Update notebooks/README.md
    patrickvonplaten authored Oct 20, 2021
    Configuration menu
    Copy the full SHA
    0bc2e54 View commit details
    Browse the repository at this point in the history
  4. up (#14079)

    patrickvonplaten authored Oct 20, 2021
    Configuration menu
    Copy the full SHA
    53dc39d View commit details
    Browse the repository at this point in the history
  5. replace assert with exception in src/transformers/utils/model_pararal…

    …lel_utils.py (#14072)
    
    * replace assert with exception in src/transformers/utils/model_parallel_utils.py
    
    * fix some code style
    
    * fix typo
    
    Co-authored-by: skpig <[email protected]>
    skpig and skpig authored Oct 20, 2021
    Configuration menu
    Copy the full SHA
    a43d935 View commit details
    Browse the repository at this point in the history
  6. Configuration menu
    Copy the full SHA
    0106826 View commit details
    Browse the repository at this point in the history
  7. Fix assert in src/transformers/data/datasets/language_modeling.py (#1…

    …4077)
    
    * replace assertion with ValueError
    
    * fix code style
    
    Co-authored-by: skpig <[email protected]>
    skpig and skpig authored Oct 20, 2021
    Configuration menu
    Copy the full SHA
    31560f6 View commit details
    Browse the repository at this point in the history
  8. Configuration menu
    Copy the full SHA
    f875fb0 View commit details
    Browse the repository at this point in the history
  9. Context managers (#13900)

    * add `ContextManagers` for lists of contexts
    
    * fix import sorting
    
    * add `ContextManagers` tests
    lvwerra authored Oct 20, 2021
    Configuration menu
    Copy the full SHA
    0270d44 View commit details
    Browse the repository at this point in the history
  10. Configuration menu
    Copy the full SHA
    49155d2 View commit details
    Browse the repository at this point in the history

Commits on Oct 21, 2021

  1. up (#14093)

    patrickvonplaten authored Oct 21, 2021
    Configuration menu
    Copy the full SHA
    e9d2a63 View commit details
    Browse the repository at this point in the history
  2. Change asserts in src/transformers/models/xlnet/ to raise ValueError …

    …(#14088)
    
    * Change asserts in src/transformers/models/xlnet/ to raise ValueError
    
    * Update src/transformers/models/xlnet/modeling_tf_xlnet.py
    
    Co-authored-by: Sylvain Gugger <[email protected]>
    
    Co-authored-by: Sylvain Gugger <[email protected]>
    westonkl and sgugger authored Oct 21, 2021
    Configuration menu
    Copy the full SHA
    9e4ea25 View commit details
    Browse the repository at this point in the history
  3. Replace assertions with ValueError exceptions (#14061)

    * Replace assertions with ValueError exceptions
    
    * Format error messages as suggested
    ddrm86 authored Oct 21, 2021
    Configuration menu
    Copy the full SHA
    3187228 View commit details
    Browse the repository at this point in the history
  4. Configuration menu
    Copy the full SHA
    f9c16b0 View commit details
    Browse the repository at this point in the history
  5. Pin PyTorch to make CI green

    sgugger committed Oct 21, 2021
    Configuration menu
    Copy the full SHA
    0f50268 View commit details
    Browse the repository at this point in the history
  6. [Examples] Add audio classification notebooks (#14099)

    * Update SEW integration test tolerance
    
    * Add audio classification notebooks
    anton-l authored Oct 21, 2021
    Configuration menu
    Copy the full SHA
    e03544a View commit details
    Browse the repository at this point in the history
  7. Fix ignore_mismatched_sizes (#14085)

    * Fix
    
    * Style
    
    * Name
    
    * Fix tests
    
    * Style
    
    * Remove embed sizes checking
    
    * Disable some tests
    
    * Fix
    
    * Apply suggestion
    qqaatw authored Oct 21, 2021
    Configuration menu
    Copy the full SHA
    234cfef View commit details
    Browse the repository at this point in the history
  8. Configuration menu
    Copy the full SHA
    f00bcea View commit details
    Browse the repository at this point in the history
  9. Configuration menu
    Copy the full SHA
    7af55d3 View commit details
    Browse the repository at this point in the history
  10. fix typo in license docstring (#14094)

    last line: "# limitations under the License." is missing
    21jun authored Oct 21, 2021
    Configuration menu
    Copy the full SHA
    d432a65 View commit details
    Browse the repository at this point in the history
  11. Configuration menu
    Copy the full SHA
    7888914 View commit details
    Browse the repository at this point in the history
  12. Configuration menu
    Copy the full SHA
    ca2ef7d View commit details
    Browse the repository at this point in the history

Commits on Oct 22, 2021

  1. up (#14116)

    patrickvonplaten authored Oct 22, 2021
    Configuration menu
    Copy the full SHA
    70f186f View commit details
    Browse the repository at this point in the history
  2. Replace assert statements with exceptions (#13871) (#13901)

    * Replace assert statements with exceptions (#13871)
    
    * Change f-strings when not needed (flake8)
    
    * Replace assert statements with exceptions (#13871)
    
    * Change f-strings when not needed (flake8)
    
    * Improve error message as suggested by reviewer
    
    * Fix identation bug
    
    * Fix style errors
    ddrm86 authored Oct 22, 2021
    Configuration menu
    Copy the full SHA
    f5a49bf View commit details
    Browse the repository at this point in the history
  3. Translate README.md to Korean (#14015)

    * Create README_ko.md
    
    * Update README.md
    
    * Update README_zh-hans.md
    
    * Update README_zh-hant.md
    
    * Update README_ko.md
    
    * Update check_copies.py
    
    * Update README_ko.md
    
    * typo
    
    * match with readme_ko
    yeounyi authored Oct 22, 2021
    Configuration menu
    Copy the full SHA
    9f53f04 View commit details
    Browse the repository at this point in the history
  4. Replace assertions with valueError Exeptions (#14117)

    * Replace assertions with valueError Exeptions
    
    * Reformatted
    jyshdewangan authored Oct 22, 2021
    Configuration menu
    Copy the full SHA
    fa4abdb View commit details
    Browse the repository at this point in the history
  5. Configuration menu
    Copy the full SHA
    16d7b70 View commit details
    Browse the repository at this point in the history
  6. Fix assertion in models (#14090)

    * replace assertions in src/transformers/models/luke/convert_luke_original_pytorch_checkpoint_to_pytorch.py
    
    * replace assertions in src/transformers/models/marian/convert_marian_to_pytorch.py
    
    * Update src/transformers/models/luke/convert_luke_original_pytorch_checkpoint_to_pytorch.py
    
    Co-authored-by: Sylvain Gugger <[email protected]>
    
    * Update src/transformers/models/marian/convert_marian_to_pytorch.py
    
    Co-authored-by: Sylvain Gugger <[email protected]>
    
    * Update src/transformers/models/marian/convert_marian_to_pytorch.py
    
    Co-authored-by: Sylvain Gugger <[email protected]>
    
    * Update src/transformers/models/marian/convert_marian_to_pytorch.py
    
    Co-authored-by: Sylvain Gugger <[email protected]>
    
    * Update src/transformers/models/marian/convert_marian_to_pytorch.py
    
    Co-authored-by: Sylvain Gugger <[email protected]>
    
    * Update src/transformers/models/marian/convert_marian_to_pytorch.py
    
    Co-authored-by: Sylvain Gugger <[email protected]>
    
    * Update src/transformers/models/marian/convert_marian_to_pytorch.py
    
    Co-authored-by: Sylvain Gugger <[email protected]>
    
    * Update src/transformers/models/marian/convert_marian_to_pytorch.py
    
    Co-authored-by: Sylvain Gugger <[email protected]>
    
    * Update src/transformers/models/marian/convert_marian_to_pytorch.py
    
    Co-authored-by: Sylvain Gugger <[email protected]>
    
    Co-authored-by: skpig <[email protected]>
    Co-authored-by: Sylvain Gugger <[email protected]>
    3 people authored Oct 22, 2021
    Configuration menu
    Copy the full SHA
    c7ccb2e View commit details
    Browse the repository at this point in the history
  7. Configuration menu
    Copy the full SHA
    05a2afc View commit details
    Browse the repository at this point in the history
  8. Rename variables with unclear naming (#14122)

    * Rename var
    
    * Add comments
    qqaatw authored Oct 22, 2021
    Configuration menu
    Copy the full SHA
    62ccbe0 View commit details
    Browse the repository at this point in the history
  9. Update TP parallel GEMM image (#14112)

    * Update TP parallel GEMM image
    
    * Delete parallelism-tp-parallel_gemm.png
    
    * Update parallelism-tp-parallel_gemm.png
    hyunwoongko authored Oct 22, 2021
    Configuration menu
    Copy the full SHA
    95bab53 View commit details
    Browse the repository at this point in the history

Commits on Oct 25, 2021

  1. Fix some typos in the docs (#14126)

    * Fix some typos in the docs
    
    * Fix a styling issue
    
    * Fix code quality check error
    h4iku authored Oct 25, 2021
    Configuration menu
    Copy the full SHA
    6b83090 View commit details
    Browse the repository at this point in the history
  2. Supporting Seq2Seq model for question answering task (#13432)

    * Add seq2seq example for QnA on SQuAD Dataset.
    
    * Changes from review - Fixing styling mistakes.
    
    * Added how to example in README, simplified the access to dataset's preprocess function.
    
    * Added tests for the seq2seq QA example.
    
    * Change dataset column name to fix tests.
    
    * Fix test command mistake.
    
    * Add missing argument 'ignore_pad_token_for_loss' from DataTrainingArguments.
    
    * Add missing argument 'num_beams' from DataTrainingArguments.
    
    * Fix processing of output predicted token ids so that tokenizer decode gets appropriate input. Updated assertion conditions on the tests.
    karthikrangasai authored Oct 25, 2021
    Configuration menu
    Copy the full SHA
    1b871e0 View commit details
    Browse the repository at this point in the history
  3. Configuration menu
    Copy the full SHA
    2ac6555 View commit details
    Browse the repository at this point in the history
  4. Fix some writing issues in the docs (#14136)

    * Fix some writing issues in the docs
    
    * Run code quality check
    h4iku authored Oct 25, 2021
    Configuration menu
    Copy the full SHA
    3e04a41 View commit details
    Browse the repository at this point in the history
  5. BartEnocder add set_input_embeddings (#13960)

    * BartEnocder add set_input_embeddings
    
    To unify the interface, add set_input_embeddings to BartEncoder.
    
    * BartEnocder add get_input_embeddings
    Liangtaiwan authored Oct 25, 2021
    Configuration menu
    Copy the full SHA
    1967c43 View commit details
    Browse the repository at this point in the history
  6. Configuration menu
    Copy the full SHA
    84b9579 View commit details
    Browse the repository at this point in the history
  7. Configuration menu
    Copy the full SHA
    3e8761a View commit details
    Browse the repository at this point in the history
  8. Update README.md

    patrickvonplaten authored Oct 25, 2021
    Configuration menu
    Copy the full SHA
    1a9381c View commit details
    Browse the repository at this point in the history
  9. Update README.md

    patrickvonplaten authored Oct 25, 2021
    Configuration menu
    Copy the full SHA
    c99a283 View commit details
    Browse the repository at this point in the history
  10. Configuration menu
    Copy the full SHA
    8560b55 View commit details
    Browse the repository at this point in the history
  11. Add TF<>PT and Flax<>PT everywhere (#14047)

    * up
    
    * up
    
    * up
    
    * up
    
    * up
    
    * up
    
    * up
    
    * add clip
    
    * fix clip PyTorch
    
    * fix clip PyTorch
    
    * up
    
    * up
    
    * up
    
    * up
    
    * up
    
    * up
    
    * up
    patrickvonplaten authored Oct 25, 2021
    Configuration menu
    Copy the full SHA
    0c3174c View commit details
    Browse the repository at this point in the history

Commits on Oct 26, 2021

  1. Add Camembert to models exportable with ONNX (#14059)

    Add Camembert to models exportable with ONNX
    
    Co-authored-by: Thomas.Chaigneau <[email protected]>
    Co-authored-by: Michael Benayoun <[email protected]>
    3 people authored Oct 26, 2021
    Configuration menu
    Copy the full SHA
    1f60df8 View commit details
    Browse the repository at this point in the history
  2. up (#14154)

    patrickvonplaten authored Oct 26, 2021
    Configuration menu
    Copy the full SHA
    e248e9b View commit details
    Browse the repository at this point in the history
  3. Add vision_encoder_decoder to models/__init__.py (#14151)

    * Add vision_encoder_decoder
    
    * Update _ignore_modules in get_model_modules()
    
    Co-authored-by: ydshieh <[email protected]>
    ydshieh and ydshieh authored Oct 26, 2021
    Configuration menu
    Copy the full SHA
    840fc8d View commit details
    Browse the repository at this point in the history
  4. [Speech Recognition] - Distributed training: Make sure vocab file rem…

    …oval and creation don't interfer (#14161)
    
    * up
    
    * better
    patrickvonplaten authored Oct 26, 2021
    Configuration menu
    Copy the full SHA
    f5ed19f View commit details
    Browse the repository at this point in the history
  5. Include Keras tensor in the allowed types (#14155)

    * Include KerasTensor in allowed types
    
    - This allows propagating symbolic tensors through TFBert models and layers' call(),
      which allows converting the subclass models to functional models.
    
    * Style pass
    
    Co-authored-by: Sergio Valcarcel Macua <[email protected]>
    Co-authored-by: matt <[email protected]>
    3 people authored Oct 26, 2021
    Configuration menu
    Copy the full SHA
    919a964 View commit details
    Browse the repository at this point in the history
  6. [megatron_gpt2] dynamic gelu, add tokenizer, save config (#13928)

    * [megatron_gpt2] dynamic gelu, add tokenizer, save config
    
    * cleanup
    
    * Update src/transformers/models/megatron_gpt2/convert_megatron_gpt2_checkpoint.py
    
    Co-authored-by: Sylvain Gugger <[email protected]>
    
    * apply suggestions
    
    Co-authored-by: Sylvain Gugger <[email protected]>
    stas00 and sgugger authored Oct 26, 2021
    Configuration menu
    Copy the full SHA
    bfd8176 View commit details
    Browse the repository at this point in the history
  7. Update README.md

    patrickvonplaten authored Oct 26, 2021
    Configuration menu
    Copy the full SHA
    9799f4e View commit details
    Browse the repository at this point in the history
  8. Add Unispeech & Unispeech-SAT (#13963)

    * unispeech
    
    * add copy from
    
    * remove hubert copy from
    
    * finish for today
    
    * add unispeech-sat
    
    * adapt more
    
    * up
    
    * up
    
    * up
    
    * up
    
    * add modeling
    
    * add tests
    
    * up
    
    * up
    
    * finish
    
    * up
    
    * Apply suggestions from code review
    
    * up
    
    * up
    
    * Apply suggestions from code review
    
    Co-authored-by: Sylvain Gugger <[email protected]>
    
    * up
    
    * up
    
    Co-authored-by: Sylvain Gugger <[email protected]>
    patrickvonplaten and sgugger authored Oct 26, 2021
    Configuration menu
    Copy the full SHA
    9f3aa46 View commit details
    Browse the repository at this point in the history
  9. [ONNX] Add symbolic function for XSoftmax op for exporting to ONNX. (…

    …#14013)
    
    * Add symbolic function for XSoftmax op for exporting to ONNX.
    
    * Fix format issues.
    
    * Fix a CI issue relative to copies.
    fatcat-z authored Oct 26, 2021
    Configuration menu
    Copy the full SHA
    3f23634 View commit details
    Browse the repository at this point in the history
  10. Fix copies

    LysandreJik committed Oct 26, 2021
    Configuration menu
    Copy the full SHA
    27c888d View commit details
    Browse the repository at this point in the history
  11. Configuration menu
    Copy the full SHA
    41dad89 View commit details
    Browse the repository at this point in the history
  12. fix typos in error messages in speech recognition example and modelca…

    …rd.py (#14166)
    
    * specify the text column name in the error message
    
    * pluralize the word fields
    mgoldey authored Oct 26, 2021
    Configuration menu
    Copy the full SHA
    42bfb83 View commit details
    Browse the repository at this point in the history
  13. Replace assertions with ValueError exception (#14142)

    Updated masked-language modeling examples in pytorch
    with convention defined by #12789
    huberemanuel authored Oct 26, 2021
    Configuration menu
    Copy the full SHA
    ebd48c6 View commit details
    Browse the repository at this point in the history
  14. switch to inference_mode from no_gard (#13667)

    * switch to inference_mode from no_gard
    faster inference
    
    * added switch to support older version of pytorch
    kamalkraj authored Oct 26, 2021
    Configuration menu
    Copy the full SHA
    8ddbfe9 View commit details
    Browse the repository at this point in the history

Commits on Oct 27, 2021

  1. Configuration menu
    Copy the full SHA
    1e53fae View commit details
    Browse the repository at this point in the history
  2. Add SEW CTC models (#14158)

    * Add SEW CTC models
    
    * Update paths
    
    * Update paths
    anton-l authored Oct 27, 2021
    Configuration menu
    Copy the full SHA
    e1dc5af View commit details
    Browse the repository at this point in the history
  3. [Gradient checkpointing] Enable for Deberta + DebertaV2 + SEW-D (#14175)

    * up
    
    * up
    
    * finish
    
    * up
    
    * final changes
    patrickvonplaten authored Oct 27, 2021
    Configuration menu
    Copy the full SHA
    6200fd7 View commit details
    Browse the repository at this point in the history
  4. Configuration menu
    Copy the full SHA
    25ceb81 View commit details
    Browse the repository at this point in the history
  5. Replace assert of data/data_collator.py by ValueError (#14131)

    * Replace assert of data_collator.py by ValueError
    
    * Replace assert of data_collator.py by ValueError
    AkechiShiro authored Oct 27, 2021
    Configuration menu
    Copy the full SHA
    e5b8ffb View commit details
    Browse the repository at this point in the history
  6. Add DistilHuBERT (#14174)

    * Add conversion
    
    * Rename
    
    * Add an integration test and remove layer_norm
    
    * Remove layer_norm from the converter
    
    * wording
    
    * Fix imports
    anton-l authored Oct 27, 2021
    Configuration menu
    Copy the full SHA
    232822f View commit details
    Browse the repository at this point in the history
  7. [TPU tests] Enable first TPU examples pytorch (#14121)

    * up
    
    * up
    
    * fix
    
    * up
    
    * Update examples/pytorch/test_xla_examples.py
    
    * correct labels
    
    * up
    
    * up
    
    * up
    
    * up
    
    * up
    
    * up
    patrickvonplaten authored Oct 27, 2021
    Configuration menu
    Copy the full SHA
    01b1466 View commit details
    Browse the repository at this point in the history
  8. Update README.md

    patrickvonplaten authored Oct 27, 2021
    Configuration menu
    Copy the full SHA
    e118db1 View commit details
    Browse the repository at this point in the history

Commits on Oct 28, 2021

  1. Update README.md

    patrickvonplaten authored Oct 28, 2021
    Configuration menu
    Copy the full SHA
    88cd82e View commit details
    Browse the repository at this point in the history
  2. [modeling_utils] respect original dtype in _get_resized_lm_head (#14181)

    * respect dtype in _get_resized_lm_head
    
    * Update src/transformers/modeling_utils.py
    
    Co-authored-by: Sylvain Gugger <[email protected]>
    
    * consistency
    
    Co-authored-by: Sylvain Gugger <[email protected]>
    stas00 and sgugger authored Oct 28, 2021
    Configuration menu
    Copy the full SHA
    123cce6 View commit details
    Browse the repository at this point in the history
  3. Add SegFormer (#14019)

    * First draft
    
    * Make style & quality
    
    * Improve conversion script
    
    * Add print statement to see actual slice
    
    * Make absolute tolerance smaller
    
    * Fix image classification models
    
    * Add post_process_semantic method
    
    * Disable padding
    
    * Improve conversion script
    
    * Rename to ForSemanticSegmentation, add integration test, remove post_process methods
    
    * Improve docs
    
    * Fix code quality
    
    * Fix feature extractor tests
    
    * Fix tests for image classification model
    
    * Delete file
    
    * Add is_torch_available to feature extractor
    
    * Improve documentation of feature extractor methods
    
    * Apply suggestions from @sgugger's code review
    
    Co-authored-by: Sylvain Gugger <[email protected]>
    
    * Apply some more suggestions of code review
    
    * Rebase with master
    
    * Fix rebase issues
    
    * Make sure model only outputs hidden states when the user wants to
    
    * Apply suggestions from code review
    
    * Add pad method
    
    * Support padding of 2d images
    
    * Add print statement
    
    * Add print statement
    
    * Move padding method to SegformerFeatureExtractor
    
    * Fix issue
    
    * Add casting of segmentation maps
    
    * Add test for padding
    
    * Add small note about padding
    
    Co-authored-by: Sylvain Gugger <[email protected]>
    NielsRogge and sgugger authored Oct 28, 2021
    Configuration menu
    Copy the full SHA
    1dc96a7 View commit details
    Browse the repository at this point in the history
  4. Configuration menu
    Copy the full SHA
    78b6a2e View commit details
    Browse the repository at this point in the history
  5. Fix SEW-D implementation differences (#14191)

    * Fix SEW-D
    
    * Update tests
    
    * isort
    anton-l authored Oct 28, 2021
    Configuration menu
    Copy the full SHA
    1251072 View commit details
    Browse the repository at this point in the history
  6. Fix EncoderDecoderModel classes to be more like BART and T5 (#14139)

    * First draft
    
    * Make tuple output more readable
    
    * Replace assertions by value errors
    
    * Make it possible to predict_with_generate for vision and speech models
    
    * Adapt Seq2SeqTrainer to work with VisionEncoderDecoder/SpeechEncoderDecoder
    
    * Add deprecation warning
    
    * Add copied from statements to vision and speech encoder decoders
    
    * Fix failing test
    
    * Apply @patrickvonplaten's suggestion
    
    * Use reshape instead of view for consistency
    NielsRogge authored Oct 28, 2021
    Configuration menu
    Copy the full SHA
    ac12a5a View commit details
    Browse the repository at this point in the history
  7. Fix EncoderDecoderModel docs (#14197)

    * Fix docs
    
    * Apply suggestions from review + fix bug
    NielsRogge authored Oct 28, 2021
    Configuration menu
    Copy the full SHA
    5f3bf65 View commit details
    Browse the repository at this point in the history
  8. Release v4.12.0

    LysandreJik committed Oct 28, 2021
    Configuration menu
    Copy the full SHA
    62bf536 View commit details
    Browse the repository at this point in the history

Commits on Oct 29, 2021

  1. Torch 1.10 (#14169)

    * Torch 1.10
    
    * torch scatter for 1.10
    
    * style
    
    * Skip tests
    ok
    LysandreJik committed Oct 29, 2021
    Configuration menu
    Copy the full SHA
    9f3f335 View commit details
    Browse the repository at this point in the history
  2. Release v4.12.1

    LysandreJik committed Oct 29, 2021
    Configuration menu
    Copy the full SHA
    e0a5154 View commit details
    Browse the repository at this point in the history
  3. Fixing image segmentation with inference mode. (#14204)

    * Fixing image segmentation for inference mode.
    
    * Update src/transformers/pipelines/base.py
    
    Co-authored-by: Patrick von Platen <[email protected]>
    
    Co-authored-by: Patrick von Platen <[email protected]>
    2 people authored and LysandreJik committed Oct 29, 2021
    Configuration menu
    Copy the full SHA
    cde7d78 View commit details
    Browse the repository at this point in the history
  4. Release: v4.12.2

    LysandreJik committed Oct 29, 2021
    Configuration menu
    Copy the full SHA
    2191373 View commit details
    Browse the repository at this point in the history

Commits on Nov 2, 2021

  1. Configuration menu
    Copy the full SHA
    ac77639 View commit details
    Browse the repository at this point in the history
  2. Bump huggingface_hub

    sgugger committed Nov 2, 2021
    Configuration menu
    Copy the full SHA
    872c4f3 View commit details
    Browse the repository at this point in the history
  3. Release v4.12.3

    sgugger committed Nov 2, 2021
    Configuration menu
    Copy the full SHA
    9ab10fc View commit details
    Browse the repository at this point in the history
  4. Add maximum check for hf hub

    sgugger committed Nov 2, 2021
    Configuration menu
    Copy the full SHA
    294a920 View commit details
    Browse the repository at this point in the history
  5. Style

    sgugger committed Nov 2, 2021
    Configuration menu
    Copy the full SHA
    3ea15d2 View commit details
    Browse the repository at this point in the history

Commits on Nov 16, 2021

  1. Fix of issue #13327: Wrong weight initialization for TF t5 model (#14…

    …241)
    
    * Fix of issue #13327: Wrong weight initialization for TF t5 model
    
    * run black formatter
    
    * fix typo
    
    * remove my name tag from comments
    
    Co-authored-by: Shirron <[email protected]>
    2 people authored and LysandreJik committed Nov 16, 2021
    Configuration menu
    Copy the full SHA
    b6b97c3 View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    c8206b4 View commit details
    Browse the repository at this point in the history
  3. Support for TF >= 2.7 (#14345)

    sgugger authored and LysandreJik committed Nov 16, 2021
    Configuration menu
    Copy the full SHA
    6bf2027 View commit details
    Browse the repository at this point in the history
  4. Configuration menu
    Copy the full SHA
    341a059 View commit details
    Browse the repository at this point in the history
  5. Experimenting with adding proper get_config() and from_config() metho…

    …ds (#14361)
    
    * Experimenting with adding proper get_config() and from_config() methods
    
    * Adding a test for get/from config
    
    * Fix test for get/from config
    Rocketknight1 authored and LysandreJik committed Nov 16, 2021
    Configuration menu
    Copy the full SHA
    e99a231 View commit details
    Browse the repository at this point in the history
  6. [Wav2Vec2] Make sure that gradient checkpointing is only run if neede…

    …d (#14407)
    
    * [Wav2Vec2] Make sure that gradient checkpointing is only run if needed
    
    * make fix-copies
    patrickvonplaten authored and LysandreJik committed Nov 16, 2021
    Configuration menu
    Copy the full SHA
    db242ae View commit details
    Browse the repository at this point in the history
  7. Fix gradient_checkpointing backward compatibility (#14408)

    * Fix gradient_checkpointing backward compatibility
    
    * Remove needless line
    
    * make sure mask prob is big enough and length small enough
    
    * Fix tests
    
    Co-authored-by: patrickvonplaten <[email protected]>
    2 people authored and LysandreJik committed Nov 16, 2021
    Configuration menu
    Copy the full SHA
    6f40723 View commit details
    Browse the repository at this point in the history
  8. Release: v4.12.4

    LysandreJik committed Nov 16, 2021
    Configuration menu
    Copy the full SHA
    527c763 View commit details
    Browse the repository at this point in the history

Commits on Nov 17, 2021

  1. Revert "Experimenting with adding proper get_config() and from_config…

    …() methods (#14361)"
    
    This reverts commit e99a231.
    Rocketknight1 authored and LysandreJik committed Nov 17, 2021
    Configuration menu
    Copy the full SHA
    a5211fc View commit details
    Browse the repository at this point in the history
  2. Release: v4.12.5

    LysandreJik committed Nov 17, 2021
    Configuration menu
    Copy the full SHA
    ef3cec0 View commit details
    Browse the repository at this point in the history

Commits on Dec 1, 2021

  1. Configuration menu
    Copy the full SHA
    b3a8de8 View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    2c63bc1 View commit details
    Browse the repository at this point in the history

Commits on Dec 7, 2021

  1. Configuration menu
    Copy the full SHA
    f83eb7e View commit details
    Browse the repository at this point in the history

Commits on Dec 8, 2021

  1. Configuration menu
    Copy the full SHA
    3356b30 View commit details
    Browse the repository at this point in the history