Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

merge main #23866

Merged
merged 74 commits into from
May 30, 2023
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
74 commits
Select commit Hold shift + click to select a range
6397b7f
Debug example code for MegaForCausalLM (#23382)
Tylersuard May 22, 2023
12ec7f0
Remove erroneous `img` closing tag (#23646)
xenova May 22, 2023
29294b0
Fix tensor device while attention_mask is not None (#23538)
zspo May 22, 2023
7bbdfd7
Fix accelerate logger bug (#23650)
younesbelkada May 22, 2023
fe34486
Muellerzr fix deepspeed (#23657)
muellerzr May 22, 2023
4ddd9de
Bugfix: LLaMA layer norm incorrectly changes input type and consumers…
TimDettmers May 22, 2023
5de2a6d
Fix wav2vec2 is_batched check to include 2-D numpy arrays (#23223)
LWprogramming May 22, 2023
6f72e71
changing the requirements to a cpu torch version that works (#23483)
sshahrokhi May 22, 2023
26a0681
Fix SAM tests and use smaller checkpoints (#23656)
Rocketknight1 May 22, 2023
b191d7d
Update all no_trainer with skip_first_batches (#23664)
muellerzr May 22, 2023
e69feab
Update workflow files (#23658)
ydshieh May 22, 2023
2f424d7
[image-to-text pipeline] Add conditional text support + GIT (#23362)
NielsRogge May 22, 2023
e30ceae
small fix to remove unused eos in processor when it's not used. (#23408)
Narsil May 23, 2023
ecc05f8
Bump requests from 2.27.1 to 2.31.0 in /examples/research_projects/de…
dependabot[bot] May 23, 2023
9bf72ae
Bump requests from 2.22.0 to 2.31.0 in /examples/research_projects/vi…
dependabot[bot] May 23, 2023
aa30cd4
Bump requests from 2.22.0 to 2.31.0 in /examples/research_projects/lx…
dependabot[bot] May 23, 2023
527ab89
Add PerSAM [bis] (#23659)
NielsRogge May 23, 2023
b687af0
Fix typo in a parameter name for open llama model (#23637)
aaalexlit May 23, 2023
abf691a
Fix PyTorch SAM tests (#23682)
ydshieh May 23, 2023
9e8d706
Making `safetensors` a core dependency. (#23254)
Narsil May 23, 2023
1fe1e3c
🌐 [i18n-KO] Translated `tasks/monocular_depth_estimation.mdx` to Kore…
HanNayeoniee May 23, 2023
71a5ed3
Fix a `BridgeTower` test (#23694)
ydshieh May 23, 2023
42baa58
[`SAM`] Fixes pipeline and adds a dummy pipeline test (#23684)
younesbelkada May 23, 2023
876d9a3
TF version compatibility fixes (#23663)
Rocketknight1 May 23, 2023
6b7d6f8
[`Blip`] Fix blip doctest (#23698)
younesbelkada May 23, 2023
3d57404
is_batched fix for remaining 2-D numpy arrays (#23309)
LWprogramming May 23, 2023
de5f86e
Skip `TFCvtModelTest::test_keras_fit_mixed_precision` for now (#23699)
ydshieh May 23, 2023
357f281
fix: load_best_model_at_end error when load_in_8bit is True (#23443)
dkqkxx May 23, 2023
003a0cf
Fix some docs what layerdrop does (#23691)
zspo May 23, 2023
33687a3
add GPTJ/bloom/llama/opt into model list and enhance the jit support …
sywangyi May 24, 2023
9d73b92
4-bit QLoRA via bitsandbytes (4-bit base model + LoRA) (#23479)
TimDettmers May 24, 2023
796162c
Paged Optimizer + Lion Optimizer for Trainer (#23217)
TimDettmers May 24, 2023
2eaaf17
Export to ONNX doc refocused on using optimum, added tflite (#23434)
MKhalusova May 24, 2023
b4698b7
fix: use bool instead of uint8/byte in Deberta/DebertaV2/SEW-D to mak…
uchuhimo May 24, 2023
767e6b5
fix gptj could not jit.trace in GPU (#23317)
sywangyi May 24, 2023
f8b2574
Better TF docstring types (#23477)
Rocketknight1 May 24, 2023
28aa438
Minor awesome-transformers.md fixes (#23453)
pagarsky May 24, 2023
d2d8822
TF SAM memory reduction (#23732)
Rocketknight1 May 24, 2023
50a56be
fix: delete duplicate sentences in `document_question_answering.mdx` …
jungnerd May 24, 2023
3d7baef
fix: Whisper generate, move text_prompt_ids trim up for max_new_token…
May 24, 2023
814de8f
Overhaul TF serving signatures + dummy inputs (#23234)
Rocketknight1 May 24, 2023
d8222be
[Whisper] Reduce batch size in tests (#23736)
sanchit-gandhi May 24, 2023
8915965
Fix the regex in `get_imports` to support multiline try blocks and ex…
dakinggg May 24, 2023
75bbf20
Fix sagemaker DP/MP (#23681)
muellerzr May 24, 2023
9850e6d
Enable prompts on the Hub (#23662)
sgugger May 24, 2023
e45e756
Remove the last few TF serving sigs (#23738)
Rocketknight1 May 24, 2023
f0a2a82
Fix `pip install --upgrade accelerate` command in modeling_utils.py (…
tloen May 25, 2023
06c28cd
Add LlamaIndex to awesome-transformers.md (#23484)
ravi03071991 May 25, 2023
7d4fe85
Fix psuh_to_hub in Trainer when nothing needs pushing (#23751)
sgugger May 25, 2023
6e4bc67
Revamp test selection for the example tests (#23737)
sgugger May 25, 2023
3416bba
[LongFormer] code nits, removed unused parameters (#23749)
ArthurZucker May 25, 2023
f04f549
Fix is_ninja_available() (#23752)
niltok May 25, 2023
4b0e7de
Bump tornado from 6.0.4 to 6.3.2 in /examples/research_projects/lxmer…
dependabot[bot] May 25, 2023
d685e33
Bump tornado from 6.0.4 to 6.3.2 in /examples/research_projects/visua…
dependabot[bot] May 25, 2023
f67dac9
[`Nllb-Moe`] Fix nllb moe accelerate issue (#23758)
younesbelkada May 25, 2023
8d28dba
[OPT] Doc nit, using fast is fine (#23789)
ArthurZucker May 26, 2023
4d9b76a
Fix RWKV backward on GPU (#23774)
sgugger May 26, 2023
d61d747
Update trainer.mdx class_weights example (#23787)
amitportnoy May 26, 2023
b7b729b
no_cuda does not take effect in non distributed environment (#23795)
sywangyi May 26, 2023
e724246
Fix no such file or directory error (#23783)
RissyRan May 26, 2023
edf7772
Log the right train_batch_size if using auto_find_batch_size and also…
muellerzr May 26, 2023
17a5553
Enable code-specific revision for code on the Hub (#23799)
sgugger May 26, 2023
4b6a5a7
[Time-Series] Autoformer model (#21891)
elisim May 30, 2023
af45ec0
add type hint in pipeline model argument (#23740)
y3sar May 30, 2023
ac224de
TF SAM shape flexibility fixes (#23842)
Rocketknight1 May 30, 2023
2faa095
fix Whisper tests on GPU (#23753)
hollance May 30, 2023
a077f71
🌐 [i18n-KO] Translated `fast_tokenizers.mdx` to Korean (#22956)
kihoon71 May 30, 2023
192aa04
[i18n-KO] Translated video_classification.mdx to Korean (#23026)
kihoon71 May 30, 2023
8673166
🌐 [i18n-KO] Translated `troubleshooting.mdx` to Korean (#23166)
0525hhgus May 30, 2023
62ba64b
Adds a FlyteCallback (#23759)
peridotml May 30, 2023
0623f08
Update collating_graphormer.py (#23862)
clefourrier May 30, 2023
6fc0454
[LlamaTokenizerFast] nit update `post_processor` on the fly (#23855)
ArthurZucker May 30, 2023
58022e4
#23388 Issue: Update RoBERTa configuration (#23863)
vijethmoudgalya May 30, 2023
af2aac5
[from_pretrained] imporve the error message when `_no_split_modules` …
ArthurZucker May 30, 2023
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
The table of contents is too big for display.
Diff view
Diff view
  •  
  •  
  •  
21 changes: 7 additions & 14 deletions .circleci/config.yml
Original file line number Diff line number Diff line change
Expand Up @@ -43,6 +43,12 @@ jobs:
else
touch test_preparation/test_list.txt
fi
- run: |
if [ -f examples_test_list.txt ]; then
mv examples_test_list.txt test_preparation/examples_test_list.txt
else
touch test_preparation/examples_test_list.txt
fi
- run: |
if [ -f doctest_list.txt ]; then
cp doctest_list.txt test_preparation/doctest_list.txt
Expand All @@ -62,19 +68,6 @@ jobs:
else
touch test_preparation/filtered_test_list.txt
fi
- run: python utils/tests_fetcher.py --filters tests examples | tee examples_tests_fetched_summary.txt
- run: |
if [ -f test_list.txt ]; then
mv test_list.txt test_preparation/examples_test_list.txt
else
touch test_preparation/examples_test_list.txt
fi
- run: |
if [ -f filtered_test_list_cross_tests.txt ]; then
mv filtered_test_list_cross_tests.txt test_preparation/filtered_test_list_cross_tests.txt
else
touch test_preparation/filtered_test_list_cross_tests.txt
fi
- store_artifacts:
path: test_preparation/test_list.txt
- store_artifacts:
Expand Down Expand Up @@ -111,7 +104,7 @@ jobs:
- run: |
mkdir test_preparation
echo -n "tests" > test_preparation/test_list.txt
echo -n "tests" > test_preparation/examples_test_list.txt
echo -n "all" > test_preparation/examples_test_list.txt
echo -n "tests/repo_utils" > test_preparation/test_repo_utils.txt
- run: |
echo -n "tests" > test_list.txt
Expand Down
15 changes: 11 additions & 4 deletions .circleci/create_circleci_config.py
Original file line number Diff line number Diff line change
Expand Up @@ -342,7 +342,6 @@ def job_name(self):
"pip install .[sklearn,torch,sentencepiece,testing,torch-speech]",
"pip install -r examples/pytorch/_tests_requirements.txt",
],
tests_to_run="./examples/pytorch/",
)


Expand All @@ -355,7 +354,6 @@ def job_name(self):
"pip install .[sklearn,tensorflow,sentencepiece,testing]",
"pip install -r examples/tensorflow/_tests_requirements.txt",
],
tests_to_run="./examples/tensorflow/",
)


Expand All @@ -367,7 +365,6 @@ def job_name(self):
"pip install .[flax,testing,sentencepiece]",
"pip install -r examples/flax/_tests_requirements.txt",
],
tests_to_run="./examples/flax/",
)


Expand Down Expand Up @@ -551,7 +548,17 @@ def create_circleci_config(folder=None):

example_file = os.path.join(folder, "examples_test_list.txt")
if os.path.exists(example_file) and os.path.getsize(example_file) > 0:
jobs.extend(EXAMPLES_TESTS)
with open(example_file, "r", encoding="utf-8") as f:
example_tests = f.read().split(" ")
for job in EXAMPLES_TESTS:
framework = job.name.replace("examples_", "").replace("torch", "pytorch")
if example_tests == "all":
job.tests_to_run = [f"examples/{framework}"]
else:
job.tests_to_run = [f for f in example_tests if f.startswith(f"examples/{framework}")]

if len(job.tests_to_run) > 0:
jobs.append(job)

doctest_file = os.path.join(folder, "doctest_list.txt")
if os.path.exists(doctest_file):
Expand Down
16 changes: 16 additions & 0 deletions .github/workflows/self-push.yml
Original file line number Diff line number Diff line change
Expand Up @@ -195,6 +195,10 @@ jobs:
git checkout ${{ env.CI_SHA }}
echo "log = $(git log -n 1)"
- name: Reinstall transformers in edit mode (remove the one installed during docker image build)
working-directory: /transformers
run: python3 -m pip uninstall -y transformers && python3 -m pip install -e .

- name: Echo folder ${{ matrix.folders }}
shell: bash
# For folders like `models/bert`, set an env. var. (`matrix_folders`) to `models_bert`, which will be used to
Expand Down Expand Up @@ -284,6 +288,10 @@ jobs:
git checkout ${{ env.CI_SHA }}
echo "log = $(git log -n 1)"
- name: Reinstall transformers in edit mode (remove the one installed during docker image build)
working-directory: /transformers
run: python3 -m pip uninstall -y transformers && python3 -m pip install -e .

- name: Echo folder ${{ matrix.folders }}
shell: bash
# For folders like `models/bert`, set an env. var. (`matrix_folders`) to `models_bert`, which will be used to
Expand Down Expand Up @@ -373,6 +381,10 @@ jobs:
git checkout ${{ env.CI_SHA }}
echo "log = $(git log -n 1)"
- name: Reinstall transformers in edit mode (remove the one installed during docker image build)
working-directory: /workspace/transformers
run: python3 -m pip uninstall -y transformers && python3 -m pip install -e .

- name: Remove cached torch extensions
run: rm -rf /github/home/.cache/torch_extensions/

Expand Down Expand Up @@ -459,6 +471,10 @@ jobs:
git checkout ${{ env.CI_SHA }}
echo "log = $(git log -n 1)"
- name: Reinstall transformers in edit mode (remove the one installed during docker image build)
working-directory: /workspace/transformers
run: python3 -m pip uninstall -y transformers && python3 -m pip install -e .

- name: Remove cached torch extensions
run: rm -rf /github/home/.cache/torch_extensions/

Expand Down
24 changes: 24 additions & 0 deletions .github/workflows/self-scheduled.yml
Original file line number Diff line number Diff line change
Expand Up @@ -119,6 +119,10 @@ jobs:
working-directory: /transformers
run: git fetch && git checkout ${{ github.sha }}

- name: Reinstall transformers in edit mode (remove the one installed during docker image build)
working-directory: /transformers
run: python3 -m pip uninstall -y transformers && python3 -m pip install -e .

- name: NVIDIA-SMI
run: |
nvidia-smi
Expand Down Expand Up @@ -176,6 +180,10 @@ jobs:
working-directory: /transformers
run: git fetch && git checkout ${{ github.sha }}

- name: Reinstall transformers in edit mode (remove the one installed during docker image build)
working-directory: /transformers
run: python3 -m pip uninstall -y transformers && python3 -m pip install -e .

- name: NVIDIA-SMI
run: |
nvidia-smi
Expand Down Expand Up @@ -221,6 +229,10 @@ jobs:
working-directory: /transformers
run: git fetch && git checkout ${{ github.sha }}

- name: Reinstall transformers in edit mode (remove the one installed during docker image build)
working-directory: /transformers
run: python3 -m pip uninstall -y transformers && python3 -m pip install -e .

- name: NVIDIA-SMI
run: |
nvidia-smi
Expand Down Expand Up @@ -268,6 +280,10 @@ jobs:
working-directory: /transformers
run: git fetch && git checkout ${{ github.sha }}

- name: Reinstall transformers in edit mode (remove the one installed during docker image build)
working-directory: /transformers
run: python3 -m pip uninstall -y transformers && python3 -m pip install -e .

- name: NVIDIA-SMI
run: |
nvidia-smi
Expand Down Expand Up @@ -315,6 +331,10 @@ jobs:
run: |
git fetch && git checkout ${{ github.sha }}
- name: Reinstall transformers in edit mode (remove the one installed during docker image build)
working-directory: /transformers
run: python3 -m pip uninstall -y transformers && python3 -m pip install -e .

- name: NVIDIA-SMI
run: |
nvidia-smi
Expand Down Expand Up @@ -361,6 +381,10 @@ jobs:
working-directory: /workspace/transformers
run: git fetch && git checkout ${{ github.sha }}

- name: Reinstall transformers in edit mode (remove the one installed during docker image build)
working-directory: /workspace/transformers
run: python3 -m pip uninstall -y transformers && python3 -m pip install -e .

- name: Remove cached torch extensions
run: rm -rf /github/home/.cache/torch_extensions/

Expand Down
1 change: 1 addition & 0 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -292,6 +292,7 @@ Current number of checkpoints: ![](https://img.shields.io/endpoint?url=https://h
1. **[ALIGN](https://huggingface.co/docs/transformers/model_doc/align)** (from Google Research) released with the paper [Scaling Up Visual and Vision-Language Representation Learning With Noisy Text Supervision](https://arxiv.org/abs/2102.05918) by Chao Jia, Yinfei Yang, Ye Xia, Yi-Ting Chen, Zarana Parekh, Hieu Pham, Quoc V. Le, Yunhsuan Sung, Zhen Li, Tom Duerig.
1. **[AltCLIP](https://huggingface.co/docs/transformers/model_doc/altclip)** (from BAAI) released with the paper [AltCLIP: Altering the Language Encoder in CLIP for Extended Language Capabilities](https://arxiv.org/abs/2211.06679) by Chen, Zhongzhi and Liu, Guang and Zhang, Bo-Wen and Ye, Fulong and Yang, Qinghong and Wu, Ledell.
1. **[Audio Spectrogram Transformer](https://huggingface.co/docs/transformers/model_doc/audio-spectrogram-transformer)** (from MIT) released with the paper [AST: Audio Spectrogram Transformer](https://arxiv.org/abs/2104.01778) by Yuan Gong, Yu-An Chung, James Glass.
1. **[Autoformer](https://huggingface.co/docs/transformers/main/model_doc/autoformer)** (from Tsinghua University) released with the paper [Autoformer: Decomposition Transformers with Auto-Correlation for Long-Term Series Forecasting](https://arxiv.org/abs/2106.13008) by Haixu Wu, Jiehui Xu, Jianmin Wang, Mingsheng Long.
1. **[BART](https://huggingface.co/docs/transformers/model_doc/bart)** (from Facebook) released with the paper [BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension](https://arxiv.org/abs/1910.13461) by Mike Lewis, Yinhan Liu, Naman Goyal, Marjan Ghazvininejad, Abdelrahman Mohamed, Omer Levy, Ves Stoyanov and Luke Zettlemoyer.
1. **[BARThez](https://huggingface.co/docs/transformers/model_doc/barthez)** (from École polytechnique) released with the paper [BARThez: a Skilled Pretrained French Sequence-to-Sequence Model](https://arxiv.org/abs/2010.12321) by Moussa Kamal Eddine, Antoine J.-P. Tixier, Michalis Vazirgiannis.
1. **[BARTpho](https://huggingface.co/docs/transformers/model_doc/bartpho)** (from VinAI Research) released with the paper [BARTpho: Pre-trained Sequence-to-Sequence Models for Vietnamese](https://arxiv.org/abs/2109.09701) by Nguyen Luong Tran, Duong Minh Le and Dat Quoc Nguyen.
Expand Down
1 change: 1 addition & 0 deletions README_es.md
Original file line number Diff line number Diff line change
Expand Up @@ -267,6 +267,7 @@ Número actual de puntos de control: ![](https://img.shields.io/endpoint?url=htt
1. **[ALIGN](https://huggingface.co/docs/transformers/model_doc/align)** (from Google Research) released with the paper [Scaling Up Visual and Vision-Language Representation Learning With Noisy Text Supervision](https://arxiv.org/abs/2102.05918) by Chao Jia, Yinfei Yang, Ye Xia, Yi-Ting Chen, Zarana Parekh, Hieu Pham, Quoc V. Le, Yunhsuan Sung, Zhen Li, Tom Duerig.
1. **[AltCLIP](https://huggingface.co/docs/transformers/model_doc/altclip)** (from BAAI) released with the paper [AltCLIP: Altering the Language Encoder in CLIP for Extended Language Capabilities](https://arxiv.org/abs/2211.06679) by Chen, Zhongzhi and Liu, Guang and Zhang, Bo-Wen and Ye, Fulong and Yang, Qinghong and Wu, Ledell.
1. **[Audio Spectrogram Transformer](https://huggingface.co/docs/transformers/model_doc/audio-spectrogram-transformer)** (from MIT) released with the paper [AST: Audio Spectrogram Transformer](https://arxiv.org/abs/2104.01778) by Yuan Gong, Yu-An Chung, James Glass.
1. **[Autoformer](https://huggingface.co/docs/transformers/main/model_doc/autoformer)** (from Tsinghua University) released with the paper [Autoformer: Decomposition Transformers with Auto-Correlation for Long-Term Series Forecasting](https://arxiv.org/abs/2106.13008) by Haixu Wu, Jiehui Xu, Jianmin Wang, Mingsheng Long.
1. **[BART](https://huggingface.co/docs/transformers/model_doc/bart)** (from Facebook) released with the paper [BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension](https://arxiv.org/abs/1910.13461) by Mike Lewis, Yinhan Liu, Naman Goyal, Marjan Ghazvininejad, Abdelrahman Mohamed, Omer Levy, Ves Stoyanov and Luke Zettlemoyer.
1. **[BARThez](https://huggingface.co/docs/transformers/model_doc/barthez)** (from École polytechnique) released with the paper [BARThez: a Skilled Pretrained French Sequence-to-Sequence Model](https://arxiv.org/abs/2010.12321) by Moussa Kamal Eddine, Antoine J.-P. Tixier, Michalis Vazirgiannis.
1. **[BARTpho](https://huggingface.co/docs/transformers/model_doc/bartpho)** (from VinAI Research) released with the paper [BARTpho: Pre-trained Sequence-to-Sequence Models for Vietnamese](https://arxiv.org/abs/2109.09701) by Nguyen Luong Tran, Duong Minh Le and Dat Quoc Nguyen.
Expand Down
1 change: 1 addition & 0 deletions README_hd.md
Original file line number Diff line number Diff line change
Expand Up @@ -239,6 +239,7 @@ conda install -c huggingface transformers
1. **[ALIGN](https://huggingface.co/docs/transformers/model_doc/align)** (Google Research से) Chao Jia, Yinfei Yang, Ye Xia, Yi-Ting Chen, Zarana Parekh, Hieu Pham, Quoc V. Le, Yunhsuan Sung, Zhen Li, Tom Duerig. द्वाराअनुसंधान पत्र [Scaling Up Visual and Vision-Language Representation Learning With Noisy Text Supervision](https://arxiv.org/abs/2102.05918) के साथ जारी किया गया
1. **[AltCLIP](https://huggingface.co/docs/transformers/model_doc/altclip)** (from BAAI) released with the paper [AltCLIP: Altering the Language Encoder in CLIP for Extended Language Capabilities](https://arxiv.org/abs/2211.06679) by Chen, Zhongzhi and Liu, Guang and Zhang, Bo-Wen and Ye, Fulong and Yang, Qinghong and Wu, Ledell.
1. **[Audio Spectrogram Transformer](https://huggingface.co/docs/transformers/model_doc/audio-spectrogram-transformer)** (from MIT) released with the paper [AST: Audio Spectrogram Transformer](https://arxiv.org/abs/2104.01778) by Yuan Gong, Yu-An Chung, James Glass.
1. **[Autoformer](https://huggingface.co/docs/transformers/main/model_doc/autoformer)** (from Tsinghua University) released with the paper [Autoformer: Decomposition Transformers with Auto-Correlation for Long-Term Series Forecasting](https://arxiv.org/abs/2106.13008) by Haixu Wu, Jiehui Xu, Jianmin Wang, Mingsheng Long.
1. **[BART](https://huggingface.co/docs/transformers/model_doc/bart)** (फेसबुक) साथ थीसिस [बार्ट: प्राकृतिक भाषा निर्माण, अनुवाद के लिए अनुक्रम-से-अनुक्रम पूर्व प्रशिक्षण , और समझ] (https://arxiv.org/pdf/1910.13461.pdf) पर निर्भर माइक लुईस, यिनहान लियू, नमन गोयल, मार्जन ग़ज़विनिनेजाद, अब्देलरहमान मोहम्मद, ओमर लेवी, वेस स्टोयानोव और ल्यूक ज़ेटलमॉयर
1. **[BARThez](https://huggingface.co/docs/transformers/model_doc/barthez)** (से École polytechnique) साथ थीसिस [BARThez: a Skilled Pretrained French Sequence-to-Sequence Model](https://arxiv.org/abs/2010.12321) पर निर्भर Moussa Kamal Eddine, Antoine J.-P. Tixier, Michalis Vazirgiannis रिहाई।
1. **[BARTpho](https://huggingface.co/docs/transformers/model_doc/bartpho)** (VinAI Research से) साथ में पेपर [BARTpho: Pre-trained Sequence-to-Sequence Models for Vietnamese](https://arxiv.org/abs/2109.09701)गुयेन लुओंग ट्रान, डुओंग मिन्ह ले और डाट क्वोक गुयेन द्वारा पोस्ट किया गया।
Expand Down
1 change: 1 addition & 0 deletions README_ja.md
Original file line number Diff line number Diff line change
Expand Up @@ -301,6 +301,7 @@ Flax、PyTorch、TensorFlowをcondaでインストールする方法は、それ
1. **[ALIGN](https://huggingface.co/docs/transformers/model_doc/align)** (Google Research から) Chao Jia, Yinfei Yang, Ye Xia, Yi-Ting Chen, Zarana Parekh, Hieu Pham, Quoc V. Le, Yunhsuan Sung, Zhen Li, Tom Duerig. から公開された研究論文 [Scaling Up Visual and Vision-Language Representation Learning With Noisy Text Supervision](https://arxiv.org/abs/2102.05918)
1. **[AltCLIP](https://huggingface.co/docs/transformers/model_doc/altclip)** (BAAI から) Chen, Zhongzhi and Liu, Guang and Zhang, Bo-Wen and Ye, Fulong and Yang, Qinghong and Wu, Ledell から公開された研究論文: [AltCLIP: Altering the Language Encoder in CLIP for Extended Language Capabilities](https://arxiv.org/abs/2211.06679)
1. **[Audio Spectrogram Transformer](https://huggingface.co/docs/transformers/model_doc/audio-spectrogram-transformer)** (MIT から) Yuan Gong, Yu-An Chung, James Glass から公開された研究論文: [AST: Audio Spectrogram Transformer](https://arxiv.org/abs/2104.01778)
1. **[Autoformer](https://huggingface.co/docs/transformers/main/model_doc/autoformer)** (from Tsinghua University) released with the paper [Autoformer: Decomposition Transformers with Auto-Correlation for Long-Term Series Forecasting](https://arxiv.org/abs/2106.13008) by Haixu Wu, Jiehui Xu, Jianmin Wang, Mingsheng Long.
1. **[BART](https://huggingface.co/docs/transformers/model_doc/bart)** (Facebook から) Mike Lewis, Yinhan Liu, Naman Goyal, Marjan Ghazvininejad, Abdelrahman Mohamed, Omer Levy, Ves Stoyanov and Luke Zettlemoyer から公開された研究論文: [BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension](https://arxiv.org/abs/1910.13461)
1. **[BARThez](https://huggingface.co/docs/transformers/model_doc/barthez)** (École polytechnique から) Moussa Kamal Eddine, Antoine J.-P. Tixier, Michalis Vazirgiannis から公開された研究論文: [BARThez: a Skilled Pretrained French Sequence-to-Sequence Model](https://arxiv.org/abs/2010.12321)
1. **[BARTpho](https://huggingface.co/docs/transformers/model_doc/bartpho)** (VinAI Research から) Nguyen Luong Tran, Duong Minh Le and Dat Quoc Nguyen から公開された研究論文: [BARTpho: Pre-trained Sequence-to-Sequence Models for Vietnamese](https://arxiv.org/abs/2109.09701)
Expand Down
Loading