Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

fixed typo #14

Closed
wants to merge 2 commits into from
Closed

fixed typo #14

wants to merge 2 commits into from

Conversation

kornosk
Copy link

@kornosk kornosk commented Nov 12, 2018

When test with SQuAD

@thomwolf
Copy link
Member

Hi,
Thanks for the PR, we don't want to add a shell script to the repo.
I will correct the typo,
Best,
Thom

@thomwolf thomwolf closed this Nov 12, 2018
@HongyanJiao HongyanJiao mentioned this pull request Sep 19, 2019
@devroy73 devroy73 mentioned this pull request Nov 10, 2019
4 tasks
stevezheng23 added a commit to stevezheng23/transformers that referenced this pull request Mar 24, 2020
* update kd qa in roberta modeling

* fix issues for kd-quac runner
amathews-amd referenced this pull request in ROCm/transformers Aug 6, 2021
Remove model specific changes for BERT and DistilBERT
jameshennessytempus pushed a commit to jameshennessytempus/transformers that referenced this pull request Jun 1, 2023
jonb377 pushed a commit to jonb377/hf-transformers that referenced this pull request Jul 27, 2023
Summary:
This pull request introduce a new way to do sharding which allow weights to be sharded in two dimensional mesh, i.e., (fsdp, tensor), and then the input to be sharded according to the fsdp dimension.

To enable it, pass --spmd_tensor_sharding 2, 2 is the tensor dimension, the fsdp dimension will be auto calculated according to num_devices // 2.

Test Plan:
Test it on a V4-8 with 2B LLaMA.
ocavue pushed a commit to ocavue/transformers that referenced this pull request Sep 13, 2023
jonb377 pushed a commit to jonb377/hf-transformers that referenced this pull request Nov 3, 2023
Summary:
This pull request introduce a new way to do sharding which allow weights to be sharded in two dimensional mesh, i.e., (fsdp, tensor), and then the input to be sharded according to the fsdp dimension.

To enable it, pass --spmd_tensor_sharding 2, 2 is the tensor dimension, the fsdp dimension will be auto calculated according to num_devices // 2.

Test Plan:
Test it on a V4-8 with 2B LLaMA.
LysandreJik pushed a commit that referenced this pull request Mar 15, 2024
* Cohere Model Release (#1)

Cohere Model Release

* Remove unnecessary files and code (#2)

Some cleanup

* Delete cohere-model directory (#3)

* Make Fix (#5)

* Pr fixes (#6)

* fixes for pr

* pr fixes for the format

* pr fixes for the format

* src/transformers/models/auto/tokenization_auto.py

* Tokenizer test (#8)

* tokenizer test

* format fix

* Adding Docs and other minor changes (#7)

* Add modeling tests (#9)

* Smol Fix (#11)

* tokenization tests are fixed

* format fixes

* fix pr doc tests

* fix pr doc tests

* fix pr doc tests

* fix pr style check

* small changes in cohere.md

* FIX: Address final comments for transformers integration (#13)

* fix modeling final nits and add proper test file

* for now leave empty tests

* add integration test

* push new test

* fix modeling cohere (#14)

* Update chat templates to use the new API (#15)

---------

Co-authored-by: ahmetustun <[email protected]>
Co-authored-by: Younes Belkada <[email protected]>
Co-authored-by: Matt <[email protected]>
lcong pushed a commit to lcong/transformers that referenced this pull request Apr 9, 2024
LysandreJik pushed a commit to LysandreJik/transformers that referenced this pull request Apr 10, 2024
itazap pushed a commit that referenced this pull request May 14, 2024
* Cohere Model Release (#1)

Cohere Model Release

* Remove unnecessary files and code (#2)

Some cleanup

* Delete cohere-model directory (#3)

* Make Fix (#5)

* Pr fixes (#6)

* fixes for pr

* pr fixes for the format

* pr fixes for the format

* src/transformers/models/auto/tokenization_auto.py

* Tokenizer test (#8)

* tokenizer test

* format fix

* Adding Docs and other minor changes (#7)

* Add modeling tests (#9)

* Smol Fix (#11)

* tokenization tests are fixed

* format fixes

* fix pr doc tests

* fix pr doc tests

* fix pr doc tests

* fix pr style check

* small changes in cohere.md

* FIX: Address final comments for transformers integration (#13)

* fix modeling final nits and add proper test file

* for now leave empty tests

* add integration test

* push new test

* fix modeling cohere (#14)

* Update chat templates to use the new API (#15)

---------

Co-authored-by: ahmetustun <[email protected]>
Co-authored-by: Younes Belkada <[email protected]>
Co-authored-by: Matt <[email protected]>
SangbumChoi added a commit to SangbumChoi/transformers that referenced this pull request Aug 22, 2024
ArthurZucker pushed a commit that referenced this pull request Sep 25, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants