Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Fix typo in subheader BertForQuestionAnswering #4

Merged
merged 1 commit into from
Nov 5, 2018

Conversation

knutole
Copy link
Contributor

@knutole knutole commented Nov 5, 2018

Should say BertForQuestionAnswering, but says BertForSequenceClassification.

@VictorSanh VictorSanh merged commit 886f595 into huggingface:master Nov 5, 2018
@VictorSanh
Copy link
Contributor

exact thanks !

qwang70 pushed a commit to DRL36/pytorch-pretrained-BERT that referenced this pull request Mar 2, 2019
@HongyanJiao HongyanJiao mentioned this pull request Sep 19, 2019
stevezheng23 added a commit to stevezheng23/transformers that referenced this pull request Mar 24, 2020
* upgrade roberta question answering based on xlnet question answering

* upgrade roberta question answering based on xlnet question answering (cont.)

* upgrade roberta question answering based on xlnet question answering (cont.)

* upgrade roberta question answering based on xlnet question answering (cont.)

* upgrade roberta question answering based on xlnet question answering (cont.)

* upgrade roberta question answering based on xlnet question answering (cont.)

* upgrade roberta question answering based on xlnet question answering (cont.)

* upgrade roberta question answering based on xlnet question answering (cont.)

* revert question answering changes in roberta/xlnet modeling

* revert question answering changes in roberta/xlnet modeling (cont.)

* revert to roberta qa simple

* revert to roberta qa simple (cont.)

* revert to roberta qa simple (cont.)

* revert to roberta qa simple (cont.)

* revert 'revert to roberta qa simple'

* update paragraph/query order for xlnet

* update paragraph/query order for xlnet (cont.)

* update answer cls layer & modulize squad output layer for roberta/xlnet

* update answer cls layer & modulize squad output layer for roberta/xlnet (cont.)

* use start/end index lookup for detokenization

* Revert "use start/end index lookup for detokenization"

This reverts commit b8807c478dee2aed01e531f188ea191e0ef4f37a.

* remove complex output layer for roberta-squad

* add back complex qa output layer for roberta-squad
LysandreJik added a commit that referenced this pull request Apr 10, 2020
* Initial commit to get BERT + run_glue.py on TPU

* Add README section for TPU and address comments.

* Cleanup TPU bits from run_glue.py (#3)

TPU runner is currently implemented in:
https://github.com/pytorch-tpu/transformers/blob/tpu/examples/run_glue_tpu.py.

We plan to upstream this directly into `huggingface/transformers`
(either `master` or `tpu`) branch once it's been more thoroughly tested.

* Cleanup TPU bits from run_glue.py

TPU runner is currently implemented in:
https://github.com/pytorch-tpu/transformers/blob/tpu/examples/run_glue_tpu.py.

We plan to upstream this directly into `huggingface/transformers`
(either `master` or `tpu`) branch once it's been more thoroughly tested.

* No need to call `xm.mark_step()` explicitly (#4)

Since for gradient accumulation we're accumulating on batches from
`ParallelLoader` instance which on next() marks the step itself.

* Resolve R/W conflicts from multiprocessing (#5)

* Add XLNet in list of models for `run_glue_tpu.py` (#6)

* Add RoBERTa to list of models in TPU GLUE (#7)

* Add RoBERTa and DistilBert to list of models in TPU GLUE (#8)

* Use barriers to reduce duplicate work/resources (#9)

* Shard eval dataset and aggregate eval metrics (#10)

* Shard eval dataset and aggregate eval metrics

Also, instead of calling `eval_loss.item()` every time do summation with
tensors on device.

* Change defaultdict to float

* Reduce the pred, label tensors instead of metrics

As brought up during review some metrics like f1 cannot be aggregated
via averaging. GLUE task metrics depends largely on the dataset, so
instead we sync the prediction and label tensors so that the metrics can
be computed accurately on those instead.

* Only use tb_writer from master (#11)

* Apply huggingface black code formatting

* Style

* Remove `--do_lower_case` as example uses cased

* Add option to specify tensorboard logdir

This is needed for our testing framework which checks regressions
against key metrics writtern by the summary writer.

* Using configuration for `xla_device`

* Prefix TPU specific comments.

* num_cores clarification and namespace eval metrics

* Cache features file under `args.cache_dir`

Instead of under `args.data_dir`. This is needed as our test infra uses
data_dir with a read-only filesystem.

* Rename `run_glue_tpu` to `run_tpu_glue`

Co-authored-by: LysandreJik <[email protected]>
LysandreJik pushed a commit that referenced this pull request Dec 9, 2020
* ci-doc-job-skip-take-4

* wip

* wip

* wip

* wip

* skip yaml

* wip

* wip

* wip

* wip

* wip

* wip

* wip

* wip

* wip

* wip

* wip

* wip

* ready to test

* yet another way

* trying with HEAD

* trying with head.sha

* trying with head.sha fix

* trying with head.sha fix wip

* undo

* try to switch to sha

* current branch

* current branch

* PR number check

* joy ride

* joy ride

* joy ride

* joy ride

* joy ride

* joy ride

* joy ride

* joy ride

* joy ride

* joy ride

* joy ride

* joy ride
wamartin-aml pushed a commit to wamartin-aml/transformers that referenced this pull request Nov 1, 2021
Remove data based dependencies in T5 for ORT
rraminen pushed a commit to rraminen/transformers that referenced this pull request Jun 3, 2022
jlamypoirier pushed a commit to jlamypoirier/transformers that referenced this pull request Apr 4, 2023
incorporate benchmarks
sim-so added a commit to sim-so/transformers that referenced this pull request Apr 23, 2023
# This is the 1st commit message:

Update docs/source/ko/tasks/summarization.mdx

Co-authored-by: Wonhyeong Seo <[email protected]>
# This is the commit message huggingface#2:

Update docs/source/ko/tasks/summarization.mdx

Co-authored-by: Wonhyeong Seo <[email protected]>
# This is the commit message huggingface#3:

Update docs/source/ko/tasks/summarization.mdx

Co-authored-by: Wonhyeong Seo <[email protected]>
# This is the commit message huggingface#4:

Update docs/source/ko/tasks/summarization.mdx

Co-authored-by: Wonhyeong Seo <[email protected]>
# This is the commit message huggingface#5:

Update docs/source/ko/tasks/summarization.mdx

Co-authored-by: Wonhyeong Seo <[email protected]>
# This is the commit message huggingface#6:

Update docs/source/ko/tasks/summarization.mdx

Co-authored-by: Wonhyeong Seo <[email protected]>
# This is the commit message huggingface#7:

Update docs/source/ko/tasks/summarization.mdx

Co-authored-by: Wonhyeong Seo <[email protected]>
# This is the commit message huggingface#8:

Update docs/source/ko/tasks/summarization.mdx

Co-authored-by: Wonhyeong Seo <[email protected]>
# This is the commit message huggingface#9:

Update docs/source/ko/tasks/summarization.mdx

Co-authored-by: Wonhyeong Seo <[email protected]>
# This is the commit message huggingface#10:

Update docs/source/ko/tasks/summarization.mdx

Co-authored-by: Wonhyeong Seo <[email protected]>
# This is the commit message huggingface#11:

Update docs/source/ko/tasks/summarization.mdx
jameshennessytempus pushed a commit to jameshennessytempus/transformers that referenced this pull request Jun 1, 2023
nikolaJovisic added a commit to nikolaJovisic/transformers that referenced this pull request Aug 22, 2023
fix binary classification for tensorflow segformer

fix binary classification for tf segformer huggingface#2
nikolaJovisic added a commit to nikolaJovisic/transformers that referenced this pull request Aug 22, 2023
This reverts commit 0b534e6.
nikolaJovisic added a commit to nikolaJovisic/transformers that referenced this pull request Aug 23, 2023
fix binary classification for tensorflow segformer

fix binary classification for tf segformer huggingface#2

fix huggingface#5

Revert "fix huggingface#5"

This reverts commit 15b5160.

Revert "fix huggingface#4"

This reverts commit 0b534e6.

fix huggingface#5

fix

fix

fix
meetpraveen pushed a commit to meetpraveen/transformers that referenced this pull request Sep 3, 2023
ArthurZucker added a commit that referenced this pull request Feb 21, 2024
* inital commit

* update

* update conversion checkpoint

* update conversion script

* nits

* some fixes

* nits

* merge

* fix permute

* nits

* fix

* nits

* nits

* nits

* fix rope

* fix both rope

* nites

* style

* make sure flax works

* fix flax init code

* fix foward

* nits

* print flax generation out

* current code

* nits

* SIIIIIIIIIIIIIIIIIII

* update

* add new tokenizer

* correct fast tokenizer

* fix conversion

* more comments

* fix modeling and conversion

* nits and nits

* nits testing

* add some tokenization tests

* add some edge cases

* add slow tests and fix them

* fixup

* fix copies for modeling

* fix copies

* add 7B slow tests

* fix

* fix

* fix tests

* make tokenizer cis go green

* styling

* last tokenizer nits

* update jax tests

* fix flax for 7b

* add jit testing 🤗

* cleanups

* isolated nit, inv_freq for rotary_emb.inv_freq

* propagate to jax

* Apply suggestions from code review

Co-authored-by: Sanchit Gandhi <[email protected]>

* adjust test

* fix conversion script

* change name

* correct file names

* update conversion script

* Fix bos and eos token ids in the model configuration (#3)

* update modelling

* update conversion script

* add static cache for gemma

* fix sdpa generate

* fix batched

* multiple fixes

* fix FA2

* final fix

* Rename a few missing strings and filenames (#4)

* merge with upstream main

* fix copies

* fix copies

* fix fixup

* fix fixup

* fix

* fix

* final tests

* fix fx gemma tests

* fix fx bf16/fp16 tests

* update slow fx tests

* fx slow tests: one logits, one generation

* move jit test standalone

* Apply suggestions from code review

* nits

* tokenizer updates

* more tokenization updates: custom GemmaSentencepieceExtrator

* style

* Update src/transformers/cache_utils.py

* Update src/transformers/models/gemma/__init__.py

* Update tests/models/gemma/test_modeling_flax_gemma.py

* small nits

* style

* update tokenization test

* fix the rotary embedding

* with style

* fix slow tests

* WARNING this commit might be very important for precisions

* Update tests/models/gemma/test_modeling_flax_gemma.py

* Update src/transformers/models/gemma/configuration_gemma.py

Co-authored-by: Lysandre Debut <[email protected]>

* Update src/transformers/models/gemma/modeling_flax_gemma.py

Co-authored-by: Lysandre Debut <[email protected]>

* small nits here and there!

* forgotten nit

* remove on the fly computation of inv_freq

* revert previous change, let's be safe and for now re-compute freq cis to make sure it's in float

* Apply suggestions from code review

Co-authored-by: Pedro Cuenca <[email protected]>

* Update src/transformers/models/gemma/convert_gemma_weights_to_hf.py

Co-authored-by: Pedro Cuenca <[email protected]>

* Update src/transformers/models/gemma/convert_gemma_weights_to_hf.py

Co-authored-by: Pedro Cuenca <[email protected]>

* Update tests/models/gemma/test_modeling_gemma.py

Co-authored-by: Pedro Cuenca <[email protected]>

* Update tests/models/gemma/test_modeling_gemma.py

Co-authored-by: Pedro Cuenca <[email protected]>

* Update tests/models/gemma/test_modeling_gemma.py

Co-authored-by: Pedro Cuenca <[email protected]>

* Update tests/models/gemma/test_modeling_flax_gemma.py

Co-authored-by: Pedro Cuenca <[email protected]>

* Update tests/models/gemma/test_modeling_gemma.py

Co-authored-by: Pedro Cuenca <[email protected]>

* Update tests/models/gemma/test_modeling_gemma.py

Co-authored-by: Pedro Cuenca <[email protected]>

* Update tests/models/gemma/test_tokenization_gemma.py

Co-authored-by: Pedro Cuenca <[email protected]>

* Update tests/models/gemma/test_tokenization_gemma.py

Co-authored-by: Pedro Cuenca <[email protected]>

* Update tests/models/gemma/test_tokenization_gemma.py

Co-authored-by: Pedro Cuenca <[email protected]>

* Update tests/models/gemma/test_tokenization_gemma.py

Co-authored-by: Pedro Cuenca <[email protected]>

* Update tests/models/gemma/test_modeling_gemma.py

Co-authored-by: Pedro Cuenca <[email protected]>

* Update tests/models/gemma/test_modeling_gemma.py

Co-authored-by: Pedro Cuenca <[email protected]>

* Update tests/models/gemma/test_modeling_gemma.py

Co-authored-by: Pedro Cuenca <[email protected]>

* Update tests/models/gemma/test_modeling_gemma.py

Co-authored-by: Pedro Cuenca <[email protected]>

* Update tests/models/gemma/test_modeling_gemma.py

Co-authored-by: Pedro Cuenca <[email protected]>

* nit conversion script link

* fix some tests

* add not doctest and pr doctest

* repo consistency

* fix last CIs 🚀

* update all readmes

---------

Co-authored-by: younesbelkada <[email protected]>
Co-authored-by: Sanchit Gandhi <[email protected]>
Co-authored-by: Pedro Cuenca <[email protected]>
Co-authored-by: Younes Belkada <[email protected]>
Co-authored-by: sanchit-gandhi <[email protected]>
Co-authored-by: Lysandre Debut <[email protected]>
ArthurZucker added a commit that referenced this pull request Feb 21, 2024
* inital commit

* update

* update conversion checkpoint

* update conversion script

* nits

* some fixes

* nits

* merge

* fix permute

* nits

* fix

* nits

* nits

* nits

* fix rope

* fix both rope

* nites

* style

* make sure flax works

* fix flax init code

* fix foward

* nits

* print flax generation out

* current code

* nits

* SIIIIIIIIIIIIIIIIIII

* update

* add new tokenizer

* correct fast tokenizer

* fix conversion

* more comments

* fix modeling and conversion

* nits and nits

* nits testing

* add some tokenization tests

* add some edge cases

* add slow tests and fix them

* fixup

* fix copies for modeling

* fix copies

* add 7B slow tests

* fix

* fix

* fix tests

* make tokenizer cis go green

* styling

* last tokenizer nits

* update jax tests

* fix flax for 7b

* add jit testing 🤗

* cleanups

* isolated nit, inv_freq for rotary_emb.inv_freq

* propagate to jax

* Apply suggestions from code review

Co-authored-by: Sanchit Gandhi <[email protected]>

* adjust test

* fix conversion script

* change name

* correct file names

* update conversion script

* Fix bos and eos token ids in the model configuration (#3)

* update modelling

* update conversion script

* add static cache for gemma

* fix sdpa generate

* fix batched

* multiple fixes

* fix FA2

* final fix

* Rename a few missing strings and filenames (#4)

* merge with upstream main

* fix copies

* fix copies

* fix fixup

* fix fixup

* fix

* fix

* final tests

* fix fx gemma tests

* fix fx bf16/fp16 tests

* update slow fx tests

* fx slow tests: one logits, one generation

* move jit test standalone

* Apply suggestions from code review

* nits

* tokenizer updates

* more tokenization updates: custom GemmaSentencepieceExtrator

* style

* Update src/transformers/cache_utils.py

* Update src/transformers/models/gemma/__init__.py

* Update tests/models/gemma/test_modeling_flax_gemma.py

* small nits

* style

* update tokenization test

* fix the rotary embedding

* with style

* fix slow tests

* WARNING this commit might be very important for precisions

* Update tests/models/gemma/test_modeling_flax_gemma.py

* Update src/transformers/models/gemma/configuration_gemma.py

Co-authored-by: Lysandre Debut <[email protected]>

* Update src/transformers/models/gemma/modeling_flax_gemma.py

Co-authored-by: Lysandre Debut <[email protected]>

* small nits here and there!

* forgotten nit

* remove on the fly computation of inv_freq

* revert previous change, let's be safe and for now re-compute freq cis to make sure it's in float

* Apply suggestions from code review

Co-authored-by: Pedro Cuenca <[email protected]>

* Update src/transformers/models/gemma/convert_gemma_weights_to_hf.py

Co-authored-by: Pedro Cuenca <[email protected]>

* Update src/transformers/models/gemma/convert_gemma_weights_to_hf.py

Co-authored-by: Pedro Cuenca <[email protected]>

* Update tests/models/gemma/test_modeling_gemma.py

Co-authored-by: Pedro Cuenca <[email protected]>

* Update tests/models/gemma/test_modeling_gemma.py

Co-authored-by: Pedro Cuenca <[email protected]>

* Update tests/models/gemma/test_modeling_gemma.py

Co-authored-by: Pedro Cuenca <[email protected]>

* Update tests/models/gemma/test_modeling_flax_gemma.py

Co-authored-by: Pedro Cuenca <[email protected]>

* Update tests/models/gemma/test_modeling_gemma.py

Co-authored-by: Pedro Cuenca <[email protected]>

* Update tests/models/gemma/test_modeling_gemma.py

Co-authored-by: Pedro Cuenca <[email protected]>

* Update tests/models/gemma/test_tokenization_gemma.py

Co-authored-by: Pedro Cuenca <[email protected]>

* Update tests/models/gemma/test_tokenization_gemma.py

Co-authored-by: Pedro Cuenca <[email protected]>

* Update tests/models/gemma/test_tokenization_gemma.py

Co-authored-by: Pedro Cuenca <[email protected]>

* Update tests/models/gemma/test_tokenization_gemma.py

Co-authored-by: Pedro Cuenca <[email protected]>

* Update tests/models/gemma/test_modeling_gemma.py

Co-authored-by: Pedro Cuenca <[email protected]>

* Update tests/models/gemma/test_modeling_gemma.py

Co-authored-by: Pedro Cuenca <[email protected]>

* Update tests/models/gemma/test_modeling_gemma.py

Co-authored-by: Pedro Cuenca <[email protected]>

* Update tests/models/gemma/test_modeling_gemma.py

Co-authored-by: Pedro Cuenca <[email protected]>

* Update tests/models/gemma/test_modeling_gemma.py

Co-authored-by: Pedro Cuenca <[email protected]>

* nit conversion script link

* fix some tests

* add not doctest and pr doctest

* repo consistency

* fix last CIs 🚀

* update all readmes

---------

Co-authored-by: younesbelkada <[email protected]>
Co-authored-by: Sanchit Gandhi <[email protected]>
Co-authored-by: Pedro Cuenca <[email protected]>
Co-authored-by: Younes Belkada <[email protected]>
Co-authored-by: sanchit-gandhi <[email protected]>
Co-authored-by: Lysandre Debut <[email protected]>
aymeric-roucher added a commit that referenced this pull request Apr 23, 2024
Support variable usage in ReactAgent
itazap pushed a commit that referenced this pull request May 14, 2024
* inital commit

* update

* update conversion checkpoint

* update conversion script

* nits

* some fixes

* nits

* merge

* fix permute

* nits

* fix

* nits

* nits

* nits

* fix rope

* fix both rope

* nites

* style

* make sure flax works

* fix flax init code

* fix foward

* nits

* print flax generation out

* current code

* nits

* SIIIIIIIIIIIIIIIIIII

* update

* add new tokenizer

* correct fast tokenizer

* fix conversion

* more comments

* fix modeling and conversion

* nits and nits

* nits testing

* add some tokenization tests

* add some edge cases

* add slow tests and fix them

* fixup

* fix copies for modeling

* fix copies

* add 7B slow tests

* fix

* fix

* fix tests

* make tokenizer cis go green

* styling

* last tokenizer nits

* update jax tests

* fix flax for 7b

* add jit testing 🤗

* cleanups

* isolated nit, inv_freq for rotary_emb.inv_freq

* propagate to jax

* Apply suggestions from code review

Co-authored-by: Sanchit Gandhi <[email protected]>

* adjust test

* fix conversion script

* change name

* correct file names

* update conversion script

* Fix bos and eos token ids in the model configuration (#3)

* update modelling

* update conversion script

* add static cache for gemma

* fix sdpa generate

* fix batched

* multiple fixes

* fix FA2

* final fix

* Rename a few missing strings and filenames (#4)

* merge with upstream main

* fix copies

* fix copies

* fix fixup

* fix fixup

* fix

* fix

* final tests

* fix fx gemma tests

* fix fx bf16/fp16 tests

* update slow fx tests

* fx slow tests: one logits, one generation

* move jit test standalone

* Apply suggestions from code review

* nits

* tokenizer updates

* more tokenization updates: custom GemmaSentencepieceExtrator

* style

* Update src/transformers/cache_utils.py

* Update src/transformers/models/gemma/__init__.py

* Update tests/models/gemma/test_modeling_flax_gemma.py

* small nits

* style

* update tokenization test

* fix the rotary embedding

* with style

* fix slow tests

* WARNING this commit might be very important for precisions

* Update tests/models/gemma/test_modeling_flax_gemma.py

* Update src/transformers/models/gemma/configuration_gemma.py

Co-authored-by: Lysandre Debut <[email protected]>

* Update src/transformers/models/gemma/modeling_flax_gemma.py

Co-authored-by: Lysandre Debut <[email protected]>

* small nits here and there!

* forgotten nit

* remove on the fly computation of inv_freq

* revert previous change, let's be safe and for now re-compute freq cis to make sure it's in float

* Apply suggestions from code review

Co-authored-by: Pedro Cuenca <[email protected]>

* Update src/transformers/models/gemma/convert_gemma_weights_to_hf.py

Co-authored-by: Pedro Cuenca <[email protected]>

* Update src/transformers/models/gemma/convert_gemma_weights_to_hf.py

Co-authored-by: Pedro Cuenca <[email protected]>

* Update tests/models/gemma/test_modeling_gemma.py

Co-authored-by: Pedro Cuenca <[email protected]>

* Update tests/models/gemma/test_modeling_gemma.py

Co-authored-by: Pedro Cuenca <[email protected]>

* Update tests/models/gemma/test_modeling_gemma.py

Co-authored-by: Pedro Cuenca <[email protected]>

* Update tests/models/gemma/test_modeling_flax_gemma.py

Co-authored-by: Pedro Cuenca <[email protected]>

* Update tests/models/gemma/test_modeling_gemma.py

Co-authored-by: Pedro Cuenca <[email protected]>

* Update tests/models/gemma/test_modeling_gemma.py

Co-authored-by: Pedro Cuenca <[email protected]>

* Update tests/models/gemma/test_tokenization_gemma.py

Co-authored-by: Pedro Cuenca <[email protected]>

* Update tests/models/gemma/test_tokenization_gemma.py

Co-authored-by: Pedro Cuenca <[email protected]>

* Update tests/models/gemma/test_tokenization_gemma.py

Co-authored-by: Pedro Cuenca <[email protected]>

* Update tests/models/gemma/test_tokenization_gemma.py

Co-authored-by: Pedro Cuenca <[email protected]>

* Update tests/models/gemma/test_modeling_gemma.py

Co-authored-by: Pedro Cuenca <[email protected]>

* Update tests/models/gemma/test_modeling_gemma.py

Co-authored-by: Pedro Cuenca <[email protected]>

* Update tests/models/gemma/test_modeling_gemma.py

Co-authored-by: Pedro Cuenca <[email protected]>

* Update tests/models/gemma/test_modeling_gemma.py

Co-authored-by: Pedro Cuenca <[email protected]>

* Update tests/models/gemma/test_modeling_gemma.py

Co-authored-by: Pedro Cuenca <[email protected]>

* nit conversion script link

* fix some tests

* add not doctest and pr doctest

* repo consistency

* fix last CIs 🚀

* update all readmes

---------

Co-authored-by: younesbelkada <[email protected]>
Co-authored-by: Sanchit Gandhi <[email protected]>
Co-authored-by: Pedro Cuenca <[email protected]>
Co-authored-by: Younes Belkada <[email protected]>
Co-authored-by: sanchit-gandhi <[email protected]>
Co-authored-by: Lysandre Debut <[email protected]>
SangbumChoi added a commit to SangbumChoi/transformers that referenced this pull request Aug 22, 2024
minor chore detr and remove obj in mask2former
gante added a commit that referenced this pull request Oct 23, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants