-
Notifications
You must be signed in to change notification settings - Fork 26.9k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Fix typo in subheader BertForQuestionAnswering #4
Merged
Conversation
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
exact thanks ! |
qwang70
pushed a commit
to DRL36/pytorch-pretrained-BERT
that referenced
this pull request
Mar 2, 2019
Closed
stevezheng23
added a commit
to stevezheng23/transformers
that referenced
this pull request
Mar 24, 2020
* upgrade roberta question answering based on xlnet question answering * upgrade roberta question answering based on xlnet question answering (cont.) * upgrade roberta question answering based on xlnet question answering (cont.) * upgrade roberta question answering based on xlnet question answering (cont.) * upgrade roberta question answering based on xlnet question answering (cont.) * upgrade roberta question answering based on xlnet question answering (cont.) * upgrade roberta question answering based on xlnet question answering (cont.) * upgrade roberta question answering based on xlnet question answering (cont.) * revert question answering changes in roberta/xlnet modeling * revert question answering changes in roberta/xlnet modeling (cont.) * revert to roberta qa simple * revert to roberta qa simple (cont.) * revert to roberta qa simple (cont.) * revert to roberta qa simple (cont.) * revert 'revert to roberta qa simple' * update paragraph/query order for xlnet * update paragraph/query order for xlnet (cont.) * update answer cls layer & modulize squad output layer for roberta/xlnet * update answer cls layer & modulize squad output layer for roberta/xlnet (cont.) * use start/end index lookup for detokenization * Revert "use start/end index lookup for detokenization" This reverts commit b8807c478dee2aed01e531f188ea191e0ef4f37a. * remove complex output layer for roberta-squad * add back complex qa output layer for roberta-squad
LysandreJik
added a commit
that referenced
this pull request
Apr 10, 2020
* Initial commit to get BERT + run_glue.py on TPU * Add README section for TPU and address comments. * Cleanup TPU bits from run_glue.py (#3) TPU runner is currently implemented in: https://github.com/pytorch-tpu/transformers/blob/tpu/examples/run_glue_tpu.py. We plan to upstream this directly into `huggingface/transformers` (either `master` or `tpu`) branch once it's been more thoroughly tested. * Cleanup TPU bits from run_glue.py TPU runner is currently implemented in: https://github.com/pytorch-tpu/transformers/blob/tpu/examples/run_glue_tpu.py. We plan to upstream this directly into `huggingface/transformers` (either `master` or `tpu`) branch once it's been more thoroughly tested. * No need to call `xm.mark_step()` explicitly (#4) Since for gradient accumulation we're accumulating on batches from `ParallelLoader` instance which on next() marks the step itself. * Resolve R/W conflicts from multiprocessing (#5) * Add XLNet in list of models for `run_glue_tpu.py` (#6) * Add RoBERTa to list of models in TPU GLUE (#7) * Add RoBERTa and DistilBert to list of models in TPU GLUE (#8) * Use barriers to reduce duplicate work/resources (#9) * Shard eval dataset and aggregate eval metrics (#10) * Shard eval dataset and aggregate eval metrics Also, instead of calling `eval_loss.item()` every time do summation with tensors on device. * Change defaultdict to float * Reduce the pred, label tensors instead of metrics As brought up during review some metrics like f1 cannot be aggregated via averaging. GLUE task metrics depends largely on the dataset, so instead we sync the prediction and label tensors so that the metrics can be computed accurately on those instead. * Only use tb_writer from master (#11) * Apply huggingface black code formatting * Style * Remove `--do_lower_case` as example uses cased * Add option to specify tensorboard logdir This is needed for our testing framework which checks regressions against key metrics writtern by the summary writer. * Using configuration for `xla_device` * Prefix TPU specific comments. * num_cores clarification and namespace eval metrics * Cache features file under `args.cache_dir` Instead of under `args.data_dir`. This is needed as our test infra uses data_dir with a read-only filesystem. * Rename `run_glue_tpu` to `run_tpu_glue` Co-authored-by: LysandreJik <[email protected]>
This was referenced Nov 12, 2020
stas00
added a commit
that referenced
this pull request
Dec 9, 2020
LysandreJik
pushed a commit
that referenced
this pull request
Dec 9, 2020
* ci-doc-job-skip-take-4 * wip * wip * wip * wip * skip yaml * wip * wip * wip * wip * wip * wip * wip * wip * wip * wip * wip * wip * ready to test * yet another way * trying with HEAD * trying with head.sha * trying with head.sha fix * trying with head.sha fix wip * undo * try to switch to sha * current branch * current branch * PR number check * joy ride * joy ride * joy ride * joy ride * joy ride * joy ride * joy ride * joy ride * joy ride * joy ride * joy ride * joy ride
2 tasks
wamartin-aml
pushed a commit
to wamartin-aml/transformers
that referenced
this pull request
Nov 1, 2021
Remove data based dependencies in T5 for ORT
4 tasks
rraminen
pushed a commit
to rraminen/transformers
that referenced
this pull request
Jun 3, 2022
Co-authored-by: Michael Melesse <[email protected]>
jlamypoirier
pushed a commit
to jlamypoirier/transformers
that referenced
this pull request
Apr 4, 2023
incorporate benchmarks
sim-so
added a commit
to sim-so/transformers
that referenced
this pull request
Apr 23, 2023
# This is the 1st commit message: Update docs/source/ko/tasks/summarization.mdx Co-authored-by: Wonhyeong Seo <[email protected]> # This is the commit message huggingface#2: Update docs/source/ko/tasks/summarization.mdx Co-authored-by: Wonhyeong Seo <[email protected]> # This is the commit message huggingface#3: Update docs/source/ko/tasks/summarization.mdx Co-authored-by: Wonhyeong Seo <[email protected]> # This is the commit message huggingface#4: Update docs/source/ko/tasks/summarization.mdx Co-authored-by: Wonhyeong Seo <[email protected]> # This is the commit message huggingface#5: Update docs/source/ko/tasks/summarization.mdx Co-authored-by: Wonhyeong Seo <[email protected]> # This is the commit message huggingface#6: Update docs/source/ko/tasks/summarization.mdx Co-authored-by: Wonhyeong Seo <[email protected]> # This is the commit message huggingface#7: Update docs/source/ko/tasks/summarization.mdx Co-authored-by: Wonhyeong Seo <[email protected]> # This is the commit message huggingface#8: Update docs/source/ko/tasks/summarization.mdx Co-authored-by: Wonhyeong Seo <[email protected]> # This is the commit message huggingface#9: Update docs/source/ko/tasks/summarization.mdx Co-authored-by: Wonhyeong Seo <[email protected]> # This is the commit message huggingface#10: Update docs/source/ko/tasks/summarization.mdx Co-authored-by: Wonhyeong Seo <[email protected]> # This is the commit message huggingface#11: Update docs/source/ko/tasks/summarization.mdx
jameshennessytempus
pushed a commit
to jameshennessytempus/transformers
that referenced
this pull request
Jun 1, 2023
1 task
nikolaJovisic
added a commit
to nikolaJovisic/transformers
that referenced
this pull request
Aug 22, 2023
fix binary classification for tensorflow segformer fix binary classification for tf segformer huggingface#2
nikolaJovisic
added a commit
to nikolaJovisic/transformers
that referenced
this pull request
Aug 22, 2023
nikolaJovisic
added a commit
to nikolaJovisic/transformers
that referenced
this pull request
Aug 23, 2023
fix binary classification for tensorflow segformer fix binary classification for tf segformer huggingface#2 fix huggingface#5 Revert "fix huggingface#5" This reverts commit 15b5160. Revert "fix huggingface#4" This reverts commit 0b534e6. fix huggingface#5 fix fix fix
meetpraveen
pushed a commit
to meetpraveen/transformers
that referenced
this pull request
Sep 3, 2023
4 tasks
ArthurZucker
pushed a commit
that referenced
this pull request
Feb 21, 2024
ArthurZucker
added a commit
that referenced
this pull request
Feb 21, 2024
* inital commit * update * update conversion checkpoint * update conversion script * nits * some fixes * nits * merge * fix permute * nits * fix * nits * nits * nits * fix rope * fix both rope * nites * style * make sure flax works * fix flax init code * fix foward * nits * print flax generation out * current code * nits * SIIIIIIIIIIIIIIIIIII * update * add new tokenizer * correct fast tokenizer * fix conversion * more comments * fix modeling and conversion * nits and nits * nits testing * add some tokenization tests * add some edge cases * add slow tests and fix them * fixup * fix copies for modeling * fix copies * add 7B slow tests * fix * fix * fix tests * make tokenizer cis go green * styling * last tokenizer nits * update jax tests * fix flax for 7b * add jit testing 🤗 * cleanups * isolated nit, inv_freq for rotary_emb.inv_freq * propagate to jax * Apply suggestions from code review Co-authored-by: Sanchit Gandhi <[email protected]> * adjust test * fix conversion script * change name * correct file names * update conversion script * Fix bos and eos token ids in the model configuration (#3) * update modelling * update conversion script * add static cache for gemma * fix sdpa generate * fix batched * multiple fixes * fix FA2 * final fix * Rename a few missing strings and filenames (#4) * merge with upstream main * fix copies * fix copies * fix fixup * fix fixup * fix * fix * final tests * fix fx gemma tests * fix fx bf16/fp16 tests * update slow fx tests * fx slow tests: one logits, one generation * move jit test standalone * Apply suggestions from code review * nits * tokenizer updates * more tokenization updates: custom GemmaSentencepieceExtrator * style * Update src/transformers/cache_utils.py * Update src/transformers/models/gemma/__init__.py * Update tests/models/gemma/test_modeling_flax_gemma.py * small nits * style * update tokenization test * fix the rotary embedding * with style * fix slow tests * WARNING this commit might be very important for precisions * Update tests/models/gemma/test_modeling_flax_gemma.py * Update src/transformers/models/gemma/configuration_gemma.py Co-authored-by: Lysandre Debut <[email protected]> * Update src/transformers/models/gemma/modeling_flax_gemma.py Co-authored-by: Lysandre Debut <[email protected]> * small nits here and there! * forgotten nit * remove on the fly computation of inv_freq * revert previous change, let's be safe and for now re-compute freq cis to make sure it's in float * Apply suggestions from code review Co-authored-by: Pedro Cuenca <[email protected]> * Update src/transformers/models/gemma/convert_gemma_weights_to_hf.py Co-authored-by: Pedro Cuenca <[email protected]> * Update src/transformers/models/gemma/convert_gemma_weights_to_hf.py Co-authored-by: Pedro Cuenca <[email protected]> * Update tests/models/gemma/test_modeling_gemma.py Co-authored-by: Pedro Cuenca <[email protected]> * Update tests/models/gemma/test_modeling_gemma.py Co-authored-by: Pedro Cuenca <[email protected]> * Update tests/models/gemma/test_modeling_gemma.py Co-authored-by: Pedro Cuenca <[email protected]> * Update tests/models/gemma/test_modeling_flax_gemma.py Co-authored-by: Pedro Cuenca <[email protected]> * Update tests/models/gemma/test_modeling_gemma.py Co-authored-by: Pedro Cuenca <[email protected]> * Update tests/models/gemma/test_modeling_gemma.py Co-authored-by: Pedro Cuenca <[email protected]> * Update tests/models/gemma/test_tokenization_gemma.py Co-authored-by: Pedro Cuenca <[email protected]> * Update tests/models/gemma/test_tokenization_gemma.py Co-authored-by: Pedro Cuenca <[email protected]> * Update tests/models/gemma/test_tokenization_gemma.py Co-authored-by: Pedro Cuenca <[email protected]> * Update tests/models/gemma/test_tokenization_gemma.py Co-authored-by: Pedro Cuenca <[email protected]> * Update tests/models/gemma/test_modeling_gemma.py Co-authored-by: Pedro Cuenca <[email protected]> * Update tests/models/gemma/test_modeling_gemma.py Co-authored-by: Pedro Cuenca <[email protected]> * Update tests/models/gemma/test_modeling_gemma.py Co-authored-by: Pedro Cuenca <[email protected]> * Update tests/models/gemma/test_modeling_gemma.py Co-authored-by: Pedro Cuenca <[email protected]> * Update tests/models/gemma/test_modeling_gemma.py Co-authored-by: Pedro Cuenca <[email protected]> * nit conversion script link * fix some tests * add not doctest and pr doctest * repo consistency * fix last CIs 🚀 * update all readmes --------- Co-authored-by: younesbelkada <[email protected]> Co-authored-by: Sanchit Gandhi <[email protected]> Co-authored-by: Pedro Cuenca <[email protected]> Co-authored-by: Younes Belkada <[email protected]> Co-authored-by: sanchit-gandhi <[email protected]> Co-authored-by: Lysandre Debut <[email protected]>
ArthurZucker
added a commit
that referenced
this pull request
Feb 21, 2024
* inital commit * update * update conversion checkpoint * update conversion script * nits * some fixes * nits * merge * fix permute * nits * fix * nits * nits * nits * fix rope * fix both rope * nites * style * make sure flax works * fix flax init code * fix foward * nits * print flax generation out * current code * nits * SIIIIIIIIIIIIIIIIIII * update * add new tokenizer * correct fast tokenizer * fix conversion * more comments * fix modeling and conversion * nits and nits * nits testing * add some tokenization tests * add some edge cases * add slow tests and fix them * fixup * fix copies for modeling * fix copies * add 7B slow tests * fix * fix * fix tests * make tokenizer cis go green * styling * last tokenizer nits * update jax tests * fix flax for 7b * add jit testing 🤗 * cleanups * isolated nit, inv_freq for rotary_emb.inv_freq * propagate to jax * Apply suggestions from code review Co-authored-by: Sanchit Gandhi <[email protected]> * adjust test * fix conversion script * change name * correct file names * update conversion script * Fix bos and eos token ids in the model configuration (#3) * update modelling * update conversion script * add static cache for gemma * fix sdpa generate * fix batched * multiple fixes * fix FA2 * final fix * Rename a few missing strings and filenames (#4) * merge with upstream main * fix copies * fix copies * fix fixup * fix fixup * fix * fix * final tests * fix fx gemma tests * fix fx bf16/fp16 tests * update slow fx tests * fx slow tests: one logits, one generation * move jit test standalone * Apply suggestions from code review * nits * tokenizer updates * more tokenization updates: custom GemmaSentencepieceExtrator * style * Update src/transformers/cache_utils.py * Update src/transformers/models/gemma/__init__.py * Update tests/models/gemma/test_modeling_flax_gemma.py * small nits * style * update tokenization test * fix the rotary embedding * with style * fix slow tests * WARNING this commit might be very important for precisions * Update tests/models/gemma/test_modeling_flax_gemma.py * Update src/transformers/models/gemma/configuration_gemma.py Co-authored-by: Lysandre Debut <[email protected]> * Update src/transformers/models/gemma/modeling_flax_gemma.py Co-authored-by: Lysandre Debut <[email protected]> * small nits here and there! * forgotten nit * remove on the fly computation of inv_freq * revert previous change, let's be safe and for now re-compute freq cis to make sure it's in float * Apply suggestions from code review Co-authored-by: Pedro Cuenca <[email protected]> * Update src/transformers/models/gemma/convert_gemma_weights_to_hf.py Co-authored-by: Pedro Cuenca <[email protected]> * Update src/transformers/models/gemma/convert_gemma_weights_to_hf.py Co-authored-by: Pedro Cuenca <[email protected]> * Update tests/models/gemma/test_modeling_gemma.py Co-authored-by: Pedro Cuenca <[email protected]> * Update tests/models/gemma/test_modeling_gemma.py Co-authored-by: Pedro Cuenca <[email protected]> * Update tests/models/gemma/test_modeling_gemma.py Co-authored-by: Pedro Cuenca <[email protected]> * Update tests/models/gemma/test_modeling_flax_gemma.py Co-authored-by: Pedro Cuenca <[email protected]> * Update tests/models/gemma/test_modeling_gemma.py Co-authored-by: Pedro Cuenca <[email protected]> * Update tests/models/gemma/test_modeling_gemma.py Co-authored-by: Pedro Cuenca <[email protected]> * Update tests/models/gemma/test_tokenization_gemma.py Co-authored-by: Pedro Cuenca <[email protected]> * Update tests/models/gemma/test_tokenization_gemma.py Co-authored-by: Pedro Cuenca <[email protected]> * Update tests/models/gemma/test_tokenization_gemma.py Co-authored-by: Pedro Cuenca <[email protected]> * Update tests/models/gemma/test_tokenization_gemma.py Co-authored-by: Pedro Cuenca <[email protected]> * Update tests/models/gemma/test_modeling_gemma.py Co-authored-by: Pedro Cuenca <[email protected]> * Update tests/models/gemma/test_modeling_gemma.py Co-authored-by: Pedro Cuenca <[email protected]> * Update tests/models/gemma/test_modeling_gemma.py Co-authored-by: Pedro Cuenca <[email protected]> * Update tests/models/gemma/test_modeling_gemma.py Co-authored-by: Pedro Cuenca <[email protected]> * Update tests/models/gemma/test_modeling_gemma.py Co-authored-by: Pedro Cuenca <[email protected]> * nit conversion script link * fix some tests * add not doctest and pr doctest * repo consistency * fix last CIs 🚀 * update all readmes --------- Co-authored-by: younesbelkada <[email protected]> Co-authored-by: Sanchit Gandhi <[email protected]> Co-authored-by: Pedro Cuenca <[email protected]> Co-authored-by: Younes Belkada <[email protected]> Co-authored-by: sanchit-gandhi <[email protected]> Co-authored-by: Lysandre Debut <[email protected]>
aymeric-roucher
added a commit
that referenced
this pull request
Apr 23, 2024
itazap
pushed a commit
that referenced
this pull request
May 14, 2024
* inital commit * update * update conversion checkpoint * update conversion script * nits * some fixes * nits * merge * fix permute * nits * fix * nits * nits * nits * fix rope * fix both rope * nites * style * make sure flax works * fix flax init code * fix foward * nits * print flax generation out * current code * nits * SIIIIIIIIIIIIIIIIIII * update * add new tokenizer * correct fast tokenizer * fix conversion * more comments * fix modeling and conversion * nits and nits * nits testing * add some tokenization tests * add some edge cases * add slow tests and fix them * fixup * fix copies for modeling * fix copies * add 7B slow tests * fix * fix * fix tests * make tokenizer cis go green * styling * last tokenizer nits * update jax tests * fix flax for 7b * add jit testing 🤗 * cleanups * isolated nit, inv_freq for rotary_emb.inv_freq * propagate to jax * Apply suggestions from code review Co-authored-by: Sanchit Gandhi <[email protected]> * adjust test * fix conversion script * change name * correct file names * update conversion script * Fix bos and eos token ids in the model configuration (#3) * update modelling * update conversion script * add static cache for gemma * fix sdpa generate * fix batched * multiple fixes * fix FA2 * final fix * Rename a few missing strings and filenames (#4) * merge with upstream main * fix copies * fix copies * fix fixup * fix fixup * fix * fix * final tests * fix fx gemma tests * fix fx bf16/fp16 tests * update slow fx tests * fx slow tests: one logits, one generation * move jit test standalone * Apply suggestions from code review * nits * tokenizer updates * more tokenization updates: custom GemmaSentencepieceExtrator * style * Update src/transformers/cache_utils.py * Update src/transformers/models/gemma/__init__.py * Update tests/models/gemma/test_modeling_flax_gemma.py * small nits * style * update tokenization test * fix the rotary embedding * with style * fix slow tests * WARNING this commit might be very important for precisions * Update tests/models/gemma/test_modeling_flax_gemma.py * Update src/transformers/models/gemma/configuration_gemma.py Co-authored-by: Lysandre Debut <[email protected]> * Update src/transformers/models/gemma/modeling_flax_gemma.py Co-authored-by: Lysandre Debut <[email protected]> * small nits here and there! * forgotten nit * remove on the fly computation of inv_freq * revert previous change, let's be safe and for now re-compute freq cis to make sure it's in float * Apply suggestions from code review Co-authored-by: Pedro Cuenca <[email protected]> * Update src/transformers/models/gemma/convert_gemma_weights_to_hf.py Co-authored-by: Pedro Cuenca <[email protected]> * Update src/transformers/models/gemma/convert_gemma_weights_to_hf.py Co-authored-by: Pedro Cuenca <[email protected]> * Update tests/models/gemma/test_modeling_gemma.py Co-authored-by: Pedro Cuenca <[email protected]> * Update tests/models/gemma/test_modeling_gemma.py Co-authored-by: Pedro Cuenca <[email protected]> * Update tests/models/gemma/test_modeling_gemma.py Co-authored-by: Pedro Cuenca <[email protected]> * Update tests/models/gemma/test_modeling_flax_gemma.py Co-authored-by: Pedro Cuenca <[email protected]> * Update tests/models/gemma/test_modeling_gemma.py Co-authored-by: Pedro Cuenca <[email protected]> * Update tests/models/gemma/test_modeling_gemma.py Co-authored-by: Pedro Cuenca <[email protected]> * Update tests/models/gemma/test_tokenization_gemma.py Co-authored-by: Pedro Cuenca <[email protected]> * Update tests/models/gemma/test_tokenization_gemma.py Co-authored-by: Pedro Cuenca <[email protected]> * Update tests/models/gemma/test_tokenization_gemma.py Co-authored-by: Pedro Cuenca <[email protected]> * Update tests/models/gemma/test_tokenization_gemma.py Co-authored-by: Pedro Cuenca <[email protected]> * Update tests/models/gemma/test_modeling_gemma.py Co-authored-by: Pedro Cuenca <[email protected]> * Update tests/models/gemma/test_modeling_gemma.py Co-authored-by: Pedro Cuenca <[email protected]> * Update tests/models/gemma/test_modeling_gemma.py Co-authored-by: Pedro Cuenca <[email protected]> * Update tests/models/gemma/test_modeling_gemma.py Co-authored-by: Pedro Cuenca <[email protected]> * Update tests/models/gemma/test_modeling_gemma.py Co-authored-by: Pedro Cuenca <[email protected]> * nit conversion script link * fix some tests * add not doctest and pr doctest * repo consistency * fix last CIs 🚀 * update all readmes --------- Co-authored-by: younesbelkada <[email protected]> Co-authored-by: Sanchit Gandhi <[email protected]> Co-authored-by: Pedro Cuenca <[email protected]> Co-authored-by: Younes Belkada <[email protected]> Co-authored-by: sanchit-gandhi <[email protected]> Co-authored-by: Lysandre Debut <[email protected]>
SangbumChoi
added a commit
to SangbumChoi/transformers
that referenced
this pull request
Aug 22, 2024
minor chore detr and remove obj in mask2former
gante
added a commit
that referenced
this pull request
Oct 23, 2024
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Should say
BertForQuestionAnswering
, but saysBertForSequenceClassification
.