forked from PaddlePaddle/PaddleNLP
-
Notifications
You must be signed in to change notification settings - Fork 0
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[taskflow]SentenceFeatureExtractionTask #1
Open
qingzhong1
wants to merge
1,380
commits into
qingzhong1:develop
Choose a base branch
from
PaddlePaddle:develop
base: develop
Could not load branches
Branch not found: {{ refName }}
Loading
Could not load tags
Nothing to show
Loading
Are you sure you want to change the base?
Some commits from the old base branch may be removed from the timeline,
and old review comments may become outdated.
Conversation
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
qingzhong1
pushed a commit
that referenced
this pull request
Sep 18, 2023
Improve prompt; fix <br> bug
* Fix the issue of P-tuning official sample error * Create codecov workflow.yml * Delete .github/workflows/workflow.yml
* modify api for pir * modify api for pir * pass none for while * modify ci test
* update * update * remove
Co-authored-by: winter-wang <[email protected]>
* Update README.md * Update README_en.md
* stage 1 * update * update * fix ci * fix ci * fix ut
* modify tensorboard requirements * modify tensorboard requirements * modify tensorboard requirement
* stage 1 * update * update * support qwen2 bf16/wint8 * add qwen2 ptq map * update * fix tune_cublaslt_gemm.cu
* modify dict include none * recover
* add yuan model * add yuan model settings * fix conflict * add readme * update format * update readme * update readme * update for lint * update for lint * update for lint * update for lint * add fp16 * update utils * fix bug * format * correct pre-commit * correct fp16 * delete rearrange * add pre_train * auto convert from torch * support sft lora * update structure * fix yuantokenizer pad * update scripts * update readme * update readme * add pad_token_id * format * format * update readme * update for review * update for review * fix bug * update to CRLF * format * support param convert * update sft config * fix modeling * fix qk fuse&spilt and fix fa * format * format * format * fix fa dtype --------- Co-authored-by: drownfish19 <[email protected]>
* add yuan2 model list * update
* set logging_step to 5 with baichuan && qwen benchmark
* support llama3.1 * update
* fix load best * update
* fix bug * update
* add precision alignment doc
* update qwen tokenizer * add test case
* Add ci for A100 and V100 device * Fix device judge * fix compare of float var * Add output of test failed * Fix cal fail test count * Remove compare_float fun with `!=` to compare * Fix output of fail tests * Run gpt-3_dygraph when related files modified * add track status( fail tests ) for every case * Add track status of loss verification failed test * Move output position of every case * modify `exit` to `return 0`
* add fp8 gen files to gitignore * append_attn support fp8 quant * Unified FP8 Network * include cuda_fp8.h * simplify qwen2 network and FusedBlockMultiTransformerFP8 * simplify llama network and code check * check fp8 params * code check * check * default config for fp8 gemm
* support optimizer state offload (#8715) * support optimizer offload * update doc * [FleetY]offload optimizer state after load optmizer state (#9352) * add offload optimizer * fix memory * [FleetY] Add reload/offload for optimizer (#9356) * add reload/offload for optimizer * fix bug --------- Co-authored-by: Guoxia Wang <[email protected]>
* mv assert to warning
* optimize fuse some kernels * optimize fuse some kernels * fix top_p reject * fix * ci * fix review * fix
* Fix exitcode bug * Fix `track_case_status` func match bug * Fix return code * Fix print_info func with exit -6 * set output format of fail tests modify verification check failed
Co-authored-by: 周天宇 <[email protected]>
* fix empty state_dict * update sharding split_parma
* add submodule * add submodule
* fix ci scripts * [CI]add recursive for submodule
* fix ci scripts * [CI]add recursive for submodule * [CI]fix
* add ktotrainer * fix oom
Co-authored-by: root <[email protected]>
* update target_lists_for_llm * add scripts/regression
* add the AutoModel for inference mode in dynamic graph * add the AutoModel for inference mode in static graph * create AutoInferenceModelForCausalLM class and polish the code * fix the return result * add the confirm_inference_model method * roback the AutoModel * modify the description * update
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
PR types
PR changes
Description