Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[taskflow]SentenceFeatureExtractionTask #1

Open
wants to merge 1,380 commits into
base: develop
Choose a base branch
from

Conversation

qingzhong1
Copy link
Owner

PR types

PR changes

Description

qingzhong1 pushed a commit that referenced this pull request Sep 18, 2023
Improve prompt; fix <br> bug
deepllz and others added 29 commits August 6, 2024 17:24
* Fix the issue of P-tuning official sample error

* Create codecov workflow.yml

* Delete .github/workflows/workflow.yml
* modify api for pir

* modify api for pir

* pass none for while

* modify ci test
* update

* update

* remove
* Update README.md

* Update README_en.md
* stage 1

* update

* update

* fix ci

* fix ci

* fix ut
* modify tensorboard requirements

* modify tensorboard requirements

* modify tensorboard requirement
* stage 1

* update

* update

* support qwen2 bf16/wint8

* add qwen2 ptq map

* update

* fix tune_cublaslt_gemm.cu
* add yuan model

* add yuan model settings

* fix conflict

* add readme

* update format

* update readme

* update readme

* update for lint

* update for lint

* update for lint

* update for lint

* add fp16

* update utils

* fix bug

* format

* correct pre-commit

* correct fp16

* delete rearrange

* add pre_train

* auto convert from torch

* support sft lora

* update structure

* fix yuantokenizer pad

* update scripts

* update readme

* update readme

* add pad_token_id

* format

* format

* update readme

* update for review

* update for review

* fix bug

* update to CRLF

* format

* support param convert

* update sft config

* fix modeling

* fix qk fuse&spilt and fix fa

* format

* format

* format

* fix fa dtype

---------

Co-authored-by: drownfish19 <[email protected]>
* add yuan2 model list

* update
* set logging_step to 5 with baichuan && qwen benchmark
* support llama3.1

* update
* fix bug

* update
DrownFish19 and others added 30 commits October 31, 2024 15:38
* add precision alignment doc
* Add ci for A100 and V100 device

* Fix device judge

* fix compare of float var

* Add output of test failed

* Fix cal fail test count

* Remove compare_float fun with `!=` to compare

* Fix output of fail tests

* Run  gpt-3_dygraph when related files modified

* add track status( fail tests ) for every case

* Add track status of loss verification failed test

* Move output position of every case

* modify `exit` to `return 0`
* add fp8 gen files to gitignore

* append_attn support fp8 quant

* Unified FP8 Network

* include cuda_fp8.h

* simplify qwen2 network and FusedBlockMultiTransformerFP8

* simplify llama network and code check

* check fp8 params

* code check

* check

* default config for fp8 gemm
)

* Add , support register tokenizer, fix some typo

* add more tests

* CustomTokenizerFast2->CustomTokenizerFastWithoutSlow

* lint
* support optimizer state offload (#8715)

* support optimizer offload

* update doc

* [FleetY]offload optimizer state after load optmizer state (#9352)

* add offload optimizer

* fix memory

* [FleetY] Add reload/offload for optimizer (#9356)

* add reload/offload for optimizer

* fix bug

---------

Co-authored-by: Guoxia Wang <[email protected]>
* optimize fuse some kernels

* optimize fuse some kernels

* fix top_p reject

* fix

* ci

* fix review

* fix
* Fix exitcode bug

* Fix `track_case_status` func match bug

* Fix return code

* Fix print_info func with exit -6

* set output format of fail tests
modify verification check failed
* fix empty state_dict

* update sharding split_parma
* add submodule

* add submodule
* fix ci scripts

* [CI]add recursive for submodule
* fix ci scripts

* [CI]add recursive for submodule

* [CI]fix
* add ktotrainer

* fix oom
* update target_lists_for_llm

* add scripts/regression
* add the AutoModel for inference mode in dynamic graph

* add the AutoModel for inference mode in static graph

* create AutoInferenceModelForCausalLM class  and polish the code

* fix the return result

* add the confirm_inference_model method

* roback the AutoModel

* modify the description

* update
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.