-
Notifications
You must be signed in to change notification settings - Fork 2.9k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Bug Fix] fix paddle multipy_fwd_func warning message #7818
[Bug Fix] fix paddle multipy_fwd_func warning message #7818
Conversation
Thanks for your contribution! |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
Codecov ReportAttention:
Additional details and impacted files@@ Coverage Diff @@
## develop #7818 +/- ##
============================================
+ Coverage 0 57.11% +57.11%
============================================
Files 0 587 +587
Lines 0 88196 +88196
============================================
+ Hits 0 50377 +50377
- Misses 0 37819 +37819 ☔ View full report in Codecov by Sentry. |
…ain/eval to align.
…ain/eval to align.
…ain/eval to align.
…ferenceModel (#7953) * Add Pipeline Parallel for PPO training. * Move new_ppo_trainer.py to ppo_trainer.py * Fix padding among batches of accumulation steps in _prepare_pipeline_inputs_func. * Fix hcg using in TP generation * Try to support generation in PP. And allow extra training args passed from main from_pratrined. * Support PP generation. * Fix PP eval by unify prediction_step * Fix reward value showing error cased by BF16 dtype when eval * fix all * Make non-PipelineParallel models use the same loss layer with PipeModel to unify. * add offload. * Use create_loss to unify Pipe and non-Pipe usage. * Add eval mode and offload level. * merge * support tp+pp * fix data split. * Fix position_ids in generation/eval/train. * fix data group. * add tp rank guard * Support rollout label data both with target length or source+target length. * Move metric calculation to rl_step to avoid comm. * fix pad * fix create group. * no print * Suppport inference model generation. * fix compatible for no eval model. * fix pp sync. * remove debug info * Refacor PPO training using StepTrainer. * Open PolicyTrainer loss logging postprocess. More StepTrainer docs. * more timer. * fix bugs. * Add EMA and PPOMetric * add tests * add unit test for rank guard. * Fix reshard zero3 and reshard infer. * Revert #7818 for llama and remove position_ids for gen/train/eval to align. * Move reload/clean/data_group to comm_utils and use guard to decorate them. * Offload sync and other data reuse fix. * Clead code * Update README * Update ppo_trainer * format code * Fix make_position_ids by 4d causal mask. * Fix nested_broadcast_tensor_with_empty import * Update eval with make_attention_mask --------- Co-authored-by: Zhong Hui <[email protected]> Co-authored-by: gongenlei <[email protected]>
PR types
Bug fixes
PR changes
Others
Description
由于主框架在 PaddlePaddle/Paddle#59518 中对于multiply输入类型不一致会输出warning日志。
在amp level=O2情况下:
hidden_states = paddle.rsqrt(variance + self.variance_epsilon) * hidden_states
中paddle.rsqrt(variance + self.variance_epsilon)
结果为float32,hidden_states
为fp16/bf16。报错日志
复现脚本: