-
Notifications
You must be signed in to change notification settings - Fork 2.9k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[AutoParallel] Add benchmark for llama-7b-dy2st. #8559
Conversation
Thanks for your contribution! |
e173eb7
to
423cd8e
Compare
Codecov ReportAll modified and coverable lines are covered by tests ✅
Additional details and impacted files@@ Coverage Diff @@
## develop #8559 +/- ##
===========================================
- Coverage 53.86% 53.86% -0.01%
===========================================
Files 620 620
Lines 97081 97110 +29
===========================================
+ Hits 52296 52304 +8
- Misses 44785 44806 +21 ☔ View full report in Codecov by Sentry. |
"per_device_eval_batch_size": 2, | ||
"tensor_parallel_degree": 1, | ||
"pipeline_parallel_degree": 1, | ||
"sharding": "stage1", |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
动态图stage1
跟静态图stage2
一致。
"tensor_parallel_degree": 1, | ||
"pipeline_parallel_degree": 1, | ||
"sharding": "stage1", | ||
"sharding_parallel_config": "enable_stage1_overlap", |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
"enable_stage1_overlap" --> "enable_stage2_overlap"
"pipeline_parallel_degree": 1, | ||
"sharding": "stage1", | ||
"sharding_parallel_config": "enable_stage1_overlap", | ||
"tensor_parallel_config": "enable_delay_scale_loss enable_mp_async_allreduce enable_mp_skip_c_identity enable_mp_fused_linear_param_grad_add", |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
"tensor_parallel_config": "enable_delay_scale_loss enable_mp_async_allreduce enable_mp_skip_c_identity enable_mp_fused_linear_param_grad_add",
-->
"tensor_parallel_config": "enable_mp_async_allreduce",
"sharding": "stage1", | ||
"sharding_parallel_config": "enable_stage1_overlap", | ||
"tensor_parallel_config": "enable_delay_scale_loss enable_mp_async_allreduce enable_mp_skip_c_identity enable_mp_fused_linear_param_grad_add", | ||
"pipeline_parallel_config": "enable_delay_scale_loss enable_release_grads disable_partial_send_recv", |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
没有开PP并行,"pipeline_parallel_config": ""设置为空即可
"weight_decay": 0.01, | ||
"bf16": true, | ||
"fp16_opt_level": "O2", | ||
"amp_master_grad": true, |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
增加:
"amp_custom_black_list": ["reduce_sum", "c_softmax_with_cross_entropy"],
"amp_custom_white_list": ["lookup_table", "lookup_table_v2"],
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
PR types
Others
PR changes
Others
Description
Add llama2-7b for test_tipc