-
Notifications
You must be signed in to change notification settings - Fork 2.9k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
rename files and add readme for llama auto_parallel #7944
Conversation
Codecov ReportAttention:
Additional details and impacted files@@ Coverage Diff @@
## develop #7944 +/- ##
========================================
Coverage 56.58% 56.58%
========================================
Files 589 589
Lines 89900 89900
========================================
+ Hits 50870 50872 +2
+ Misses 39030 39028 -2 ☔ View full report in Codecov by Sentry. |
hidden_states=outputs.hidden_states, | ||
attentions=outputs.attentions, | ||
) | ||
# loss = None |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
注释的直接删除?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
后面提PR完善代码
layer.down_proj.weight.scale_(factor) | ||
if isinstance(layer, LlamaAttentionAuto): | ||
factor = 1 / math.sqrt(2 * self.config.num_hidden_layers) | ||
layer.o_proj.weight.scale_(factor) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
删除?
# masked_lm_loss = self.loss_func(prediction_scores.astype("float32"), masked_lm_labels.unsqueeze(2)) | ||
# # skip ignore_index which loss == 0 | ||
# masked_lm_loss = masked_lm_loss[masked_lm_loss > 0].astype("float32") | ||
# loss = paddle.mean(masked_lm_loss) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
删除?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
PR types
Others
PR changes
Docs
Description
rename files and add readme for llama auto_parallel