Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

fix for token_type_ids #149

Open
wants to merge 3 commits into
base: release/v1.18
Choose a base branch
from

Conversation

quic-mamta
Copy link
Contributor

@quic-mamta quic-mamta commented Oct 10, 2024

Fix token type ids in pytorch inputs for falcon 180B

Signed-off-by: Mamta Singh <[email protected]>
@@ -53,6 +53,7 @@ def prepare_pytorch_inputs(self):
input_ids = inputs["input_ids"]
batch_size, input_len = input_ids.shape
inputs.pop("attention_mask")
inputs.pop("token_type_ids", None)
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

don't we need to do this in prepare_ort_inputs and also in text_generation script?
Did you test the model that has this issue with infer CLI API?

Copy link
Contributor Author

@quic-mamta quic-mamta Oct 10, 2024

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I had tested with export command provided in the jira but 180B model export started but failed due to RAM issue. added the changes for ort and aic runs as well. and verified infer command for single layer model.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants