Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

tag_encoder and text_decoder #191

Open
Stephen-K1 opened this issue Jul 1, 2024 · 1 comment
Open

tag_encoder and text_decoder #191

Stephen-K1 opened this issue Jul 1, 2024 · 1 comment

Comments

@Stephen-K1
Copy link

Hi, thanks for open sourcing your great work !

When reading the codes, I'm confused by the next-token prediction in calculating the loss_t2t, and I don't understand why the first four (prompt_length) labels are ignored (set as -100) during training. So I start to read the inference code hoping to figure this out. However, I found that both inference_ram.py and inference_ram_openset.py did not use the tag_encoder and text_decoder during inference, which makes me more confused. So I want to kindly ask that:

  1. Can you explain the next-token prediction in calculating the loss_t2t and why some labels are set as -100?
  2. Why tag_encoder and text_decoder are not used in the inference?

Thanks in advance !

@Stephen-K1
Copy link
Author

Alright, I think I can figure this out by reading inference_tag2text.py. Thanks for sharing the codes anyway.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant