-
Beta Was this translation helpful? Give feedback.
Answered by
rewbs
Nov 17, 2023
Replies: 1 comment 1 reply
-
Hi! The problem is your tensorrt compiled model had a maximum prompt token length. By default it'll be 75 or 150. At some point during the generation you hit a frame where the full prompt exceeds that max. So you can work around this by shortening your prompts or not having composable diffusion overlaps. The best I could do I think is add functionality that shows you the maximum token length across all frames, so you can confirm it's below your TRT max before you generate. |
Beta Was this translation helpful? Give feedback.
1 reply
Answer selected by
Zafflower
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Hi! The problem is your tensorrt compiled model had a maximum prompt token length. By default it'll be 75 or 150. At some point during the generation you hit a frame where the full prompt exceeds that max. So you can work around this by shortening your prompts or not having composable diffusion overlaps.
The best I could do I think is add functionality that shows you the maximum token length across all frames, so you can confirm it's below your TRT max before you generate.