Skip to content

Parseq generation stops in the middle when using TensorRT #231

Answered by rewbs
Zafflower asked this question in Q&A
Discussion options

You must be logged in to vote

Hi! The problem is your tensorrt compiled model had a maximum prompt token length. By default it'll be 75 or 150. At some point during the generation you hit a frame where the full prompt exceeds that max. So you can work around this by shortening your prompts or not having composable diffusion overlaps.

The best I could do I think is add functionality that shows you the maximum token length across all frames, so you can confirm it's below your TRT max before you generate.

Replies: 1 comment 1 reply

Comment options

You must be logged in to vote
1 reply
@Zafflower
Comment options

Answer selected by Zafflower
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants