You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
With more and more and larger and larger diffusion transformer models coming up, I think it makes sense to support the direct loading of quantized models from quanto.
Currently, the workflow to load a quantized model with quanto is simple:
This issue has been automatically marked as stale because it has not had recent activity. If you think this still needs to be addressed please comment on this thread.
Please note that issues that do not follow the contributing guidelines are likely to be ignored.
With more and more and larger and larger diffusion transformer models coming up, I think it makes sense to support the direct loading of quantized models from quanto.
Currently, the workflow to load a quantized model with quanto is simple:
We should be able to just do:
Currently,
transformer=transformer
would fail because there's no mapping here:diffusers/src/diffusers/pipelines/pipeline_loading_utils.py
Line 67 in b1f43d7
Thoughts @yiyixuxu @DN6?
The text was updated successfully, but these errors were encountered: