-
Notifications
You must be signed in to change notification settings - Fork 26.9k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add support for the unclip (Variations) models, unclip-h and unclip-l #8958
Add support for the unclip (Variations) models, unclip-h and unclip-l #8958
Conversation
Getting this error when xformers is not installed (using SDP attention):
|
Working now without xformers, thank you MrCheeze. |
Confirming its working also on my trained model on the SD noclip diffusers And here is the original result from SD |
Add support for the unclip (Variations) models, unclip-h and unclip-l
…ariations-model"" This reverts commit 5ef2f58.
…Cheeze/variations-model""" This reverts commit 242d8f1.
…ariations-model"" This reverts commit 5ef2f58.
I was getting the same issue, until I removed the --medvram flag and now it works. When I checked out the source code, there was a place where the model is supposed to call an embedder method. However, an embedder method is not being assigned to the sd_model object when the model is allowed to use only less VRAM. https://github.com/AUTOMATIC1111/stable-diffusion-webui/wiki/Optimizations |
Describe what this pull request is trying to achieve.
This adds support for stable-diffusion-2-1-unclip checkpoints that are used for generating image variations. (See also.)
It works in the same way as the current support for the SD2.0 depth model, in that you run it from the img2img tab, it extracts information from the input image (in this case, CLIP or OpenCLIP embeddings), and feeds those into the model in addition to the text prompt. Normally you would do this with denoising strength set to 1.0, since you don't actually want the normal img2img behaviour to have any influence on the generated image.
One thing I did not implement is any way to use this functionality but starting from random noise like txt2img does - which would probably generate more varied variations. This would be good for future work.
Additional notes and description of your changes
Key changes:
checkpoints/karlo_models/ViT-L-14_stats.th
. I opted to hotpatch the config to point atmodels/karlo/ViT-L-14_stats.th
instead; and to check the file into the repo itself, since it's very small - thoughts on this approach?Environment this was tested in
Windows, NVIDIA GTX 1660 6GB
Screenshots or videos of your changes