Some interesting things to try #802
Replies: 9 comments 17 replies
-
@lllyasviel Using similar technique for executing first few steps with SD-based models can also yield interesting results. There are some interesting finetuned SD models available that do pretty well at resolutions close to 512x512, and they can be used to provide initial concept of the image (let's say first 10-30% of the steps), that then can be used as input for further SDXL (or SDXL base + SD refiner) processing - in general it would work like built-in img2img stage with high denoise value. This way you can for example use output from SD like this: and turn it into this (using original SDXL base and refiner): I was playing with hybrid workflows like that when no ControlNets for SDXL were available - I was applying CN 1.1 to SD 1.5 at initial stage, and then processed it further through img2img with SDXL models. But SDXL base model (or any finetuned variant of it) should probably stay in the flow, as it seems to be best choice for providing middle steps for resolutions close to 1024x1024. So let's say roughly start for SDXL base could be at 20%, and switch at around 70% of the steps. Optimal switch value might of course vary (depending on used sampler, scheduler, seed, prompt, and models) - your example here is good illustration of that :). |
Beta Was this translation helpful? Give feedback.
-
Another interesting thing that might be worth trying is SSD-1B model (GitHub, Hugging Face). It's a distilled (about 4.5 GB) variant of SDXL with 1) some layers removed and 2) some additional training provided. According to authors it allows achieving results close to SDXL with about 60% speedup. Support for it was added in ComfyUI yesterday. My initial observations:
It might be be useful alternative for people with weaker GPUs (provided for example via another preset in Fooocus). |
Beta Was this translation helpful? Give feedback.
-
This is really quite a cool method to generate images! I definitely increases the realism! I really like this workflow method! |
Beta Was this translation helpful? Give feedback.
-
@lllyasviel can you drop the preset json file that you are using for the above image generation, as i am have been trying to do the same but i am getting this error ValueError: Error while deserializing header: HeaderTooLarge
File corrupted: /content/Fooocus/models/checkpoints/realisticVisionV51_v51VAE.safetensors I have added my preset file below. When i am making the refiner my custom preset json |
Beta Was this translation helpful? Give feedback.
-
Has anybody gotten Realistic Vision v51 to work with FaceSwap imagePrompt? I admit, the quality and realism of the output is impressive - however, it's not playing nice, nor taking the FaceSwap input. Thoughts? |
Beta Was this translation helpful? Give feedback.
-
Looks like the refiner totally ignores the prompt, I've had men turn into women, Tyra Banks turn asian, animals turn into people, etc. If you want a good laugh try using "Ape close up" as prompt. |
Beta Was this translation helpful? Give feedback.
-
I have troubles using faceswap when using this method to refine an image. It basically jumps in, and deletes the face, and swaps it alltogether. Any tips ? |
Beta Was this translation helpful? Give feedback.
-
Hello, sorry for stupid questions. |
Beta Was this translation helpful? Give feedback.
-
As you know, fooocus uses a very special method for SD1.5 as refiner swap. Since recent versions, the "switch at" can be set in advanced.
Playing with this is lots of fun. We are still experimenting this and will document it in a more formal way after it is ready.
I think I should put an early example
Below is "woman, street", seed 1234, run.bat default parameters (default sdxl + refiner)
then you download Realistic Vision V5.1 (it is a SD1.5) https://civitai.com/models/4201/realistic-vision-v51
Then you set it as refiner
Then set refiner switch to 0.4
Then you get this
just omg - what the h
I also tried dreamsharperv8 but Realistic Vision looks better for realistic images
Note that we are still experimenting this and the swapping algorithm may be changed in the future. (maybe the latent image saturation will be reviewed again, probably)
This swapping algorithm is only available in Fooocus.
Beta Was this translation helpful? Give feedback.
All reactions