People like my img2img workflow so it wasn’t much work to adapt it to just be a headswap workflow for different uses and applications compared to full character transfer.

Its very simple and very easy to use.

Only 3 variables need changing for different effects.

  • Denoise up or down

  • CFG higher creates more punch and follows the source image more closely in many cases

  • And of course LORA strength up or down depending on how your lora is trained

Once again, models are inside the workflow in a text box.

Here is the workflow: https://pastebin.com/z2nbb1ev

Extra Tip: You can run the output back through again for an extra boost if needed.

EG: Run 1 time, take output, put into the source image, run again

ty

EDIT:

I haven’t tried it yet, but i’ve just realised you can probably add an extra mask in the segment section and prompt ‘body’ and then you can do a full person transfer without changing anything else about the rest of the image or setting.


💬 Discussion r/StableDiffusion (193 points, 31 commentaires) 🔗 Source