AI Nudify TUTORIAL THREAD - /b/ (#936977344) [Archived: 396 hours ago]

AssetsWiz
7/12/2025, 2:02:39 AM No.936977344
Tutorial Header
Tutorial Header
md5: e0389fcf58453f3efcc805574cfd081b๐Ÿ”
THIS IS NOT A NUDIFY THREAD! DO NOT POST REQUESTS IN THIS THREAD!
This thread is a tutorial, from start to finish on how to install stable diffusion with ComfyUI and then use it for inpainting to make nudify images.
I will be explaining settings, processes, nodes, workflows, checkpoints and basically everything else you need to get started
Replies: >>936977407
AssetsWiz
7/12/2025, 2:04:24 AM No.936977407
Tutorial 1
Tutorial 1
md5: 315ed62c3ea37fb21e7108b8de442726๐Ÿ”
>>936977344 (OP)
So you want to get started with AI? Not a problem. Most PCs should be able to handle this, it just takes a bit to get set up, and that's what I'm here to explain.
A quick disclaimer, my screenshots are all from Linux but I'll be posting the instructions for both Linux and Windows so if it looks a little different don't get discouraged.
The very first thing you're going to need to do is open up a command terminal.
On windows hit the windows key and type "cmd" then select comand prompt from the results. On Linux you just click the icon
Replies: >>936977420
AssetsWiz
7/12/2025, 2:04:47 AM No.936977420
Tutorial 2
Tutorial 2
md5: 0c627c287f542704493aa74b97f4f822๐Ÿ”
>>936977407
Now we need to clone the repo to your machine.
If you have a specific folder you want to save all this to you'll have to navigate to that folder first.
Let's say you want to put everything in a folder called AI that you store in the root of your C drive. Then you just create an empty folder and use the command:
cd ai

I put mine in the root directory for now to make this tutorial easier.
Once you're in the folder where you want to put everything you need to download the repo to that folder. Use the command:
git clone https://github.com/comfyanonymous/ComfyUI.git
Replies: >>936977450 >>936981652
AssetsWiz
7/12/2025, 2:05:21 AM No.936977450
Tutorial 3
Tutorial 3
md5: 335e60a30d45f9fb4639c4423d29e03f๐Ÿ”
>>936977420
now you simply need to use the command:
cd comfyui

This will make your active directory the folder that everything is stored in. You're going to need to remember this location because you're going to need to go to this folder any time you want to start up your AI.
So if, as I mentioned earlier, you wanted to store it in a folder called AI, you would first need to run cd ai then cd comfyui, or you should be able to run cd ai/comfyui
Replies: >>936977461
AssetsWiz
7/12/2025, 2:05:42 AM No.936977461
>>936977450
OK, I'm going to skip some of the screenshots for the command terminal stuff.
Next you run:
python -m venv venv

This creates a virtual environment (a temporary computer that exists inside of your computer).
Now the commands get different from Linux and Windows. Windows Command:
venv\Scripts\activate.bat

Linux Command:
source venv/bin/activate

This actually starts up the virtual environment and is a command you'll need to run any time you want to start up your AI
Replies: >>936977475
AssetsWiz
7/12/2025, 2:05:59 AM No.936977475
>>936977461
Now we need to install some tools that the AI needs in order to run.
Windows or Linux command:
pip install torch torchvision torchaudio --extra-index-url https://download.pytorch.org/whl/cu121

After that finishes running you then need to run this command:
pip install -r requirements.txt

After those finish you're basically set up and ready to go, at least if all you wanted was to generate images based on text. Inpainting (the thing we do when nudifying) takes a little more setup, but don't worry, we're mostly done with the command terminal
Replies: >>936977497
AssetsWiz
7/12/2025, 2:06:27 AM No.936977497
Tutorial 4
Tutorial 4
md5: fd6e5f99114a0c6e79feb541dfdc5a34๐Ÿ”
>>936977475
Alright, this is the last screenshot of the command terminal stuff but it's an important one. Any time you want to start up your AI you're going to need to run a few commands.
You need to get to the directory comfyUI is stored at, you need to start the virtual environment, and you need to actually run the program within the virtual environment.

On Windows it'll look like:
cd comfyui
venv\Scripts\activate.bat
python main.py

On Linux it'll look like:
cd comfyui
source venv/bin/activate
python main.py

Once it finishes running just open your web browser and go to the address that it gave you in the command terminal (since we're running this locally instead of setting up a web server it should start with http://127.0.0.1:)
Replies: >>936977534
AssetsWiz
7/12/2025, 2:07:18 AM No.936977534
Tutorial 5
Tutorial 5
md5: c7a28d613f6c41f4c656abaa69b7947c๐Ÿ”
>>936977497
OK, so we're actually done with the technical part of the setup. Now we just need to make the workflow and download a model.
When you go to interface for your AI you're going to see a premade tutorial workflow already set up for you. If you want to test it out just enter a prompt and hit queue and it'll generate an image based on the prompt.
The default model isn't great, and this workflow doesn't do anything but generate an image based on a prompt, but if you want to make sure you've set everything up hit the button and try it out
Replies: >>936977559
AssetsWiz
7/12/2025, 2:08:01 AM No.936977559
Tutorial 6
Tutorial 6
md5: 921f15f2386f5139b2e78d785bfd619b๐Ÿ”
>>936977534
Before we get to the inpainting though, I'm going to show you *why* we go through all this trouble. Suppose you just want to load an image in, give it a prompt, and then let the AI do it's thing. Skipping over how to set this up (for now) this is a basic image to image workflow that will allow you to do just that: Load an image, apply a prompt, hit run and watch the magic. So let's try just that.
Replies: >>936977584
AssetsWiz
7/12/2025, 2:08:44 AM No.936977584
Tutorial 7
Tutorial 7
md5: 7c8beaa4e4c1308e0b614f5299867736๐Ÿ”
>>936977559
Oh... What the fuck happened? Well a lot of things were wrong with this approach, but in essence, the ai doesn't really know what it's looking at. So if you just say nude it will arbitrarily decide what is supposed to be nude on that picture and then make something similar. There's also a setting called denoise which controls how much the ai changes of the original picture. I set the denoise at .50 so that it would still be somewhat recognizable while making visible changes. If I had set it closer to 1.0 then it would have replaced almost the entire image. So If you want it to make a significant change like removing clothes, you need your denoise set higher, but that also warps the picture more. So we need extra instructions for the ai to know what to change and where before it can make anything even close to what we want. Also it took a solid 2 and a half minutes for my computer to generate it. Not bad for an old laptop, sure, but still frustrating to have to wait so long to see if it worked. We'll fix all of that though, just follow along.
Replies: >>936977603
AssetsWiz
7/12/2025, 2:09:10 AM No.936977603
Tutorial 8
Tutorial 8
md5: cd3a4eef7f2b8d6f48a5342b38728119๐Ÿ”
>>936977584
Now... instead of using the default I'm going to show you how to build a new workflow. (if you want to skip some of the fancier stuff I do then you can just get the template inpainting workflow but clicking "Workflows>Browse Templates", then you can skip the next couple of steps)

This is a basic starter inpainting workflow, I'll go over each group and what it does, but feel free to use this as a reference for setting yours up
Replies: >>936977638
AssetsWiz
7/12/2025, 2:09:48 AM No.936977638
Tutorial 9
Tutorial 9
md5: 0dc9bdabc9084ce7fc0ebde66a87a734๐Ÿ”
>>936977603
First up we have the Checkpoint/VAE group, this is essentially the start and finish of the processing your image. Since nothing really changes on these I group them off to the side on their own.
The Load Checkpoint node loads in the model that you'll be using and the VAE Decode node takes the processed static from the image and turns it into the finished image.
Replies: >>936977691
AssetsWiz
7/12/2025, 2:11:38 AM No.936977691
Tutorial 10
Tutorial 10
md5: 0183fcb808e194181b47f324bb2858d8๐Ÿ”
>>936977638
Next up we have the load image/mask group. This is essential for inpainting.
The Load Image node brings the image you want to work on into the ai.
The Feather Mask blurs the edges of the area you want to work on so that your edits blend into the finished image smoother.
The VAE encode breaks the image into static for the AI figure out how it's constructed in the first place. This specific version also tells it to grow the mask which basically means that it looks at stuff slightly outside of the selected area.

Don't worry, I'll explain more about image masks in a minute
Replies: >>936977702
AssetsWiz
7/12/2025, 2:12:00 AM No.936977702
Tutorial 11
Tutorial 11
md5: 31d8a67a12df76c8e0b0edbaaca6cde0๐Ÿ”
>>936977691
Now we have Prompts. Pretty self explanatory here. The positive prompts tell the ai what you want to add and the negative prompts tell the ai what you want to avoid.
For inpainting you almost never use negative prompts. I find some limited use out of it if I find that it's trying to draw a full body and I just want it to draw a torso, but more often than not I'll get a better result by just rewording the positive prompt.
The negative prompt is more useful when you're making a brand new image from text, but for the most part you don't want to touch it unless you're having trouble with a particular part or area (like if you've tried several times and it just keeps fucking up the hands so you put hands into the negative prompt and it should hide them behind something)
Replies: >>936977724
AssetsWiz
7/12/2025, 2:12:41 AM No.936977724
Tutorial 12
Tutorial 12
md5: d6ebd2d5047da6ea0b4c397e9abe473f๐Ÿ”
>>936977702
The last one I'm going to go over is the settings and post processing group.
I have an image blur node set to a low level in order to hide those edit lines but it's not 100% necessary. The main thing we're going to pay attention to is the Settings node (which is named KSampler by default but I renamed it for clarity) and I'm going to go over each of these with a quick explanation:

Seed: The random numbers that it uses to generate static. You don't need to worry about it for now
Control After Generate: Used to generate the next seed. You don't need to worry about it for now
Steps: the amount of the image that it's trying to generate at any one time. Higher numbers take longer, lower numbers look worse. I find 18-30 is a great range to stay in.
CFG: How close the AI is going to try and match the prompt you give it. Adjust this carefully, half a point at a time.
Sampler name: Again, a thing dealing with randomness and static. Don't touch it unless the model you download tells you to.
Scheduler: Similar to the last setting but this one can actually make a big difference. If the model you downloaded suggests a Scheduler then use that, but otherwise try them until you find one that looks right. Normal works in most situations.
Denoise: Like I mentioned earlier, the denoise is what's turning it from the static noise back into an image. A higher denoise setting replaces more of the original image, a lower denoise replaces less. For inpainting I generally set it between .85 and .95.
Replies: >>936977752
AssetsWiz
7/12/2025, 2:13:14 AM No.936977752
Tutorial 13
Tutorial 13
md5: f89831902af22700f756c7703a7dad4c๐Ÿ”
>>936977724
Alright, you also need to have an image save node but we're starting with the basic image save for now which is self explanatory so I didn't include a screenshot.
Only one more bit of setup before we get to the nudifying.
As I mentioned earlier the default model is kinda crap so I would suggest getting a new one off of CivitAI dot com.
My favorite for beginners is epiCRealism XL.
You'll need to sign up for an account but it's free so quit complaining.
Replies: >>936977768
AssetsWiz
7/12/2025, 2:13:37 AM No.936977768
Tutorial 14
Tutorial 14
md5: 99f51e838d19ce7d4d9085865ea2d2f9๐Ÿ”
>>936977752
After you've downloaded the model you'll see a file called
epicrealism_v10-inpainting.safetensors
Get that file and you're going to want to move it over to the checkpoints folder. You can just drag and drop it or copy it over.
To find the checkpoints folder you'll need to open up your ComfyUI folder in the file manager...
Replies: >>936977787
AssetsWiz
7/12/2025, 2:14:02 AM No.936977787
Tutorial 15
Tutorial 15
md5: 8ee61ac5824d01e8ab315c8715f0e2dc๐Ÿ”
>>936977768
Then open the models folder and you'll see the checkpoints folder. This is going to be the of the only folders we interact with today, but if you try out new models and new types you might also need to put something in one of the older folders, like Loras (foreshadowing)


Again, if you ever have questions about how to install a particular model, check and see if the creator posted any instructions. The more helpful ones will tell you how to set it up and what settings work best. Refresh your page on the ComfyUI interface and you can select the new model from the checkpoints node and we're ready to go.
Ready to start inpainting?
Replies: >>936977809
AssetsWiz
7/12/2025, 2:14:25 AM No.936977809
Tutorial 16
Tutorial 16
md5: 590ca2a59219a1140abcb89e84b6549c๐Ÿ”
>>936977787
Alright. So, after you've loaded your image in the load image node, right click on that preview image and select "Open in Mask Editor"
Replies: >>936977830
AssetsWiz
7/12/2025, 2:14:57 AM No.936977830
Tutorial 17
Tutorial 17
md5: b56eb7dd775e1f1e06f2f156d2dc522c๐Ÿ”
>>936977809
Now you want to "Mask" the areas you want to change, so assuming you want to remove a bikini you draw over that bikini and hit save.
That said, there's a few things that will making your final results a little better.
Draw the mask to include just outside the borders of what you want to change. I use a lower opacaity setting to better blend the edges of the mask along with a round brush and I set the mask to negative to make it easier to see what I've selected.
Replies: >>936977852
AssetsWiz
7/12/2025, 2:15:48 AM No.936977852
Tutorial 18
Tutorial 18
md5: d9e6b45499d9bd787bfe41b73f4904ae๐Ÿ”
>>936977830
After you draw and save your mask, enter your prompt and hit run to start the queue.
For demonstration I just entered the prompt as "nude" to show what happens when you try that.
Replies: >>936977915
AssetsWiz
7/12/2025, 2:17:23 AM No.936977915
Tutorial 19
Tutorial 19
md5: dec3a663df93aec7aa9a42f943bd08fb๐Ÿ”
>>936977852
Honestly, I've seen way worse, but we can do a lot better. For one thing while that blur can help hide the edges of the edit, it's a noticeable downgrade. For another, most AI needs to be told if it's anything other than a bog standard full body picture facing front straight to the camera. Some of that can be addressed with describing your subject's position/skintone/bodytype/etc with the prompt as well as the quality of the light. You'll eventually get some intuition as to what prompts work best for what kinds of images. Finally, while it didn't come up with this source image, this workflow is completely incapable of handling images larger than 1024x1024 which is frustrating as hell.
Replies: >>936977938
AssetsWiz
7/12/2025, 2:18:09 AM No.936977938
>>936977915
If you're OK with that level of quality then you're basically done, but there's a lot of things this basic workflow can't handle and you're going to run up against those limitations constantly.

That said, you do now know the basics of how to nudify.
All the generated attempts got saved in the ComfyUI/output folder so I can share it or move it to wherever I'd like.
You can go try it out yourself if you'd like or you can stick around to learn about how to get better, more efficient results.
Replies: >>936977956
AssetsWiz
7/12/2025, 2:18:39 AM No.936977956
Tutorial 20
Tutorial 20
md5: a84a180ad48ed820acc6f5eb5e1f9828๐Ÿ”
>>936977938
Now we're going to start talking about how to get better results overall and how to integrate specialized models called loras.
A checkpoint is the large model of images that the ai works off of to generate images. A lora is a specialized subset that can give better detail on specific areas. I like downloading new loras off of civitai and trying them out to see how well they work. Since they're generally smaller it's not uncommon for me to go looking for a new one if I get a request for something specific. The smallest checkpoint I have is 2.1gb, the largest Lora I have is under 700mb. Obviously when you download a Lora you'll save it to the Lora folder like you did the checkpoint. The only other thing to note is to make sure that the lora was designed to work with the base model of the checkpoint. If you're going to use the checkpoint I recommended you'll want to filter your searches to loras designed to work with SD 1.5.
This is an upgraded version of that earlier workflow that uses an advanced save function and a couple Lora nodes.
Replies: >>936977973
AssetsWiz
7/12/2025, 2:19:05 AM No.936977973
Tutorial 21
Tutorial 21
md5: 45ff0fdbb0360b979b9af701248eb8f7๐Ÿ”
>>936977956
So let's explain how these Lora settings work. The lora takes in the model and clip from the checkpoint and processes the model to refine it before passing along the clip to the prompt and the KSampler (I renamed that Settings earlier for simplicity). The setting for Strength Model is how much it should rely on the base model, and the setting for Strength Clip is how much it should rely on the Lora. These don't need to add up to 100% or anything. I usually keep a skin texture lora that helps to make skin look more natural and will change a second one between a handful of others depending on what I specifically want to see (pubic hair, nipples poking through fabric, etc).
Replies: >>936977981
AssetsWiz
7/12/2025, 2:19:24 AM No.936977981
Tutorial 22
Tutorial 22
md5: d91e0042a82540d8873f2f803b87ff3b๐Ÿ”
>>936977973
The other big change is to replace the basic Save Image with Save Image Extended.
There's a lot of settings on here and you can mouse over them to get a description of what each does, but I'm going to just point out the two that you should know right now. The first one is output_ext which allows you to select the file format you want your images to be output as. The second is quality which determines the final quality of the image you output. Why add this? Because /b/ has a ridiculously low restriction of 2mb for image uploads. Adding this in allows you to control how large the final output file is without having to compress the image or go through any other ridiculous measures.
Replies: >>936978000
AssetsWiz
7/12/2025, 2:19:46 AM No.936978000
Tutorial 23
Tutorial 23
md5: b13b4a3cf391a5d240f2d7dda3857320๐Ÿ”
>>936977981
One thing to note though is that this is a custom node (not made by me) that you'll need to install through the custom nodes manager. I honestly don't have the energy to walk you through how to do that since this is the only custom node we have and it's not a thing everyone needs. If you want to try and install it look for this button and see search for save image in the custom nodes manager.
Replies: >>936978034
AssetsWiz
7/12/2025, 2:20:37 AM No.936978034
Tutorial 24
Tutorial 24
md5: 7d145c5ea210afbe7cb6867b0381f1fa๐Ÿ”
>>936978000
Ok so that explained what Loras do and how to set them up but why? Does it really make that big of a difference? Well if you want something *really* specific, you might need a lora or 2 to get it to look right. This one used a combination of pregnancy and transparent clothes. You can decide for yourself if it's worth the effort.
Again, this is another good place to jump off if you're happy with your results. There's still a few limitations but we're getting much better results than before in my opinion and we're able to get some specialized looks in too. If you want to go have some fun and play around with it then enjoy. The next phase is how to get even cleaner results, faster, from larger images. But it takes a lot more setup, so be prepared for more technical stuff.
Replies: >>936978053
AssetsWiz
7/12/2025, 2:21:13 AM No.936978053
Tutorial 25
Tutorial 25
md5: 9d52125779293ef67e505d9295442015๐Ÿ”
>>936978034
This is the upgraded workflow I've been using recently. It includes an auto rescaler to handle larger images without distorting them. It also includes a latent composite mask to help the AI use more of the original image while still drawing over it. Essentially if you want better results that more closely match the posing, lighting, skintone, and proportions you're going to need something like this.
Replies: >>936978061
AssetsWiz
7/12/2025, 2:21:36 AM No.936978061
Tutorial 26
Tutorial 26
md5: 7fe832970a7abfcf3693565bcefaf74c๐Ÿ”
>>936978053
First I'm going to go over the Mask/Image Composite group since it's relatively easy to set up. All you need is a latent noise mask node (essentially an image filled with static that the AI will draw on and then composite together with the original image), and a Latent Composite Masked node which is what combines the noise, the mask, and the original image together before sending it off to the KSampler for processing. This will, in general, help the AI to understand the original image better by letting it look at the entire image, not just the masked part.
Replies: >>936978082
AssetsWiz
7/12/2025, 2:22:15 AM No.936978082
Tutorial 27
Tutorial 27
md5: 4070985e93649dae1ac1c394b36841f3๐Ÿ”
>>936978061
Now 'm going to go over my Auto-Resize group. This uses a few nodes from the ComfyMath extension so you might need to enable that in the custom nodes manager. When the image is first loaded in, I pass it to a Get Image Size node which reads the original resolution, NearestSDXLResolution which finds the closest resolution that will still fit in what the stable diffusion model is capable of creating while maintaining the original aspect ratio, and it also passes the image to ResizeAndPadImage along with the output from NearestSDXLResolution to take the original image and turn it into something the AI can actually handle. If you don't go through all that and try to edit a large image then it will get confused when it tries to go beyond those boundaries and will wind up repeating the prompt multiple times in multiple areas. If you've gotten results that wouldn't look out of place in a cronenberg film then that's probably why. Of course we don't want the image to stay at that small resolution which is where the upscaler comes in after the processor finishes but before it saves. I personally find that the bicubic setting works best in general, but feel free to try the others.
Replies: >>936978100
AssetsWiz
7/12/2025, 2:22:50 AM No.936978100
Tutorial 28
Tutorial 28
md5: 7438fd1f03f95095d46674605341d72a๐Ÿ”
>>936978082
This one is super important to get right because you'll not only get better results overall, but because it's not trying to inpaint over a larger resolution image it will also finish much faster. That image I loaded in at the beginning that took nearly 3 full minutes just to process? I ran it again on this workflow and got this back in 43 seconds. The tradeoff is that the more it has to upscale an image on the backend the more pixelation and warping you'll notice so if you try and process a 4k image or something you might have to give it a few tries to look right. Also, depending on the resizing and aspect ratio you'll notice a black border on some images. Still trying to fix that. However, for the purposes of an image board with a 2MB file size cap it works great.

That's normally the area where I stop but, as promised the last time I ran one of these threads, I'm also going to teach you how to build a workflow for quick and easy cumshopping/facial edits (it has more uses than that, but this is a lot of what I've been testing lately).
Replies: >>936978145
AssetsWiz
7/12/2025, 2:24:07 AM No.936978145
Tutorial 29
Tutorial 29
md5: 7a98e86a89c3092d330a8399c2fe7e3b๐Ÿ”
>>936978100
Ok, first thing you need is a good picture to work off of. This isn't an inpainting workflow, we're not worrying about masks or feathering or any of that, this is just load the image, adjust settings if needed, and hit run. A good picture for this specifically is at a decently sized resolution (at least 1024x1024) and shows the face clearly. Since this is for face edits generally speaking the more of the frame the face takes up, the better. I grabbed this one earlier off another thread for testing purposes (no I don't know or care who this is).
Replies: >>936978210
AssetsWiz
7/12/2025, 2:26:06 AM No.936978210
Tutorial 30
Tutorial 30
md5: 680da8df04e30a3a2ff19a589deb071b๐Ÿ”
>>936978145
Here's the new-ish workflow we're building (feel free to copy and paste from the inpainting workflow and make changes as needed). As you can see there's some new nodes and we've got several lora nodes as well. I'll go over the new stuff and why it's organized this way, but in this case I'm going to give you the specific Loras I'm using in order of most to least important.
Replies: >>936978236
AssetsWiz
7/12/2025, 2:26:51 AM No.936978236
Tutorial 31
Tutorial 31
md5: 55862356235bde6ca6ab852b94b6ea41๐Ÿ”
>>936978210
First up is Cum Facial 55. It has a built in depth mask which is *very* good at detecting the contours of faces. Can this sort of thing be done without this Lora? Sure, plenty of others do the same thing. Do I want to try out the hundreds of others that do similar things to find one to replace one I already like? Hell no.
Replies: >>936978250
AssetsWiz
7/12/2025, 2:27:14 AM No.936978250
Tutorial 32
Tutorial 32
md5: 1eb6876d38c85024af18b057c08299eb๐Ÿ”
>>936978236
Next we have MS Real Lite Bukkake. While this generally does increase the amount of cum on the face by a decent amount I find that I get much better results from a wider variety of skintones by integrating this which is the main reason I keep this one on.
Replies: >>936978287
AssetsWiz
7/12/2025, 2:28:05 AM No.936978287
Tutorial 33
Tutorial 33
md5: 26885427b297a612c965dadfe5516606๐Ÿ”
>>936978250
Next up is Running Makeup. This is where you'll see it go from a few cumshots to a proper drenching. One thing to note with this one is that, especially at higher clip strength, you'll notice the facial expression changing more.
Replies: >>936978310
Anonymous
7/12/2025, 2:28:25 AM No.936978298
Thank you so much AssetWiz, do you have a discord I can message you on for any trouble I run into while trying to make my own gens?
Replies: >>936978354
AssetsWiz
7/12/2025, 2:28:40 AM No.936978310
Tutorial 34
Tutorial 34
md5: 544069a5cc048e467328886a00dd3e32๐Ÿ”
>>936978287
Finally I have MS Real Lite Cum On Tongue. I've tried a number of different loras for cum in the mouth and so far this is the best one I've found. Will update this tutorial later if I've found better.
Also keep in mind that you can mix and match these however you like, I am by no means an authority on the available models for this sort of thing so if you find that you're not getting the specific results you like go ahead and take a look at what else is out there and experiment with it (I'll actually go over later how to better control your results when testing out new loras).
Replies: >>936978327
AssetsWiz
7/12/2025, 2:29:15 AM No.936978327
Tutorial 35
Tutorial 35
md5: 864fc7f7c5ad6b98ed6c144c2814c4a7๐Ÿ”
>>936978310
OK, so that's the loras we're working with. You can use the same epicrealism from the inpainting workflow and it works great for this The only other thing I want to quickly note before I start talking about the new nodes and how to set this up is about settings. I've already talked a little about them before but image to image without inpainting is a little tricky. Since we don't want to change the structure of the face set your denoise low. Very low. 0.1 or lower. Literally even turning it up by 0.05 is enough to warp the face of your target in some cases you'll also want to be careful with the clip strength of your loras. I usually keep them between 0.6 and 0.7. With that out of the way let's talk about the new groups and nodes. This is the starter Lora group which I just called Lora 1. It *is* important that you group together a lora and a basic string field for this, I'll explain why later. In the string field you'll want to put the trigger phrase for the lora itself. Most of them are activated by a specific prompt, generally listed on the download page on civitai. If you're using civitai to download check the Trigger Words section on the right. For Cumfacial 55 it's cumfacial55. Naturally. But they're often different based on how the creator constructed it in the first place, so always check the documentation.
Replies: >>936978373
AssetsWiz
7/12/2025, 2:29:50 AM No.936978354
>>936978298
I do, but I don't want to give it out right now since I tend to get a lot of requests through there. Post yours and I'll add you later
Replies: >>936978493
AssetsWiz
7/12/2025, 2:30:24 AM No.936978373
Tutorial 36
Tutorial 36
md5: c4d376bb0f8103d96598901c4ee445be๐Ÿ”
>>936978327
The second lora group (and 3rd and 4th) has a Concatenate node instead of the basic string node. This adds the string a and string b together and outputs them as a single string. So if you feed the output of the last groups string into the input of this groups concatenate node string a, then put the lora trigger into string b, it will output a single string containing the trigger words for each lora you've connected. You'll also need something to separate these so it doesn't just mash them into a single word, that's the delimiter. In that field you'll want to type a comma. Each group after the first is set up the same way with only the last one being connected slightly differently. In essence you want to take the output of Model and Clip from the Load Checkpoint node, feed it into the first group's Lora inputs, and then take those Model and Clip outputs and feed them into the next lora's inputs, until you get to the last which feeds the last clip into the clip text encode (positive and negative) and the last model into the ksampler. For the strings you take the first string, into the second, combine them, and so on until you eventually send the final output string into the Positive Text encode (don't send it to the negative or it will cancel itself out). I usually color code my text encoders so I don't mix up which is connected where.

I understand that's a lot of setup, but that's essentially it. There's the resizer group that we went over back in the inpaint and the save image extended, but again, every other node is something we've already gone through. I actually built this one by saving a copy of my inpainting workflow and removing a few nodes before adding in some extra lora groups. So we've got all this facial stuff set up, how well does it work. Let's load up that test image and give it a try.
Replies: >>936978447
AssetsWiz
7/12/2025, 2:32:25 AM No.936978447
Tutorial 37
Tutorial 37
md5: 8ca89e571e7b291dd660f41850ba77a6๐Ÿ”
>>936978373
Now that's a lot of cum. Some would say too much. Also it has her sticking her tongue out and it's a little awkward. But that's exactly why I had you assign the loras to groups to make the next part easier.
Replies: >>936978492
AssetsWiz
7/12/2025, 2:33:47 AM No.936978492
Tutorial 38
Tutorial 38
md5: 7872cd9cc644096ad1b34414f2692ae6๐Ÿ”
>>936978447
One of the neat features of ComfyUI that you can take advantage of is to Bypass nodes allowing you to maintain connections without having the underlying effect of the node taking place. So if, for instance. I want to scale back the amount of cum from a trial by drowning to a more realistic squirt on the face, I can do that by bypassing the lora nodes and string nodes that are making the overall effect more intense or adding things like a tongue that I might not want for this image. And because they're in groups I can simply right click on the group I want to temporarily bypass and select Bypass Group Nodes. If I want to turn it back on later I do the same thing but select Set Group Nodes to Always. This allows me to quickly turn on and off whole sections of prompts, loras, entire portions of my workflow if I want. I mostly use groups to keep my workflow better organized and for this quick bypassing feature but there's genuinely so much more you can do with this.
Replies: >>936978536
Anonymous
7/12/2025, 2:33:51 AM No.936978493
>>936978354
Thank you so much wizard. This looks very intimidating I wonโ€™t lie, but the will to goon may help me prevail in getting something setup. For someone with the average computer literacy, how long can I expect until I can get some results going?
>Discord
megacuckson
Replies: >>936978722
AssetsWiz
7/12/2025, 2:34:54 AM No.936978536
Tutorial 39
Tutorial 39
md5: dd725c4ac08c032b1d2ee38277ec00d5๐Ÿ”
>>936978492
Now to show you how dramatic of a change you can get by stacking those loras I temporarily set my Ksampler's "Control after Generate" setting to "Fixed" so I could use the exact same random seed to generate a second image but with a different set of loras/prompts. This is how you can test out how a lora works by using "fixed" to test it side by side in an on/off fashion or by using "increment" which allows you to generate several images with more subtle changes in the exact same style. Using the exact same seed as before I bypassed all but the cumfacial 55 group. Let's see the difference
Replies: >>936978581
AssetsWiz
7/12/2025, 2:35:50 AM No.936978581
Tutorial 40
Tutorial 40
md5: 97e237636b5589d834e70c6bbbd2b74d๐Ÿ”
>>936978536
Much more understated. Honestly I might have wanted to bump up the weight just a bit just to make it more visible but you can see how it's a massive difference while still having a noticeable effect on the original image.
And as you might have guessed, while this last part has been focused on cumshopping the central concept of [base image > several loras > finished image] allows for a ton of flexibility depending on the loras you have.

Now there's still areas to improve on all this. For one thing, I need to adjust the auto resizer and to fix those issues. For another I can't do outpainting (adding new areas to an image). And naturally I want to go ahead and start making my own loras. But that's for another day.
AssetsWiz
7/12/2025, 2:39:25 AM No.936978722
>>936978493
If you follow the whole thing step by step, use the screenshots for references on how to connect things, and use google (or hell even chatgpt or gemini) to help you get past any error messages that might crop up in case I missed something, you should be able to get this up and running within an hour or two.
I get that I'm a software developer and so using command prompts and concepts like concatenating strings are kinda second nature for me, but I did my best to explain this clearly so that you should at least be able to follow along even if you're bad with computer shit
Replies: >>936978827
AssetsWiz
7/12/2025, 2:40:42 AM No.936978775
Normally I would stick around for more questions but I gotta go. My wife just got home
Replies: >>936978899
Anonymous
7/12/2025, 2:41:59 AM No.936978827
>>936978722
Iโ€™ll let you know how it goes for me when you add me back and I get an afternoon to set aside time to set everything up. Iโ€™m mainly interested in the cum/bukkake gens, the tutorial you provided produced a great gen from the original pic. Itโ€™s exactly the kind of stuff Iโ€™m wanting to make
Anonymous
7/12/2025, 2:43:43 AM No.936978899
>>936978775
Donโ€™t forget to add me please
Anonymous
7/12/2025, 3:53:48 AM No.936981652
>>936977420
bump?
Anonymous
7/12/2025, 4:39:17 AM No.936983391
I guess it's different with automatic1111? Or it's mostly the same, but without the node setting?