SDXL ComfyUI Stability Workflow - What I use internally at Stability for my AI Art

Sdílet
Vložit
  • čas přidán 27. 07. 2023
  • Since we have released stable diffusion SDXL to the world, I might as well show you how to get the most from the models as this is the same workflow I use on a daily basis at stability.ai. In this video I show you some of the basics on how to get the model from the models to generate your best AI artwork from our models. You will need some of the custom nodes over at civit, but you can choose the package that works best for you, as they are all pretty similar.
    We will start with a basic workflow and then complicate it with a refinement pass, but then we will add in another special twist I am sure you will enjoy. #stablediffusion #sdxl #comfyui
    Grab some of the custom nodes from civit.ai: civitai.com/tag/comfyui
    Grab the SDXL model from here (OFFICIAL): (bonus LoRA also here)
    huggingface.co/stabilityai/st...
    The refiner is also available here (OFFICIAL):
    huggingface.co/stabilityai/st...
    Additional VAE (only needed if you plan to not use the built-in version)
    huggingface.co/stabilityai/sd...
  • Krátké a kreslené filmy

Komentáře • 298

  • @dxnxz53
    @dxnxz53 Před 18 dny +1

    it blew my mind that you can load an entire workflow from the image! thanks for the great content.

  • @Pfaeff
    @Pfaeff Před 10 měsíci +25

    Why are there width and height values in the CLIPTextEncoderSDXL and what is the difference between width and target_width and why is one of them 4096?

    • @courtneyb6154
      @courtneyb6154 Před 9 měsíci +4

      Great questions and hopefully Scott can take the time to explain. Building out the workload is a great first step, but not knowing what everything does so that you can fine tune it is lame.

  • @TedWillingham
    @TedWillingham Před 10 měsíci +33

    I would love if you could go over some of those settings in advanced detail - like "oh, I fiddle with more conditioning steps when I want to X", etc. There are so many superstitious people out there giving bunk advice that your level-headed breakdown would be super valuable!

    • @sedetweiler
      @sedetweiler  Před 10 měsíci +15

      Great idea! I will have to ponder where to start! :-)

  • @me.shackvfx5911
    @me.shackvfx5911 Před 10 měsíci +32

    I've grown to understand and enjoy comfy UI more that the one i was using before thanks to your videos.I really appreciate you and the effort you put into making these tutorials. One of these days you can show us how to train sdxl 1 or it's lora with our faces . Thanks :)

    • @sedetweiler
      @sedetweiler  Před 10 měsíci +8

      Great to hear! Training will be coming soon! Cheers!

  • @novantha1
    @novantha1 Před 10 měsíci +3

    Huh. I wonder what would happen if you had dedicated models for a variety of tasks (hands, eyes, hair, reflections, contrast, and so on) and fed a few steps from each of them in a daisy chain until you got to the first "true" sampler...
    Truly the possibilities are endless; thanks for the food for thought and the hard work!

    • @sedetweiler
      @sedetweiler  Před 10 měsíci +1

      That's a great idea, and we do have those as loras. It's fun to combine them to help get what you want.

    • @tripleheadedmonkey6613
      @tripleheadedmonkey6613 Před 10 měsíci +1

      That is an interesting idea. The multitude of experts approach is proving to be the more effective of what we have developed recently.
      Not too mention that you could also combine this with prompt blending syntax to ensure that each part of the processing is focusing entirely on one subject in the prompt while still maintaining an overall mixed composition.
      If for simplification purposes you set up 5 samplers, each with an equal number of steps, 4 for the limbs and 1 for the head/torso. Then you set up a prompt blending which focuses 20% of the processing on each limb etc. it may even have better results.

    • @tripleheadedmonkey6613
      @tripleheadedmonkey6613 Před 10 měsíci

      And yeah using LORA chains would mean that we could have a separate model output for each limb, while maintaining the same initial model. Allowing for less resources used at the same time compared to multiple dedicated models.

    • @tripleheadedmonkey6613
      @tripleheadedmonkey6613 Před 10 měsíci

      I think I'm going to play around with this now actually xD Minus the dedicated limb lora of course.

  • @iiiCorrosiveiii
    @iiiCorrosiveiii Před 10 měsíci

    thank you so much, i've become really proficient with A1111 and moving to comfyui was a big switch, so your help with how the workflows work in comfyui has made it just as easy as using A111 for me.

  • @lakislambrianides7619
    @lakislambrianides7619 Před 10 měsíci +1

    This is a great video congrats. Very informative very thorough and you left no doubts. Can't wait for the next step!

  • @14MTH3M00N
    @14MTH3M00N Před 5 měsíci +2

    Love your disgust for the negative prompts lists haha. relatable stuff

    • @sedetweiler
      @sedetweiler  Před 5 měsíci +3

      (((((((((extra arms!))))))))) :-)

  • @henryabela2672
    @henryabela2672 Před 10 měsíci

    Thank You so much! videos like these are a blessing and help people to get into it more professionally.

  • @larryross9380
    @larryross9380 Před měsícem

    Perhaps things have changed since this was published nine months ago, because this workflow just gave me dark, abstract images. But I learned a lot about how to build out a workflow! Thanks!5

  • @angryDAnerd
    @angryDAnerd Před 10 měsíci +4

    Excellent tutorial, thanks! I got SDXL up and running with the refiner. If you have the time I'd like to see you make a video explaining how Stable Diffusion works and explain exactly what the program is doing as it sends the data through the nodes in Comfy so I can have a greater conceptual understanding of what is happening. Believe me I could watch hours of technical stuff lol.

  • @wschieber
    @wschieber Před 10 měsíci

    Thanks, Scott. I was really looking for something like this to get started with SDXL in ComfyUI.

  • @rsunghun
    @rsunghun Před 10 měsíci +2

    I was waiting for it. These are very difficult for ordinary people to figure out how to use it. Thank you for the video!

  • @shallowandpedantic2320
    @shallowandpedantic2320 Před 10 měsíci +1

    Thanks. If you're looking for recommendations, a video focused on comparing upscalers and incorporating upscaling into this kind of workflow might help people. Seems like a nice next step. Appreciate what you've shared so far.

  • @Ziixu
    @Ziixu Před 9 měsíci

    Great tutorial!
    This is my first time using ComfyUI and this video helped me a lot, tyvm!

  • @imperfectmammal2566
    @imperfectmammal2566 Před 10 měsíci +1

    Thank you so much! Even though I couldn’t understand much, it helped me get started with comfy.

    • @sedetweiler
      @sedetweiler  Před 10 měsíci

      You’re welcome 😊 Just keep working with it and it will start to click into place.

  • @archielundy3131
    @archielundy3131 Před 4 měsíci +1

    A million thanks for these. As finicky and frustrating as the program is for beginners, your calm expertise is just what's needed.

  • @gbresaleking
    @gbresaleking Před 10 měsíci

    Thanks for this ! I was using comfy and using refiner at the beginning and it was coming up with really wacky pictures, now I can use it correctly thx

  • @Feelix420
    @Feelix420 Před 10 měsíci

    I'm mind blown! never thought of using comfyui ever but seems like I'm sold over this video. very nice sir and thank you for sharing your knowledge

    • @sedetweiler
      @sedetweiler  Před 10 měsíci +1

      Glad you liked it! It will also teach you a lot more about how things work, which I always feel is a good idea.

    • @Feelix420
      @Feelix420 Před 10 měsíci

      @@sedetweiler absolutely! downloading it now hehe

  •  Před 5 měsíci

    Thanks Scott, as a beginner your videos are great. Very well explained and "easy to learn than a bunsh of others.

  • @lionhearto6238
    @lionhearto6238 Před 10 měsíci

    all the info i was looking for. great video. thank you

  • @hleet
    @hleet Před 10 měsíci +1

    WOW ! that's a super tutorial of ComfyUI there ! Thanks. I never know that there was this new addition of clipnode for SDXL !
    The only drawback that I find in ComfyUI is the way it manage the workflows. I mean when you want to change your original workflow, you need to save a local file, and if you want to do something else (like inpainting) you have to redo ALL your workflow and save it to a file to recall your workflow and switch by loading one workflow or another depending on what you want to do. Definitly not fond of this way of managing workflows. They could have done some kind of "favorite" workflow. Like 5 or more "workflow ready" that you could custom afterwards and save your "favorite custom workflow" and switch whenever you like. it would skyrocket the use and adoption of comfyui !

    • @sedetweiler
      @sedetweiler  Před 10 měsíci +1

      I just drop the json you get from using "save" into the interface and it loads. But,*do agree that would be nice.

    • @hleet
      @hleet Před 10 měsíci

      @@sedetweiler ooh ! Nice another tip ! Drag and drop the json just works too ! I might be able to explore more versatile stuff with comfyui now :)

  • @russellmm
    @russellmm Před 10 měsíci

    quickly becoming my goto channel. keep up the great work

  • @ImAlecPonce
    @ImAlecPonce Před 10 měsíci

    Thanks!!! these boxes are actually starting to make sense

  • @PieterLaroy
    @PieterLaroy Před 10 měsíci

    My first steps into ComfyUI, and it's the kind of thing I really like 🙂

  • @DarnSylon
    @DarnSylon Před 9 měsíci

    I kept to the two sampler idea but introduced separate prompts for the first sampler. It creates an interesting variation. Especially if you put different colors in each positive prompt. For example, make a white bridge and a black animal. I was hoping to create a background and a foreground. But the twist turned out pretty cool anyway.

  • @eliasdelreal3863
    @eliasdelreal3863 Před 10 měsíci

    You are da man! Thank You so much for this tutorial!!!

  • @benkilowhiskey
    @benkilowhiskey Před 14 dny

    Great watch, thank you!

  • @Yggdrasil777
    @Yggdrasil777 Před 10 měsíci +2

    I have been in love with ComfyUI since I found it (coming from Unreal Blueprints, very familiar system). I am currently working out some torch issues with my current system, but I generate whenever I can. It is great to see you building out the workflow and explaining the nodes that you use and why. Very informative and THANKS for the tip with the shift-click to copy nodes AND connections. NICE!

    • @sedetweiler
      @sedetweiler  Před 10 měsíci

      Great to hear! I am really happy with the nodes, but I hope they really update to things like docking, etc. Cheers!

    • @digitalbear3831
      @digitalbear3831 Před 10 měsíci +1

      Same here since I come from Houdini, just love the node spagetti

    • @sedetweiler
      @sedetweiler  Před 10 měsíci +1

      Yusss! I also used Houdini as well as Substance Designer and I am hoping to get into nested nodes here as well. Cheers!

  • @paullangdon1852
    @paullangdon1852 Před 10 měsíci

    Great video, very usefull. I am struggling to do a workflow from load image and the put through an ultimate upscale node.

  • @zerosequels
    @zerosequels Před 8 měsíci

    Thanks Scott, your tutorials are great

  • @CMak3r
    @CMak3r Před 10 měsíci +7

    Prompt switching can be realized with additional KSampler that will render first steps with completely different prompt. For example you may want to create triangle composition, or a symmetrical image, and it can be done at early steps of a generation. Good for abstract art. And also I like that in ComfyUI it's seed can be fixed while base model and refiner will be generating on different seeds

    • @zacharykrevitt7560
      @zacharykrevitt7560 Před 7 měsíci

      good idea! just tried this out and it worked in an ineresteing way. Essentially prompting an init image

  • @JRis44
    @JRis44 Před 9 měsíci +1

    This is getting fun! Cant wait to work on img to img tomorrow after work!

  • @ColbstaD
    @ColbstaD Před 6 měsíci

    Thank you for this! I've created my own custom workflow based on this one with lots of inputs --> primitives to change stuff quickly.

  • @conquestofninjacats
    @conquestofninjacats Před 10 měsíci +1

    Love your videos :D just started using ComfyUI with SDXL. Having a lot of fun so far!

  • @MonzonMedia
    @MonzonMedia Před 10 měsíci

    Appreciate this Scott, you helped me fill in the blanks! I was wondering how the primitive nodes were used, it was driving me nuts! Hahaha! I was able to add an extra step to add an upscale process and it works very well! Looking forward to more.

    • @MonzonMedia
      @MonzonMedia Před 10 měsíci +1

      By the way, is there a way to create an image gallery somehow? Sort of like how invoke Ai is set up?

    • @sedetweiler
      @sedetweiler  Před 10 měsíci +1

      I just use Bridge, since I already have an Adobe sub and it's better than most gallery apps. Comfy really isn't good for that type of thing at this point.

  • @nicolasmarques7460
    @nicolasmarques7460 Před 7 měsíci

    Amazing tutorial, thanks for sharing!

  • @potusuk
    @potusuk Před 6 měsíci

    yay finally got it working :) learning fast thank you

  • @clonosaurios
    @clonosaurios Před 7 měsíci +1

    Thank you for your video! I learnt that comfyui is awesome :)

  • @tomaseriksson5430
    @tomaseriksson5430 Před 7 měsíci

    thanks for this! looks so crazy to a beginner but I followed and it's a great place to start generating images

    • @sedetweiler
      @sedetweiler  Před 7 měsíci

      It really is! You have officially leveled up, as this specific type of workflow shows you how things work, not how to use a specific UI. Cheers!

  • @lukeovermind
    @lukeovermind Před 10 měsíci +1

    fantastic! I am looking at some advance workflows, however with no real explanations how they work. I want to use it but I dont know what some of the nodes and flows do! However I found alot of value from your vids and at this stage I am happy to just play and learn comfy and put of creating art projects/ideas with SDXL for the time being.
    That 3rd Sampler is neat! I tried to see if you can use latent upscale method in your previous video with SDXL base and refiner, didnt work but that is the beauty of comfy! You get to try stuff

    • @sedetweiler
      @sedetweiler  Před 10 měsíci +1

      I also think it is a pretty great way to learn how all of this works together. It really is limitless!

  • @TomMaiaroto
    @TomMaiaroto Před 10 měsíci +2

    I'm new to ComfyUI all and really love your videos. Thanks! Maybe this is obvious to folks, but one thing I recently learned was the ability to condition after one KSampler ran so you can continue to refine your final image. It ended up being an alternative (or another tool in the toolbelt) to inpainting. I wasn't just refining, I was adding to or dramatically changing the final image - all without losing the "base" starting point that was all "locked down" in that the seed was fixed, the cfg and steps didn't change, etc. So it was a very non-destructive compositional workflow. If I wanted to add an object to the image, I could do that through a second prompt that was applied to a second KSampler.
    I could also introduce new LoRAs later on in those steps. I'm going to continue to experiment with this strategy and go through this more than once. So instead of a long prompt followed by a smaller corrective one, do more of a build up of prompts. Start simple and continue to add on to it so that elements within the image can be independently adjusted, removed, or re-arranged. Again, a more compositional approach during image generation to hopefully reduce the amount of work in post (or a series of very similar images that can be worked together in post processing). This could get a bit messy too, but maybe not if they are arranged left to right in a linear fashion building up the scene.

    • @sedetweiler
      @sedetweiler  Před 10 měsíci

      That's great! It is a lot of fun adding into the pipeline. It's what we do internally as well when testing models and playing with new ideas. Cheers!

  • @skylightikab443
    @skylightikab443 Před 7 měsíci +1

    Thanks for no nativ english speaker this was a good tutorial. It was very helpful! :)

  • @Adreitz7
    @Adreitz7 Před 10 měsíci +5

    Thanks for this look at the setup that Stability uses internally. I'm not so familiar with Comfy, but I've been using and enjoying SDXL through Invoke, which has a similar Nodes capability. I have a few questions and comments:
    1. What are the Original and Target W/H actually doing for the CLIP conditioning nodes and what is the logic to setting those values? I played around with it, testing various combinations, and the only thing I could confidently say is that setting Original W/H smaller than 1024 causes the image to become blurry. I couldn't see any specific benefit to any other value, as I tried 1024, 4096, and 40960 for Original and between 64 and 40960 for Target -- setting different values made the image different, but not obviously better or worse. I settled on just setting them the same as the output image dimensions.
    2. Why are there two prompt inputs for the base text encoder node when you provide the same input to both? Invoke calls one input the prompt and the other the style. What effects are caused by, e.g. separating your prompts into a prompt and a style and sending them independently to the two inputs, switching the inputs (so prompt goes to the "style" input and vice versa), setting them both the same, or leaving one or the other blank? I've found that if I prompt the base model for a roller coaster in the first input, I get a roller coaster. But if I prompt "roller coaster" for the first input and "photograph" for the second, I get anything BUT a roller coaster -- ruined buildings, abstract paintings, etc.
    3. Connected with #2, Invoke's refiner conditioning node only includes a "style" input, but I've found that only giving it a style prompt can cause the refiner to do weird things (like making architecture look like it's made of tent fabric).
    4. You've indicated that initializing the noise with the refiner is an interesting idea, which it is, but have you seen any consequences other than just making the images different? Does it provide any actual benefit?
    5. I've experimented with higher resolution SDXL generations. I'm on a Mac and there are some apparent generation bugs with Invoke on MPS (about 1856 square and above it becomes debilitating). But I've noticed that my scenes at higher resolution (photographic sci-fi style architecture) tend to become wide angle and taken from a high vantage point, almost as if the resolution setting is correlated with the position and zoom of the virtual camera. Has Stability done any experiments at higher resolutions than 1024x1024?
    6. Is there a benefit or danger to sending the same noise seed to both the base and refiner?

    • @bobbyboe
      @bobbyboe Před 6 měsíci +1

      Good questions... I also would like to know the answer. Did you understand the concept of why there is a field of dimensions in a node that is supposed to provide only text?

  • @ysy69
    @ysy69 Před 7 měsíci

    ComfyUI is truly about fine tuning the way one approaches the creation of an image using AI

  • @MikevomMars
    @MikevomMars Před 10 měsíci +1

    Using SDXL with a 2060 SUPER 8GB + ComfyUI and it works great 👍Turned out that Comfy is 15x(!!!) faster than A1111 for the same tasks using SDXL! Also I never got any errors with ComfyUI while A1111 always gives me "NaN tensor" errors when working with SDXL. For SDXL, Comfy is a MUST! So I`m looking forward for more tutorials on ComfyUI.

    • @sedetweiler
      @sedetweiler  Před 10 měsíci

      Great to hear!

    • @technoprincess95
      @technoprincess95 Před 10 měsíci

      I am on the opposite; I am only be able to use Colab. When using A111, it's very fast and smooth, allowing me to work with videos and other tasks, including upscaling to 8k,comfuyi just pictures. However, with Comfuyi on Colab, I encounter issues such as disconnects and running out of ram after just 5 images, im using workflow of olivio sakiras, using base ckpt sd1.0 + refiner sd 1.0 . overheard refiner consumes lots of Ram

    • @MikevomMars
      @MikevomMars Před 10 měsíci

      @@technoprincess95 I`d double check your workflow used then. Also did you try to run ComfyUI locally? You don`t need a decent GPU to do so. Even an old 8GB NVIDIA will do it.

  • @reekster30
    @reekster30 Před 9 měsíci

    wow - great tutorial dude. I've only recently got into comfy and wondered why all the controlNETs were failing last week :D All new ones install thanks to your videos and loving all the sdxl videos... fun times ahead (but I really need a pc gaming rig for speed) haha
    Out of interest - what kind of set up for a pc would you recommend for quicker generation/processing? massive 128gb RAM and like a RTX4090? :D
    thanks for your videos - amazing

  • @florentraffray1073
    @florentraffray1073 Před 2 měsíci +1

    Thanks for these tutorials, great to have an in depth dive into the UI.
    I'm a little confused about the start/end steps and steps in the KSampler.
    In your second sampler in the chain of them, if you start at step 3 and do 12 steps, wouldn't that leave you at step 15 for your starting point in the next one?

    • @sedetweiler
      @sedetweiler  Před 2 měsíci

      There are some advantages to skipping steps in some cases. It all has to do with the residual noise.

  • @A.polon.i.a
    @A.polon.i.a Před 2 měsíci

    Great video Scott, I wonder could you explain how to change the image size? What do I have to alter to produce an image of 832 x 1216 for example? Or point me to a future video that explains it, as I'm only on ep.2 Thanks💖

  • @badmojjo
    @badmojjo Před 10 měsíci +1

    I am definitely going to search for a good upscale workflow on your channel.

  • @demoran
    @demoran Před 10 měsíci

    So I was following the guidance here, and found that UniPC and the 2M variants will barf on you when the refiner steps are higher than the base steps. I tried with the 12/20 pair you've demoed here, and got an image with nasty vertical streaks in it. It was fine at 20/20, but barfed again at 20/50.

  • @DarnSylon
    @DarnSylon Před 9 měsíci +2

    When you added the third or 'pre-sampler', why did you not pass the noise information as you had done with the first of the two samplers? I messed with that setting on the first two and didn't notice much of a change. Thank you for the videos and instructions. They are extremely helpful. And you suggest not to add things like extra fingers to the negative prompt. What is your method of not getting extra fingers or limbs, etc?

  • @renderartist
    @renderartist Před 10 měsíci

    Thank you! 👍🏼

  • @matthewharrison3813
    @matthewharrison3813 Před 10 měsíci +1

    Thanks for the great video. Could you please talk more on the clip encoder width and height and target width and height? What do they do and is there any documentation? Why are you using a different value for the target than the base?

    • @4richis
      @4richis Před 6 měsíci

      I would love to see and answer to this as well

  • @Aaabii
    @Aaabii Před 10 měsíci

    Thank you very much. ı prefer comfyUI over A1111 and you are my go to channel for my purposes.

  • @bigbo1764
    @bigbo1764 Před 9 měsíci +1

    I’m curious, how would I implement a lora in this setup? I tried inserting 2 lora nodes after the checkpoint nodes and connecting them like I would in SD 1.5, but it seems to not be registering the existence of my Lora and just skipping over it. My checkpoints are connected to the Lora nodes only, except for the VAE, which is used for the decoding, what am I doing wrong and how exactly do I fix this?

  • @tripleheadedmonkey6613
    @tripleheadedmonkey6613 Před 10 měsíci +1

    One question I had. Is there any reason why you recommend using the VAE from the refiner, when there is only 1 version of the VAE (barring custom fixes for FP16) publicly available?
    If I choose to merge the fixed FP16 base VAE with the refiner, am I getting the same experience as you are (besides fp16-fp32 differences) ?

  • @San4itos
    @San4itos Před 7 měsíci

    You set Base KSampler to return the leftover noise but there is no leftover noise because it does all of its steps. Then Refiner adds its own noise and process it further. You may see it in
    Base preview. I guess if you turn off the leftover noise from Base, result is gonna be the same. What you need to do to pass the leftover noise to the Refiner is to use for example 20 steps but end on step 12. Then disable add noise function on Refiner KSampler.

  • @case0h
    @case0h Před 10 měsíci

    thank you!

  • @JohnSundayBigChin
    @JohnSundayBigChin Před 8 měsíci +1

    Hi Scott, im rewatching the whole series again, you have done a good job. I have a question in this particular episode with the Sampler...why do you have the possibility of using the denoise within the KSampler but not with the advanced KSampler? Do they work differently?

    • @sedetweiler
      @sedetweiler  Před 8 měsíci +1

      it was to simplify things. when you start at a later step with the advanced sampler, you are "skipping" some of the pieces you do not want to denoise, so it is the same thing but harder to explain.

    • @JohnSundayBigChin
      @JohnSundayBigChin Před 8 měsíci +1

      When you make Img2Img in one of the videos I saw that you used the common Ksampler because you needed the denoiser. Now everything is much clearer to me, thank you very much for answering.

  • @deafponi
    @deafponi Před 4 měsíci

    Hi there Scott, thank you for the excellent tut. I must admit though, my robots did not look anything close to how refined yours came out. I wonder if I missed anything somewhere...

  • @user-kn6vx2xz4e
    @user-kn6vx2xz4e Před 8 měsíci

    very good

  • @MisterKerstov
    @MisterKerstov Před 10 měsíci +1

    Thank for this really concise and helpful tutorial. Just one thought, you did not enable the "return with leftover noise" for the "initial conditioning" node. Wouldn't it make sense to do so?

    • @sedetweiler
      @sedetweiler  Před 10 měsíci +1

      It actually returns so much that things go sideways. Give it a try. I have not found that to work well.

  • @Smashachu
    @Smashachu Před 6 měsíci +1

    Hmm i'm messing around with rendering the first 2-3 steps as something that i know SDXL is trained very well in so for example a brown horse racing for a positive prompt on the first 3 frames, then using a negative prompt for the Brown, with the new color being purple with a (purple horse:1.3). It's been working very well especially for harder to generate things, it's like it's erasing the colors and redrawing it now that there's a rough shape. I'd love to see how it will workout in combination with controlnet to maintain consistency in textures and shapes.

    • @sedetweiler
      @sedetweiler  Před 6 měsíci

      That method can also help with LoRA images that are not as strong as you prefer. It's a great workflow. 🥂

  • @courtneyb6154
    @courtneyb6154 Před 9 měsíci

    Excellent video Scott. If you could do some of us a favor and go into detail about what everything is and how it works within the cliptextencode nodes then that would be of tremendous value. I have scoured the net and am only able to find limited info about the options and nothing i have found has explained how or why they work. Building out the workflow is a great first step but not knowing how to fine tune is lame 😂Thanks!!!!

  • @ysy69
    @ysy69 Před 7 měsíci

    Very helpful. Have you experimented with learned with using multiple Ksamplers? Are you still keeping at least 3 at different steps today as part of your workflow?

  • @spiralofhope
    @spiralofhope Před 7 měsíci +1

    I was able to follow the tutorial well. I'm a bit confused at the three separate seeds. I can adjust the first (the conditioner/initializer) and get changed results, do I care about the others? In a previous video you said it wouldn't matter much for that context. Is that also true here?

  • @davewills6121
    @davewills6121 Před 9 měsíci

    Excellent tutorial, learned a whole lot in a short time. Why is it that the Refined images are indeed sharper, they seem to loose some of the more acute details. EXAMPLE: undersea shot without Refiner shows a murky underwater world with subtle light refraction and a sense of DOF, refiner seems to strip that away, leaving a sterile shot with little atmosphere?. Kudos

    • @sedetweiler
      @sedetweiler  Před 9 měsíci

      There is some balancing going on for sure!

    • @davewills6121
      @davewills6121 Před 9 měsíci

      @@sedetweiler Im just an eager noob getting his toes wet. PS: Can i trouble you for one question: If i pick one image from say my ''history'', can i build from that, run further batches based off of that one image, so that i can fine tune my results?

  • @gameplayfirst6548
    @gameplayfirst6548 Před 10 měsíci +2

    How is there any noise left during handover to the refiner, if you don't use the "end_at_step" parameter? Don't you get images without any noise from the base sampler if you don't limit the end in any way?
    Your base preview image confirms that you don't have any noise left after the base, which doesn't match the workflow described in the SD-XL documentation.
    And why do you overlap steps? For example you do 12 steps in base, but start at step 12 in refiner, instead of starting at step 13.

  • @GamingDaveUK
    @GamingDaveUK Před 10 měsíci

    Very useful video and great to see videos that are not automatic 1111 lol

  • @vanstonecypher
    @vanstonecypher Před 10 měsíci

    Thank you so much for the tutorial. It really helped with some basic knowledge that was not obvious as a new user of ComfyUI (double-clicking to get a list of nodes for example.) A couple of question though. Why are the width and height of the CLIPTextEncodeSDXL nodes are set to 4096? What does this mean since the output is still 1024?

    • @sedetweiler
      @sedetweiler  Před 9 měsíci

      It is the resolution CLIP was conditioned at prior to scaling. I tend to use it and prefer the result.

  • @ianwilliams7740
    @ianwilliams7740 Před 8 měsíci

    on that third sampler you added you kept the return with leftover noise to disable.. does that mean you use up all the noise in those 3 early steps? what's the thought in not setting that to enable??

  • @97DarkSkull
    @97DarkSkull Před 10 měsíci

    How would you precess sequences of images (I'm going or similar) with Comfyui?

  • @ysy69
    @ysy69 Před 10 měsíci +1

    Hi Scott, really appreciate your giving us the most recent update on SDXL. Do you know how to fine tuned a model using SDXL 1.0 and Dreambooth? Is this something you can create a tutorial video for us?

    • @sedetweiler
      @sedetweiler  Před 10 měsíci +2

      That is coming soon. It is going to be easier to train, results wise, but still getting methodology together.

    • @ysy69
      @ysy69 Před 10 měsíci

      @@sedetweiler 🙏🙏looking forward to… do you know if the new dataset should be set at minimum at 1024 by 1024?

  • @97DarkSkull
    @97DarkSkull Před 10 měsíci

    How would you process sequences of images (I'm going or similar) with Confyui?

  • @ethanhorizon
    @ethanhorizon Před 25 dny

    Thanks for the tutorial! Is the "noise seed" in Ksampler Advanced same as "seed" in Ksampler? You set noise seed as 4, what's the meaning of the number? What if I left it as zero?

  • @kenjix7316
    @kenjix7316 Před 9 měsíci

    why exactly dont we change the return leftover noise for the first sampler (the 2nd refiner one added at the end)?

  • @TomSweeney-ov8qs
    @TomSweeney-ov8qs Před 9 měsíci

    Do you have any videos (or recommendations for other videos) that go in depth on debunking the negative prompt urban legends you mention?

    • @sedetweiler
      @sedetweiler  Před 9 měsíci

      No, but I should make one. It's just terrible what people pass on as the perfect negative. Do they think the model was trained on "bad anatomy" and "extra fingers?"

  • @kabochsaygo6976
    @kabochsaygo6976 Před 7 měsíci

    Is it possible for you to do a tutorial showing the ComfyUI ->Models folder structure, and what goes into each of them? I manually installed the manager with no issues. But other things such as diffusers, embeddings, clip_vision, etc. are unknown to me. And a lot of things on huggingface can't be found within the manager. Thanks. PS: Just getting started with SDXL and using Comfy. So going through your videos one at a time.

  • @TissaUnderscore
    @TissaUnderscore Před 2 měsíci

    Should i use a refiner for a custom model? for example if i use juggernaut xl?

  • @Aksuloid
    @Aksuloid Před 9 měsíci

    Would it be possible for you to make a video explaining the ideas behind negative prompting? You've mentioned it twice in the ComfyUI playlist now, and I am quite interested to know why and what one would actually put in there.

    • @sedetweiler
      @sedetweiler  Před 9 měsíci +1

      It was really only an issue with SD 1.5, and not so much with SDXL. Consider for a moment that the model has no clue how many fingers a person has, so using a negative prompt of ((((((extra fingers))))) is just silly. People making it higher priority with all the "(" is like yelling English to someone that doesn't speak the language, cause that helps. The model was not trained on what "extra fingers" are, and that is just one example.

    • @Aksuloid
      @Aksuloid Před 9 měsíci

      @@sedetweiler That makes alot of sense, thank you. From what I've gathered then, its going to be more useful for removing unwanted objects/themes/concepts from an image (for example), or steering the model away from unwanted styles, instead of fixing small anatomical mistakes, etc.

  • @dreaminspirer
    @dreaminspirer Před 10 měsíci +1

    thanks so much for the video.
    I'm having BASE Steps and TOTAL Steps Primitives. So I'm trying to use a Primitive node to feed the PRERUN steps to 1st Refiner (let's call it PRERUN KSampler) but i bumped into a problem.
    - Feeding "steps" into PRERUN Ksampler is fine but I can not feed this "steps" INT to "start at step" for the BASE KSampler . they're both INT, but perhaps ComfyUI considers "steps" and "start/end at step" are different types. 😒
    - The other way around is feeding "end at step" for PRERUN and feeding this value to "start at step" for BASE and feed all KSampler with same "steps" value. But for some reason, the PRERUN Ksampler needs to be fed with exact amount of steps otherwise the result is nothing but NOISE. 😒
    please help , thanks again.

    • @sedetweiler
      @sedetweiler  Před 10 měsíci +2

      I have also noted that, and I think it is a bug. That should work just fine. I got around it by using a math node, since that was the end goal anyway.

    • @dreaminspirer
      @dreaminspirer Před 10 měsíci

      @@sedetweiler thats exactly what i found. Derfuu VAR nodes and MATH nodes did the trick without any problem.
      Having said that, i found PRERUN step should not be more than 3 or it's all crap :)
      Thanks again and pls keep sharing with us the quirky tricks to play with Comfyui

  • @alexlindgren1
    @alexlindgren1 Před 7 měsíci

    I'm aware that SD don't take account of spatial relationships, but I want to be able to replace for example a sofa in an existing image with an image of another sofa, but not sure on how to take on that challenge with SD, do you have any suggestions where to start? I don't want to manually mask each image, but I want the AI to recognize what part of the image is a sofa and mask it for me, I should just provide the image of the sofa and the "base image" of the livingroom.

  • @kick851
    @kick851 Před 13 dny

    for the cliptextencodesdxl
    if my latent image is 768x1280 do i still use 4096 for width and height and what about the target width and height

  • @Zizos
    @Zizos Před 9 měsíci

    I just did download the official base and refiner but it seems I've got the VAE version from somewhere else in the past.
    What's the difference? I get that the VAE is built-in to the model. Does this mean you get to delete the VAE Decode node or some other node?
    Can you just keep the VAE version and follow your workflow with no difference in results or at least no negative results in quality?
    As in the last step you showed you can 1st generate a blank latent and then into the base and refiner... Seems like you can do all sorts of tricks like that to experiment with the resulting image. I wonder if it makes sense. If I get it right, it seems that the latent creates a base noise ignoring the models so that you can just get something a bit out of the box (model). Is that right?
    Thank you for the tutorial. I have lot's of stuff to learn.

  • @RobertWildling
    @RobertWildling Před 5 měsíci

    Hmmm... at around 14:15, when you add the first refiner with the 3 steps, shouldn't the last refiner's "start_at_step" be changed to 15?

  • @RamonGuthrie
    @RamonGuthrie Před 10 měsíci

    Is there a way to change the folder path where the images are saved to easily? 📂

  • @parsley8188
    @parsley8188 Před 10 měsíci

    Nice

  • @latent-broadcasting
    @latent-broadcasting Před 9 měsíci

    I'm trying to mix this with your ControlNet tutorial and for some reason the output image has lot of artifacts and glitches. I believe it has something to do with the VAE but I'm new to this and I wouln't know

  • @martdiamond6109
    @martdiamond6109 Před 10 měsíci

    Do you also use an upscaler with SDXL? All the Comfyui examples I've seen never include it, so I'm just wondering how that would look in this workflow?

    • @sedetweiler
      @sedetweiler  Před 10 měsíci +1

      Ya, you can use any upscaler and use them repeatedly. It's way more flexible than AUTO1111. I will do a video on this super soon. Cheers!

  • @AltimaNEO
    @AltimaNEO Před 10 měsíci

    So I've been using a workflow that was on the comfy up Github in their examples page. I'm struggling with trying to figure out how many steps I should be giving the refiner?

    • @sedetweiler
      @sedetweiler  Před 10 měsíci

      I would start with 32 in the base and 8 in the refiner

  • @GuitarWithMe100
    @GuitarWithMe100 Před 10 měsíci +1

    Im still confused on what the ClipTextEncodeSDXL does? and how does the value 4096 affect it?

    • @sedetweiler
      @sedetweiler  Před 10 měsíci

      That was the initial conditioning prior to scaling, so we just prefer that for the refiner.

  • @xevenau
    @xevenau Před 7 měsíci +1

    Quick question, why was the last Ksampler added without a preview mode?

    • @sedetweiler
      @sedetweiler  Před 7 měsíci

      It wasn't on purpose. I just add them for their maths, not the previews.

    • @xevenau
      @xevenau Před 7 měsíci +1

      @@sedetweiler thank you!

  • @Darkwing8707
    @Darkwing8707 Před 10 měsíci +3

    Why did you choose 4096 for the height and width in the conditioners?

    • @digitalbear3831
      @digitalbear3831 Před 10 měsíci +1

      I'd like to know that one too

    • @AdamDesrosiers
      @AdamDesrosiers Před 10 měsíci +2

      also would like to know what these conditioners numbers do. And somehow, I've been happier with outputs when I set those number to 2048. But why? I don't know what those are doing.

    • @sedetweiler
      @sedetweiler  Před 10 měsíci +1

      The refiner was initially conditioned at that size prior to scaling, so we tend to use that size.

    • @jonnyfat
      @jonnyfat Před 10 měsíci

      @@sedetweiler Thanks for this tutorial - great reference. Great to have tutorials on this by someone who knows what they're talking about :-) I picked up on the size thing too - so it's 4096 for the base and 1024 for the refiner? Thanks!

    • @petec737
      @petec737 Před 3 měsíci

      @@sedetweiler "we tend to use that size" isn't really an answer. The only reason you'd have those numbers different is if you want to CROP a portion of the image..so in your case it's like wanting to crop out a 4096x4096 OUT OF a 1024x1024 image; which obviously is not how math works :)

  • @karenreddy
    @karenreddy Před 10 měsíci

    Even the simpler setup is convoluted. I've worked with shader graphs, so it's alright, but I can see how this has a bad learning curve for many. I just don't see the big gain in using this setup for this utility quite yet.
    Stability should allow it to be "baked" into a simple GUI, so you can create a front end with different graphs, then not mess with it much, unless you want to add more pieces to the front end. Saving this front end would allow it to be shared with beginners and make it easy to get into, the complexity would be hidden until they're ready to explore.
    What is the advantage to conditioning with the refiner first?

  • @imperfectmammal2566
    @imperfectmammal2566 Před 10 měsíci +1

    Can you tell me how to use the Loras offset that came with sdxl in comfy

    • @sedetweiler
      @sedetweiler  Před 10 měsíci +2

      Yes, I will post a video on that and it is SUPER easy to do! Cheers!

  • @artistaartificial5635
    @artistaartificial5635 Před 10 měsíci

    What if generate latent image with a fine-tuned 1.5 model and then use the refiner? Or first custom 1.5 img2img SDXL 1.0 and finally refiner?

    • @sedetweiler
      @sedetweiler  Před 10 měsíci

      Interesting, but you might run into issues with the sizes. However, you would always do an img2img with the first result and run that into the refiner.

  • @lechatsportif124
    @lechatsportif124 Před 5 měsíci

    Couple of things, isn't it recommended for the refiner to actually be started at 80% of total steps? Also, is conditioning via the refiner really a thing or did you just kind of mess around with it? You didn't select pass on noise, so I'm not sure what that means.
    Thank you for the tutorials, they are great!

  • @scottmahony4742
    @scottmahony4742 Před 4 měsíci

    models, the refiner, etc. Where can I find definitions for all these variables?

  • @dkf-nl1703
    @dkf-nl1703 Před 9 měsíci

    @14:20, doesn't the second sampler go up to step 15? And as a result, shouldn't the third sampler start at 15? And thanks for a great video!

    • @sedetweiler
      @sedetweiler  Před 8 měsíci +1

      They are exclusive, the step start is correct.