Stable Diffusion Samplers - Which samplers are the best and all settings explained!

Поделиться
HTML-код
  • Опубликовано: 28 май 2024
  • This video will teach you everything you need to know about samplers in Stable Diffusion including how things change for SDXL and which samplers are which for ComfyUI.
    First, I break down the differences between the samplers by comparing processing speed, convergence, and the minimum steps required. Afterwards, show which samplers result in similar inputs and use the results from the comparison to recommend a few samplers for you to use.
    After that, we get into the really nerdy stuff. I show you what every single advanced setting in Automatic1111 does including Eta, Sigma Noise, Sigma Churn, the Scheduler, and more!
    Intro - 00:00
    Changes for SDXL - 00:42
    ComfyUI Samplers vs Automatic1111 - 01:19
    Processing Speed - 01:41
    Convergence - 02:13
    Minimum Steps - 03:10
    Output Groups - 04:29
    Group 1 - 04:54
    Group 2 -06:43
    Group 3 - 07:44
    DPM Adaptive - 08:57
    DPM Fast - 10:17
    Advanced Sampler Settings - 10:47
    Eta - 10:59
    UniPC Settings - 12:00
    Sigma Noise and Sigma Churn - 13:30
    Schedulers, Sigma Noise, and Rho - 14:54

Комментарии • 57

  • @nightynight5990
    @nightynight5990 2 месяца назад +2

    Dude, I usually dont write comments, but this guide is awesome as hell. The explanation with the "a" in the name and the benchmarks PLUS your "good image" disclaimer were extremely well done and professional. (I am currently at minute 3, but had to post a comment first lol) I hope you get more subscribers and viewers. Thank you for your service!

  • @OriBengal
    @OriBengal 9 месяцев назад +6

    Whoa! First explanation of all the sampler that goes in depth... usually it's "just try them all out and see what you like" :)

  • @TheColonelJJ
    @TheColonelJJ 17 дней назад +1

    Thank you VERY much for the time and effort you put into this! Most excellent!

  • @k225
    @k225 9 месяцев назад +12

    All samplers can be used with SDXL in ComfyUI. The DDIM, PLMS, UniPC sampler limitation is an Automatic1111 problem.

    • @Because_Reasons
      @Because_Reasons 9 месяцев назад

      Correction, they can now be used in Auto's new commit (release candidate) SDXL update.

  • @HarryMcKenzy
    @HarryMcKenzy 4 месяца назад +2

    that was awesomely presented ! thanks for the good work !

  • @onety-one9087
    @onety-one9087 8 месяцев назад +3

    A very useful video. It's rare these days that I see new info in a SD video, but I didn't know half the samplers took twice the time to generate. That could save a huge amount of time using Deforum. Thanks! ❤

  • @bluemurloc5896
    @bluemurloc5896 9 месяцев назад +2

    this info is pure gold

  • @pupernickel
    @pupernickel 5 месяцев назад

    THANKS for the video. you did a very good job here!

  • @haktan7482
    @haktan7482 8 месяцев назад

    bro,i love your works,thanks for the video

  • @galefraney
    @galefraney 8 месяцев назад +1

    Fantastic video!!!

  • @valerymoyseenko
    @valerymoyseenko Месяц назад

    According to Wikipedia, Leonhard Euler and Karl Heun are Germans, in their language EU reads as OY - OYLER and HOYN - it comes from pronunciation of Greek ϒυ - ύψιλον. It would make sense to pronounce the names based on the one principle, and the way their mom and dad called them.

  • @jibcot8541
    @jibcot8541 9 месяцев назад +2

    Thanks, a very in-depth look at the samplers, thanks for explaining all that, a lot to wrap your head around. 99% of the time I use 30 steps of DPM++ 2M Karras and just occasionally run the final output with a high steps Heun if it is something really special as I prefer the way it looks in all those grids and has a great quality.

  • @Artischokk
    @Artischokk 5 месяцев назад

    Thanks! very useful video!

  • @AIAngelGallery
    @AIAngelGallery 9 месяцев назад

    just wow! very in-depth tutorial of samplers!! thx a lot for your effort

  • @sunviewer7
    @sunviewer7 3 месяца назад

    Best in depth video.

  • @jamarti
    @jamarti 9 месяцев назад +3

    Your videos are great and informative, but it would be helpful if you added a check mark or arrow in the places you talk about. Then the video would be easier to watch.

  • @luke2642
    @luke2642 6 месяцев назад

    Great video. So much useful info. If you have a spare weekend you could repeat all this with the new LCM lora on 0.5, 0.75 and 1.0... You get great results in 3 steps on DPM SDE. Or maybe wait until the proper LCM scheduler is implemented in auto1111? Even before LCM, many samplers give great results at cfg 1, 1.5 or 2 even with just 5 steps, which I feel is often overlooked.

    • @luke2642
      @luke2642 6 месяцев назад

      Even without the LCM lora, try these on 4 steps:
      DPM2 cfg 1.5
      DPM2 a cfg 1.5
      DPM++ 3M SDE cfg 1.5
      DPM2 Karras cfg 2
      DPM++ 2S a Karras cfg 2
      DPM++ SDE Karras cfg 3

    • @luke2642
      @luke2642 6 месяцев назад

      Then on 5 steps, more samples become usable at these low cfgs:
      DPM++ 3M SDE 1
      DPM2 a 1.5
      DPM2 2
      DPM2 Karras 2
      DPM++ 2M Karras 2
      Euler a 2
      Euler 2
      LMS Karras 2
      UniPC 2
      DPM++ 2S a Karras 2.5
      DPM++ SDE Karras 3
      DPM++ SDE 3

    • @luke2642
      @luke2642 6 месяцев назад

      You could continue this process, producing a graph to illustrate the cfg vs steps trade off... There's a usable range and optimal cfg for each step count, for each sampler. I've not seen it properly graphed in a paper or RUclips video, if you're looking for more inspiration!

    • @luke2642
      @luke2642 6 месяцев назад

      And how did you get the wall clock computation time out? Is there a benchmark tool that spits out the data?

    • @luke2642
      @luke2642 6 месяцев назад

      Your fantastic video on cfg is here, the legend at 6:50 is great, but perhaps more could be investigated in the bottom left corner below 12 steps :-) There's also the cfg fix in the form of dynamic thresholding... so many variables! ruclips.net/video/kuhO9zAzetk/видео.htmlsi=A61TTCeWHdetui6x

  • @MrPicklesAndTea
    @MrPicklesAndTea 8 месяцев назад

    Well, this taught me a bit but I'll probably just stick to DDIM for text to image, then DPM++ 2S Karras for image to image/inpainting.

  • @RyokoChanGamer
    @RyokoChanGamer 9 месяцев назад

    Hello how are you? I really like your videos and your didactics, lately I'm looking for knowledge to create loras, these last 2 days I read and saw several things about the subject, but I'm still lost because in each place I see something teaching or informing something different .. I would really appreciate it if you could make a video about this subject, if possible... I have no doubt that you will explain things clearly... I'm using Kohya and the biggest problem is in the parameters tab, they are many settings and I don't know exactly what the main ones do, I've done dozens of tests, but I still get lost in options like Optimizer, LR Scheduler, and mainly, Network Rank (Dimension) and Network Alpha, the latter I think have a big weight in the result, just like Epochs x Steps... keep up with your beautiful didactic work

  • @Ai-Kyuu
    @Ai-Kyuu 9 месяцев назад +1

    adding loras and embeddings will continue to change the image significantly

  • @psb962
    @psb962 9 месяцев назад

    Question: when comparing steps required to get acceptable image between SD1.5 and SDXL, how were the SDXL steps divided between base and refiner samplers?

    • @siliconthaumaturgy7593
      @siliconthaumaturgy7593  9 месяцев назад +1

      For this testing, all base steps. Refiner vs base will effect subjective quality and for the best subjective quality, you will want to go beyond the minimum steps anyway.
      My criteria for a decent image was all artifacts disappearing and the image not looking out of focus or burnt, not best subjective quality.

  • @DoubleBob
    @DoubleBob 9 месяцев назад +1

    Could you make a similar video about the additional samplers in ComfyUI? Samples like dpmpp_3m_sde_gpu

    • @siliconthaumaturgy7593
      @siliconthaumaturgy7593  9 месяцев назад +1

      Wow. They waited until right after I made a new video to release new samplers.
      I saw a quick comparison on Reddit indicating dpmpp_3m_sde_gpu is in the same group as the rest of the SDE samplers. There is a conversion table for A1111 vs Comfy sampler names near the start of the vid
      Also, seeds for A1111 and Comfy are not the same (GPU vs CPU noise) so you won't get identical results even with the same seed. I think the gpu SDE samplers in comfy might be an exception, but I didn’t verify

  • @coloryvr
    @coloryvr 6 месяцев назад

    Big FANX for this Video! I have the strong feeling that with some samplers, long and very detailed prompts require more steps...(?)
    Has anyone else noticed a connection between prompt complexity, Sampler and number of steps?
    Happy colored Greetinx

  • @krystiankrysti1396
    @krystiankrysti1396 9 месяцев назад +8

    You forgot about one important thing- decreasing CFG to 3 will let you generate in 6 steps, CFG value and number of steps required are very dependant, steps number must be about twice the CFG value or more, if its less then you get artifacts... which makes this whole test very incomplete. From my testing, DPM++ 2M Karras , SDE and Euler a for 9 images on 3090, are al about - 7.7 sec with 12 steps 512res but i dont like euler a soft texture most of the time unless i need nontextured images, so for 9 images at 6 steps you get about 5 seconds to generate them

    • @BenderdickCumbersnatch
      @BenderdickCumbersnatch 5 месяцев назад +1

      That's not true at all.
      Steps = How many times it refines the image.
      CFG = How much it tries to force the noise to match the prompt rather than filling it with something else that it sees. Lower CFG = more realism since it uses more of the neural network's own knowledge. Higher CFG = More and more "cooked" image if you go too far. About 6.5 is the highest I ever do, and 3 is the lowest I ever do.
      They have nothing to do with each other. They are completely independent.

  • @darkjanissary5718
    @darkjanissary5718 6 месяцев назад

    What about DPM++ 3M versions? Are they faster or better??

  • @timothyvass
    @timothyvass 8 месяцев назад

    Is it possible to save one output for each step beyond X for ancestral samplers?

  • @IIStaffyII
    @IIStaffyII 9 месяцев назад

    Interrasting, I've never truly understood the samplers that well. In regards to dpm being cfg focused might it be possible that it works better compared to other samplers when you would require a high cfg value for whatever reason?

  • @theoc1009
    @theoc1009 9 месяцев назад

    new samplers now in the SD WebUI update

  • @jasemali1987
    @jasemali1987 5 месяцев назад

    This is the best video I saw about samplers.
    However, I really wish to have an in-depth comparison of them based on CFG value. I use huge ass prompts that full of every word I can think of that is related to the subject. It is like an essay in the positive and another in the negative, with a punch of negative embeddings. I usually need a very high CFG number (above 20) with hundreds of steps, but I get dark saturated contrasty images.
    I'm still experimenting with the samplers for generating the images and also for scaling. It is very hard to compare when I have no idea what all of those samplers do, what they can't do, what they can do with extra steps.
    Any suggestions for an extreme case of very large prompt with very large CFG with open number of steps ?

    • @ForeverTemplar
      @ForeverTemplar 5 месяцев назад

      I'm finding some models don't really take well to novel-length prompts. Usually the creator will state that but sometimes...generate a thousand images and maybe get a hundred that actually apply to my intended parameters.

    • @jasemali1987
      @jasemali1987 5 месяцев назад +1

      @ForeverTemplar I used to do that, generate many photos and pick from them, but no matter what number I generate, I always find the images lacking so much from the prompt, so I end up, adding so much words eventually.
      I made a workflow that inject new words in the prompt in 6 stages, where the old words is carried out in the same promt but the new words are kept in front. The stages start with main stage, then, composition stage, shapping stage, effects stage, detailing stage. The last stage is upscaling with the accumulated prompt.
      The CFG and steps increase as the stages increase, while the denoise decreases as the stages take place. It is a manual convergent method that gives me control on the outcome in the early stages. I'm still refining it, but I'm getting pretty consistent results at this point.

    • @whatthe573
      @whatthe573 2 месяца назад

      ​@@jasemali1987This is interesting. How do you do this?

  • @MaxPayne_in
    @MaxPayne_in 9 месяцев назад +1

    what you use most of the time kindly share .

  • @Chris3s
    @Chris3s 9 месяцев назад

    Was DPM++ 2M Karras also tested above 20 steps?

    • @krystiankrysti1396
      @krystiankrysti1396 9 месяцев назад +1

      From my testing, DPM++ 2M Karras , SDE and Euler a for 9 images on 3090, are al about - 7.7 sec with 12 steps 512res but i dont like euler a soft texture most of the time unless i need nontextured images, also decreasing CFG to 3 will let you generate in 6 steps, CFG value and number of steps required are very dependant, steps number must be about twice the CFG value or more, if its less then you get artifacts

    • @Chris3s
      @Chris3s 9 месяцев назад +1

      @@krystiankrysti1396 have you tested with xformers and the slider option in settings? (forgot the name)

    • @krystiankrysti1396
      @krystiankrysti1396 9 месяцев назад +1

      @@Chris3s I turned off xformers, have only v1 optimisation in bat file enabled, the reason is i get NaN exception with xformers that i cant fix by swapping to another model and back, i get same speds without and with xformers so i dont need them really and yes i tested it to compare if i lose something by disabling them.

  • @demoran
    @demoran 9 месяцев назад +2

    The gpu and non-gpu versions produce different results

  • @lukas5220
    @lukas5220 9 месяцев назад +2

    This is incredibly good, valuable content, just as always :)
    Could you maybe do a video deep dive on the refiner? i rarely get better results with it. :( it's even more useless with 3rd party models.

  • @msampson3d
    @msampson3d 9 месяцев назад

    Apparently the latest automatic1111 added a bunch of new samplers. Are you howling in anguish? :P

    • @siliconthaumaturgy7593
      @siliconthaumaturgy7593  9 месяцев назад +2

      Definitely poor timing on my part haha.
      Though I'm use to videos going out of date within weeks at this point. Just how it is.

    • @msampson3d
      @msampson3d 9 месяцев назад

      @@siliconthaumaturgy7593 this is true. You tend to manifest changes into automatic1111 with your videos.
      We'll have to see if we can find ways to use this special power to add handy features to the app. Make a video about mitigating memory leaks so they fix that! :)

  • @tbc76
    @tbc76 2 месяца назад

    Extremely helpful video, thanks

  • @omegablast2002
    @omegablast2002 9 месяцев назад +3

    hahahah not obsessed with that "chinless dude" hahaha