Stable Diffusion - Prompt 101

Поделиться
HTML-код
  • Опубликовано: 18 авг 2023
  • This is a beginner's guide to crafting prompts in Stable Diffusion.
    #ai #stablediffusion #aiart

Комментарии • 52

  • @DrDaab
    @DrDaab 8 месяцев назад +5

    Wow, this is far better than anything about Prompts that I have looked at. Excellent Job ! Thanks !

  • @WifeWantsAWizard
    @WifeWantsAWizard 9 месяцев назад +14

    Thanks for making this video. A few things:
    (0:55) Lock the seed (manually set it rather than letting it randomize) so that "silver hair" gives you the same shot but with silver hair.
    (5:38) The picture did completely change. You have two choices: a) never go above "1.2", or b) shift the guidance by one or two.
    (7:50) Look carefully at the "0.7" version. You'll see she's inside of a building. This shows you that the seed wants to do indoors. Change "fire" to "burning building" to get a great shot.
    (8:18) A weight of one is useful for grouping multi-word terms. "(16th Century Sloop:1.0)" works significantly better than just "16th Century Sloop" where the adjectives can float.
    (12:50) Notice that only "underwater portrait" is a different shot; the rest are just versions of the same. That's because descriptors appended to the end are naturally weighted less, and so are just offering "noise" to the original. It just so happens the seed in question has "underwater" data outputs in can draw upon that was ignored originally but now is revisited.
    (12:59) Tell me why "concept art" would do anything. They're all "concepts" and they're all "art". That phrase only adds "noise", which is only helpful if your guidance value is too low.
    (22:43) There is no such thing as "resolution" markers. As evidenced by...
    (24:17) "Is there a huge difference? Not really." People throw all the "resolution markers" in and pray. When they get a good result, they claim victory and when things get worse, they assume it's something else. "8K" does not exist (1024x1024 max) and "Unreal Engine" is a video game platform. Why would things "look better" in video game format where we devs throw out detail on purpose to make the game run better?

    • @3diva01
      @3diva01 9 месяцев назад +2

      These are great tips and most are consistent with my own experiences with prompting with A1111. Thank you for taking the time to share your advice!

    • @NotThatComplicated
      @NotThatComplicated  9 месяцев назад +9

      Thank you for the detailed feedback. Getting to your questions - yes, I agree that "concept art" didn't/doesn't have much impact. I just wanted to get a cross-section of attributes I see getting used. Second, you are correct, 4k, 8k, and the other resolutions aren't actually putting the image in those resolutions. When the model was trained 4k and 8k images were sometimes used. Those images tended to have more details and sharper features. Using say "4k" isn't generating a 4k image, but rather you are telling the AI to make a clean crisp image similar to what you'd see in a high-resolution image. Sometimes it has nice results, sometimes it does next to nothing. I just wanted to introduce another tool people could add to their toolbox.

    • @WifeWantsAWizard
      @WifeWantsAWizard 9 месяцев назад +2

      @@NotThatComplicated Thank you for the reply. Here's the error the Internet has told you and here's why "8K", as you say, "sometimes has nice results". I build my own datasets using DreamBooth. All incoming images must be a) 1:1 ratio, and b) 512x512 max. No 4K. Second, dataset makers don't add resolution to the prompt because every image we use is 512x512, so that's just wasted typing.
      So why does it "sometimes" work? If your guidance is too low and you add more characters (even 2 letters) to your prompt for subsequent images, the results will get better. Same is true of too high a guidance. If you remove words from your prompt, you now are at proper guidance and the result improves.

    • @dylc3373
      @dylc3373 8 месяцев назад +1

      ​@WifeWantsAWizard hey, not sure if this is really the place to ask, but you both seem really smart so I figure nothing ventured, nothing gained.
      I generated an image, got a good picture of a character, and I want to use that image of a character in various ways. Different pose, different background, full body, etc. Do you two have any tips or advice to go about that, aside from inpaint to change/adjust the background on the image? Would that be more of a photoshop sort of job? Am I missing a useful tool or extension?

    • @WifeWantsAWizard
      @WifeWantsAWizard 8 месяцев назад

      @@dylc3373 Two pieces of advice:
      1) generating characters on a gray background and then removing the gray in Photoshop works great for me because I'm 20+ years with PS, however...
      2) in the same way that using "(Actor 1 Name|Actor 2 Name:0.5)" can prevent randomness in your subject, you can also use "(Location Name:1.1)" to attempt to lock down the background in Stable Diffusion.
      The more unique the better. For instance, "(Hagia Sophia:1.1) hallway" is a very consistent background across all seeds versus "(Hogwarts:1.1) hallway" which will have zero consistency.

  • @TheColonelJJ
    @TheColonelJJ 3 месяца назад

    Vary helpful, Thanaks!

  • @ctrlartdel
    @ctrlartdel 9 месяцев назад +3

    Greaaaaattt video! Good luck! Keep it up this channel will grow fast. Nice, soothing voice to the point information.

    • @NotThatComplicated
      @NotThatComplicated  9 месяцев назад

      Thank you for the kind words. I've got a lot of things to work on, but hope to live up to the praise.

  • @user-df6oq1ij5o
    @user-df6oq1ij5o 6 месяцев назад

    Highly recommend installing the ADetailer extension for face and hand fixing

  • @n16161
    @n16161 5 месяцев назад

    This is great, thanks for the info and your voice is easy to listen to.

  • @RobertWildling
    @RobertWildling 4 месяца назад

    What a great tutorial! Thank you very much! Subscribed! - Looking forward to your next videos.

  • @midnightanimal4844
    @midnightanimal4844 8 месяцев назад

    Ngl, the first image looks great too, compared to the abrerrations i get even using a lot of prompts and good descriptions

  • @antonpictures
    @antonpictures 9 месяцев назад +4

    🎯 Key Takeaways for quick navigation:
    00:00 📝 Exploring how to organize prompts for Stable Diffusion.
    01:01 📝 Adding details to prompts for more specific outputs.
    02:41 📝 Adjusting attributes' weights to influence image generation.
    04:14 📝 Exploring different mediums: portrait, digital painting, and ultra-realistic illustration.
    05:58 📝 Fine-tuning attributes like fire weight to achieve desired effects.
    07:03 📝 Examining the impact of different artistic styles on image generation.
    09:05 📝 Balancing weights between attributes to prevent canceling effects.
    11:08 📝 Exploring how resolution markers like "portrait" affect generated images.
    13:21 📝 Comparing the influence of different resolutions on image style.
    16:02 📝 Using artists' styles to create different artistic interpretations.
    19:15 📝 Considering the impact of different styles on generated images.
    21:13 📝 Exploring alternate resolutions like "unreal" in the prompt.
    23:18 📝 Comparing effects of various resolutions on image style.
    25:00 📝 Adding depth of field to a prompt for enhanced image quality.
    26:19 🔥 Using alternate resolutions and styles to enhance image details.
    27:01 💡 Exploring color, lighting, and effects to add more depth to images.
    28:26 🌟 Comparing different lighting and effects options for image generation.
    29:20 🎯 Selecting the best combination of settings for the final image.
    29:35 🎬 Summarizing the tutorial and plans for a companion video.
    Made with HARPA AI

  • @ianwilliams7740
    @ianwilliams7740 8 месяцев назад

    I enjoyed the video and am looking forward to the companion one.

  • @wernerblahota6055
    @wernerblahota6055 8 месяцев назад

    You won yourself another subscriber. Very informative and most of all very well explained. Keep up the good work. Looking forward to your next video.

  • @KarimMarbouh
    @KarimMarbouh 9 месяцев назад

    oobabooga demon mode

  • @synthoelectro
    @synthoelectro 9 месяцев назад

    seriously helpful, been working with SD since beta 2 last year.

  • @DrAmro
    @DrAmro 9 месяцев назад

    keep the good work bro, thank you, perhaps you should continue with mastering the prompt to know how to prevent color bleeding, minimize the token, using the capital word's secrets and how to use the additional prompt of the hires and the ADetailer to our side .. thank you again for sharing your knowledge.

    • @NotThatComplicated
      @NotThatComplicated  9 месяцев назад

      After Detailer (adetailer) is on my list of things to cover. I had a few suggestions to do a more advanced prompt video. Look for it soon.

  • @fathanimanassaffat5746
    @fathanimanassaffat5746 8 месяцев назад

    got another subscriber here, thankyou so much for making this video

  • @BetterHumanz
    @BetterHumanz 9 месяцев назад

    Awesome mate cheers I learnt alot.

  • @AiVisualArtist
    @AiVisualArtist 9 месяцев назад

    Very very informative.

  • @Warblade1979
    @Warblade1979 9 месяцев назад

    Helpful, Very good.

  • @gamingchamp2424
    @gamingchamp2424 9 месяцев назад

    Such a fine video

  • @luminoucid
    @luminoucid 9 месяцев назад

    thank you!

  • @froilen13
    @froilen13 8 месяцев назад

    new subscriber here. 1 suggestion for future videos, you can make them shorter even tik tok length for tips and fast searching.

    • @NotThatComplicated
      @NotThatComplicated  8 месяцев назад

      I want to get through a 101 series, for those that need the long form, but the plan is to absolutely make short for quick tips.

  • @MrSongib
    @MrSongib 9 месяцев назад

    most custom model it can generate up to 768:768, 960:640 or 1024:640 (double head may occur).
    basically double resolution can break the image, in this case 1024p. my rule for SD 1.5 max is 768:1024.

    • @NotThatComplicated
      @NotThatComplicated  9 месяцев назад

      Yeah, sometimes I generate straight to 1024 by 1024, but for the sake of the tutorial I keep it at the default 512 by 512 for the sale of time. I think I did 768 by 1024 in my first tutorial. Funny enough, SDXL was trained in 1024 by 1024 images and you often get an undesirable result if you process at any less than that. Thanks for the feed back!

  • @TheMagista88
    @TheMagista88 5 месяцев назад

    HI, how do I get that upscaler? thank you!

  • @jessecool561
    @jessecool561 5 месяцев назад

    I am running SDXL.... Does the weight adjustments carry over SDXL or is this just a 1.5 prompt?

  • @corybanticfire
    @corybanticfire 8 месяцев назад +1

    On the mediums, are the words for the prompt exactly the same as your labels on the resulting images?

    • @NotThatComplicated
      @NotThatComplicated  8 месяцев назад

      Yes, the labels on the images were actually generated by the search and replace script. The group of images with the words across the top was all generated by the program at the time of creation.

  • @Dex_1M
    @Dex_1M 8 месяцев назад

    prompt editing 101 when ?

    • @NotThatComplicated
      @NotThatComplicated  8 месяцев назад

      What would you like to see in a prompt editing 101 video?

  • @gbutube8
    @gbutube8 8 месяцев назад +1

    First time ive heard the "prompt breakdown". Thank you. Note worthy that SD knew what you meant by "woman". 🤣

  • @kovanjola
    @kovanjola 9 месяцев назад

    Hi, is it possible to help, When I updated (stable Diffusion) it all Extensin stopped working , Especially (roope and conterolNet )

    • @NotThatComplicated
      @NotThatComplicated  9 месяцев назад

      First thing I'd like you to check...go into extensions, go to your "installed" tab and make sure you are on the most current version of both roop and controlnet.

  • @LouisGedo
    @LouisGedo 8 месяцев назад

    👋

  • @stray1969
    @stray1969 9 месяцев назад +1

    You won yourself another subscriber. Very informative and most of all very well explained. Keep up the good work. Looking forward to your next video.