Is Stable Diffusion Actually Better Than Dall-e 2?

Поделиться
HTML-код
  • Опубликовано: 11 сен 2024
  • Stable Diffusion's models just got published!
    The internet is about to be chaotic for quite some time lol.
    I also wrote this way before time, so some stuff might be a bit more inaccurate compared to the latest info they just released.
    Stable Diffusion by Stability AI
    [Project Page] stability.ai/b...
    [Code] github.com/Com...
    Dall-e 2
    [Project Page] openai.com/dal...
    A comparison board By made by @Ethan_smith_20
    miro.com/app/b...
    This video is supported by the kind Patrons & RUclips Members:
    🙏Andrew Lescelius, Chris LeDoux, Dan Kennedy, 正毅 水野, Shawn77134, Panther Modern, Jake Disco, Demilson Quintao, Tony Jimenez, Shuhong Chen, Hongbo Men, happi nyuu nyaa, Carol Lo, Deagan, Mose Sakashita, Miguel, Bandera, Tono Cabanelas, Gennaro Schiano, gunwoo, Daniel Skaale, Ravid Freedman
    Special thanks to these people for helping me out with this video:
    EthanSmith2000
    Shawn77134
    JustinJohn
    [Discord] / discord
    [Twitter] / bycloudai
    [Patreon] / bycloud
    [PayPal] paypal.me/bycl...
    [Profile & Banner Art] / pygm7
    [Video Editor] Askejm
    Thumbnail Taken from u/Kaarssteun

Комментарии • 301

  • @bycloudAI
    @bycloudAI  2 года назад +251

    Like this comment if you still don't have dalle 2 access
    edit: as of Sep 27th, I still don't have access. Am considering reapplying for the 3rd time LOL
    edit 2: the moment I made my first edit they opened access for dalle 2 nice. Everyone can use it now!

    • @jobz9150
      @jobz9150 2 года назад +7

      I did get access but I find that Midjourney is wayy better than Dall. E - 2

    • @olegmoki
      @olegmoki 2 года назад

      I hope it will take less than a year to get the access

    • @jobz9150
      @jobz9150 2 года назад +3

      @@olegmoki oh I got it in a month. It depends a lot on your email ID and social media accounts. If they seem legit, you'll get it pretty fast. If they seem too incognito and stuff you might never get it

    • @Infinite3D
      @Infinite3D 2 года назад

      i got it on 3 separate accounts

    • @jobz9150
      @jobz9150 2 года назад +1

      Midjourney is wayy better though, in my opinion.

  • @SpeakingofAI
    @SpeakingofAI 2 года назад +205

    Nice comparison! The new stable diffusion weights (v1-4) are out today. Apparently the results are better. And still free!

    • @CrazyFanaticMan
      @CrazyFanaticMan 2 года назад +6

      I've got the v1-4 from huggingface setup and working on my machine but is it possible to use the seed for a generated photo that you really liked and make variations of it?

    • @factsbykidd4765
      @factsbykidd4765 2 года назад

      @@CrazyFanaticMan yes, if your using miniconda it will tell you the seed of the image as it’s running, you can also set a manual seed using -seed ##

    • @rodriguez4363
      @rodriguez4363 Год назад

      Where could I find that model?

    • @poingly
      @poingly Год назад

      Exciting.

  • @Znegil
    @Znegil 2 года назад +148

    The strength of Dall-e2 is it's understanding of interactions between things, others struggle with this A LOT.
    A test I usually do is. "A woman showered with purple paint" or "a woman covered with purple paint".
    Dall-e2 creates really believable pictures from this, because it understands the interaction of objects. Other models so far show a woman with some parts of her body look like the have been recolored in pink. If you want simple pictures where objects don't interact with each other, stable AI can create really impressive stuff.

    • @zaidlacksalastname4905
      @zaidlacksalastname4905 2 года назад +7

      my first prompt was a guy riding a goat over a swamp under a pink sun so ig i picked the wrong prompt lol

    • @BlakeEM
      @BlakeEM 2 года назад +8

      The word "shower" can't be used in Midjourney so I tried "covered" and the images had no pink in them and looked realistic (using the new --testp model). I tried it in Stable Diffusion and got more realistic images than I did from DALL-E and I didn't even need to change any settings. Interaction between words depends on the settings for imagination vs coherence. You set it for more imagination when you want ideas combined or less for separate. When a model is set to be imaginative it may make the skin look pink if it tries to combine it with the color of the paint.

    • @pedrogorilla483
      @pedrogorilla483 2 года назад +6

      The prompt is only one part of the process is stable diffusion, you need to play with seeds, variations, weights of keywords, etc.

    • @michael_
      @michael_ Год назад

      @@BlakeEM --test and --testp actually use Stable Diffusion

    • @6-dpegasus425
      @6-dpegasus425 Год назад

      I've done similar tests with stable diffusion, specifically with the purplesmart AI image generator, and it is seemingly far more powerful at understanding interactions as long as you give enough detail to the prompt. DALLE 2 seems to excel in mainly accepting a sentence prompt with the minimal needed detail while stable diffusion needs as much clarification on the prompt and often its negative, but it's quality potential seems to be far higher

  • @defect1
    @defect1 2 года назад +117

    tutorial on how to setup stable diffusion?

    • @jordinne2201
      @jordinne2201 2 года назад +7

      its on huggingface

    • @xyzonox9876
      @xyzonox9876 2 года назад +11

      Look up the yt channel TinTingin he makes the installation straightforward and even includes the option to use lower vram nvidia gpus.

    • @toncortiella1670
      @toncortiella1670 2 года назад +2

      ​@@xyzonox9876 I didn't found it, did you spelled it wrong?

    • @askplays
      @askplays 2 года назад +4

      @@toncortiella1670 its actually TingTingin*

    • @mo6790
      @mo6790 2 года назад

      ruclips.net/video/0zQyCihHjQU/видео.html

  • @SamiTheAnxiousBean
    @SamiTheAnxiousBean 2 года назад +50

    as someone that has Access to Dall•E too, Honestly it was a bit underwhelming, especially going from stiff like CrAIyon which isn't as good looking output wise but has a way higher detabase even knowing certain film or game charcaters unlike Dall•E which completely fails at that, aswell as just being boring with it's prompts
    Honestly glad you're bringing stable defusion up!
    (Also people shouldn't be surprised that Dall E can't handle certain artstyles or just has no idea on what you're talking about because it seriously feels like the sample images were just stock photos and picked by someone...who is very out of touch)

    • @Deepak-gt9wd
      @Deepak-gt9wd 2 года назад +5

      yah! my DALL E seems to not even understand studio ghibli

    • @leandrodfcorreia2
      @leandrodfcorreia2 2 года назад +2

      Craiyon has awful image quality, still it seems to understand concepts better.

  • @Zanroff
    @Zanroff Год назад +13

    I downloaded Stable Diffusion. Been messing with it. I'm not smart, so I've been modifying other people's prompts. It's extremely impressive what it makes. 1 of 15 are stunning. I like that I don't run out of credits to use it too.

  • @arandomfox999
    @arandomfox999 2 года назад +35

    This basically means that Dall E 2 is probably very restricted by its dataset and has superior language model characteristics.
    I wonder how it would perform after being trained on an insanely large and diverse dataset.

  • @JCTechCorner
    @JCTechCorner 2 года назад +45

    Can't wait for text-to-image synthesis to be part of our everyday life and communication for enhanced image creativity

    • @hombacom
      @hombacom 2 года назад +8

      Without context it's very random and inconsistent to interpret a text

    • @JCTechCorner
      @JCTechCorner 2 года назад +8

      @@hombacom well that's why with time, anything can become better

    • @kiwi6421
      @kiwi6421 2 года назад +5

      I want to be able to generate gifs on inside jokes on the spot

    • @series1054
      @series1054 2 года назад +4

      ENHANCED MEMES.

  • @blankmoment2
    @blankmoment2 2 года назад +29

    I just want to say that people bashing on SD need to remember one thing. Stable diffusion is Open - Source - Code. This meant that the user could dissect it and then spawn something beyond SD originally. While other are locked into alot of policy, restriction . . . It had less chance to further itself. So 1 on 1 now for SD, then SD take an L in some aspect but in term of the future SD are standing on more solid ground.

  • @samuelkibunda6960
    @samuelkibunda6960 2 года назад +174

    One of the worst aspect to come from image generation are people calling themselves "prompt engineer"

    • @libiroli
      @libiroli 2 года назад +9

      Sorry why bad? I'm still learning about this space

    • @samuelkibunda6960
      @samuelkibunda6960 2 года назад +74

      @@libiroli calling yourself an engineer just because you wrote words in a search bar is a big insult to engineer's, it's the equivalent of someone spending 20 min searching something on Google and calling themselves an engineer. You're not building, solving or creating anything when you enter a string of words in an image generation software!

    • @itsfadixx
      @itsfadixx 2 года назад +5

      @@samuelkibunda6960
      you still made the prompt tho, ur the one to choose the art, image , who cares

    • @samuelkibunda6960
      @samuelkibunda6960 2 года назад +23

      @@itsfadixx So similar to commissioning an artist where does the engineering part come in?

    • @DansuB4nsu03
      @DansuB4nsu03 2 года назад +5

      @@samuelkibunda6960 I don't know, basic logical ability that most animals, including humans, possess?
      (sarcasm)

  • @jobz9150
    @jobz9150 2 года назад +20

    I absolutely love Midjourney. It's really a capable AI. It has very good artistic and color sense and has a good grasp of composition.

  • @AIpha7387
    @AIpha7387 Год назад +1

    Tim Sweeney unveiled the Unreal Engine for free and said: 'If You Love Something, Set It Free.'
    Although it cannot be imposed on others, such great choices used to make the world leap forward.

  • @carleyprice3138
    @carleyprice3138 2 года назад +7

    the level of censorship of dalle-2 makes me despise it.. what a shame. stable diffusion is the future.

  • @user-js7ud9du2y
    @user-js7ud9du2y Год назад +6

    Stable Diffusion being free is worth the 10/10

  • @kaltsssit
    @kaltsssit 2 года назад +2

    It's crazy how you can just tell when someone is actually passionate about the tech they're talking about, rather than just reporting on new stuff for views. Instantly subscribed 😎

  • @BlakeEM
    @BlakeEM 2 года назад +13

    I've used all 3 extensively. I've used DALL-E for a few weeks, I pay for unlimited Midjourney access, and I run Stable Diffusion on my computer.
    I tested the running in space prompt on all 3 and here is what I learned.
    DALL-E: I did nothing, got amazing sci-fi images, maybe better than the ones you show.
    Midjourney: I didn't get a man until I set --stylize 1250 or lower, default is 2500.
    Stable Diffusion: I didn't get a man consistently until I set --cfg_scale / --scale 16 or higher and I had to up my --ddim_step / --steps to 100 or higher to fill in the extra detail required for the image. Default scale is 7 (or 7.5?) and steps are 50.
    DALL-E ls definitely less imaginative. Even this image shows him on the ground like standard sci-fi, while both MJ and SD show a man actually running around the atmosphere of a planet. Most of the differences are down to the data/artists they learned from, the default settings for imagination vs coherence, how long the file is processed (quality/steps), size of the image generated compared to the size of the training images, and the weights placed on the model. They also all read prompts completely different and put higher weights on words at the front of the prompt in different ways. MJ prompt is by far the most powerful since you can place custom weights and there is even a --no command, the new beta is very impressive and is at least equal to the other two.

    • @TrevorMatthews
      @TrevorMatthews 2 года назад

      Thankyou for your info. This is exactly what I'd like to do - run it in my lab at home. Servers, RAM, and video cards/memory won't be a problem, I just wasn't sure where to start. This vid kind of showed me the differences and now based on your comments I'm hoping to get stable diffusion up and running. Do you have a link to a guide or doc you used to do the setup? Docker, VM or bare metal?

    • @asymmetrix7049
      @asymmetrix7049 2 года назад +1

      @@TrevorMatthews There's a help post pinned on r/stablediffusion that has a bunch of info and links

    • @BlakeEM
      @BlakeEM 2 года назад +3

      ​@@TrevorMatthews For some reason my replies keep not showing up in RUclips after I post them.
      Check out TingTingin's video on "Stable Diffusion How To Install And Run" for how to do it at command line with all the steps in detail. There is also "How To Install Stable Diffusion Web UI On Your Computer For Free" by Aitrepreneur. You could follow only the later video, however you will need to create your own environment using "conda env create -f environment.yaml" from the stable diffusion folder using the Anaconda Prompt to create the "ldm" environment that isn't in that video. You need to install Anaconda as well. I run it on my Windows 11 gaming PC (w/ RTX3080 10GB), no need for a VM, but you could. There are also two .ckpt files, don't use the bigger 7GB one if you have under 8GB of video memory, use the 4GB file. Good luck!

    • @TrevorMatthews
      @TrevorMatthews 2 года назад

      @@BlakeEM Thanks appreciate you being persistent, thanks!

  • @sadaneduardo4391
    @sadaneduardo4391 2 года назад +10

    so they spent the 1b from musk/microsoft on stock images...

  • @nefwaenre
    @nefwaenre Год назад +6

    i got my Dall-e2 invite this July and honestly, i used to think Mid Journey and heck even stable diffusion is better. While to me, stable is still better BUT the edit image feature of Dalle is a game changer. i... can't even tell you how amazing it has been! It's a shame i only get 15 credit/mo.

    • @Dusty2455433
      @Dusty2455433 Год назад +5

      The amazing thing about being open source, is that someone out there already made a GUI for Stable Diffusion with the same edit functionality!

    • @perrymanso6841
      @perrymanso6841 Год назад +4

      @@Dusty2455433 of course, SD is gonna eat the rest of AI's in no time. Open Source = gazillions of developers working for free.

  • @Simsonlover222
    @Simsonlover222 Год назад +1

    i played so much with stable diffusion that would cost me ALOT and im just jusing it for a few days AND its crazy... its so crazy... the moment i realized what AI art generator is and how it works...what its capable of... and what i found really interesting how good and sometimes surprisingly well can adopt to your prombt... not every picture is perfect but when it is... it just is :D even pictures that look like real life... no animator could put a yellow swimming suit so good on angela merkel and put her on the mount everest than these AI tools... and to see what randomness it kicks in sometimes...
    stablediffusion can make as good pictures as dall e... especially when u run this native... and the fact that it is largely uncensorde and the fact that you can download any software of it and its open source (real open source, not just some halfway open source)... even if it got censored in the future, you could always run this locally there will be ways
    you cant imagine how impressed i was
    it made me stay awake more than 24 hours and putting prombts in...
    you get alot of experience and learn how the ai works...

    • @OkOk-tu1jo
      @OkOk-tu1jo Год назад

      I use stable diffusion and it's kinda trash at understanding the prompt but it still creates fairly good images

  • @walidflux
    @walidflux 2 года назад +4

    Fuck DALL-E 2 we don't need something we can't get

  • @Darkjayson82
    @Darkjayson82 2 года назад +6

    You can actually get some decent anime style images from stable diffusion but it relies heavly on the prompts used unlike in dalle 2 but it looks like there is greater variance in stable diffusion. Also there is some anime theme finetuning happening in the scene as well.
    Now piece of advice, your generated art lives and dies by its seed. Even if the settings and prompts used are perfect if you get the wrong seed number the image that comes out will look bad. This is why you should run your prompts a few times with new seeds just to see if its the seeds that is bad or the prompts and settings. In fact one time I ran a prompt and settings for over a 100 images with random seeds and only got 2 good ones out of it. Luck seams to be a very important factor as well. Settings, Prompts and luck.

    • @boot5672
      @boot5672 Год назад

      Kind of new to this. Can u explain seeds? What r they and what do they do in the system?

    • @Darkjayson82
      @Darkjayson82 Год назад

      @@boot5672 The seed is just a random number to give the software a point to start from, they also use it in other ways to direct the image generation. You see if you use the same model, same settings, same prompts and same seed you get the same image, well most of the time sometimes a bit of change in hardware even temperature can change the outcome but the result will always be close to the image you generated before.
      So to generate different images they use a seed number and each one will produce different images.
      This way you can have the same model, settings and prompts but get different images from what you entered. Further more even if you have a good model, settings and prompts you might not get a good image because you got a bad seed. This is why you should always run a few images before changing prompts or settings if your not getting good results. In fact you see those great images people post they are more than likely the best result from tens to hundreds of generated images using the same settings and prompts just different seeds.
      If you ever played minecraft or any other kind of game with procedurally generated worlds they used seeds to change the world, use the same seed you get the same world.

  • @AIntune
    @AIntune Год назад +1

    8 month later Stable diffusion is a wild beast

  • @DodaGarcia
    @DodaGarcia 2 года назад +1

    I love the throwaway "both are bad at counting as usual" lol, so ironic when we're talking about advanced computing

  • @quosswimblik4489
    @quosswimblik4489 Год назад +1

    They are both limited by the biases in the art available. However AI in the future could be able to extract a data trait library avoiding more bias issues.

  • @Y0UT0PIA
    @Y0UT0PIA 2 года назад +7

    It's interesting to see the contrast between your 'objective' comparison of the generators' capabilities along different dimensions (which is, of course, the sensible way of getting at the question of which is more technically advanced), and our, I think shared, subjective feeling that stable diffusion is just miles above Dall-E in terms of generating aesthetically pleasing images and creating something fun from meme prompts, and that this is what actually matters in terms of applicability - at least at this stage.
    Dall-E is *impressive*, but I still couldn't see it generating the kinds of images that someone would actually want to commission. Stable diffusion absolutely does have that capability.

  • @memegazer
    @memegazer 2 года назад +6

    To me it is like the comparison between midjourny and dal-e mini/craiyon
    Both have their strong and weak points...but ultimately one is open source and they other is not.

  • @elismart13
    @elismart13 Год назад +1

    my said "five g b" but puts 4gb on screen... which is it? cause 5 is such an odd vram number

  • @MrAwesomePoopz
    @MrAwesomePoopz Год назад +7

    OpenAI: criticizes other learning AIs for not being free and open source.
    ALSO OpenAI: hides hides their codes under a bunker in mars and charges abusive prices for what is literally a meme generator.

    • @perrymanso6841
      @perrymanso6841 Год назад

      SD charges prices how??

    • @CharinVZain
      @CharinVZain Год назад +1

      @@perrymanso6841 SD is made by Stability AI, DALL-E is made by Open AI

  • @Younex
    @Younex 2 года назад +1

    A wise man once said: "Nothing beats free" :)

  • @Arne_Boeses
    @Arne_Boeses 2 года назад +11

    Since SD's public release yesterday, I've been comparing Dalle with SD. SD takes the L most of the time. I'm happier with my Dalle creations. I don't know if StabilityAI released the same model as they had before, but it is kinda disappointing tbh. Let's hope it's just me being terrible at prompt writing. Nice video btw!

  • @KuraSourTakanHour
    @KuraSourTakanHour Год назад +1

    I really like to test Dall-e with nonsensical prompts like "A moon-coloured dog diving into a melting mirror as pixies dance in kimonos"
    I've honestly gotten some fantastic artworks, sure sometimes it flops but sometimes has great results with interpretations you didn't expect

  • @Plagueheart
    @Plagueheart Год назад +2

    Open Source will always progress better than closed software as an entire community can contribute. Greatest example when picking a Linux Distro is to choose which repo is more popular with commits and activity simply because you don't wanna be outdated on security features and hardware drivers

  • @muhsetio
    @muhsetio 2 года назад +3

    Interesting. So can we conclude that for ppl who want artistic pleasure (artist for example), stability ai have better output.
    But for ppl who want more realistic and exact generation, dall E 2 is still on top.

  • @Amipotsophspond
    @Amipotsophspond 2 года назад +2

    4:35 "7 apples" count the stems on the apples and you see the Ai got it correct, because all apples must have a stem and so the apple like fruit around the 7 apples are not apples.

  • @SylvesterAshcroft88
    @SylvesterAshcroft88 Год назад

    Novel Ai also specialises in anime style art, after being trained on danbooru, and that uses stable diffusion's code set afaik.

  • @shocknawe
    @shocknawe Год назад

    3:07 - Is there any other A.I. Implementation that does that besides DALL-E?
    The entire “get a photo you took and put something in it” is what interests me the most.

  • @draken5379
    @draken5379 2 года назад +2

    People dont know how to use stable diff properly. The input is everything.

  • @virtualfg
    @virtualfg 2 года назад +2

    The fact that stable diffusion is free and open source makes it 1000% better, right now a script is already made for there to be an edit option for stable diffusion and someone is even making a entirely different stable diffusion model in order to make only anime characters and anime art in general which is the most weeb thing I heard in a while but it's open source what can you say

  • @rjameslower
    @rjameslower 2 года назад +2

    We need installation tutorial! O Collab!

  • @Fahad-gf1wx
    @Fahad-gf1wx 2 года назад

    It’s okay to feel that why. I felt like that sowhat at the beginning. Just use the love you have for making soft to overpower that feeling

  • @twitterglobalarmy
    @twitterglobalarmy Год назад +1

    Dall-E is cool but I've recently experimented with Blue Willow and I'm blown away. Those who can use these AI tools smartly will be the winners

    • @jopansmark
      @jopansmark Год назад

      Why are you commenting only on theme of BW?

  • @AtariWow
    @AtariWow 2 года назад +9

    After using both, I don't really like Dalle-2. And honestly, I think Midjourney produces my favorite results out of the three. It takes a lot less fiddling with the prompt to get good results.

    • @NadDew
      @NadDew 2 года назад +1

      but it's only works on discord, right?

    • @AtariWow
      @AtariWow 2 года назад +1

      @@NadDew yes, as of now Midjourney is only available through their discord bot.

  • @_Pyroon_
    @_Pyroon_ Год назад +1

    Okay someone's gotta say it, when will the gov get into this in the name of preventing ai generated cp (and probably many things they won't tell us about)? I could see them banning completely uncensored ai

  • @Lulink013
    @Lulink013 2 года назад +2

    So apparently you can't run SD on AMD GPUs (unless you use Linux/use a workaround involving Linux)?

  • @MondoMurderface
    @MondoMurderface 2 года назад +1

    I never used Dall-e beyond that multi image demo people played with. Stable Diffuse is nuts though.. I can get it to render just about anything if I take my time and plan my prompts and settings. Except hands.. Or limbs that interact..

  • @LutraLovegood
    @LutraLovegood 2 года назад +1

    But how good is SD at niche subjects? How well does it handle fan art? Animal anatomy? Spatial awareness?

  • @TheSheepLP
    @TheSheepLP 2 года назад +1

    7:58 providing an artist generally helps to get better results with dalle-2

  • @freaky_bro_
    @freaky_bro_ 2 года назад

    aah, i just got into making soft rn and tNice tutorials is so helpful and your voice is so sootNice tutorialng btw! thanks for tNice tutorials tutorial

  • @kaosssfx
    @kaosssfx 2 года назад

    I just wanna learn even more now- it looks so cool o.O

  • @woszkar
    @woszkar 2 года назад

    2:29 - What is this tool and where can I try it? Looks powerful.

  • @_lonath_
    @_lonath_ 2 года назад +1

    I knew this was coming and was waiting for you video before trying out stable diffusion
    Thanks for the vid

  • @fastfitc1
    @fastfitc1 2 года назад

    Hi Bycloud, im the one that made the meme in the video, OccultFusion I did get banned making that meme 🤣🤣

  • @notigor325
    @notigor325 Год назад

    Imo dalle 2 feels like an artists replacement, and stable diffusion more like an artists tool... In dalle 2 you can input a very general prompt and get very creative results, meanwhile in stable diffusion you need to think of the placement of objects in the image yourself, and describe it to the ai, like you would describe it to a person, and at that point it can generate really good looking pictures

  • @aydennguyen2802
    @aydennguyen2802 2 года назад

    Don’t be afraid, just embrace it.

  • @nightwintertooth9502
    @nightwintertooth9502 2 года назад +1

    Enjoy the free weights! More will roll out as the model continues to train.

  • @newwonderer
    @newwonderer Год назад

    Oh Lord, only 2 months passed and upgrade on Stable Diffusion (like 1.4) has much difference between what is presented in the video
    holyshit how this industry is running ...

  • @hangry3102
    @hangry3102 2 года назад +2

    I'm glad that there's gonna be a free and actually GOOD image generator that doesn't censor itself.

  • @faysalahmed8660
    @faysalahmed8660 Год назад

    Again thank you for this great video,❤️

  • @alefmiguel8820
    @alefmiguel8820 2 года назад

    Thank you, I've been looking for a long time

  • @DroBuilds
    @DroBuilds 2 года назад

    Awesome video.. I was wondering why dalle wasn't making very good anime stuff

  • @TAKEASHORTBREAK1
    @TAKEASHORTBREAK1 Год назад

    From my personal experience, stable diffusion makes it easier to get stunning results, especially when combining text to image and image to image.

  • @aaronstanley6914
    @aaronstanley6914 2 года назад

    Love this video it actual comparison specs and history of the ai thank you

  • @Mr_AciD
    @Mr_AciD Год назад

    @bycloud Would you have a link to the Genekogan tool you are showing?

  • @salmonsushi47
    @salmonsushi47 2 года назад +1

    9:09 summary of the video.

  • @nilaier1430
    @nilaier1430 2 года назад +2

    I feel like Laion dataset is kinda bad overall. In my opinion, that's the case when more ≠ good. If we would just have something like Dall-E dataset, but with all this copyright stuff added like in Laion dataset, then, that would be a great mix.
    Also one thing I want to add, Laion 5b dataset doesn't have a lot of characters in it. I've been looking for many different vtubers from Hololive, VShojo and Nijisanji, but couldn't find any of them, while much smaller datasets such as Conceptual Captions, Conceptual 12M and YFCC100M have them, as it was proven when generating images using craiyon (Dall-E mini).

    • @user-fi8ud8sm6y
      @user-fi8ud8sm6y 2 года назад

      They didn't use that dataset they filtered it to have better images

    • @nilaier1430
      @nilaier1430 2 года назад

      @@user-fi8ud8sm6y This sentence literally contradicts itself

    • @user-fi8ud8sm6y
      @user-fi8ud8sm6y 2 года назад

      @@nilaier1430 The problem you said was that the datset was too large they filtered it down which means it's smaller than the 5b significantly so in fact it was sub 2b so they threw out most of the images when I say they didn't use 5b I mean they didn't use the 5billion images they instead used a subset of that which was much smaller I.E not the same dataset

    • @nilaier1430
      @nilaier1430 2 года назад

      @@user-fi8ud8sm6y Yeah, true. They've used only Laion 2B-en, that's, in fact, just a Laion 5b dataset, but only with images that are titled with English captions. So, it's not a big difference. It's still a part of 5b dataset.

    • @user-fi8ud8sm6y
      @user-fi8ud8sm6y 2 года назад

      @@nilaier1430 they trained an ai model with votes on discord to recognize aesthically pleasing images then used that to rate the images and selected from a subset of those stability has announced that they'll eventually be releasing the dataset via LAION called LAION aestheics which would be the dataset that SD used its not out yet though regardless if your point was that size of the dataset was a problem well from the 5billion images you thought it was it was a few billion images less my point was that the dataset wasnt large like you thought it was

  • @MetroTO.
    @MetroTO. Год назад

    Imagine if istock bought Dall e 2 and now the ai generated images will have the same watermark of istocks's that really covers everything

  • @core3gamegd587
    @core3gamegd587 2 года назад

    An ai video editing tool?
    Oh my god I LPVE THIS IT WILL SAVE MY LIFENASNJDQJMAK

  • @DinoMoonCreations
    @DinoMoonCreations Год назад

    at this time.. stock images website do not have to worry. best dalle can do at the moment is 1024 by 1024 images... I need better resolution for my projects. so I keep my stock images.

  • @miserablevfx
    @miserablevfx 2 года назад

    is dalle 2 now available for all? idk i got the access tho and i made a couple of pictures and it looks very good

  • @amycao814
    @amycao814 Год назад

    Can I have the link of the last pictures in comparison?

  • @mfdim
    @mfdim Год назад +1

    has anyone noticed how simular dalle and stable diffusion logos are?

  • @deegeeooh
    @deegeeooh 2 года назад

    For a relative noob to this prompteous partee I salute you for this videjo, But seriously.

  • @djangel3108
    @djangel3108 Год назад

    Say, are there any free AI that can take an image and draw images similar to the one it is prompted with like DALL•E 2?

  • @mwwhited
    @mwwhited 11 месяцев назад

    My biggest issue with the big AI companies is how much content they steal for free to feed their models versus how much they charge. Since they don’t pay for source materials their models should be free for the public.. after all, they stole our data to build these systems.

  • @Valcuda
    @Valcuda 2 года назад

    I have access to DALLE 2, and it's so annoying at times.
    I try describing one of my characters to it to see what it'll generate, and then remember the word blood is censored.
    So instead, I tell it to generate a red viscous liquid, and that somehow works.
    I also once asked it to generate an egg with a photo realistic face, with a mustache, and can definitely agree with it looking like a stock image.
    Luckily, that worked in my favor since I was then easily able to put it on a 3D Egg Model and terrorize my friends in VR Chay with it

  • @iftikharhussain2443
    @iftikharhussain2443 Год назад

    ChatGPT Answwer :
    Stable Diffusion and DALL-E 2 are both state-of-the-art language generation models developed by OpenAI, they have different capabilities and use cases.
    DALL-E 2 is a powerful image generation model that can generate high-resolution images from natural language text prompts. It can generate a wide range of images, from photorealistic images to abstract and artistic images, and it can generate images of things that don't exist in the real world.
    Stable Diffusion is a more recent language generation model that has been developed to generate more coherent and consistent text. It is a more powerful version of GPT-3, It can be used for tasks such as writing articles, creating poetry, and more.
    It's hard to say which model is better as it depends on the specific use case and requirements. If you need to generate images, then DALL-E 2 would be the better choice, while if you need to generate text then Stable Diffusion is the better option.
    In general, it's recommended that you evaluate both models and see which one performs better for your specific use case. OpenAI provides various ways to access their models and test them, you can use the OpenAI API to generate images and text in your application and compare the results.

  • @dealznfilmz7069
    @dealznfilmz7069 Год назад

    You made a video without having access to dalle 2....I'm still watching this video without access to a computer.....

  • @aminazli
    @aminazli Год назад

    Dall-e 2 is better in understanding the prompt

  • @AshT8524
    @AshT8524 2 года назад +4

    Today I used Stable Diffusion for few hours it's not at all close to Dall-E or Midjourney. hopefully it'll be in some time. the images it creates are not that good yet. it is pretty fast to run on a PC tho

    • @alicat4906
      @alicat4906 2 года назад +1

      You have to guide it a lot more. MidJourney is good images out of the box, but Stable Diffusion tends to need more prompt guidance

    • @chadcrypto2675
      @chadcrypto2675 2 года назад +2

      it was much better when it was in the discord server. It feels like they dumb it down on release. Just like Dall-E 2

    • @AshT8524
      @AshT8524 2 года назад

      @@alicat4906 hmm ok I'll test it out more.

    • @user-fi8ud8sm6y
      @user-fi8ud8sm6y 2 года назад

      @@chadcrypto2675 nothing changed at all you literally can run the same prompt from the discord and get the exact same image

  • @dadlord689
    @dadlord689 Год назад

    Some people are drawing photo realistically like a machine - try to ban that. It is quite absurd to ban image generation.

  • @oeunsothearin9696
    @oeunsothearin9696 2 года назад

    crystal clear is a nice drink

  • @allhack7660
    @allhack7660 2 года назад

    it's finally ti. i'm ready to learn.

  • @AffinityXP
    @AffinityXP 2 года назад

    this is an amazing video!

  • @khalatelomara
    @khalatelomara Год назад

    I actually made a video about how to use this in blender by CEB SD addon , and it is incredible with architecture !
    it has also many uses other than text prompt like 3d from 2d feature , 1 click and you have a 3d model
    you can turn on captions as the video is in arabic 😁😁

  • @NekzLvL
    @NekzLvL 2 года назад

    I wish they make low end version of it though.

  • @ojosamuel9422
    @ojosamuel9422 2 года назад

    Your video is very helpful.. thank you so much

  • @user-ik8vy1rg8f
    @user-ik8vy1rg8f 2 года назад

    I chuckled in your 6 second intro.

    • @user-ik8vy1rg8f
      @user-ik8vy1rg8f 2 года назад

      Nice motion graphics the first minute in. Your presentation abilities are definitely sharpening.

    • @Askejm
      @Askejm 2 года назад

      @@user-ik8vy1rg8f new editor lmao

  • @erock.steady
    @erock.steady 2 года назад

    ok so correct me if i'm wrong, but my experience is that it certainly is not free. what am i missing?

  • @LucasDrouot
    @LucasDrouot Год назад +1

    Stable Diffusion has a lot of problems with animals' tails or long legs! (for example, with flamingos: ruclips.net/user/shortsdvgboKbetbs or with cats: ruclips.net/user/shortslwBx1-rvjcY)

  • @hipjoeroflmto4764
    @hipjoeroflmto4764 2 года назад

    We can train stable diffusion now with our own images

  • @Gounesh
    @Gounesh 2 года назад

    There’s a GUI version, which is just click and run. It’s exe though, be aware for trojans.

  • @mrrbrown86
    @mrrbrown86 Год назад +1

    Question can you use this and sell a image that u creat??🤔

    • @OkOk-tu1jo
      @OkOk-tu1jo Год назад

      As long as you didn't use a copywrited image while making your image

  • @yalheenalaurensanchez600
    @yalheenalaurensanchez600 2 года назад

    GR8 IDEA TYTY

  • @GeorgeJung19856
    @GeorgeJung19856 Год назад +1

    u can upload ur own pictures and manipulate them with dall-e. It won.

  • @kaoset1237
    @kaoset1237 Год назад

    I feel like stable defusion won't stay free forever.

  • @oatcookies_
    @oatcookies_ 2 года назад

    Thanks Brother

  • @faltutp9386
    @faltutp9386 Год назад +1

    Great video but it starts at 4:23

  • @SoarkProductions
    @SoarkProductions 2 года назад

    17 minutes than i've seen in years.

  • @jcims
    @jcims 2 года назад +1

    It's way better than DALL-E at coming up with interior design ideas.

    • @float32
      @float32 2 года назад

      Oh wow, that’s awesome. You could take a picture of your room and have it inpaint furniture and decorations!

  • @sped325
    @sped325 2 года назад

    I got an invitation to Dall E but they never sent me an sms to verify my account and the invitation ticket just expired today. Now as not very computer savy guy am trying to find how to get one of Stable diffusion builds on my laptop. There's so many different links and stuff it's confusing

    • @blankmoment2
      @blankmoment2 2 года назад

      I dont know but Stable diffusion is quite demanding. I had only seen it work ok-ish with 8gb VRAM while people with less than that struggle alot.