@@theAIsearch Just tried it. The first attempt, despite telling it not to put random figures in the background or create mutant hands, or multiple limbs, did just that. The second attempt surprisingly obeyed some commands, but did the same, as did the third attempt. It's just dumb.
The funniest thing about SDXL is that a My Little Pony model had become the gold standard for anime generation and hence forth most anime XL models are based on Pony XL
@@hipjoeroflmto4764 The author did a diligent job in training pretty much all the popular anime characters and did a great job in tagging the training data correctly, and as a result you can give keywords to numerous character references, styles and emotions and get great results out of the box.
I'm sick of censorship. I couldnt even do a hostage scene in luma, it kept knocking it back. Just ridiculous if your into horror stuff of wanna create a horror/thriller trailer.
Finally something is happening in the world of picture generation! From more than a year only existed SD and SDXL . Lately we got SD 3 ( lol) , Kolors, AuraFlow.
Just tried it. The first attempt, despite telling it not to put random figures in the background or create mutant hands, or multiple limbs, did just that. The second attempt surprisingly obeyed some commands, but did the same, as did the third attempt. It's just dumb.
Technical artist here. Blender for 3D and vector graphics for 2D. It's straightforward. Relying on AI might actually consume more of your time and effort.
@@losing_interest_in_everything Generative AI is really for dummies. I mean, come on. In order to get anything decent out of it, you have to create the artwork you want first, feed it to the machine, wait for its output, then correct what you get back, paying for your trouble. No decent artist will bite.
@@handlenumber707 Yes! In my company, I assist HR department when we have to recruit 3D artists, sound designers, and technical artists. Since the rise of AI, my job has become a lot harder. In 2018, I rejected people because their skills didn't fit our needs. Now, I reject people because they have no skills. The worst are the arrogant ones who think they can be our "thinking guys." These candidates make my day better because I get to crush their hopes of being "prompt engineers." This year, I've already rejected around 20 of them, and it was deserved!
I am a landscape architect and I have been researching the best model to help me! I have found that Midjourney so far is the best for me but after watching your video I am confident in trying SD for landscape design images!
SDXL is king because it's unfiltered It can do things better, to try and filter things in any sense removes its ability's and strengths. Always go unfiltered.
I also don't care that much about zero shot. I can break up the composition, do infill and outfill, lighting and background separately. There's a certain minimum number of steps you need to do yourself otherwise it's not really creative. IMHO YMMV.
High prompt adherence and NSFW restriction are mutually exclusive. The more precisely you can steer the model the easier it is to drive around their speedbumps.
In my testing, I've found most of them don't work well with natural human language. They won't understand wording like "next to it" or "by its side". The AI processor reads "next to it" and just places the coconut and/or drink next to whatever, maybe the tree, maybe the water, maybe the plant to the side. The generation doesn't recognize "it" as being associated with anything like the human brain would. Sometimes it may just get it right by chance, other times like the SDXL and SD3 it sees "coconut" and "drink" separately, randomizes the placement of the item. Using comma delimited with direct wording I found is still the best method, even with new tools. Don't use alternate descriptors like "there", "it", or similar wording, and don't use negatives like "no item" in the positive prompt. Polar bear wearing Hawaiian shirt, wearing sunglasses, sitting in hammock tied between palm trees, beach scene, coconut drink with red umbrella on ground
Yeah that is very true - very quickly run into those models not doing what I plan to do besides all the hype. AI should really be more pro programmed logic - it is not more.
Even if you don't even plan to generate NSFW pictures, it likes like you have to wait for user-trained versions of AuraFlow and SD3 if you want to generate something that involves some form of human limbs.
Thanks for your video. I watched just couple of minutes but when the day comes, and I need a picture of zebra playing Ice piano on the hilltop, I will definitely watch the rest…
OMG! This Ai takes the cake for me for one reason.. I asked it to generate a centaur and it delivered first try, TWICE! and with all features I asked for! Do you realize just how crazy this is? if not, ask any other AI image generator the same thing. I dare you lol
Just tried it. The first attempt, despite telling it not to put random figures in the background or create mutant hands, or multiple limbs, did just that. The second attempt surprisingly obeyed some commands, but did the same, as did the third attempt. It's just dumb. EDIT: There's no negative prompt box. Doesn't really matter as I removed those instructions and it still failed.
@@handlenumber707When you say you told it to not include those things, do you mean like in a “negative prompt” field, or that you put like, “no [thing you don’t want]” in the prompt ?
Just tried it. The first attempt, despite telling it not to put random figures in the background or create mutant hands, or multiple limbs, did just that. The second attempt surprisingly obeyed some commands, but did the same, as did the third attempt. It's just dumb.
Thank you so much for your fantastic review comparison! I was wondering about this new model, and obviously it looks pretty fantastic when it comes to following the prompt! I have been trying to do cartoons style stories, and this will obviously help tremendously! Thanks again :-)
do you mind sharing the prompt? i find that for simple prompts sdxl does a much better job. aura excels in understanding very complex prompts like the ones i demoed
In your opinion, what is the good ones for generating scenes with multiple characters consistently? I find that most struggle with comprehension even for a single character. They focus too much on rendering quality and it takes hundreds of attempts and tweaking to produce the right content in the picture. Most of time we achieve it by luck and it is hardly reproducible.
The answer is black stripes if I remember correctly. There is a rare genetic anomaly in zebras that removes the formation of their stripes, which makes them white
I tried the image, Portrait of Darth Vader enjoying some cocktails with girls in bikini on a beach. It is not bad, but at least we have some glimpses of the guy having some time away from the Emperor. Nice.
I wonder how we can get the gear going for AuraFlow. I have so much SDXL (Pony actually) Lora saved, the switch will be hard. I'll personally will follow the lead of the NSFW creators ;)
SD 1.5 is superior still, if what you want to do in NSFW. SDXL was LESS censored, and therefore it understand basic anatomy. But not to the point where it could fool anyone. Custom models made from SD 1.5 are much better at this still.
Refined models tend to not follow prompts as well as their source models, hence the prompt not being that much followed with RealVisXL as it's one of the realistic ones which are less "creative"
the model is really good for alpha but its generations seem like stacked images someone would cuttout and make in photoshop and it cant do nsfw meaning slightly exposed skin like legs or arms breaks anatomy
You used SDXL-Lightning based fine tuned model, but it follows prompts less than the base SDXL at higher steps/CFG scale. You are also using a fine tuned model, that is not as creative as the base SDXL model, due to overfitting of a limited training data set. CalicoMix is SD1.5 and not SDXL, as it states in this video. Is AuraFlow based on the SDXL architecture (like Pony Diffusion)? If so, you may want to use the specific SDXL text encoder nodes so you get the best coherency at the edges when generating different aspect ratios. So far, I haven't been very impressed by these models trained on very limited data sets. They lack creativity, because the data is so limited. They ignore words that are not in the training and negative prompts don't work as well, because they don't put the things you don't want in the training images.
oh. in the SDXL base model they did/do still have NSFW images/things in the base model? where as any SD AI models after that do not? interesting. a shame that removing them made it so much worse at understanding people.
with midjourney, you dont get failed human bodies and wrong styles. it seems to me opensource is clearly lagging far too behind proprietary models, unlike chatbots where llama3 works correctly
So the lesson I got is; use Auraflow to create an image that actually follows your prompt. Then use that image as a controlnet in SDXL to make a realistic image from it. 😁
@@fritt_wastaken hmmm dunno dude, midjourney can make insanely high quality images if your good at prompting with it. A good number or loras pretty much use MJ images for training data.
This will not replace SD 1.5. or SDXL) because it's heavily censored. So it may replace SD 3 (which is also terrible - I guess censorship requires more resources than progress in companies like this)
@@theAIsearch But what could be a use case: Using this new model just for prompt understanding and as a base for img2img workflows. I've also achieved great results with MJ base images and img2img (or img2vid) workflows in comfyUI using those images as part(!) of the input.
Great channel. Ive been following since the start. AuroFlow looks promising, but no way will it take on SD. In order for these image gens to be really reliable on a professional level, they will need to integrate tools to modify iteratively. You can't truly work on prompts alone.
So if you didnt use comfy ui , it would just be hard to run this right? like you would need to clone more of the code on github and run it through VS code ..?
WELL ACKSHULLY 😂 have you seen baby zebras that are black with white spots? Yeah, zebras are black with white stripes, so technically, SD3 got it right 😂
"A Very Tricky Font" or just a poorly written one? I agree, Auraflow is much better at creating images from shitty prompts. You literally do not tell the AI that you want the hammock tied to the two trees and then complain about it.
Hello anyone knows an AI that can generate 2D spritesheet like spritesheet to animate a horse walking or a bird flying? I need it for make simple game. Thanks in advance 🙏🙏
With how bad these models are at anatomy, what point is there in using them? Legit question. What do people use these image generators for when they can't properly generate a human body.
They're cheap, and people who can't draw think bad drawings look good. For one off images, pinups, celebrities and general throwaway artwork. the undemanding will find this useful. For actual artisans it's a source of mirth. Just tried it. The first attempt, despite telling it not to put random figures in the background or create mutant hands, or multiple limbs, did just that. The second attempt surprisingly obeyed some commands, but did the same, as did the third attempt. It's just dumb.
2:12 Uhm, no. Just no! There is definitely not an infinite number of different images for one prompt. That's just not how math works. You'll never have infinite possibilities for a finite set of bits. For n bits, there are 2^n possibilities. You would need infinite resolution or infinite color depth for infinite possibilities. Please don't use the word infinite like that when you said "in theory" before, which kind of gives it all a scientific flavor.
Thanks to our sponsor Abacus AI. Try their new ChatLLM platform here: chatllm.abacus.ai/?token=aisearch
👋
Am I the only one that noticed that CalicoMix is not SDXL? If only we had an AI assistant to double check these things!
Ir is crazy how this just came out of the box with so much potential, really wanna see what this will be in some time.
yep, looking forward to what the open source community can build from this
@@theAIsearch Just tried it. The first attempt, despite telling it not to put random figures in the background or create mutant hands, or multiple limbs, did just that. The second attempt surprisingly obeyed some commands, but did the same, as did the third attempt. It's just dumb.
The funniest thing about SDXL is that a My Little Pony model had become the gold standard for anime generation and hence forth most anime XL models are based on Pony XL
OH, so that's what they mean when they say pony based
Makes sense those mlp characters have giant anime eyes
So it should be good at gmod images too then
@@hipjoeroflmto4764 The author did a diligent job in training pretty much all the popular anime characters and did a great job in tagging the training data correctly, and as a result you can give keywords to numerous character references, styles and emotions and get great results out of the box.
@@hipjoeroflmto4764 its because the person that made the model has a good training dataset, nothing to do with eye size
Following you since months
Love your work
Thanks for keeping us updated about the latest AI technology
Much appreciated!
a new KING .. cant do NSFW .........
until... you make it
😭
Breh
> open source image gen
> cant do NSFW
The universe is morphing itself to make this statement false as we speak
I'm sick of censorship. I couldnt even do a hostage scene in luma, it kept knocking it back. Just ridiculous if your into horror stuff of wanna create a horror/thriller trailer.
Finally something is happening in the world of picture generation!
From more than a year only existed SD and SDXL . Lately we got SD 3 ( lol) , Kolors, AuraFlow.
Don't forget Playground
Right on! As a solo indie game dev, its incredible what I can do with AI tools now versus even just a year ago. Total... game changer.
Just tried it. The first attempt, despite telling it not to put random figures in the background or create mutant hands, or multiple limbs, did just that. The second attempt surprisingly obeyed some commands, but did the same, as did the third attempt. It's just dumb.
@@handlenumber707skill issue tbh
Technical artist here. Blender for 3D and vector graphics for 2D. It's straightforward. Relying on AI might actually consume more of your time and effort.
@@losing_interest_in_everything Generative AI is really for dummies. I mean, come on. In order to get anything decent out of it, you have to create the artwork you want first, feed it to the machine, wait for its output, then correct what you get back, paying for your trouble. No decent artist will bite.
@@handlenumber707 Yes! In my company, I assist HR department when we have to recruit 3D artists, sound designers, and technical artists. Since the rise of AI, my job has become a lot harder. In 2018, I rejected people because their skills didn't fit our needs. Now, I reject people because they have no skills. The worst are the arrogant ones who think they can be our "thinking guys." These candidates make my day better because I get to crush their hopes of being "prompt engineers."
This year, I've already rejected around 20 of them, and it was deserved!
the stable diffusion zebra would totally be something we'd see on an album cover
I am a landscape architect and I have been researching the best model to help me! I have found that Midjourney so far is the best for me but after watching your video I am confident in trying SD for landscape design images!
Best of luck! SD with controlnet is great for landscape design
12:00 The way it separated coconut and drink is quite funny😂
SDXL is king because it's unfiltered
It can do things better, to try and filter things in any sense removes its ability's and strengths.
Always go unfiltered.
I also don't care that much about zero shot.
I can break up the composition, do infill and outfill, lighting and background separately.
There's a certain minimum number of steps you need to do yourself otherwise it's not really creative. IMHO YMMV.
it's filtered, just less heavily. If you try to make NSFW, you'll be much better off with a custom model based on SD 1.5
I don't understand why not train the model with the NSFW images and then censor them after generation. Would make for far better models.
Less risk of letting some through.
Cool. Waiting for the beginners tutorial.
birb
So hype for a new competitor in the field
finally!
High prompt adherence and NSFW restriction are mutually exclusive. The more precisely you can steer the model the easier it is to drive around their speedbumps.
In my testing, I've found most of them don't work well with natural human language. They won't understand wording like "next to it" or "by its side". The AI processor reads "next to it" and just places the coconut and/or drink next to whatever, maybe the tree, maybe the water, maybe the plant to the side. The generation doesn't recognize "it" as being associated with anything like the human brain would. Sometimes it may just get it right by chance, other times like the SDXL and SD3 it sees "coconut" and "drink" separately, randomizes the placement of the item.
Using comma delimited with direct wording I found is still the best method, even with new tools. Don't use alternate descriptors like "there", "it", or similar wording, and don't use negatives like "no item" in the positive prompt.
Polar bear wearing Hawaiian shirt, wearing sunglasses, sitting in hammock tied between palm trees, beach scene, coconut drink with red umbrella on ground
Interesting, thanks for sharing!
Yeah that is very true - very quickly run into those models not doing what I plan to do besides all the hype.
AI should really be more pro programmed logic - it is not more.
Even if you don't even plan to generate NSFW pictures, it likes like you have to wait for user-trained versions of AuraFlow and SD3 if you want to generate something that involves some form of human limbs.
If it is open source, then I hope the comunnity would be able to finetune unlike SD 3.
im expecting a ton of fine tuned models for auraflow to come out soon!
Thanks for your video. I watched just couple of minutes but when the day comes, and I need a picture of zebra playing Ice piano on the hilltop, I will definitely watch the rest…
I swear to God this is the best AI channel. Peace be upon you, good sir
Thanks!
SDXL baby!
+1
+1
Link 🔗 please
OMG! This Ai takes the cake for me for one reason..
I asked it to generate a centaur and it delivered first try, TWICE! and with all features I asked for! Do you realize just how crazy this is? if not, ask any other AI image generator the same thing. I dare you lol
yes, this one is great for prompt following
Just tried it. The first attempt, despite telling it not to put random figures in the background or create mutant hands, or multiple limbs, did just that. The second attempt surprisingly obeyed some commands, but did the same, as did the third attempt. It's just dumb. EDIT: There's no negative prompt box. Doesn't really matter as I removed those instructions and it still failed.
@@handlenumber707When you say you told it to not include those things, do you mean like in a “negative prompt” field, or that you put like, “no [thing you don’t want]” in the prompt ?
@@drdca8263 There's no negative prompt box. Doesn't really matter as I removed those instructions and it still failed.
Pony diffusion makes centaurs just fine.
Ask my Lillia folder ..What?
Just tried it. The first attempt, despite telling it not to put random figures in the background or create mutant hands, or multiple limbs, did just that. The second attempt surprisingly obeyed some commands, but did the same, as did the third attempt. It's just dumb.
skill issue
@@fritt_wastaken It's like some con artist trying to con an actual artist to pay to hand over his work. It might work on a child, perhaps, maybe.
@@handlenumber707these channels don't want bad or realistic feedback, they just want quick positive build up, don't trust them for anything reliable.
@@joelrobinson5457 I trust NO RUclips channel. They exist be corrected only.
It's still very early in its training, thus the 0.1. Given that it's a 6.8B param model, almost triple SD3, fine tuning should be a beast.
None of these are worth a damn until we can train them locally. That is when the magic happens.
Generation uses a tiny fraction of compute needed for training. You wouldn't live long enough to train anything decent in this domain on your home pc.
@@user-qn6kb7gr1d Then it is a dead mofo.
Great video and model. THX for sharing. The most important information comes at about 20 min in the video btw :D
Thanks for watching!
It's not that good currently, but I'm glad to see another open source model come out.
Thank you so much for your fantastic review comparison! I was wondering about this new model, and obviously it looks pretty fantastic when it comes to following the prompt! I have been trying to do cartoons style stories, and this will obviously help tremendously! Thanks again :-)
You're welcome, and thanks for sharing!
25:30 CalicoMix is a SD1.5 finetune, not SDXL. Wich is even more embarassing for SD3 and AuraFlow.
So, AuraFlow for initial generation, SDXL for refining
I hope someone makes something like Fooocus for Aura flow.
Dont forget that SD with likely SD3(?) is going to have a non so open source friendly structure, so may as well test dalle-3 too.
I feel like SD just focused on realism while AuraFlow on accuracy
this thing can't generate a simple image of an EMPTY room! SD 2.1 can do this easily...
love it
I tested it and it's not that great. It didn't follow my prompt at all. All others I tested follow my prompt.
do you mind sharing the prompt? i find that for simple prompts sdxl does a much better job. aura excels in understanding very complex prompts like the ones i demoed
In your opinion, what is the good ones for generating scenes with multiple characters consistently?
I find that most struggle with comprehension even for a single character. They focus too much on rendering quality and it takes hundreds of attempts and tweaking to produce the right content in the picture. Most of time we achieve it by luck and it is hardly reproducible.
All good except one point. Does a Zebra have White Stripes, or Black Stripes? Why?
😵💫
The answer is black stripes if I remember correctly.
There is a rare genetic anomaly in zebras that removes the formation of their stripes, which makes them white
dazzle camo
Zebra is black with white stripes. Why? Evolution. To be precise camouflage, thermoregulation, and some other things.
I tried the image, Portrait of Darth Vader enjoying some cocktails with girls in bikini on a beach. It is not bad, but at least we have some glimpses of the guy having some time away from the Emperor. Nice.
I am here to study cognition. I need to elicit fascination from a compelling subject.
I wonder how we can get the gear going for AuraFlow. I have so much SDXL (Pony actually) Lora saved, the switch will be hard.
I'll personally will follow the lead of the NSFW creators ;)
pony is awesome! it's a shame that auraflow is censored
Why didn't you try higher guidance setting with the zebra?
Playground 2.5 should be a better alternative, with good anatomy images.
SD 1.5 is superior still, if what you want to do in NSFW. SDXL was LESS censored, and therefore it understand basic anatomy. But not to the point where it could fool anyone. Custom models made from SD 1.5 are much better at this still.
looks promising good video
thanks!
Refined models tend to not follow prompts as well as their source models, hence the prompt not being that much followed with RealVisXL as it's one of the realistic ones which are less "creative"
That was great
thanks
the model is really good for alpha but its generations seem like stacked images someone would cuttout and make in photoshop and it cant do nsfw meaning slightly exposed skin like legs or arms breaks anatomy
is there more sources to convert images to text is very helpful
this guy is Fireship on god
You were using 50 steps for the first couple of images but only 28 after you switched to hf, that may have made images less detailed
Do you know about AI that describes what it sees in the footage?
Why does zebra don't have a tail yet😂?
You used SDXL-Lightning based fine tuned model, but it follows prompts less than the base SDXL at higher steps/CFG scale. You are also using a fine tuned model, that is not as creative as the base SDXL model, due to overfitting of a limited training data set. CalicoMix is SD1.5 and not SDXL, as it states in this video.
Is AuraFlow based on the SDXL architecture (like Pony Diffusion)? If so, you may want to use the specific SDXL text encoder nodes so you get the best coherency at the edges when generating different aspect ratios. So far, I haven't been very impressed by these models trained on very limited data sets. They lack creativity, because the data is so limited. They ignore words that are not in the training and negative prompts don't work as well, because they don't put the things you don't want in the training images.
following you from alot of months, just want to let you know you are a fkin GOAT in AI teaching.
oh. in the SDXL base model they did/do still have NSFW images/things in the base model? where as any SD AI models after that do not? interesting. a shame that removing them made it so much worse at understanding people.
Thank you!!!
You're welcome!
just started watching. i dont know it looks bad to me
with midjourney, you dont get failed human bodies and wrong styles.
it seems to me opensource is clearly lagging far too behind proprietary models, unlike chatbots where llama3 works correctly
So the lesson I got is; use Auraflow to create an image that actually follows your prompt. Then use that image as a controlnet in SDXL to make a realistic image from it. 😁
Do you have an example of that workflow?
awesome idea!
@@nicktumi he just said the workflow lol, if u didn't notice I think you got other things to worry about
@@KDawg5000 duo you have a error for comfyui for this example?
@@hipjoeroflmto4764 y r u gA
good video
Hi anything you suggest for for lip sync ai tool open-source, my use case is i have a video in english i want to convert it into hindi lipsync
I use midjourney, how does midjourney stack up to this? Is midjourney outdated now or considered low end?
It's still considered to be the best/easiest to use, I believe
Midjourney blows this out of the water easily.
mj is great but it's closed & paid. i'm only comparing open source here
@@Astro-uv1xq Midjourney was never better than SD.
Easier to use, yes. But not really useful
@@fritt_wastaken hmmm dunno dude, midjourney can make insanely high quality images if your good at prompting with it. A good number or loras pretty much use MJ images for training data.
16GB? No smaller version?
thanks
stable cascade base model is better than sd3 and sdxl base model, probably should've used that instead
This will not replace SD 1.5. or SDXL) because it's heavily censored. So it may replace SD 3 (which is also terrible - I guess censorship requires more resources than progress in companies like this)
thanks for sharing. agree that censorship is a major limitation here
@@theAIsearch But what could be a use case: Using this new model just for prompt understanding and as a base for img2img workflows. I've also achieved great results with MJ base images and img2img (or img2vid) workflows in comfyUI using those images as part(!) of the input.
Great channel. Ive been following since the start. AuroFlow looks promising, but no way will it take on SD. In order for these image gens to be really reliable on a professional level, they will need to integrate tools to modify iteratively. You can't truly work on prompts alone.
Thanks for sharing!
Replace Stable Diffusion ? 😮
Only for some cases
It's comical we use anime thumbnails than ai ones
i guess you still need controlNet and sdxl.
So if you didnt use comfy ui , it would just be hard to run this right? like you would need to clone more of the code on github and run it through VS code ..?
correct, the only way that was mentioned in their blog was comfyui. you could try duplicating and tweaking one of the online spaces
I say the next vid is about llama 3.1 and a lot about 405b
WELL ACKSHULLY 😂 have you seen baby zebras that are black with white spots? Yeah, zebras are black with white stripes, so technically, SD3 got it right 😂
May be someone can try fine-tuning it for NSFW task?
I don't think it is that good if sd3 was uncensored and with fine tuning will crush all of them 😢
its a safetensor file, wonder if i can just plug it into my comfyui folder and use it from there?
Try that! ;-)
No "safetensor" is a file format, it have nothing to do with the training data. SD 1.5 is actually the best still for anatomy.
Still no nsfw
😢
is abacus good?
"A Very Tricky Font" or just a poorly written one? I agree, Auraflow is much better at creating images from shitty prompts. You literally do not tell the AI that you want the hammock tied to the two trees and then complain about it.
Can foocus ai platform use sdxl? Is there a open source for sdxl?
almost every existing software does support sdxl
It makes poor images of tuxedo mask compared to others I use.
AROUND A WEEK AGO WEEK AGO
Hello anyone knows an AI that can generate 2D spritesheet like spritesheet to animate a horse walking or a bird flying? I need it for make simple game. Thanks in advance 🙏🙏
Sdxl’s yoga pose wasn’t perfect the woman’s foot is flipped 😂
The king has no clothes lol
😏
training on synthetic data is sketchy at best.
copilot ai is the best
It run on DALL-e 3 💀
SD3 often looks like the worst of both worlds, I don't see the point of this model
true
With how bad these models are at anatomy, what point is there in using them? Legit question. What do people use these image generators for when they can't properly generate a human body.
They're cheap, and people who can't draw think bad drawings look good. For one off images, pinups, celebrities and general throwaway artwork. the undemanding will find this useful. For actual artisans it's a source of mirth. Just tried it. The first attempt, despite telling it not to put random figures in the background or create mutant hands, or multiple limbs, did just that. The second attempt surprisingly obeyed some commands, but did the same, as did the third attempt. It's just dumb.
Ideogram wins by a longshot
Sorry bro, SDXL is still king
yep, that's my preferred model for now
i really enjoy your content, i can't wait to hear what you have in store for us about Flux :D
Thanks! Coming soon!
Why girl punches man?
she has androphobia
and now we have flux
yep. bye auraflow
one more
2:14 : not infinite. There are only finitely many possible images of a particular resolution and color depth.
given the seed count think its some like half a quintillion.
The white parts on a zebra are stripes too, not only the black ones, so i guess sdl kinda did ehat you asked kk
Haha. Yeah that tripped me up too
🤯
Welhome
2:12 Uhm, no. Just no! There is definitely not an infinite number of different images for one prompt. That's just not how math works. You'll never have infinite possibilities for a finite set of bits. For n bits, there are 2^n possibilities. You would need infinite resolution or infinite color depth for infinite possibilities.
Please don't use the word infinite like that when you said "in theory" before, which kind of gives it all a scientific flavor.
It's fine. Seed in theory can be from 0 to infinity. Some generations would not be unique and that's it.
EARLY GANG
😃