Thank you for being one OF THE FEW that don't sensationalize your video to get views,. Just honest straight forward. Don't change....you will prevail much farther than all the others :)
Thank you for always making this easy for those of us who are not programmers. I breathed a sigh of relief when I found this video amongst the lesser install vids where they always assume everybody has a professional-level knowledge of doing pulls and tweaking Python settings. So thank you!
I was getting really poor results, and I had a tough time trying to find out why. Thanks to you, I realized I was generating images in 512x512, as I did on SD1.5. I appreciate it! 🙌
Thank you so much for mentioning that the A1111 needed updating to work with the new SDXL, that fixed my problem with it not working. Nerdy Rodent, did a great video, but made no mention of this!
Always love your videos. Easy to follow and I like how you explain what you are doing while you're doing it....makes things much more understandable. Thanks for all your hard work!!! Love your accent too! Haha!
This model checkpoint doesnt load for me, always get back to de last one i have used... you know why? Failed to load checkpoint, restoring previous size mismatch for model.diffusion_model.output_blocks.8.0.skip_connection.bias: copying a param with shape torch.Size([320]) from checkpoint, the shape in current model is torch.Size([640]).
11:15 you don't HAVE to use 1024x1024. In my testing, SDXL can generate images as low as 768x768 without suffering severe quality loss. This is useful for lower VRAM systems, like my 8GB laptop, since it allows you to generate 768x1024 (portrait) or 1024x768 (landscape) images while saving a little bit of vram. I have even done 1280x720 images featuring a kitten wearing knight armor with very little quality loss compared to 1280x1024.
How are you doing it? Its apparently not working on my 8gb Rtx 3070, is there any other way to make it work? For me it only works if i write 2 word prompts and just makes awful results
Is speed of 768x768 image generation the same with sdxl1.0 vs sd1.5 or is it noticeably longer? Will you be able to upscale that 768x768 to 1024x1024 in img2img using the refiner with similar result?
@@CoffeeAddictGuy SDXL in my testing is about 25-30% slower than SD1.5, but it's hard for me to get exact numbers since I have to use --medvram, which does change the performance a bit. Also, yes, you can render 768x768 with the base model and upscale it with the refiner to get better details in img2img. If you go to Civit and search @frankenburger you can find my test images that are a result of this method along with their metadata
@@Daddy.please97 I'm not sure what you mean by awful results, but I'm not using a1111 any differently with SDXL than I was with SD1.5. Without being able to see your settings, I'd like to suggest going over to Civit and searching @frankenburger. I posted sample images that were rendered at 768x768 using the base model (and then upscaled them using the refiner) and included their metadata for your reference.
I've created a simple extension that uses the refiner for the hires fix pass but it requires a minimum of 32gb ram (not vram). But lets hope that we'll get a native way to use the refiner in A1111
Great seeing you improve as you create each episode. Following you for quite some time now and it's obvious you really put a lot of effort in your moderation skills. This said, this episode was one of the most entertaining ones I ever saw on your channel. Keep it up! You rock! 👍🏻🚀
The refiner model should not be used for img2img. It's made to work with LEFTOVER NOISE from the base model. The refiner does not work very well on Gaussian noise added to a fully completed image. You need to wait for auto1111 to support the refiner model to use the it correctly, or switch to a comfyui workflow that uses it correctly. You can see it working in this video, adding some detail, but it has little understanding of the image so it also morphs the skin texture completely.
I’m not understanding your description of the proper use for the refiner model. I also don’t see what you say it’s doing to skin texture. Honestly both images at 15:02 have a very airbrushed / unrealistic painted style with little to no texture in the skin. Not the most impressive example image.
Looks interesting so far. Will be interesting to see how it will eventually work out with deforum and similar extensions once they start to get updated.
Olivio, based on your instructions I installed Invoke AI 3.0. Stable Diffusion XL 1.0 works quite well even though my GPU is not great, only 12GB RAM - the image generation is fast. I use Midjourney prompts and in some cases achieve even better results. Thank you for the tips.
@OlivioSarikas The charts were generated with results from blind image generation. People could vote on thousands of images. The choices were side by side and nobody knew which models were being used to generate the images. This is how those results were generated in the graph.
I know, but that still doesn't tell us much about how good the really is. Of course it is better than SD 1.5, but we still have to see if the trained models from the community are far better or just a little better. Of course also the improvements will be less big, but more precises over time.
Finally! I was waiting for this video. I tried updating A1111 today to v1.5 on my own but it works way slower for me for some reason. Even for normal SD1.5 checkpoints. I was hoping to see you talk about the command line arguments and tips for them but I guess you skipped that part.
You're probably running out of memory... it slows way down if the total video memory used is bigger than your GPU's vram. SDXL might *run* on 8GB of VRAM, but it's not happy until you have 16GB, at least right now. I think I've read that comfy UI does better, so you could try that.
@@Steamrick I have 12GB of vram but I wasn't even talking about SDXL. The regular SD1.5 checkpoints all work 5 times longer after update to A1111v1.5 than what I was using before which is A1111v1.3.
@@Macieks300I had this happen before and auto1111 wasn't using the gpu. I had to update the python torch files ro latest version, it was ok after. Search for a tutorial on it.
For me, the best way of using SDXL so far is to set up a second copy of Auto1111 alongside my normal version - run it clean with no extensions installed, and the --no-half-vae --opt-sdp-attention --medvram command line options. Works like a charm, and it's pretty fast as long as you have the right drivers installed.
@@okachobe1 I have no clue what that means... Do you mean zip the current SD folder I have so that it can't be affected, install SD again, and then what? Can I unzip my folder containing my original SD and use one for SDXL and one for the previous?
I'm lucky this year. Another superb discovery of a channel 🙂 Olivio, I have one question - can I use my own image source to generate a graphic where Im situated on a different planet and drinking some fine gin? 😅
Very informative video, thanks. Unfortunatley as a linux user I have tested this on 2 seperate machines on automatic1111 and had consistant errors with loading the refiner model. Hope that it gets sorted soon as it looks great :)
There were discussions about the SDXL 1.0 VAE and how it created some strange artifacts, a lot of people seem to recommend to use the SDXL 0.9 VAE with SDXL 1.0 to avoid those issues. Maybe worth a try to see if you still get those problems with the eyes when not using face restore.
So far I didn't notice any difference when I use SDXL 1.0 VAE compared to when I use None. I have Restore Faces checked. If I don't then faces suddenly all turn blue and distorted in the last Step using the 1.0 VAE. Gotta try 0.9 VAE then.
You can see why Midjourney want v6 out as it will intrepert text better than v5, which seems to be one of the big features of SDXL. MJ is still generally better even at v5 but SDXL is very very close now.
Tnx for the update, love this channel. Question: I'm at hugging face and i see a file uploaded two days ago, should I install "1.0_0.9vae.safetensors" or "1.0.safetensors"
You and Sebastian are my favorites. The real AI OG's in the industry. How can I trust you to not be an AI Olivio? Real question. Where do you see what version of SD you're running? I keep trying to update, but it says I'm already up to date, fine, but how my SDXL isn't working. It just load :(
Amongst other places, it'll spit out the version number in the cmd dialogue right beneath the python version and all of the version numbers are listed at the bottom of the A1111 UI. I really don't see how you could possibly have difficulty finding them?!
I had that problem and the solution I found was to do a fresh install with the latest A1111 webUI. Even if it says you're up to date, you aren't. You'll know for definite as it will tell you it's 1.5.0 or 1.5.1 where Steamrick mentions. You don't need to get Python again but remember to add it's path to the webui-user.bat. @Steamrick The A111 webui doesn't show it's version number for all installs. For older installs, those UI numbers you mention didn't show the UI version, they started with Python. 🙂
Excellent tutorial and information @OlivioSarikas but I was wondering, and if is not too much to ask... What is the Automatic 1111 that you are running, because I don't have some of the slide controllers (like the Clip Skip), and option windows (SD VAE), that you showed in this video. Is this part of a default installation of Automatic 1111, or are there extensions that you have to install? One more time thank you for your excellent videos. Take Care.
Noticed my WebUI was just an unpacked zip-file, so I went to start over with cloning down the repo this time to easier keep it updated. Taking a while, but hopefully it will be worth it. The safetensors also take a long time to download so I imagine a lot of people are downloading them to enjoy themselves with AI art.
Great stuff really helpful - as a photographer and designer getting into AI generation its amazing to get these supporting tutorials - For me my desktop AMD PC with NVIDIA Super 1650 is very slow using Automatic 1111 it takes around 20 minutes to get one image generated and so this is not practical for me so im looking at options of updating spending £300 on new GPU or other alternatives
@@Resmarax hi yes it was - super slow - its faster to use somehting like clipdrop but its rather restrictive - will see about spending out on a new GPU and read up if my AMD processor has any influence
I tried the prompt, "a perfectly normal man looking at his perfectly normal hands"... SDXL produced a handsome man with seven fingers and three thumbs!
When I select the base model I see the following error and the model auto reverts to previous loaded model: Loading weights [31e35c80fc] from F:\stable-diffusion-webui\models\Stable-diffusion\sd_xl_base_1.0.safetensors Failed to load checkpoint, restoring previous + size mismatch for model.diffusion_model.output_blocks.8.0.skip_connection.bias: copying a param with shape torch.Size([320]) from checkpoint, the shape in current model is torch.Size([640]).
I saw this behavior when selecting the base VAE model named sd_xl_base_1.0_0.9vae.safetensors in Automatic 1111. But using sd_xl_base_1.0.safetensors doesn't cause that for my PC.
Thank you for the video, you saved me a minute of rendering by using this "HACKER" method. I was struggling for this issue, since I have 8 gb VRAM card, and the rendering takes like 1,5 minute to render on the base model, which led me to use ComfyUI which is much faster and stabler. But it needs some learning to add nods for refiner and Lora etc.. But you saved me man thank you! I am waiting for your videos on how to train our own Loras and models for this specific model, is it the same like before? Is there any change? All this will keep me busy watching all upcoming videos about this new version.
Thank you Olivio for your great support and detailed explanation. One question, when trying to download the SDXL base model and the SDXL Refiner, I can see that today exists now the versions including VAE for both models. Is there any suggestion from your side to select the VAE or not VAE version to download and use?. Thank you.
Well one thing not mentioned is how much GPU you need. I have RTX 2060 Super with 8G VRAM. WIll it do? Should i use something else? I don't actually care that much about anatomical correctness, photorealism and number of fingers, or about high resolutions above 768x768, but i want interesting varied results.
thank you for this video I got error when using SDXL size mismatch for model.diffusion_model.output_blocks.8.0.skip_connection.bias: copying a param with shape torch.Size([320]) from checkpoint, the shape in current model is torch.Size([640]) please help
Updated to A1111 1.8.0, when loading SDXL check point getting error. size mismatch for model.diffusion_model.output_blocks.8.0.skip_connection.bias: copying a param with shape torch.Size([320]) from checkpoint, the shape in current model is torch.Size([640]). I'm assuming something broke related to SDXL in this latest A111 update. Any suggestions ?
It refuses to allow me to select SDXL 1.0 or the refiner. It takes a long time and then kicks it back to the last model I had loaded. Anyone else having this problem? Tons of size mismatches and torch errors in cmd window.
Great video, right now I'm enjoying SDXL a lot and getting some very decent stuff generated. Though things will 'go nuclear' once we have a series of custom models out and ControlNet works with it. My only fear is VRAM. I can't wait to do Dreambooth with this, but I assume that if we're using 1024x1024 as source images then VRAM requirements are going to skyrocket?
The images you are showing I'm wandering if it's just my screen but they appear to have vertical lines on all around the edges of the characters also through the characters. Is this a shortcoming of the AI or perhaps the way its transferred to this media? Also the image at 2:50 an improvement would be for the AI to render in the foot prints from the dog running and jumping in the sand, at the speed it would have to be running there would in a real pic be prints left in sand and also more than likely sand flung out from the dogs back paws. Just some observations.
Oh my Gosh, it can spell correctly...... I tried this"James Dean as a character in a comic book magazine, front cover, very colorful, highly detailed, super sharp, with the words JAMES DEAN on cover"
Thanks for your content. It has been so helpful. I have been a pro photographer for more than 35 years and this AI image creating is one of the most interesting things that has happened during my career. It will one new thing that I will be using in my professional work. I am so happy that I got to see this development in image making. I tested the SDXL 1.0 model and I am having hard time to get a natural looking skin. The images looks nice and in most cases the anatomy is next to perfect, but the skin looks too plastic. I personally like the old fashion natural looking skin. Any ideas?
I experiment with sdxl a bit and found that refiner model can also be used for upscale (I used Ultimate SD), I set tile size to 1024x1024. It can not only upscale image, but add a lot of details to image.
SDXL is huge, heavy and it's already excluding lots of users with low/mid-end machines - the world is big and there are users with all kinds of hardware and internet connections. There are LoRas with almost 1GB being produced. Yeah, "show must go on", but I see developers jumping on XL wagon and quickly dropping SD1.5 with some concern.
Hey Olivio, I'm looking since long time for an audio TTS AI which I can run locally (similar to A1111 or even CLI) with german voices. Do you know what I could use? Would be perfect I there would be a way to train my own voice model, too.
Thanks for the info Olivio, I am trying to understand if there is a way we can use one of the model to batch process multiple images with the same treatment to get similar results on all the images of the same batch, any guidance would be highly appreciated.
Is it normal for the model to take 5 minutes to load? It works fine once it's loaded, renders in about 20s for 1024x1024 but base model and refiner model can both take up 5min to actually load which is annoying. Automatic1111 and all models are installed on an ssd so idk. Is this the same for everyone else?
Thank you for being one OF THE FEW that don't sensationalize your video to get views,. Just honest straight forward. Don't change....you will prevail much farther than all the others :)
thank you :)
@@OlivioSarikas But i wanted you to click bait me with uncensored stuff and telling me its better than anything you've ever seen. ;)
Totally agree...really starting to cringe over those videos
Ironic. This used to be Affinity Photo channel but when AI showed up, almost over night he saw where views are. So... "dont change AGAIN" maybe :D
Thank you for always making this easy for those of us who are not programmers. I breathed a sigh of relief when I found this video amongst the lesser install vids where they always assume everybody has a professional-level knowledge of doing pulls and tweaking Python settings. So thank you!
Olivio, I love your energy
thank you :)
Thanks! I was looking for someone who is actually talking about it and not just trying to hype it up.
that refiner trick is awesome. Very cool discovery. Always appreciate you Olivio
My pleasure!
I was getting really poor results, and I had a tough time trying to find out why. Thanks to you, I realized I was generating images in 512x512, as I did on SD1.5. I appreciate it! 🙌
I loved the Hacker Olivio, lol , doing great work as usual.
Thank you. Hacker BOI might come around more often :)
On a lark I tried the Comfy UI install and I'm very glad i did. Beautiful.
I installed Comfy two days ago and I'm pretty addicted with some of the custom stuff I'm doing. It's great.
SDXL doesn't have controlnet yet.
Thank you so much for mentioning that the A1111 needed updating to work with the new SDXL, that fixed my problem with it not working. Nerdy Rodent, did a great video, but made no mention of this!
you are welcome :)
Always love your videos. Easy to follow and I like how you explain what you are doing while you're doing it....makes things much more understandable. Thanks for all your hard work!!! Love your accent too! Haha!
Thank you, my friend
This model checkpoint doesnt load for me, always get back to de last one i have used... you know why?
Failed to load checkpoint, restoring previous
size mismatch for model.diffusion_model.output_blocks.8.0.skip_connection.bias: copying a param with shape torch.Size([320]) from checkpoint, the shape in current model is torch.Size([640]).
#### Links from the Video ####
SDXL 1.0 Announcement: stability.ai/blog/stable-diffusion-sdxl-1-announcement
SDXL 1.0 Base and Lora: huggingface.co/stabilityai/stable-diffusion-xl-base-1.0/tree/main
SDXL 1.0 Refiner: huggingface.co/stabilityai/stable-diffusion-xl-refiner-1.0/tree/main
Stability Image: twitter.com/StableDiffusion/status/1684254689250902025
Nerdy Rodert Image: twitter.com/NerdyRodent/status/1684506233334538246
OrctonAI Images: twitter.com/OrctonAI/status/1684344552654610434
Great tip for that 1111 update in the .bat file, it wasn't working for me with loads of errors but does now! Thanks!
11:15 you don't HAVE to use 1024x1024. In my testing, SDXL can generate images as low as 768x768 without suffering severe quality loss. This is useful for lower VRAM systems, like my 8GB laptop, since it allows you to generate 768x1024 (portrait) or 1024x768 (landscape) images while saving a little bit of vram. I have even done 1280x720 images featuring a kitten wearing knight armor with very little quality loss compared to 1280x1024.
How are you doing it? Its apparently not working on my 8gb Rtx 3070, is there any other way to make it work? For me it only works if i write 2 word prompts and just makes awful results
Is speed of 768x768 image generation the same with sdxl1.0 vs sd1.5 or is it noticeably longer?
Will you be able to upscale that 768x768 to 1024x1024 in img2img using the refiner with similar result?
@@Daddy.please97write --medvram in start bat file, that worked for me
@@CoffeeAddictGuy SDXL in my testing is about 25-30% slower than SD1.5, but it's hard for me to get exact numbers since I have to use --medvram, which does change the performance a bit.
Also, yes, you can render 768x768 with the base model and upscale it with the refiner to get better details in img2img. If you go to Civit and search @frankenburger you can find my test images that are a result of this method along with their metadata
@@Daddy.please97 I'm not sure what you mean by awful results, but I'm not using a1111 any differently with SDXL than I was with SD1.5. Without being able to see your settings, I'd like to suggest going over to Civit and searching @frankenburger. I posted sample images that were rendered at 768x768 using the base model (and then upscaled them using the refiner) and included their metadata for your reference.
Thanks man been waiting to get into A1111 for a while this got me there
First videos of yours I've seen. Instant sub and like. GREAT WORK!
Pro Tip: Use Happy Diffusion if you don't have a powerful GPU.
I've created a simple extension that uses the refiner for the hires fix pass but it requires a minimum of 32gb ram (not vram). But lets hope that we'll get a native way to use the refiner in A1111
@@user-jm4cd5sd1x a1111 need much more ram for sdxl then comfy ui on my system..
Great seeing you improve as you create each episode. Following you for quite some time now and it's obvious you really put a lot of effort in your moderation skills. This said, this episode was one of the most entertaining ones I ever saw on your channel. Keep it up! You rock! 👍🏻🚀
The refiner model should not be used for img2img. It's made to work with LEFTOVER NOISE from the base model. The refiner does not work very well on Gaussian noise added to a fully completed image. You need to wait for auto1111 to support the refiner model to use the it correctly, or switch to a comfyui workflow that uses it correctly. You can see it working in this video, adding some detail, but it has little understanding of the image so it also morphs the skin texture completely.
I’m not understanding your description of the proper use for the refiner model. I also don’t see what you say it’s doing to skin texture. Honestly both images at 15:02 have a very airbrushed / unrealistic painted style with little to no texture in the skin. Not the most impressive example image.
@@johnnyc.31SD next has option for using refiner or not. It's (probably) coming to A1111 soon. img2img is not the way it's meant to be used.
Yes, but this is the only way to use it in A1111 for now. So it's better than nothing for people who don't want to use a different UI.
Yea it does not feel it does much in which UI is it supported to work correctly ?
@@openroomxyz StableSwarmUI or comfyui
Looks interesting so far. Will be interesting to see how it will eventually work out with deforum and similar extensions once they start to get updated.
Olivio, based on your instructions I installed Invoke AI 3.0. Stable Diffusion XL 1.0 works quite well even though my GPU is not great, only 12GB RAM - the image generation is fast. I use Midjourney prompts and in some cases achieve even better results. Thank you for the tips.
awesome!
@OlivioSarikas The charts were generated with results from blind image generation. People could vote on thousands of images. The choices were side by side and nobody knew which models were being used to generate the images. This is how those results were generated in the graph.
I know, but that still doesn't tell us much about how good the really is. Of course it is better than SD 1.5, but we still have to see if the trained models from the community are far better or just a little better. Of course also the improvements will be less big, but more precises over time.
Finally! I was waiting for this video. I tried updating A1111 today to v1.5 on my own but it works way slower for me for some reason. Even for normal SD1.5 checkpoints. I was hoping to see you talk about the command line arguments and tips for them but I guess you skipped that part.
You might try to delete you venv folder and restart A1111. It takes a while to set everything up again. also try to use --xformers in the command args
@@OlivioSarikas Same here speed for me is terrible takes between 4 and 5 minutes for one image even with xformers activated
You're probably running out of memory... it slows way down if the total video memory used is bigger than your GPU's vram.
SDXL might *run* on 8GB of VRAM, but it's not happy until you have 16GB, at least right now. I think I've read that comfy UI does better, so you could try that.
@@Steamrick I have 12GB of vram but I wasn't even talking about SDXL. The regular SD1.5 checkpoints all work 5 times longer after update to A1111v1.5 than what I was using before which is A1111v1.3.
@@Macieks300I had this happen before and auto1111 wasn't using the gpu. I had to update the python torch files ro latest version, it was ok after. Search for a tutorial on it.
For me, the best way of using SDXL so far is to set up a second copy of Auto1111 alongside my normal version - run it clean with no extensions installed, and the --no-half-vae --opt-sdp-attention --medvram command line options. Works like a charm, and it's pretty fast as long as you have the right drivers installed.
--xformers might help with the speed as well with that setup
How do you go abouts making a second copy of A1111? Did you just reinstall A1111 in a new directory?
@@gulfblue zip the original folder and then install the new install
@@okachobe1 I have no clue what that means...
Do you mean zip the current SD folder I have so that it can't be affected, install SD again, and then what? Can I unzip my folder containing my original SD and use one for SDXL and one for the previous?
Thank you! Needed some easy to understand update info for Automatic1111
Thank you for your consistently informative videos, so easy to understand and full of great stuff!
My pleasure!
I'm lucky this year. Another superb discovery of a channel 🙂 Olivio, I have one question - can I use my own image source to generate a graphic where Im situated on a different planet and drinking some fine gin? 😅
Been waiting for this video, thank you!
Thank you. Sorry was busy today with another project
I'm absolutely loving it!
Very informative video, thanks. Unfortunatley as a linux user I have tested this on 2 seperate machines on automatic1111 and had consistant errors with loading the refiner model. Hope that it gets sorted soon as it looks great :)
There is nothing unfortunate about being a Linux user ;) Which distro?
Was a joy to watch and very informative and clear, well done video, create something cool
Ngl the end was amazing
There were discussions about the SDXL 1.0 VAE and how it created some strange artifacts, a lot of people seem to recommend to use the SDXL 0.9 VAE with SDXL 1.0 to avoid those issues. Maybe worth a try to see if you still get those problems with the eyes when not using face restore.
So far I didn't notice any difference when I use SDXL 1.0 VAE compared to when I use None.
I have Restore Faces checked. If I don't then faces suddenly all turn blue and distorted in the last Step using the 1.0 VAE.
Gotta try 0.9 VAE then.
@Cutieplus is it when editing the .bat file with a notepad?
they reupload the SDXL VAE several hours after they publish original SDXL VAE, i thought it already fixes that strange artifact.
Fantastic enjoyable content. Had to subscribe ❤ Keep em coming. 👍
Thanks for the mention , always great informative videos 🙏
Your advice has been very tasty 😊, Thanks!
Love the sunglasses for ‘hacker mode.’ 😂
You can see why Midjourney want v6 out as it will intrepert text better than v5, which seems to be one of the big features of SDXL. MJ is still generally better even at v5 but SDXL is very very close now.
Yeah 🎉 official release 🙌
8 minutes in and you finally get to what I came here for. I can look at websites on my own 😂
Tnx for the update, love this channel. Question: I'm at hugging face and i see a file uploaded two days ago, should I install "1.0_0.9vae.safetensors" or "1.0.safetensors"
As always I am amazed at your content and explanations - when do you ever sleep?
You and Sebastian are my favorites. The real AI OG's in the industry. How can I trust you to not be an AI Olivio?
Real question. Where do you see what version of SD you're running? I keep trying to update, but it says I'm already up to date, fine, but how my SDXL isn't working. It just load :(
Amongst other places, it'll spit out the version number in the cmd dialogue right beneath the python version and all of the version numbers are listed at the bottom of the A1111 UI. I really don't see how you could possibly have difficulty finding them?!
I had that problem and the solution I found was to do a fresh install with the latest A1111 webUI. Even if it says you're up to date, you aren't. You'll know for definite as it will tell you it's 1.5.0 or 1.5.1 where Steamrick mentions. You don't need to get Python again but remember to add it's path to the webui-user.bat.
@Steamrick The A111 webui doesn't show it's version number for all installs. For older installs, those UI numbers you mention didn't show the UI version, they started with Python. 🙂
thank you. at the very bottom of the browser page OR at the start in the CMD window when you load
Excellent tutorial and information @OlivioSarikas but I was wondering, and if is not too much to ask... What is the Automatic 1111 that you are running, because I don't have some of the slide controllers (like the Clip Skip), and option windows (SD VAE), that you showed in this video. Is this part of a default installation of Automatic 1111, or are there extensions that you have to install?
One more time thank you for your excellent videos. Take Care.
check settings
no extensions, you need to enable them in your settings
great video. Explaining it very well.
hey i have a problem that when i want to select this model a1111 didn't let me do it someone has this problemand know how to fix it?
Noticed my WebUI was just an unpacked zip-file, so I went to start over with cloning down the repo this time to easier keep it updated. Taking a while, but hopefully it will be worth it. The safetensors also take a long time to download so I imagine a lot of people are downloading them to enjoy themselves with AI art.
A great video as always, with the plus of hacker Olivio 😂
WOW! Very cool! BIG FANX & Colored Greetinx!
Thank you :)
Great stuff really helpful - as a photographer and designer getting into AI generation its amazing to get these supporting tutorials - For me my desktop AMD PC with NVIDIA Super 1650 is very slow using Automatic 1111 it takes around 20 minutes to get one image generated and so this is not practical for me so im looking at options of updating spending £300 on new GPU or other alternatives
That sounds a bit excessive. Was this with SDXL using 1024x1024 resolution?
@@Resmarax hi yes it was - super slow - its faster to use somehting like clipdrop but its rather restrictive - will see about spending out on a new GPU and read up if my AMD processor has any influence
Great video and amazing look with glass !! 😆
Thank you!! 😁
hey for me it uses way to mutch system ram like above 32 gb and then it crashes
For me too.
Well, the Refiner Model that is.
60GB+ then Crash or Error Message.
@@fr0zen1isshadowbanned99 with comfy ui i dont get the proplem
you are an absolute legend
I tried the prompt, "a perfectly normal man looking at his perfectly normal hands"... SDXL produced a handsome man with seven fingers and three thumbs!
Nicely done as always!
Thank you :)
When I select the base model I see the following error and the model auto reverts to previous loaded model: Loading weights [31e35c80fc] from F:\stable-diffusion-webui\models\Stable-diffusion\sd_xl_base_1.0.safetensors
Failed to load checkpoint, restoring previous + size mismatch for model.diffusion_model.output_blocks.8.0.skip_connection.bias: copying a param with shape torch.Size([320]) from checkpoint, the shape in current model is torch.Size([640]).
I saw this behavior when selecting the base VAE model named sd_xl_base_1.0_0.9vae.safetensors in Automatic 1111. But using sd_xl_base_1.0.safetensors doesn't cause that for my PC.
@@stormjackthanks for your comment, I am getting this when selecting sd_xl_base_1.0.safetensors
stable-diffusion-xl-refiner-1.0 when i select this my RAM usage go 100% and fail to load model
same here, did you fixed it? mine takes forever to load and takes like 3 minutes to generate one image at 1 it/s
@@ZayxSt no I deleted the SDXL modes
Thanks! Awesome video! Can you please release your automatic 1111 layout (e.g extensions, models, favorite settings etc)?
Thank you for the video, you saved me a minute of rendering by using this "HACKER" method. I was struggling for this issue, since I have 8 gb VRAM card, and the rendering takes like 1,5 minute to render on the base model, which led me to use ComfyUI which is much faster and stabler. But it needs some learning to add nods for refiner and Lora etc..
But you saved me man thank you!
I am waiting for your videos on how to train our own Loras and models for this specific model, is it the same like before? Is there any change? All this will keep me busy watching all upcoming videos about this new version.
You are welcome :) It's not supposed to be used that way, but i think the results can be pretty nice too :)
Great vid BTW,subbed :)
So what is the Resolution of the Renders?
Thank you Olivio for your great support and detailed explanation. One question, when trying to download the SDXL base model and the SDXL Refiner, I can see that today exists now the versions including VAE for both models. Is there any suggestion from your side to select the VAE or not VAE version to download and use?. Thank you.
SDXL is indeed amazing 🎉❤
Well one thing not mentioned is how much GPU you need. I have RTX 2060 Super with 8G VRAM. WIll it do? Should i use something else? I don't actually care that much about anatomical correctness, photorealism and number of fingers, or about high resolutions above 768x768, but i want interesting varied results.
Amazing! Thanks you Olivio
Is Controlnet already available for SDXL 1.0 with Automatic 1111? Couldn't find anything on the web so far
I was getting poor quality results and errors most of the times. Now I know why. I had my vae set to 84k. Lets try out with new knowledge. Thank you!
awesome, happy i could help
thank you for this video
I got error when using SDXL
size mismatch for model.diffusion_model.output_blocks.8.0.skip_connection.bias: copying a param with shape torch.Size([320]) from checkpoint, the shape in current model is torch.Size([640])
please help
Updated to A1111 1.8.0, when loading SDXL check point getting error. size mismatch for model.diffusion_model.output_blocks.8.0.skip_connection.bias: copying a param with shape torch.Size([320]) from checkpoint, the shape in current model is torch.Size([640]). I'm assuming something broke related to SDXL in this latest A111 update. Any suggestions ?
It refuses to allow me to select SDXL 1.0 or the refiner. It takes a long time and then kicks it back to the last model I had loaded. Anyone else having this problem? Tons of size mismatches and torch errors in cmd window.
same here, but with both archives
@@erickromano5030 Cool. I hope someone has an answer.
I have the same error, long list of missmatch messages, can't find a solution :(
@Cutieplus yes, I am
Same errors hoping for a solution
I have a RTX 3090 , aside from xformers do you have any other command prompts you recommend I add? Thanks
Hey Olivio, thanks for making the tutorial. How do you get the Clip Skip and the VAE selector up there?
Settings -> User Interface -> Quick Settings List
Add "CLIP_stop_at_last_layers" and "sd_vae"
Apply settings -> restart automatic1111
Great video, right now I'm enjoying SDXL a lot and getting some very decent stuff generated. Though things will 'go nuclear' once we have a series of custom models out and ControlNet works with it.
My only fear is VRAM. I can't wait to do Dreambooth with this, but I assume that if we're using 1024x1024 as source images then VRAM requirements are going to skyrocket?
Do I need to install the SDXL base model in order to use SDXL checkpoints from CivitAI?
am i missing something? when i try to select the sdxl model it never lets me use it. Always just goes back to whatever model i had selected before.
The images you are showing I'm wandering if it's just my screen but they appear to have vertical lines on all around the edges of the characters also through the characters. Is this a shortcoming of the AI or perhaps the way its transferred to this media? Also the image at 2:50 an improvement would be for the AI to render in the foot prints from the dog running and jumping in the sand, at the speed it would have to be running there would in a real pic be prints left in sand and also more than likely sand flung out from the dogs back paws. Just some observations.
Hello, is it possible to get it running with Vlad automatic? So far I haven't seen any tutorials for it.
Hello, how can I get the box to choose VAE ??
Oh my Gosh, it can spell correctly...... I tried this"James Dean as a character in a comic book magazine, front cover, very colorful, highly detailed, super sharp, with the words JAMES DEAN on cover"
SDXL 1.0 Refiner will go into which FOlder ?
I'll start using SDXL when ControlNet is available for it.
For me, SD XL is just producing low quality noise... no idea what im doing wrong. im using Euler a and 1024x1024 res...
Instantly subscribed bro. Your content is definitely for me!
Thank you for this. I do not have the SD VAE bar though. Why is that?
have a look here: ruclips.net/video/BKHWJ_b3h-s/видео.html&lc=Ugy9j83wfHxgdDVac_x4AaABAg
i tried the same prompts with different model / sdxl and sd1.5 and they turn out different but i would not say one is better than the other...
Thank you, it is great. I saved your video.
Is there any video from you about the step-by-step installation of A1111?
Thanks for your content. It has been so helpful. I have been a pro photographer for more than 35 years and this AI image creating is one of the most interesting things that has happened during my career. It will one new thing that I will be using in my professional work. I am so happy that I got to see this development in image making.
I tested the SDXL 1.0 model and I am having hard time to get a natural looking skin. The images looks nice and in most cases the anatomy is next to perfect, but the skin looks too plastic. I personally like the old fashion natural looking skin. Any ideas?
I experiment with sdxl a bit and found that refiner model can also be used for upscale (I used Ultimate SD), I set tile size to 1024x1024.
It can not only upscale image, but add a lot of details to image.
Hi, I am working a face regenerating model, can we connect to discuss together?
@@user-vs3qg4zs8s Hi.
Sure.
Although I'm pretty bad at speaking English 😁
SDXL is huge, heavy and it's already excluding lots of users with low/mid-end machines - the world is big and there are users with all kinds of hardware and internet connections. There are LoRas with almost 1GB being produced. Yeah, "show must go on", but I see developers jumping on XL wagon and quickly dropping SD1.5 with some concern.
According to my experiments, version 0.9 still gives better results. (Tested on comfyui)
Hey Olivio, I'm looking since long time for an audio TTS AI which I can run locally (similar to A1111 or even CLI) with german voices. Do you know what I could use?
Would be perfect I there would be a way to train my own voice model, too.
i don't do audio ai, sorry
Thanks for the info Olivio, I am trying to understand if there is a way we can use one of the model to batch process multiple images with the same treatment to get similar results on all the images of the same batch, any guidance would be highly appreciated.
will automatic add correct way to use model + refiner?
"Well call me Bob and butter me sideways" is the best line in any RUclips video this century, and this is FACT 😀
How will this one fare with lower end users? 2060 ti super and lower? I have to use --lowvram in the arguments. Will I need xformers etc.?
I tried to upload sdxl with cagliostro but it didn't work.. does it need some new colab notebook? (my pc is too week for all of these)
Is it normal for the model to take 5 minutes to load? It works fine once it's loaded, renders in about 20s for 1024x1024 but base model and refiner model can both take up 5min to actually load which is annoying. Automatic1111 and all models are installed on an ssd so idk. Is this the same for everyone else?
Yh same here
Same problem on my side
Killer vid! Question: what is that “apply colour correction to img2img” setting you’ve got - is it an extension?
i used that for a batch render in img2img tutorial to create videos there instead. haven't used it since
not usable for me with Auto 1111, comfy takes 1-2 minutes in auto its 20minutes for 1 image :/
wow, that is a huge difference!
@@OlivioSarikas yes but im running on a lower end gpu, but the difference is huge. GTX 1070, on comfy everything works fine.