These tutorial series has really SHOWN (and not just said) how powerful Comfy is. People know it is powerful but never really got a deep explanation like these series. Mateo, true legend. Don't need to watch other youtubers regurgitating stuff other youtubers have said. THESE are the true tutorials!
I'm coming from A1111/Forge and trying like hell to learn Comfy. Your videos showed me just how insanely capable comfyui is and helped me understand what's actually going on in the process of generating an image. Can't believe I delayed learning this UI for so long -- it was daunting, but I finally feel like I'm beginning to understand it. Very valuable content!
Going to add one more comment here about how much I learned from this! The CFG rescale trick is going to be really useful, I had no idea it could improve multiple conditioning like that. Thank you ^_^
Amazing, most simple character tutorial ever. please make more tutorial about multi-IDs . Like : - making 2+ characters fight scene - putting the characters in a portrait with "multi area conditioning" nodes - applying instant ID to objects that can be held by characters (swords, smartphones, etc.) Thank you so much man, your instant ID just solve some character consistency problem in comic creation community.
take care of yourself, because, as I'm sure you can tell, there's an incredibly positive response to your content. it's probably exciting for you but please take it easy and don't burn out because you're just THAT much better than every other content creator on this topic. we need you! lol
im an a1111 user. these connecting nodes, adding bunch of different control-net models just scaring me. I see that is another level. However thats a perfect tutorial! Congrats my friend.
Thank you again Matt3o for another Best-in-Class tutorial, as a developer and as a ComfyUI teacher! You have no equal in this space. Even if there are also a few other good channels, your's IS THE BEST. Period.
At 5:18, you say that there are far better models for photorealism. Can you recommend any for Comfyui in SDXL? I have a single photo of someone and can't create a Lora from that, but want to use it for generating other realistic photos.
It’s really difficult not to learn something tremendously useful by watching your master tutorials. It’s simply amazing how much knowledge you own from IA world and especially from ComfyUI. I have no words to thank you enough!
instant id works with 1.5/v2 models, if you DON't use the loras, make sure the model is LCM compatible and use LCM lora, cache/retrieve your embeddings with v2 turned on (even if its not) and use LCM sampler, ddim_uniform/kerras scedular (i'm using it right now, and have been for a week) ;)
@latentvision I’m not sure the batch images node actually would make a difference in the sampler, because the sampler input is an image, not an image batch. So in the end, I think all this does is cycle through the batch, but in the end, it would sample whatever the last image was in the batch. I’m guessing in order to truly end up with a little LoRA, one would have to figure out a conditioning weighting batch operation. That would be tremendously useful, but likely result in heavy resource load.
Thank you for your help. It would be slightly better if the video had more zoom (lower resolution), enough to identify the model used on a cellphone screen. Thank you, I couldn't find the solution until I came across this video.
Great work. Please guide me to install DLib to be used for ComfyUI instantid. 1.Where to create the folder in the ComfyUI folder 2. How to install then
Let's generate a lot to see if we can have better hands ... (no-hands-picture-shows-up) Oh this is nice ! ;p Hands are so uncool in AI generation. Love your videos, it's just terribly hard to understand when you start with AI and comfyui. So in a nutshell, InstantID is cooler for its style purpose of the final image ? At the end I think that you have to learn every new thing that comes up and make your own choice. I'm still on SD1.5 as most of Fanart models are on that one in civitai, SDXL is not really my cup of tea. Cascade is very fast and nice but I miss my SD1.5 ... Waiting for SD3 and hopefully will be easier to make your own lora faster and easier.
It's a very amazing tutorial, thanks.❤ I wonder why in 11:50, you used "condition combined" instead of chaining "positive and negative" conditions of 1st InstantID to 2nd InstantID? 😅
instantid is a kind of uncommon controlnet, at the moment that is the best solution, I need to check if I can send the text prompts directly into attention, that would simplify things
It's me again, with a question this time! You are using an image for masking (at 4:48) and you keep the same image through the whole video, even using an inversion mask for the 2nd IpAdapter. I'm not quite how this works : are you supposing the mask you are drawing will more or less be where the face of the mona lisa will be, to take only the rest of the painting? Wouldn't it be simpler to c/c the reference image, draw a perfect mask on the Mona Lisa''s face and invert it? Thank you!
Matteo, could you cover such functionalities like Patch Attention and the Controlnet Attention? I think only you can help us with it and show the use cases for that, would be very interesting.
After installing the package, I cannot access the nodes needed for the program to work. Being human and new at this, I feel that I may have neglected a detail or managed to put something in the wrong place. any assistance would be greatly appreciated.
Hi Mateo, in the last Apply instantID advanced part, you said that we can choose to fine-tune ip_weight and cn_strength. Do we set the continuation degree of the reference image style by adjusting the ip_weight, and set the proximity to the facial features and prompt words of the reference image by adjusting cn_weight? Thanks.
@latentvision Dear Matteo , I am trying to use Depth-anything (Marigold depth estimation) within your InstantID workflow like the one used in the InstantID Huggingface space , but unfortunately I am facing error because the depth-anything model is based on SD1.5 , could you please help me about that , I can give you my workflow to check it . And as usual your tutorials are very helpful and easy to learn from it , Thanks a lot. 👍
I don't have the "Apply IPAdapter" or the "Load IPAdapter model" nodes and I have IPAdapter install and I see lots of other related nodes. Had something changed in recent versions? am I missing something?
@@latentvision Thanks! I thought I was losing my mind, I spent half of the day trying to figure out why I don't have the nodes everyone is using. Looking forward for the documentation.
very good tutorial, but the custom nodes mentioned installation are very hard and painful to not have errors, thus you can't follow the result on this tutorial. im stuck on faceanalysis
Amazing video. How to do when we are not able to pass under the 1.2 for euclidean and 0.8 for cosine? I used your workflow. It works perfectly with your inpu images but with my personal inputs, I don't have the same results. My 3 images for the InstanID and the IPAdapter are wide and the face enter in a 640x640px...THANK YOU in advance.
previus error was solve downloading all models with instantID... but new error came when try the ksampler run Error occurred when executing KSampler: 'NoneType' object has no attribute 'shape' SOLVED: need an XL checkpoint... srry.
You mentioned InstantID is to add styling to someone's image. And you said there are far better photorealistic face swapping models. Which are the best ones in your opinion, to do the best, most realistic faceswapping? Thank you!
I have to say InstantID is quite effective, especially for creating side faces from only frontal references, which the other face models often refuse to draw. I do run into a weird problem though. It appears my InstantID workflow wants to work at much lower CFG (~2) than what the model would usually prefer (4~6), so I need to either sacrifice background complexity by running very low CFG or risking burning the image with high CFG. I wonder if this is specifically an InstantID issue?
@@latentvision I guess the problem is that the checkpoint normally works quite well at CFG=5, but when connected with the Instant ID node, it produces images with weird colours unless I turn down the CFG...
@@mithrillis CFG 5 is very low already. InstantID requires CFG 4-6 if the checkpoint works at 7-8. So if the starting point is 5 it makes sense that you need to lower it to 2-3. Again try with rescalecfg
Mateo I'm thinking this works best for advanced users of comfy. I'm still reviewing the set up, which is quite complex. However you get what you put in when it comes to the analytics of the spaghetti pipelines and TBH you really are in control of what you want as the final out put. Not for those looking for a quick fix in the production of AI rendering. I remained subbed (for the learning experience and nuance)
Amazing!! Will there be an implementation for sd1.5? These models for SDXL are flying way over my 8gb vram system. Tried it on colab though. Vram tops at 14 gigs.😂
Another really interesting video. If you don't mind me asking, is this all in python? I'm learning how to code and it would be interesting to try to apply some of it to comfyUI to earn more about AI on my coding journey. Love your content, keep it coming!
Hi! THX for you awesome video! Quick question. You used "conditioning combine" to link multiple InstantID in parallel, but what about connecting them in series instead?
each instantid controlnet instance has to be patched with the person's embeds. If you connect them in series the embeds get averaged. I'll see if there's some sorcery I can do to make it work, but at the moment this seems to be working.
Is there a faster way to import batch images for the input reference image instead of adding the batch node which has only 2 input sockets? Thanks in advance
You are an absolute master. I've learned so much. However I still can't understand why all my output is a face close up despite trying to show "three-quarters view" or full head and chest or even full body. What am I doing wrong? Thank you!
Thank you for the most useful information on comfyUI. Your lessons are out of competition. Unfortunately, I still can't figure out why there is some kind of noise at the output (without any errors in console), noticed that if you connect the clip text encode directly to ksampler bypassing instantid, then everything is fine with the image, but the face is not too similar. And I noticed that it works much faster than other instantid before him, even faster than Forge in webui
@vision Thx!! Update really solved the problem. It works very fast on my 8gb graphics card! Now it really works as regular generation instead of 10+ minutes
These tutorial series has really SHOWN (and not just said) how powerful Comfy is. People know it is powerful but never really got a deep explanation like these series.
Mateo, true legend. Don't need to watch other youtubers regurgitating stuff other youtubers have said. THESE are the true tutorials!
Love it. Learned more in 15 minutes than I have in a month of struggling with basics. Thank you!
I'm coming from A1111/Forge and trying like hell to learn Comfy. Your videos showed me just how insanely capable comfyui is and helped me understand what's actually going on in the process of generating an image. Can't believe I delayed learning this UI for so long -- it was daunting, but I finally feel like I'm beginning to understand it. Very valuable content!
We respect and support every lecture you give.Kudos to you for your overwhelming knowledge.
God, there was so much to unpack in this video...I'm going to be coming back to this for months to understand it.
Going to add one more comment here about how much I learned from this! The CFG rescale trick is going to be really useful, I had no idea it could improve multiple conditioning like that. Thank you ^_^
By far the best ComfyUI tutorials on YT. Thank you so much 🙏
100%
100%
This guy is the goat- writes the nodes and gives useful and detailed instruction on how to use them.
Agreed
I also agree.
100%
Thank you Mateo, your channel is definitely my favorite RUclips channel on the topic.
The man knows how stable diffusion's brain works, your the best instructor I've ever seen Mateo, keep the great work
Thank you! I finally got this working as I followed along. Lookin forward to more :)
Don't have the time or motivation to dive into this, but a tutorial of this quality deserves kudos even from someone like me. :)
Amazing, most simple character tutorial ever.
please make more tutorial about multi-IDs . Like :
- making 2+ characters fight scene
- putting the characters in a portrait with "multi area conditioning" nodes
- applying instant ID to objects that can be held by characters (swords, smartphones, etc.)
Thank you so much man, your instant ID just solve some character consistency problem in comic creation community.
Another excellent video, thank you
I spent an hour going through your video. thank you, very helpful.
Whew! Amazing as usual. Thank you!
Amazing videos man. Love u.
take care of yourself, because, as I'm sure you can tell, there's an incredibly positive response to your content. it's probably exciting for you but please take it easy and don't burn out because you're just THAT much better than every other content creator on this topic. we need you! lol
LOL... okay, I'll take a break 😄
@@latentvision noooooooo too long ;) you are the GOAT!
I haven’t watched any comfy ui vids yet but dang this looks fun from an experimenting perspective
Your lectures are always fantastic!
im an a1111 user. these connecting nodes, adding bunch of different control-net models just scaring me. I see that is another level. However thats a perfect tutorial! Congrats my friend.
Thank you for your tools development and videos you make, it is amazingly helpful to improve skills and keep moving industry forward!
Well, yes. You rock. You give us the *what*, and the most important *why*.
Ты просто боженька чувак! Это невероятно круто! Твое видео в 15 минут надо разбирать неделю!))) Спасибо большое за твою работу!
Thank you again Matt3o for another Best-in-Class tutorial, as a developer and as a ComfyUI teacher!
You have no equal in this space. Even if there are also a few other good channels, your's IS THE BEST. Period.
At 5:18, you say that there are far better models for photorealism.
Can you recommend any for Comfyui in SDXL?
I have a single photo of someone and can't create a Lora from that, but want to use it for generating other realistic photos.
Complimenti Matteo, every Comyui node that you do, is amazing :)
It’s really difficult not to learn something tremendously useful by watching your master tutorials. It’s simply amazing how much knowledge you own from IA world and especially from ComfyUI. I have no words to thank you enough!
Amazing work!
Thank you!
Thank you so much for your hard work. Your tools and tutorials are incredible. You are truly awesome.
Great work as usual, thanks!
instant id works with 1.5/v2 models, if you DON't use the loras, make sure the model is LCM compatible and use LCM lora, cache/retrieve your embeddings with v2 turned on (even if its not) and use LCM sampler, ddim_uniform/kerras scedular (i'm using it right now, and have been for a week) ;)
what? how? Do you have a comfy workflow you can share?
what is v2?
@mordokai597 can u plz share it
Also I have all the IPAdapter stuff installed but there is no Apply IPAdapter?
Holy moly I learned a ton with your video. Thank you!
Another great video packed with valuable insights 🎉 - Thank you Matteo!
you are a genius teacher
this is just brilliant. thanks a lot.
@latentvision I’m not sure the batch images node actually would make a difference in the sampler, because the sampler input is an image, not an image batch. So in the end, I think all this does is cycle through the batch, but in the end, it would sample whatever the last image was in the batch.
I’m guessing in order to truly end up with a little LoRA, one would have to figure out a conditioning weighting batch operation. That would be tremendously useful, but likely result in heavy resource load.
the face embeds are averaged
At 5:08 you said there are far better models for photo realism. What were you talking about specifically and what are the models?
check my "face science" video ruclips.net/video/oBKcjY-JO3Y/видео.htmlsi=hF0zDCAuFlZn05sp
That's what I was waiting for! Grazie mille
best comfy related all time guide.
very informative.thanks.
Can you make one for photorealistic results, please? with all the optimal settings
Sei il migliore, grazie! Ciao!
Thank you for your help. It would be slightly better if the video had more zoom (lower resolution), enough to identify the model used on a cellphone screen. Thank you, I couldn't find the solution until I came across this video.
Thank you Matteo! love your work.
Great work. Please guide me to install DLib to be used for ComfyUI instantid.
1.Where to create the folder in the ComfyUI folder
2. How to install then
I learn more from your videos than all other AI videos combined.
Thank you very much 😀 that was helpful.
Any recommendation for photo realistic model, I tried the same workflow with epicgasm timeline 3:33 but ksampler was failing.
realistic vision is generally good
I wait for your videos. I have subscribed to quite a few channels, but yours is the only one with 🔔 on. Thanks! 🙏
Keep up the good work!
Let's generate a lot to see if we can have better hands ... (no-hands-picture-shows-up) Oh this is nice ! ;p
Hands are so uncool in AI generation.
Love your videos, it's just terribly hard to understand when you start with AI and comfyui.
So in a nutshell, InstantID is cooler for its style purpose of the final image ? At the end I think that you have to learn every new thing that comes up and make your own choice.
I'm still on SD1.5 as most of Fanart models are on that one in civitai, SDXL is not really my cup of tea. Cascade is very fast and nice but I miss my SD1.5 ... Waiting for SD3 and hopefully will be easier to make your own lora faster and easier.
Very great tutorial. Thank you very much!
I am getting a some loss of identity when using image kps for head positioning
do you have solutions for that? I have the same problem
Thanks so much for this ❤
Great! I was looking for this on your channel, an hour ago!
Thank you matteo 😊
One day, I wish to be as incredible as you. Thank You.
Y'all are correct. Protect this man at all costs.
LOL, thanks for the laugh :D
It's a very amazing tutorial, thanks.❤
I wonder why in 11:50, you used "condition combined" instead of chaining "positive and negative" conditions of 1st InstantID to 2nd InstantID? 😅
instantid is a kind of uncommon controlnet, at the moment that is the best solution, I need to check if I can send the text prompts directly into attention, that would simplify things
👍 That would be very effective. 🎉
It's me again, with a question this time!
You are using an image for masking (at 4:48) and you keep the same image through the whole video, even using an inversion mask for the 2nd IpAdapter.
I'm not quite how this works : are you supposing the mask you are drawing will more or less be where the face of the mona lisa will be, to take only the rest of the painting?
Wouldn't it be simpler to c/c the reference image, draw a perfect mask on the Mona Lisa''s face and invert it?
Thank you!
the face position is pretty much fixed and the attention mask is rather loose anyway, so yeah it works :)
Matteo, could you cover such functionalities like Patch Attention and the Controlnet Attention? I think only you can help us with it and show the use cases for that, would be very interesting.
Thank you for sharing part of your knowledge :)
This is masterpiece mate 🙏
You are the best!! ❤
After installing the package, I cannot access the nodes needed for the program to work. Being human and new at this, I feel that I may have neglected a detail or managed to put something in the wrong place. any assistance would be greatly appreciated.
this is fire 🔥 thanks!
hey there, i cant instal the custom node currently, any idea why? from comfy UI manager
Amazing tutorial!! ...I have a question: how to make preview image for only the specify node (with out running Queue Prompt)?
Hi Mateo, in the last Apply instantID advanced part, you said that we can choose to fine-tune ip_weight and cn_strength. Do we set the continuation degree of the reference image style by adjusting the ip_weight, and set the proximity to the facial features and prompt words of the reference image by adjusting cn_weight? Thanks.
they work together. ipadapter is a helper for instantid. instantid takes care of both the pose and about 70% of the likeliness.
Thanks for your quick response, I will think about it further more to understand it better.@@latentvision
@latentvision Dear Matteo , I am trying to use Depth-anything (Marigold depth estimation) within your InstantID workflow like the one used in the InstantID Huggingface space , but unfortunately I am facing error because the depth-anything model is based on SD1.5 , could you please help me about that , I can give you my workflow to check it .
And as usual your tutorials are very helpful and easy to learn from it , Thanks a lot. 👍
Which model is needed in the Load InstantID model node?
I love your vids, But i cannot get indightface to work on my computer, any suggestions?
insightface is a tough cookie. check this thread github.com/cubiq/ComfyUI_IPAdapter_plus/issues/162
I don't have the "Apply IPAdapter" or the "Load IPAdapter model" nodes and I have IPAdapter install and I see lots of other related nodes.
Had something changed in recent versions? am I missing something?
yeah a new version was released today. I'll post the documentation later.
@@latentvision Thanks!
I thought I was losing my mind, I spent half of the day trying to figure out why I don't have the nodes everyone is using.
Looking forward for the documentation.
@@latentvision Hi! can you please tell me when you can publish the documentation? or have you already done it?
@@AntonRybalkin check my latest video :)
very good tutorial, but the custom nodes mentioned installation are very hard and painful to not have errors, thus you can't follow the result on this tutorial. im stuck on faceanalysis
Great video. I wish I could subscribe twice.
I'm having issues with InstantID Face Analysis node, Could you please give me solution for this error? Thankyou
What base model you have used ?
Can you share link to download the model
Look like it's not working for me, with the latest version of ComfyUI (Portable 0.3.10)
hi I need to know what is exactly the control net model you used here and where to download it?
please check the repository, there's the link there
Aaaaamazing once again!!! ❤
Amazing video. How to do when we are not able to pass under the 1.2 for euclidean and 0.8 for cosine? I used your workflow. It works perfectly with your inpu images but with my personal inputs, I don't have the same results. My 3 images for the InstanID and the IPAdapter are wide and the face enter in a 640x640px...THANK YOU in advance.
that's very hard to say without checking the actual images and workflow. Join my discord I can try to help there
Thanks !
6:22 : recale CFG
Is it possible to use InstantId but replace Inisightface with another face analysis library?
the model has been trained with insightface, the FaceAnalysis tool can use dlib but that is a different thing
previus error was solve downloading all models with instantID... but new error came when try the ksampler run
Error occurred when executing KSampler:
'NoneType' object has no attribute 'shape'
SOLVED: need an XL checkpoint... srry.
Please do more videos on forge (automatic 1111)
You mentioned InstantID is to add styling to someone's image. And you said there are far better photorealistic face swapping models. Which are the best ones in your opinion, to do the best, most realistic faceswapping? Thank you!
if you don't need a lot of control over the style (and facial expression) check my FACE SCIENCE video!
I have to say InstantID is quite effective, especially for creating side faces from only frontal references, which the other face models often refuse to draw. I do run into a weird problem though. It appears my InstantID workflow wants to work at much lower CFG (~2) than what the model would usually prefer (4~6), so I need to either sacrifice background complexity by running very low CFG or risking burning the image with high CFG. I wonder if this is specifically an InstantID issue?
it depends on the checkpoint, if you need 2 it's seriously overtrained. Anyway you can use rescaleCFG
@@latentvision I guess the problem is that the checkpoint normally works quite well at CFG=5, but when connected with the Instant ID node, it produces images with weird colours unless I turn down the CFG...
@@mithrillis CFG 5 is very low already. InstantID requires CFG 4-6 if the checkpoint works at 7-8. So if the starting point is 5 it makes sense that you need to lower it to 2-3. Again try with rescalecfg
Thanks a lot. I can not find the "face embeds distance" mode unfortunately. has it been removed?
it's linked in the video description
Mateo I'm thinking this works best for advanced users of comfy. I'm still reviewing the set up, which is quite complex. However you get what you put in when it comes to the analytics of the spaghetti pipelines and TBH you really are in control of what you want as the final out put. Not for those looking for a quick fix in the production of AI rendering. I remained subbed (for the learning experience and nuance)
check the "basics" tutorials if this is too much :)
Amazing!! Will there be an implementation for sd1.5? These models for SDXL are flying way over my 8gb vram system. Tried it on colab though. Vram tops at 14 gigs.😂
Another really interesting video. If you don't mind me asking, is this all in python? I'm learning how to code and it would be interesting to try to apply some of it to comfyUI to earn more about AI on my coding journey.
Love your content, keep it coming!
yes, mostly python (and thanks! 😄)
Hi! THX for you awesome video!
Quick question. You used "conditioning combine" to link multiple InstantID in parallel, but what about connecting them in series instead?
each instantid controlnet instance has to be patched with the person's embeds. If you connect them in series the embeds get averaged. I'll see if there's some sorcery I can do to make it work, but at the moment this seems to be working.
Is there a faster way to import batch images for the input reference image instead of adding the batch node which has only 2 input sockets? Thanks in advance
Load Images (Path). It's part of the VHS extension
You are an absolute master. I've learned so much. However I still can't understand why all my output is a face close up despite trying to show "three-quarters view" or full head and chest or even full body. What am I doing wrong? Thank you!
with instantID you can set a reference for the pose
Is there a precompiled version of insightface or do I have to install Visual Studio c++?
5:08 What better models are you talking about?
inswapper for example and refinement with faceid
@@latentvision What about ReActor?
@@adrianfels2985 it's inswapper
@@latentvision Ah, sorry. Thanks you!
Thank you for the most useful information on comfyUI. Your lessons are out of competition.
Unfortunately, I still can't figure out why there is some kind of noise at the output (without any errors in console), noticed that if you connect the clip text encode directly to ksampler bypassing instantid, then everything is fine with the image, but the face is not too similar.
And I noticed that it works much faster than other instantid before him, even faster than Forge in webui
you probably just need to update comfyui
@vision Thx!! Update really solved the problem. It works very fast on my 8gb graphics card!
Now it really works as regular generation instead of 10+ minutes
Cannot execute because a node is missing the class_type property.: Node ID '#11' How to fix it, thank you.
how does this compare with fooocus image prompt