Image to Video with Nvidia Cosmos in ComfyUI!

Поделиться
HTML-код
  • Опубликовано: 31 янв 2025

Комментарии • 92

  • @TUSHARGOPALKA-nj7jx
    @TUSHARGOPALKA-nj7jx 6 дней назад

    This is quite amazing. Keep making more amazing video on open source image to video models. Amazing times!

    • @NerdyRodent
      @NerdyRodent  6 дней назад

      Just did another one… Apache 2.0 this time 😍

  • @ImmacHn
    @ImmacHn 12 дней назад +6

    This is something that I have been saying from the start, pure learning from videos or images won't do, modeling a 3d world, now that's where it's at.

  • @CGFUN829
    @CGFUN829 12 дней назад +5

    hey man love your work, also your voice is amzing and calming.

  • @swannschilling474
    @swannschilling474 12 дней назад +2

    Thanks for this one Nerdy! 😊

  • @synthoelectro
    @synthoelectro 12 дней назад +6

    everything from flux to SD has been worked out on my 4GB Nvidia since the early days using virtual memory, it's slow but it works.

    • @ARMASPIRIT
      @ARMASPIRIT 12 дней назад +3

      @@synthoelectro hi, please I need to use virtual memory, is there any tutorial?

    • @synthoelectro
      @synthoelectro 11 дней назад

      @@ARMASPIRIT I'm using windows 10, simply go to control panel look on your choice of search engine online for the answer about how to change swap memory or virtual memory in windows 10 and before.

  • @theairchitect
    @theairchitect 8 дней назад +1

    you always rocks 🤘 thanks🙏

  • @Statvar
    @Statvar 11 дней назад +3

    I'm getting a speed of 423.54s/it with the default res and length at 20 steps. This with with 16gb vram. Why is it terribly slow for me?

  • @ControlTheGuh
    @ControlTheGuh 10 дней назад

    I've been looking for a solution where we can identify keyframes, select the item/cloth we want to change automatically, then inpaint and finally interpolate between those frames.
    The interpolation part is very interesting to me, I wonder what that would look like with similar keyframes

  • @boythee4193
    @boythee4193 2 дня назад

    awesome video. the text to video worked right out of the gate, but the img to video is missing a custom node cosmosimgtovideolatent that the node manager does not see

    • @boythee4193
      @boythee4193 2 дня назад

      turns out that comfyui needed updating, but doing it via manager did not work. git pull in the comfyui dir works in my case "comfyuiportable/ comfyui" cmd there and type git pull in the terminal

  • @pauljones9150
    @pauljones9150 9 дней назад

    The new CaPa paper for mesh generation mentioned it will fit onto ControlNet pretty well. Wonder if that's going to go crazy with 3d Printing or not. We only have the code rn, no demos, so might a "in a few months" thing

  • @dansu1503
    @dansu1503 12 дней назад +1

    Nice ending song!

  • @banzaipiegaming
    @banzaipiegaming 11 дней назад +2

    Hi I keep getting a ksampler error: "Expected size for first two dimensions of batch2 tensor to be: [154, 768] but got: [154, 1024]." and I didn't alter the base workflow at all so not sure why this is happening

  • @РоманСырватка
    @РоманСырватка 11 дней назад +2

    Good job, it works on my 3080 10gb very well (7B model). Nothing crashes due to lack of memory.
    i2v 704x704@65 frames ~ 7min, 1024x1024@65 frames ~ 37min.

    • @BuntAsHell
      @BuntAsHell 9 дней назад

      Hi. I ran the example from the Text 2 Video workflow on a 4070ti (12gb) and it took 37 minutes. Any tips on speading that up?

    • @РоманСырватка
      @РоманСырватка 9 дней назад +1

      @@BuntAsHell 1. When starting the ComfyUI, make sure that xformer - enabled ( read console )
      2. In the nvidia panel, enable CUDA in the global settings - use the system memory reserve. (I have 64 GB)
      Maybe it will help.

    • @Vokoca
      @Vokoca 5 дней назад

      Can confirm that it worked on my 3080 as well, albeit it took ~30 mins at 704x704@121 frames. Feels too slow to experiment with, but the result was surprisingly consistent.

  • @heyselcuk
    @heyselcuk 7 дней назад +2

    7:45 i am trying to figure out since 3 hours, how you made this workflow. i wish i could download this or maybe better explanation :(

    • @boythee4193
      @boythee4193 2 дня назад

      he just rearranged the workflow from the author. as for the chain at the end, i haven't figured that out yet

  • @freddiebull5869
    @freddiebull5869 7 дней назад

    Hi there! Great video. This is exactly something that I needed for a task I was asked to do but I am brand new to SD and I was wondering if this workflow would run fine on a 16gb vram gpu or would it need 24? Thanks in advance!

    • @NerdyRodent
      @NerdyRodent  7 дней назад

      Yes, the requirements are provided in the video

  • @CHARIOTangler
    @CHARIOTangler 12 дней назад +18

    35 minutes for a very artifact-y 5 second render on a 3090ti that screams bloody murder the whole time? No, thank you. As long as you're not using it locally, it might be ok.

    • @bazio5592
      @bazio5592 12 дней назад +2

      3090 in 2025 bro, change job 😂

    • @ronbere
      @ronbere 12 дней назад +7

      @@bazio5592 wow super the relevance of the answer.. Buy yourself a life no?

    • @_rkrd
      @_rkrd 11 дней назад +4

      @@bazio5592 what an out of touch comment

    • @NerdyRodent
      @NerdyRodent  11 дней назад +4

      @CHARiOTangler that doesn’t sound right as my 3090 is faster…

    • @VioFax
      @VioFax 9 дней назад +3

      @@bazio5592 3090 isn't that bad of a card bro. Not everybody needs to fund Jensen's jacket collection. 3090 is still a $900 card. And still better than a 4060~4070 with vram...
      Someone sounds like they are trying to justify purchasing their 40 series so close to the 50 series release and/or compensating for something by bashing other peoples hardware...
      You replace a GPU like it's a glass of milk? Good for you be proud of your superiority or w/e. But not everybody, CAN, wants to, or needs to do that.
      Unless you are rocking a rack of H100's I'd get off my high horse. None of the gaming hardware like even the 5090 is worth a damn for more than a quarter anyway. So enjoy your 5090 it will be considered bargain bin trash soon too.

  • @OnigoroshiZero
    @OnigoroshiZero 12 дней назад +4

    I can't wait for consumer NPUs to become available in 2-3 years from now, as GPUs are not scaling along with model capabilities, and I don't have much hope for optimizations that will make them viable for real-time local use (video game emulation).

    • @VioFax
      @VioFax 9 дней назад

      NPU's are mostly monitoring hardware. Not for user playtime.

  • @tekkaman000
    @tekkaman000 11 дней назад +1

    Does it work with portable comfyUI?

  • @amakaqueru33
    @amakaqueru33 8 дней назад

    I just get the error "compute_indices_weights_linear" not implemented for 'Half'

  • @HikingWithCooper
    @HikingWithCooper 8 дней назад

    Where do the videos save? I'm getting single pngs in the default folder but no video. Tried adding the Video Save node but that doesn't work.

  • @my_username_was_taken
    @my_username_was_taken 6 дней назад

    Bummer, for some reason it just crashes my ComfyUI...I have all the models downloaded and everything, using a 3090Ti. ... nevermind ... I forgot I was training a LoRA at the same time lol. Weird I didn't see any kinda message about running out of memory.

  • @AkshatDobhal-m3h
    @AkshatDobhal-m3h 11 дней назад +5

    Hey man, love your videos. However I wanted to ask that the worflow youre using is not the same as in the example workflow. could you provide us with the workflow shown in your video. thanks a lot! Much love

  • @Lythian123
    @Lythian123 11 дней назад

    for me it is generating only a black video. 121 frames just pure black. anyone else had that problem and how can i fix it? :)

  • @INVICTUSSOLIS
    @INVICTUSSOLIS 11 дней назад +2

    Me and my mac M2 feeling that we are missing out.

  • @PapayaPositive
    @PapayaPositive 11 дней назад

    Fun times! 🐭

  • @Posky_fl
    @Posky_fl 12 дней назад

    Anyone knows how does it compare to Hunyuan?

  • @DeconvertedMan
    @DeconvertedMan 12 дней назад +4

    Its data to produce data so data can be data! O_o;

  • @homfes
    @homfes 11 дней назад

    I can't wait to try this a soon-to-be-mine RTX 5070. They say it's as fast as an RTX 4090.

    • @dogvandog
      @dogvandog 11 дней назад +1

      5070 is equal to 4090 only in games with AI frame generation. In Raw tests 5070 will be 15% faster than 4070super.

    • @Statvar
      @Statvar 11 дней назад +1

      @@dogvandog I'm pretty sure that comment is just bait. I'm hoping I'm not wrong. I'd be concerned for him if he wasn't joking

  • @Doctor_Random
    @Doctor_Random 11 дней назад +1

    darn it - I wish it could work for my lowly 8GB GPU :(

  • @rauliss1
    @rauliss1 11 дней назад +2

    ello, Empty Cosmos Latent Video seems to be missing, how can I fix it?

    • @NerdyRodent
      @NerdyRodent  7 дней назад +1

      Check out the first few seconds of the video for more information!

  • @jonorgames6596
    @jonorgames6596 12 дней назад +1

    Note to self: Avoid drinking bewerages while watching ai-vidoes.

  • @kariannecrysler640
    @kariannecrysler640 12 дней назад +3

    My nerdy friend🤘💜!!!!
    🌳🦋💃🌍💃🦋🌳

  • @heyselcuk
    @heyselcuk 3 дня назад

    can you please make a video only for the loop? in minute 8:00 zoom in to all nodes and explain how to connect them

  • @GMP007
    @GMP007 10 дней назад

    It's good. But hopefully ComfyUI reached the level of Hailuo Video Generator.

  • @juanjesusligero391
    @juanjesusligero391 12 дней назад +8

    Oh, Nerdy Rodent! 🐭🎵
    He really makes my day! ☀
    Showing us AI, 🤖
    in a really British way! ☕🎶

  • @Eldritch_
    @Eldritch_ 3 дня назад +2

    it says that CosmosImageToVideoLatent doesnt exist, any fix ?

    • @NerdyRodent
      @NerdyRodent  3 дня назад +2

      Check 0:00 😉

    • @Eldritch_
      @Eldritch_ 2 дня назад +1

      @@NerdyRodent ye dw I ended up restarting comfy and it worked lol

  • @erdbeerbus
    @erdbeerbus 11 дней назад +1

    Hej, great info, but ... CosmosImageToVideoLatent this node is red outlined and gives an error:
    Cannot execute because a node is missing the class_type property.: Node ID '#83'

    • @erdbeerbus
      @erdbeerbus 11 дней назад

      update was not solving that issue, also cosmos is missing at the loadclip node

    • @erdbeerbus
      @erdbeerbus 11 дней назад

      better use GIT PULL! ;)

  • @sdgtr4
    @sdgtr4 12 дней назад +3

    Really need new gpu

  • @felipealmeida5880
    @felipealmeida5880 11 дней назад

    Extremely slow running locally. Not yet.

  • @wakegary
    @wakegary 12 дней назад

    can we just pause everything so my 4090 stays good enough to curb my obsession with this entire scene? cause i kinda want that 5090 and it's a bit pricey, my four-legged g? i dont care if its faster, this is gollum level shit, i just want the benchmark json... I'm concerned that i'll upload my face and it will somehow just Bad Egg me like willy wonka for being too unstable 4 stable. I will say that the engineer brains dont do well with marketing/design sometimes, and I love how this particular hobby kinda forces them together. Interesting to see non-artistic brains get in the Picasso Mech Suit and mess around in awkward ways. That's kind of AI in a nut shell to me. Not quite right in that 'you seem a biiiiit too drunk to drive but it's nothing personal' way. uncanny valley is deeper than I ever though but the parties down here are wicked

  • @K-A_Z_A-K_S_URALA
    @K-A_Z_A-K_S_URALA 11 дней назад +1

    круто!

  • @ritpop
    @ritpop 12 дней назад +1

    the voice track says the original is english (united states) lol

  • @LilAirpod-goat
    @LilAirpod-goat 8 дней назад

    y2mate just popped this shit up

  • @USBEN.
    @USBEN. 12 дней назад

    I just want LTX 1.0 with paid video quality. These hunyuan and cosmos too heavy.

  • @f4ture
    @f4ture 11 дней назад

    “The best model”... not really. Hunyuan far surpasses it in terms of rendering quality, speed, flexibility, lora support, etc... If you're using the Kijai's wrapper you can go even further with optimizations. Not to mention the forthcoming arrival of their I2V version. I think this model will soon be forgotten unless they come out with a new more accomplished version.

  • @HikingWithCooper
    @HikingWithCooper 10 дней назад

    Slow @ 10 minutes on a 4090? Meanwhile, I've been generating an 85 frame video in Hunyuan for about an hour so far. Can't wait to try slow!
    UPDATE: The Nvidia model isn't any faster and the quality is much worse than Hunyuan's. Tried with both models and neither produced worthwhile results. I don't know where they get the "50x faster" idea. Must be marketing by the guy at Nvidia saying the 5070 is the same as a 4090.
    UPDATE2: While this model's txt2vid is bad, the img2vid is the best I've seen in a local gen. That's not saying it's "good" because, well, it's not. It has a tendency to create extreme motion with massive morphing no matter the prompt or input image. It has an absolute lack of understanding of basic physics and even trying to make a character walk is a horror show. I'm actually pretty surprised NVidia would put their name on something of this quality. But it is also actually pretty fast so that's nice. Hopefully I can find some sort of settings / prompts to make this model useful.

  • @cxs001
    @cxs001 6 дней назад

    der workflow funktioniert nicht, weil 2 nodes in ComfyUI fehlen!

  • @DeconvertedMan
    @DeconvertedMan 12 дней назад +2

    NERDSS!!!!

  • @shinchin384
    @shinchin384 12 дней назад

    Another "free" product. We all know it will change into a really "expensive" one finally

    • @benveasey7474
      @benveasey7474 11 дней назад

      do you think SD will remain free and Comfy will be expensive eventually?

  • @norbu_la
    @norbu_la 12 дней назад

    It's "we nerds can", not "us nerds can". You wouldn't say "us can". Why does literally everybody get this wrong?

    • @SyntheticVoices
      @SyntheticVoices 12 дней назад +7

      no one cares

    • @Elwaves2925
      @Elwaves2925 12 дней назад +9

      Says the person using 'literally' incorrectly.

    • @Statvar
      @Statvar 11 дней назад +2

      @@Elwaves2925 Bro all y'all are some nerds lmao

    • @ApexArtistX
      @ApexArtistX 11 дней назад

      Just like people using God instead of god .. internet does not care about grammar

    • @Elwaves2925
      @Elwaves2925 11 дней назад

      @ Educated dear boy/girl, educated and thank you for your compliment.

  • @mattm7319
    @mattm7319 9 дней назад

    hunyuan or ltx locally just smokes this, you can upscale vids with topaz video upscaler $299

  • @IshanJaiswal26
    @IshanJaiswal26 6 дней назад

    Watching with rtx 3060

  • @JoelleLB-xm1pi
    @JoelleLB-xm1pi День назад

    getting this error saying CosmosImageToVideoLatent is missing, but I downloaded the diffusion models and everything. Missing custom nodes dont show anything how can I fix this?

    • @NerdyRodent
      @NerdyRodent  День назад +1

      If you don’t have the built-in node, then it’s probable that you aren’t using a current version, but are instead using a very old version. What you will find is the previous versions don’t have the same functionality as the current versions. This means that whenever something new comes out, you won’t have it in your old version. Obviously this doesn’t just apply here, but generally throughout all software.