New Image2Video. Stable Video Diffusion 1.1 Tutorial.

Поделиться
HTML-код
  • Опубликовано: 12 фев 2024
  • SVD 1.1 Tutorial in ComfyUI. How to easily create video from an image through image2video.
    Detailed text & image guide for Patreon subscribers here: / turn-with-svd-1-98390700
    Download SVD model here huggingface.co/stabilityai/st...
    Basic ComfyUI Workflow comfyanonymous.github.io/Comf...
    How to install ComfyUI • How to install and use...
    Prompt styles for Stable diffusion a1111, Comfy & Vlad/SD.Next: / sebs-hilis-79649068
    Get early access to videos and help me, support me on Patreon / sebastiankamph
    Chat with me in our community discord: / discord
    My Weekly AI Art Challenges • Let's AI Paint - Weekl...
    My Stable diffusion workflow to Perfect Images • Revealing my Workflow ...
    ControlNet tutorial and install guide • NEW ControlNet for Sta...
    Famous Scenes Remade by ControlNet AI • Famous Scenes Remade b...
  • ХоббиХобби

Комментарии • 108

  • @sebastiankamph
    @sebastiankamph  3 месяца назад +4

    Detailed text & image guide for Patreon subscribers here: www.patreon.com/posts/turn-with-svd-1-98390700

    • @electronicmusicartcollective
      @electronicmusicartcollective 3 месяца назад

      pls add the Forge URL, peace

    • @nemesisone8927
      @nemesisone8927 5 дней назад

      i wonder if we will be able to select an area that moves in a direction instead of the whole image moving like a camera, this way i could make water move for 4 seconds and dump it into video AI and slow the image down and increase the frames per second and also pump it up to HD cool thing with Video AI is that it fills in the gaps in turn you get longer videos smooth and in slowmotion.

  • @ShocktorGaming
    @ShocktorGaming 3 месяца назад +1

    Excellent video - I will have to try this out tonight!

  • @adrianmunevar654
    @adrianmunevar654 3 месяца назад +7

    And today came Sora and boom! 💥
    Let's see what the SD team and the others do. Let the competition increase for our benefit 💪🏻😅

    • @N3_r
      @N3_r 3 месяца назад

      😂😂😂

  • @kaiz0099
    @kaiz0099 3 месяца назад +7

    Great stuff, more ruined fooocus tutorials please. Im also in love with it

    • @sebastiankamph
      @sebastiankamph  3 месяца назад +2

      I do like it, but there's not a lot to cover on it :D

  • @JefHarrisnation
    @JefHarrisnation 3 месяца назад +1

    Thanks for this. Question. In the Manager I see you have a process running view, is that a plug in?

  • @ken-cheenshang6829
    @ken-cheenshang6829 3 месяца назад

    thanks for sharing!!!

  • @sn0wbr33z3
    @sn0wbr33z3 3 месяца назад +7

    Waiting for someone to train commercially free models

    • @user-hy1rz9og1d
      @user-hy1rz9og1d 3 месяца назад

      As a Newbie in this space may I ask what’s the downpoint of this? I mean that you can’t sell your videos made using their model right? Or there’s more in their strictly agreement?

    • @ijayraj
      @ijayraj 3 месяца назад +1

      My work flow is not similar as him, do you have the same or different, althought i've perfectly installed everything but still having errors, what can i do in this ?

  • @Maltebyte2
    @Maltebyte2 3 месяца назад +1

    Atm is it only capable of outputting a 1sec clip?

  • @Avalon19511
    @Avalon19511 3 месяца назад

    What I have noticed about all these image2vid is the length seems to be the problem without having to do any editing, I think I'll wait to mess with these when the length is no longer a issue

  • @AlexandrePorto
    @AlexandrePorto 2 месяца назад

    Can you write text prompts for the videos? And can you extend the videos from the last frame and concatenate?

    • @sebastiankamph
      @sebastiankamph  2 месяца назад

      Not by default in SVD. I have seen some workflows that expand on svd with prompting but then it's just running the output with denoising.

  • @bladechild2449
    @bladechild2449 3 месяца назад +5

    comfy hasn't worked for half a year for me and I kinda gave up on it. just super finicky to get any and all of its many parts working in tandem. I kinda hope a lot of this stuff comes to fooocus. Just soooo much easier.

    • @sebastiankamph
      @sebastiankamph  3 месяца назад

      I understand that fully!

    • @ijayraj
      @ijayraj 3 месяца назад

      My work flow is not similar as him, do you have the same or different, althought i've perfectly installed everything but still having errors, what can i do in this ?

  • @screamingstarprime3503
    @screamingstarprime3503 3 месяца назад +4

    I’m having fun with it. Stumbled upon a trick to use LCM acceleration too. So far, just boils down to seed luck

    • @HistoryIsAbsurd
      @HistoryIsAbsurd 3 месяца назад

      Okay i thought i was crazy lol I had gotten this working one day too but was dumb and lost the JSON file for it and never could get it working again

    • @ijayraj
      @ijayraj 3 месяца назад

      I can't find the files to Download could you please let me know where is it ?

  • @pefisplace8483
    @pefisplace8483 2 дня назад

    Could someone explaine me, why i don´t have the "video combine" I mean the last Window where I can see the clip?? Thx for help

  • @lorenzo_susca
    @lorenzo_susca 2 месяца назад

    All very clear, except one thing, in what menu is the node "Video Combine"?

    • @sebastiankamph
      @sebastiankamph  2 месяца назад

      VHS - Video helper suite. Custom node.

  • @holly11111111
    @holly11111111 Месяц назад

    love it! one question, I like how it moves my animation but the faces are distorted, What parameters can I change or play?

    • @sebastiankamph
      @sebastiankamph  Месяц назад

      For SVD, none. You'd need a much more advanced workflow that re-renders the faces. It's all about rendering lots and being lucky.

  • @fluffycatmedia
    @fluffycatmedia 3 месяца назад +1

    if doing this in A1111 - which file would i put the following to work with low vram?
    pipe.enable_model_cpu_offload()
    pipe.unet.enable_forward_chunking()
    frames = pipe(image, decode_chunk_size=2, generator=generator, num_frames=25).frames[0]

    • @sebastiankamph
      @sebastiankamph  3 месяца назад

      What do you mean?

    • @ijayraj
      @ijayraj 3 месяца назад +1

      bro i'm havign eror but don't know coding could you pleaes help me.

  • @Lv7-L30N
    @Lv7-L30N 3 месяца назад

    Thank you

  • @lilillllii246
    @lilillllii246 2 месяца назад

    Thank you. Text prompt is not possible?

  • @rezahasny9036
    @rezahasny9036 3 месяца назад +1

    hello sir, can you give me the tutor how to creat video combine node on there ?

  • @adrianfels2985
    @adrianfels2985 День назад

    Am I right with the assumption that you have 0 control over the camera movement? The workflow simply decides what the "camera" has to do by analyzing the input?

  • @benharris144
    @benharris144 3 месяца назад +1

    Love your videos sebastian - Always to the point and thoroughly explained.
    Do you have any plans to cover the Forge WebUI?

    • @sebastiankamph
      @sebastiankamph  3 месяца назад +2

      Thank you! I just might. What do you think about it?

    • @benharris144
      @benharris144 3 месяца назад

      For me personally, the speed at which it generates SDXL images on my 2080 is so much faster than in a1111.
      Aside from that, I'm pretty new to all this and am still trying to figure out how to install it so it works with all my existing checkpoints/LoRas.@@sebastiankamph

    • @ijayraj
      @ijayraj 3 месяца назад

      My work flow is not similar as him, do you have the same or different, althought i've perfectly installed everything but still having errors, what can i do in this ?

  • @user-pw7ko2jd4j
    @user-pw7ko2jd4j 3 месяца назад

    Could you please do a tutorial for automatic1111?

  • @Deminii
    @Deminii 3 месяца назад

    can u tell name of extension this info system workin at queue prompt and loading % generation?

    • @amnzk08
      @amnzk08 3 месяца назад

      Yeah would like to know too!

  • @justanothernobody7142
    @justanothernobody7142 3 месяца назад +2

    Does anyone know what the VRAM requirements are for these models?

    • @ijayraj
      @ijayraj 3 месяца назад

      mnimum 4 gb

    • @justanothernobody7142
      @justanothernobody7142 3 месяца назад +1

      @@ijayraj Ok thanks. I read somewhere that it was 8, 4 seems a bit low to me. I have a 2070 so I have 8 anyway so will give it a go at some point.

    • @ijayraj
      @ijayraj 3 месяца назад

      @@justanothernobody7142 sure good luck bro

  • @VIpown3d
    @VIpown3d 3 месяца назад

    AI videos are awesome and absolutely frightening. My god how far they've come

    • @sebastiankamph
      @sebastiankamph  3 месяца назад +2

      Getting better each new version, just like in the early days of image generation :)

  • @twilightfilms9436
    @twilightfilms9436 3 месяца назад

    Sebastian, have you tried to replicate the image enhancement of Krea or Magnific? I spoke with the CEO of Magnific and he told me they have SDXL running under the hood…..I was just curious to know why no one has ever tried to do it……anyways, thanks for the video!

    • @sebastiankamph
      @sebastiankamph  3 месяца назад

      What you mean is they are replicating image enhancements of stable diffusion ;). It's just SD in a marketing package. No magic.

    • @ijayraj
      @ijayraj 3 месяца назад

      My work flow is not similar as him, do you have the same or different, althought i've perfectly installed everything but still having errors, what can i do in this ?

  • @chucknorris8704
    @chucknorris8704 3 месяца назад

    Can this be used with A1111 locally???

  • @DiwaSelvendran
    @DiwaSelvendran 3 месяца назад +1

    Can you make an video AMD installation of SD

    • @sebastiankamph
      @sebastiankamph  3 месяца назад +1

      I don't have an amd card, and I don't want to fake it :D

  • @inLofiLife
    @inLofiLife 14 дней назад

    can I use this for commercials? on yt?

  • @jimhowkings1974
    @jimhowkings1974 3 месяца назад +2

    hi. is there a way to tell it what we want ? or control the animation in any way?

    • @aegisgfx
      @aegisgfx 3 месяца назад

      No and thats the main problem, at this rate it'll be years before anyone has any fine level of control over what comes out of text to video. To me it looks like this tech is pretty much dead on arrival, and I dont see it getting better anytime soon.

    • @Wobbothe3rd
      @Wobbothe3rd 3 месяца назад

      ​@aegisgfx it will improve at a rapid pace just like all the other AI tech.

    • @aegisgfx
      @aegisgfx 3 месяца назад

      @@Wobbothe3rd ya Ive been waiting for that, seen no improvements so far

    • @Voltisvoltis
      @Voltisvoltis 14 дней назад

      @@aegisgfx And then Sora, Vidu and Google's video model got announced.

  • @97BuckeyeGuy
    @97BuckeyeGuy 3 месяца назад

    I hate it when an "upgrade" isn't a clear upgrade. Thank you.

  • @AISign-Lookbook
    @AISign-Lookbook 3 месяца назад

    can we use prompt for change?

    • @sebastiankamph
      @sebastiankamph  3 месяца назад +1

      Sadly not with this.

    • @uncleben2019
      @uncleben2019 3 месяца назад +1

      I’m guessing you can do text-to-image and then chain that to this workflow to achieve something similar. But yeah it’s just Image to Video for now.

  • @CVSiN
    @CVSiN 3 месяца назад +1

    Doesnt work well at all for me.. used your defaults and I get a model that looks deformed in animation..

  • @piechcik52
    @piechcik52 3 месяца назад

    Is possible to make longer video?? ~10 seconds

  • @ThaMoonwalkerFSDBetaChannel
    @ThaMoonwalkerFSDBetaChannel 3 месяца назад

    So i can use this to bring my images to life?

  • @ArcaneRealities
    @ArcaneRealities 3 месяца назад +1

    while the hamburger did look good it wasn't realistic to the scene unless u where spinning the camera and the hamburger was on a spinning plate - its to bad they can't give us access to a keyword system to control motions for foreground and background - seems like if they did simulations on the results by injecting words into the mix they could find the nodal points that generate consistent action within the results

  • @FlyingLotus
    @FlyingLotus 3 месяца назад

    nice ! is this possible with automatic1111?

  • @kaareej
    @kaareej 2 месяца назад

    I did not see where you put the prompts.

  • @nikgrid
    @nikgrid 3 месяца назад

    @sebastiankamph I've been messing about with Forge...it is FAST. Look forward to your video mate.

  • @nicoli3143
    @nicoli3143 3 месяца назад

    It's already good enough to replace entire production teams for commercials...

    • @ijayraj
      @ijayraj 3 месяца назад

      My work flow is not similar as him, do you have the same or different, althought i've perfectly installed everything but still having errors, what can i do in this ?

  • @nadrantalhaalhady1747
    @nadrantalhaalhady1747 3 месяца назад +1

    What is sowtware name?

    • @arothmanmusic
      @arothmanmusic 3 месяца назад

      He's showing Stable Diffusion in ComfyUI.

    • @Elwaves2925
      @Elwaves2925 3 месяца назад +1

      To be accurate it's Stable Video Diffusion 1.1

  • @Ollegruss_Music
    @Ollegruss_Music 14 дней назад

    Фэнкс!

  • @KardinalMoses
    @KardinalMoses 3 месяца назад

    I'm the 666th person to have liked this video :D may it bring you luck

  • @CamiloMonsalve
    @CamiloMonsalve 2 месяца назад

    Sorry, but I'm having this error:
    Error occurred when executing LoadImage:
    cannot identify image file 'H:\\ComfyUI\\ComfyUI_windows_portable\\ComfyUI\\input\\image_sample_1024x576_V01.png'
    File "H:\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 151, in recursive_execute
    output_data, output_ui = get_output_data(obj, input_data_all)
    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
    File "H:\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 81, in get_output_data
    return_values = map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True)
    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
    File "H:\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 74, in map_node_over_list
    results.append(getattr(obj, func)(**slice_dict(input_data_all, i)))
    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
    File "H:\ComfyUI\ComfyUI_windows_portable\ComfyUI
    odes.py", line 1459, in load_image
    img = Image.open(image_path)
    ^^^^^^^^^^^^^^^^^^^^^^
    File "H:\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\PIL\Image.py", line 3339, in open
    raise UnidentifiedImageError(msg)

  • @meadow-maker
    @meadow-maker 3 месяца назад

    and did you know the the 'gullible' isn't even in the Oxford English Dictionary? 🤣

  • @robertgoldbornatyout
    @robertgoldbornatyout 3 месяца назад +2

    Hi Sebastan very helpfull upload thanks, but please when are you going to do a help vid all about the new Forge UI - 75% faster than Automatic 1111, and also will the new Forge work on AMD gpus all other useless youtube uploaders seem to all forget about all us, usesers, who don't wont to spend or can not spend lots of money on new hardware just to test out for example the new forge, hope you understand thanks,
    your channel has always been the best for all my AI Stable Diffusion help thanks, 💯💯💯👍👍👍

    • @sebastiankamph
      @sebastiankamph  3 месяца назад +1

      Good suggestion! And thank you :)

    • @robertgoldbornatyout
      @robertgoldbornatyout 3 месяца назад +1

      Thanks Sebastian you're the one who got me into A.I. art But A1111 is still a bit slow on my AMD, would be fantastic for all us penny less AMD users if we all could start using Forge. @@sebastiankamph

    • @ijayraj
      @ijayraj 3 месяца назад

      My work flow is not similar as him, do you have the same or different, althought i've perfectly installed everything but still having errors, what can i do in this ?

  • @aegisgfx
    @aegisgfx 3 месяца назад +2

    The pace of text to video tech has been really slow compared the text to image stuff. I have to say all the txt-video solutions I have seen are incredibly unremarkable, especially compared to the txt-image models which are all phenomenal. Unless the ball gets rolling on txt-video pretty fast, Im calling this technology dead on arrival.

    • @bladechild2449
      @bladechild2449 3 месяца назад

      Image stuff that isn't the new dall e 3 / GPT tech has been practically stale for a year now. Nothing really has improved, and with the release of SDXL it took a bit of a hit because training your own loras and shit for it is as big a pain as it was in the early days of training for SD 1.5

    • @sebastiankamph
      @sebastiankamph  3 месяца назад +2

      Slower, yes. And video consistency is also much harder to achieve than a still image. Looking back a year from now, I'm sure the comparison will be breathtaking :)

    • @Bentler
      @Bentler 2 месяца назад

      ​@bladechild2449 I thought it was just me. I took a break for a year and came back and I have been struggling to find anything thats really new.. 2022-early23 was moving so fast.

  • @luman1109
    @luman1109 3 месяца назад +7

    It's still not great

  • @ijayraj
    @ijayraj 3 месяца назад +1

    Great info man, thanks for keeping us upated about new stuffm i followed the whole video but still getting this erroe
    This one in command prompt
    """
    0%| | 0/20 [00:02

    • @sebastiankamph
      @sebastiankamph  3 месяца назад +1

      You don't have enough video/gpu memory. It says out of memory there sadly.

    • @ijayraj
      @ijayraj 3 месяца назад +1

      @@sebastiankamph I've 4 GB RTX 3050, can't we solve this error?

    • @ijayraj
      @ijayraj 3 месяца назад +1

      @@sebastiankamph i wants to use this feature 😔

    • @sebastiankamph
      @sebastiankamph  3 месяца назад

      You can use a cloud solution like Thinkdiffusion.com@@ijayraj

    • @ijayraj
      @ijayraj 3 месяца назад +1

      @@sebastiankamph will try it, thanks man, thanks you very much for the support.