Deterministic Image Editing with DDPM Inversion, DDIM Inversion, Null Inversion and Prompt-to-Prompt

Поделиться
HTML-код
  • Опубликовано: 2 янв 2025

Комментарии • 7

  • @FangruiHuang
    @FangruiHuang 20 дней назад

    why you can directly change x_t-1 to x_t+1 in 29:21.

    • @gabrielmongaras
      @gabrielmongaras  15 дней назад

      It depends on if you do a forward step or an inversion step in the opposite direction using the formulas above. Both come from x_t though.

  • @EkShunya
    @EkShunya 5 месяцев назад +1

    great one, really liked it
    thanks

  • @陈兆伟-s5w
    @陈兆伟-s5w 4 месяца назад

    How is the equality in DDPM established in 17:49?

    • @gabrielmongaras
      @gabrielmongaras  4 месяца назад +1

      Looks like I forgot to write out the square root over the first term. As for the inner term that got turned into a fraction, I just multiplied sqrt{1-a_t} by the fraction sqrt{1-a_t}/sqrt{1-a_t}.

  • @ml-ok3xq
    @ml-ok3xq 2 месяца назад

    congrats on writing a paper! i notice that another recent paper from NVIDIA uses a unit vector for attention (nGPT) where the dot product is naturally equal to cosine as the lengths are one. are these two works related to each other in any way?

    • @gabrielmongaras
      @gabrielmongaras  2 месяца назад

      Thanks!! I only read through the nGPT paper briefly, but I think nGPT was trying to make softmax attention/transformers more expressive and efficient by changing a few things. They do normalize before they apply the softmax function, making the logits a cosine similarity between -1 and 1. However they keep the softmax operation which forces the model to stay quadratic in terms of complexity. The paper I worked on removed the softmax function which allowed the attention mechanism to be changed into an RNN which is linear in complexity.