Types of Attention in NLP and Transformer Multi-Head Attention Explained.

Поделиться
HTML-код
  • Опубликовано: 13 янв 2025

Комментарии • 9

  • @mosesmaned2151
    @mosesmaned2151 Год назад

    your videos are game changing for sure thank you very much please you are a life saver

  • @wilfredomartel7781
    @wilfredomartel7781 10 месяцев назад

    😊😊❤amazing tutorial man

  • @maryjoudaki
    @maryjoudaki 2 года назад +1

    your videos are great. thanks a lot ! 🤩

  • @wolfisraging
    @wolfisraging 3 года назад

    Amazing, waiting for more videos!!

    • @wolfisraging
      @wolfisraging 3 года назад

      @@machinelearningwithpytorch Amazing!

  • @veenasarda2841
    @veenasarda2841 3 года назад +1

    How are query, key and values calculated

  • @v1hana350
    @v1hana350 2 года назад

    What is the meaning of fine-tuning and Pre-trained in Transformers?

  • @medozeus2404
    @medozeus2404 2 года назад

    I Am not able hear that's sounds of pen when it's moving on paper that's only thing made me leave the tutorial

  • @imjaekyu
    @imjaekyu 2 года назад

    I think your entire softmax values are wrong, for example, the first entry of a first row should be 0.55 not 1. (e^0.9)/ (e^0.9)+(e^0)+(e^0) = 0.55123.