Комментарии •

  • @harshder4509
    @harshder4509 Год назад +1

    Awesome, Your explanation style is amazing. I understand it easily then ever. Thank you.

  • @mkamp
    @mkamp 2 года назад +4

    Awesome, as always.
    Great to see how you restrict the visual elements to support your message, like the colors. Also, very helpful to see you adding the dimensions as inline comments. It makes following along code much easier. Kudos.

  • @yizhouchen652
    @yizhouchen652 2 года назад +3

    I was looking for such tutorial for GPT last week but did not find a good one. Today your video just showed up, what a legend.

  • @beizhou2488
    @beizhou2488 2 года назад +2

    Another informative video! I am gonna need to watch it a couple of more times to digest all the knowledge. Thank you very much buddy.

  • @imdadood5705
    @imdadood5705 2 года назад +1

    Found this on Reddit. Seems like I was already a subscriber. I have stumbled upon a gold mine. I am still a beginner!!! These will be useful

  • @roomo7time
    @roomo7time 2 года назад +2

    immense thanks for your explanation and useful resource links!!!!!!!

  • @NeuralHackswithVasanth
    @NeuralHackswithVasanth 2 года назад +2

    Legend video. Expecting two videos from your side sir one on BERT implementation and other one is how do you implement a research paper code just by seeing the GitHub code and research paper on our own. These two would be really helpful. Hope i can get that

  • @MSEEchen
    @MSEEchen Год назад

    I first came to your git repo, then I found you here... Thanks for the explanation

  • @dizoner2610
    @dizoner2610 Год назад

    Thanks a lot for this perfect (I want to emphasise) Perfect video! 👍

  • @Neptutron
    @Neptutron 2 года назад +2

    This is awseome! Thank you!

  • @alivecoding4995
    @alivecoding4995 Год назад +3

    What tool did you use for making these nice presentstion slides? It looks very clean.

  • @evakreplova1534
    @evakreplova1534 2 года назад +2

    :-)) great video !

  • @muhtasham32
    @muhtasham32 2 года назад +2

    great video! thanks for sharing

  • @evab.7980
    @evab.7980 2 года назад +1

    👏👏👏

  • @tinyentropy
    @tinyentropy Год назад

    you are just amazing 😊

  • @AmeerHamza-xm5ro
    @AmeerHamza-xm5ro 2 года назад

    Thanks, keep up the good work.

  • @nehetnehet8109
    @nehetnehet8109 2 года назад +1

    Nice haircut:)great video!

  • @chickenp7038
    @chickenp7038 2 года назад +2

    why can’t we use the nn.Transformer Encoder and use the mask with that?

    • @mildlyoverfitted
      @mildlyoverfitted 2 года назад +1

      Great question:) I think we could! I just tried to implement as many things from scratch as possible to show what is "under the hood":)

    • @chickenp7038
      @chickenp7038 2 года назад +1

      @@mildlyoverfitted okay i tried it and weirdly increasing the number of encoder layers results in a higher minimum loss. super weird. i replaced the encoder with your decoder and it works perfectly.

  • @archeryisfun
    @archeryisfun 2 года назад

    Is the difference between HuggingFace gelu and PyTorch gelu due to the fact that the activation function is..... a Gaussian process?

    • @mildlyoverfitted
      @mildlyoverfitted 2 года назад +1

      Hmm, it seems like PyTorch and HuggingFace use different approximations. See paperswithcode.com/method/gelu . Anyway, it is a detail and maybe I shouldn't have bothered:)

  • @beizhou4025
    @beizhou4025 2 года назад

    Are you interested in making tutorial about data processing and training process of GPT model?

    • @mildlyoverfitted
      @mildlyoverfitted 2 года назад +1

      Yeh, potentially! The problem with training models like GPT is that one is probably better off using existing frameworks (e.g. transformers) rather than writing everything yourself. So the format would have to change from "from scratch" to "how to use a 3rd party framework". There is nothing wrong about that of course. It is just that I don't think I could do a better job of explaining it than the official documentation:) But I will definitely consider making videos like these too:)

    • @beizhou4025
      @beizhou4025 2 года назад +1

      @@mildlyoverfitted I am thinking to train a GPT-like model. Thank you for your suggestion. I will consider using the transformers model to do it. And it will also be great if you could do a tutorial about it.

  • @racocat1947
    @racocat1947 2 года назад +1

    BERT tutorial pls

  • @taibaili9074
    @taibaili9074 2 года назад

    BERT tutorial pls~