Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer

Поделиться
HTML-код
  • Опубликовано: 1 окт 2024
  • This video explores the T5 large-scale study on Transfer Learning. This paper takes apart many different factors of the Pre-Training then Fine-Tuning pipeline for NLP. This involves Auto-Regressive Language Modeling vs. BERT-Style Masked Language Modeling and XLNet-style shuffling, as well as the impact of dataset composition, size, and how to best use more computation. Thanks for watching and please check out Machine Learning Street Talk where Tim Scarfe, Yannic Kilcher and I discuss this paper!
    Machine Learning Street Talk: / @machinelearningstreet...
    Paper Links:
    T5: arxiv.org/abs/...
    Google AI Blog Post on T5: ai.googleblog....
    Train Large, Then Compress: arxiv.org/pdf/...
    Scaling Laws for Neural Language Models: arxiv.org/pdf/...
    The Illustrated Transformer: jalammar.github...
    ELECTRA: arxiv.org/pdf/...
    Transformer-XL: arxiv.org/pdf/...
    Reformer: The Efficient Transformer: openreview.net...
    The Evolved Transformer: arxiv.org/pdf/...
    DistilBERT: arxiv.org/pdf/...
    How to generate text (HIGHLY RECOMMEND): huggingface.co...
    Tokenizers: blog.floydhub....
    Thanks for watching! Please Subscribe!

Комментарии • 27

  • @connor-shorten
    @connor-shorten  4 года назад +7

    2:00 Pushing the NLP State-of-the-Art
    2:40 Text-to-Text Framework
    3:28 Factors of Variation Explored
    5:00 Value of Pre-Training
    5:25 Attention Masking
    6:18 Architecture Results
    7:02 Denoising Objectives
    8:47 Span Corruption Strategy
    9:45 Self-Supervised Learning Study Overview
    11:14 Datasets
    12:24 Dataset Size
    12:56 Fine-Tuning Strategy
    14:25 Task Imbalance
    15:20 Pre-Train, then Fine-Tune
    16:26 How should we use extra computation?
    18:47 Scaling up to 11B parameters
    19:30 What Didn’t Make the List
    22:08 Context-Free Question Answering

  • @vatsalkrishna5627
    @vatsalkrishna5627 Год назад +3

    I never expected to learn so much from one single video. Amazing work presenting the paper in such a nuanced way!

  • @emanuelgerber
    @emanuelgerber 3 года назад +5

    Thank you! This helped me a lot to understand all the different aspects of T5

    • @connor-shorten
      @connor-shorten  3 года назад

      Thanks Emanuel, really glad to hear that!

  • @rohitghule9437
    @rohitghule9437 4 года назад +3

    Why so fast

  • @forcedlevy
    @forcedlevy 3 года назад +2

    Watch at 0.75 speed

  • @BiancaAguglia
    @BiancaAguglia 4 года назад +1

    You're getting better and better at explaining these papers, Connor. Great job. Also, I enjoyed the conversation on the Machine Learning Street Talk channel. Looking forward to seeing more videos there too. 😊
    I've decided to start studying NLP in a more organized manner (right now I have some intuition about how it works, but not much theoretical or practical knowledge.) I'll be watching your NLP videos when I need a productive break from my studies. 😊
    P.S. I'm embarrassed to admit that only today I found out your first name was Connor. For some reason I thought it was Henry.

  • @justinmilner8
    @justinmilner8 Год назад

    Is 'deshuffling' really an accurate description of the XLNet pre-training objective? To me, deshuffling indicates prediction of the order of tokens within the text - which is not matching with my understanding of XLNet's pretraining objective.

    • @justinmilner8
      @justinmilner8 Год назад

      Yes I can confirm now that the deshuffling objective referred to in the T5 paper is not referencing XLNet's permutation masking objecive. (Deshuffling it is cited to SummAE in the T5 paper)

  • @TimScarfe
    @TimScarfe 4 года назад +3

    Amazing job Connor!

  • @heinsaar
    @heinsaar 4 года назад

    Thanks for sharing! It would be wonderful if you could get a better mic though. The laptop mic has a very unpleasant echo.

  • @tommykelly6840
    @tommykelly6840 2 года назад

    What is the difference between iid mask tokens and Bert Style mask tokens

  • @L33TNINJA51
    @L33TNINJA51 4 года назад +3

    A little hard to follow as someone who hasn't learned much about AI, but still enjoy your videos!

    • @hoangnhatpham8076
      @hoangnhatpham8076 4 года назад

      The target audience of these kinds of videos isn't supposed to be someone who hasn't learned much about AI anyway.

    • @L33TNINJA51
      @L33TNINJA51 4 года назад

      @@hoangnhatpham8076 I guess I just need to get to learning and stop being just a fanboy :)

    • @bikideka7880
      @bikideka7880 4 года назад +1

      I think it is hard to follow the videos for someone who do not encounter with research papers(like me).

  • @MakerBen
    @MakerBen 4 года назад +1

    Thanks for posting this! This is super helpful!

    • @connor-shorten
      @connor-shorten  4 года назад

      Thank you so much! Glad you found this useful!

  • @---kt8cs
    @---kt8cs 4 года назад

    Thank you, sir, your videos are gold!

  • @SantoshGupta-jn1wn
    @SantoshGupta-jn1wn 4 года назад

    These videos are amazing, thanks Henry

  • @taku8751
    @taku8751 4 года назад

    amazing

  • @salimbo4577
    @salimbo4577 3 года назад +1

    how much time does it take you guys to read a research paper and what parts do you read. because everytime i try to read one i strart loosing focus, any tips pls help

    • @zeinramadan
      @zeinramadan 3 года назад +3

      read the paper in 3 different passes:
      1) In your first pass, start with reading the following sections within the paper: title, abstract and figures.
      2) The second pass entails you reading the following sections: introduction, conclusion, another pass through figures and scan through the rest of the content.
      ⁃ The introduction and conclusion section of a paper contains clear and concise information on the content of the paper and a summary of any findings. The information presented in this section usually dismisses any supplementary information and only key information are included. This is beneficial to you as a reader as you get the vital information required to proceed to the other sections within the paper.
      3) The third pass of the paper involves reading the whole sections within the paper but skipping any complicated maths or technique formulations that might be alien to you
      ⁃During this pass, you can also skip any terms and terminologies that you do not understand or aren’t familiar.
      Check out this article by Andrew Ng about how to efficiently read papers: towardsdatascience.com/how-you-should-read-research-papers-according-to-andrew-ng-stanford-deep-learning-lectures-98ecbd3ccfb3

  • @dislike__button
    @dislike__button 2 года назад

    I still don't understand how did they combine training on the C4 dataset and all the task specific datasets (squad etc).
    What role did the C4 datset play? How did they turn the raw text data of C4 into a input output task to train on?
    Would be grateful if someone could explain, thanks.