Lesson 2: Byte Pair Encoding in AI Explained with a Spreadsheet

Поделиться
HTML-код
  • Опубликовано: 10 фев 2025
  • In this tutorial, we delve into the concept of Byte Pair Encoding (BPE) used in AI language processing, employing a practical and accessible tool: the spreadsheet.
    This video is part of our series that aims to simplify complex AI concepts using spreadsheets. If you can read a spreadsheet, you can understand the inner workings of modern artificial intelligence.
    🧠 Who Should Watch:
    Individuals interested in AI and natural language processing.
    Students and educators in computer science.
    Anyone seeking to understand how AI processes language.
    🤖 What You'll Learn:
    Tokenization Basics: An introduction to how tokenization works in language models like Chat GPT.
    Byte Pair Encoding (BPE): Detailed walkthrough of the BPE algorithm, including its learning phase and application in language data tokenization.
    Spreadsheet Simulation: A hands-on demonstration of the GPT-2's tokenization process via a spreadsheet model.
    Limitations and Alternatives: Discussion on the challenges of BPE and a look at other tokenization methods.
    🔗 Resources:
    Learn more and download the Excel sheet at spreadsheets-a...

Комментарии • 19

  • @rshaikh05
    @rshaikh05 Месяц назад

    wow! what an explanation. it cannot get better than this. thank you so much.

  • @lewisbentley3177
    @lewisbentley3177 11 месяцев назад +3

    This is amazing! Thank you for making this. Really looking forward to your next video about text and position embeddings

  • @michaelmalone7614
    @michaelmalone7614 11 месяцев назад +4

    The quality of this video has blown my mind! Really great stuff and looking forward to your next video. Thank you!

  • @somdubey5436
    @somdubey5436 6 месяцев назад

    seeing this video made me feel I am not even close to say I know excel. Your understanding of concept is really really deep as implementing something like GPT-2 in excel requires one to have thorough understanding of all the concepts. Hats off to you.

  • @paulwillis88
    @paulwillis88 10 месяцев назад +1

    This is an incredible video. The way you describe these advanced AI concepts is awesome. I'd love to see more, maybe on the GAN technology that Sora uses

  • @BryanSeigneur0
    @BryanSeigneur0 10 месяцев назад

    So clear! Great instruction!

  • @fastler2000
    @fastler2000 10 месяцев назад

    Incredibly brilliant. Words fail me. Thank you for sharing this, it helps me enormously in understanding AI.
    On a side note: How ingenious can you handle Excel? Unbelievable!

  • @jjokela
    @jjokela 10 месяцев назад

    Really cool stuff, it really helps me to understand how the BPE works. Looking forward to your follow-up videos!

  • @rpraver1
    @rpraver1 Год назад +4

    Very good explanation, are you going to go into positional embedding?

    • @Spreadsheetsareallyouneed
      @Spreadsheetsareallyouneed  11 месяцев назад +7

      thank you! yes just haven't had time to get around to it yet. Embeddings will be the next video. Not sure if I'll do token and positional embeddings in the same video or will break it up into two parts.

  • @MStrong95
    @MStrong95 10 месяцев назад +1

    Large language models and AI in general seems to do a good job of compressing and then turning back into an approximation of the input. Is this a byproduct of nural networks in general or just specific subsets? Could you make a large language model or a lot of purpose built AI that are good for various different compression situations and more often than not perform better than current compression algorithms?

  • @Cal1fax
    @Cal1fax 11 месяцев назад +1

    I learned a ton from your video

  • @defface777
    @defface777 10 месяцев назад +1

    Very cool! Thanks

  • @ricp
    @ricp 11 месяцев назад +1

    great explanations

  • @lordadamson
    @lordadamson 10 месяцев назад

    amazing work. please keep going :D

  • @kennrich213
    @kennrich213 10 месяцев назад

    Great video. How exactly are the scores calculated from the possible pairs? You said a standard filter? Could you explain more?

  • @JohnSmith-he5xg
    @JohnSmith-he5xg 10 месяцев назад

    Why does having a large Embedding Table matter? Can't it just be treated as a lookup into the Table (which should be pretty manageable regardless of size)? Do we really have to perform the actual matrix multiply ?

  • @magickpalms4025
    @magickpalms4025 11 месяцев назад +1

    Why did they include reddit usernames in the training? Considering the early days where people would have extreme/offensive names as a meme, that is just asking for trouble.

  • @ameliexang7543
    @ameliexang7543 Год назад

    promo sm