Stanford CS25: V4 I Demystifying Mixtral of Experts

Поделиться
HTML-код
  • Опубликовано: 29 янв 2025

Комментарии • 8

  • @yimahuang1916
    @yimahuang1916 Месяц назад

    great❤❤❤

  • @marknuggets
    @marknuggets 8 месяцев назад +1

    Cool format, Stanford quickly becomes my favorite blogger lol

  • @r.d.7575
    @r.d.7575 4 дня назад

    47:00 Anyone knows the paper that suggests that learning to route isn't any better than random routing ?

  • @crwhhx
    @crwhhx 5 месяцев назад

    6:22 here “xq_LQH = wq(x_LD).view(L, N, H)” should be “xq_LQH = wq(x_LD).view(L, Q, H)” right?

  • @何孟飞
    @何孟飞 8 месяцев назад +1

    where to get slides

  • @acoustic_boii
    @acoustic_boii 8 месяцев назад +1

    Dear Stanford online recently I have completed product management course from Stanford online but i haven't got the certificate help me please how will I get the certificate

    • @Ethan_here230
      @Ethan_here230 8 месяцев назад +1

      Wait u will get it
      - Ethan from Stanford

  • @gemini_537
    @gemini_537 7 месяцев назад

    Gemini 1.5 Pro: The video is about demystifying mixture of experts (MoE) and Sparse Mixture of Experts (Smoe) models.
    The speaker, Albert Jang, who is a PhD student at the University of Cambridge and a scientist at Mistral AI, first introduces the concept of dense Transformer architecture. Then he dives into the details of Smoes. He explains that Smoes are a type of neural network architecture that can be more efficient than standard Transformers by using a gating network to route tokens to a subset of experts. This can be useful for training very large models with billions of parameters.
    Here are the key points from the talk:
    * Mixture of Experts (MoE) is a neural network architecture that uses a gating network to route tokens to a subset of experts.
    * Sparse Mixture of Experts (Smoe) is a type of MoE that can be more efficient than standard Transformers.
    * Smoes use a gating network to route tokens to a subset of experts, which can be more efficient than training a single large model.
    * Smoes are well-suited for training very large models with billions of parameters.
    The speaker also discusses some of the challenges of interpreting Smoes and the potential for future research in this area. Overall, the talk provides a good introduction to Smoes and their potential benefits for training large language models.