Large Language Models: How Large is Large Enough?

Поделиться
HTML-код
  • Опубликовано: 8 сен 2024
  • Explore IBM watsonx → ibm.biz/IBM-wa...
    When it comes to large language models, one of the first things you may read about is the enormous size of the model data or the number of parameters. But is "bigger" always better? In this video, Kip Yego, Program Marketing Manager explains that the truthful answer is "It depends". So what are the factors that will drive this decision? Kip breaks them down and explains one-by-one how to decide.
    Get started for free on IBM Cloud → ibm.biz/sign-u...
    Subscribe to see more videos like this in the future → ibm.biz/subscri...
    #llm #ai #generativeai #ml

Комментарии • 21

  • @ttjordan81
    @ttjordan81 8 месяцев назад +2

    Thank you, this is the information I was searching for. I was explaining the concept in theory to someone. The idea was to use smaller models that are trained for specific domains. By eliminating or reduce all the other domains, the model should perform better and reduce messy results.

  • @thatdudewiththething
    @thatdudewiththething 8 месяцев назад +4

    These videos are fantastic!
    Thank you so much for making them available :D

  • @dominiquecoladon8343
    @dominiquecoladon8343 8 месяцев назад +1

    Well done video, Get kip to more of these please.

  • @YvesNewman
    @YvesNewman 8 месяцев назад

    Great video Kip!
    At the moment it seems that bigger equals better. Time to change that perception accordingly

    • @sherpya
      @sherpya 8 месяцев назад

      already the trend, see mixture of experts concept

  • @donson3326
    @donson3326 2 месяца назад

    0:16 🤣 You tell me.

  • @LaurenFrazier-ch4kn
    @LaurenFrazier-ch4kn 8 месяцев назад

    Great video, super informative!

  • @gjjakobsen
    @gjjakobsen 8 месяцев назад

    The MBA in me says, beyond some point, the trade-off isn't worth it. Then again, that's probably what they said about the Apollo mission.

  • @7rich79
    @7rich79 8 месяцев назад

    Thank you, that was informative. One question I have is how you determine domain specificity, and perhaps potential lost opportunity?
    For example, using financial services tasks as in your example. If you ask someone working in finance about what insights they'd be looking for, tax or perhaps transfer pricing may not be what they consider as part of their domain. However, transfer pricing and tax could have a huge impact on what finance should consider when taking decisions. How do you ensure the domain specificity is not too narrow?

    • @julioberas2106
      @julioberas2106 7 месяцев назад

      I believe anything remotely related to the domains should be included in the training data. He didn't talk about the training data size, but I believe it should still be very big (but smaller than a general one)

  • @Alice8000
    @Alice8000 5 месяцев назад

    5:38 THANK YOU BRO. Definitely feel more confident after hearing that.

  • @fortuneosho8137
    @fortuneosho8137 8 месяцев назад

    It really depends

  • @IsaacFoster..
    @IsaacFoster.. 8 месяцев назад

    My llm's so large, it reaches almost every 1 and 0 it can write on; you can literally call it a "wipe"

  • @tyrojames9937
    @tyrojames9937 8 месяцев назад

    INTERESTING. 😀

  • @nirmal7103
    @nirmal7103 8 месяцев назад

    How can we find a domain specific models or how to train them?

    • @ttjordan81
      @ttjordan81 8 месяцев назад +1

      I think that's the next business idea, lol... At this point, pick an industry, and create specific domain model! It's a race! Also, specific domain Vector Databases will be needed!

  • @aberobwohl
    @aberobwohl 8 месяцев назад

    I see no point whatsoever in comparing a domain specific finetuned model to a non finetuned model to draw conclusions or suggest any insights doing this.

  • @deathlife2414
    @deathlife2414 8 месяцев назад

    Lets go phi. chroot chroot chroot

  • @warsin8641
    @warsin8641 8 месяцев назад

    The Bloke

  • @Alice8000
    @Alice8000 5 месяцев назад

    bro you got worse handwriting than me!!! Good info though. lol

  • @TheBiffsterLife
    @TheBiffsterLife 6 месяцев назад

    Kip, that’s a very poor analogy.