What Makes Large Language Models Expensive?

Поделиться
HTML-код
  • Опубликовано: 29 май 2024
  • Explore watsonx.ai → ibm.biz/IBM_watsonx_ai
    Amidst the buzz surrounding the promising capabilities of large language models in business, it's crucial not to overlook a practical concern: cost. In this video, Jessica Ridella, Jessica Ridella, IBM’s Global Sales Leader for the watsonx.ai generative AI platform, delves into seven pivotal factors crucial for understanding generative AI in the enterprise. She explores elements influencing cost, such as model size and deployment options, while also shedding light on potential cost-saving strategies like harnessing pre-trained models. By the video's conclusion, you'll gain a comprehensive understanding of the factors influencing costs and discover optimal strategies for the efficient utilization of large language models in your enterprise.
    Get started for free on IBM Cloud → ibm.biz/sign-up-now
    Subscribe to see more videos like this in the future → ibm.biz/subscribe-now

Комментарии • 70

  • @aqynbc
    @aqynbc 5 месяцев назад +5

    Another excellent videos that makes you understand the fundamentals of an otherwise complicated subject.

  • @emil8367
    @emil8367 5 месяцев назад +1

    Very interesting and useful. Thanks for explaining so many topics !

  • @benthiele
    @benthiele 5 месяцев назад

    Incredibly helpful video. Please make more!

  • @saikatnextd
    @saikatnextd 3 месяца назад

    Thanks Jessica for this video, really eye opening and introspective at the same time.......

  • @attainconsult
    @attainconsult 5 месяцев назад

    this is a great start to costing running models, I think you need to think/explain more along the lines of business i.e. adding in all biz file/google/365 docs, biz emails, other biz data sales cash flow, stock usage, forecasting usage of consumables lettuces coffee... all the things biz work off

  • @seanlee2002
    @seanlee2002 5 месяцев назад

    Excellent explanation. A great understanding of how AI works

  • @shaniquedasilva1856
    @shaniquedasilva1856 4 месяца назад +3

    Great video Jessica and so informative!! I’m working on a project now implementing Gen AI (gen fallback, generators). Identifying proper use cases are so important to yield the best results while thinking about the # of LLM calls.

    • @unclenine9x9
      @unclenine9x9 3 месяца назад

      Yes, we need to select the suitable LLMs for pickings up the request with cost effective way. Thus the cost of operation should be lowered.

  • @user-mt5zb7qs7q
    @user-mt5zb7qs7q 5 месяцев назад

    Great explanation Jessica

  • @aymerico11
    @aymerico11 4 месяца назад

    Very good video thanks a lot !

  • @KP-sg9fm
    @KP-sg9fm 5 месяцев назад +8

    Can you make a video talking about smaller more effecient models (Orca, Phi II, Gemini Nano, etc)
    Do they have a future, and if so, what does it look like?
    Will more sota models leverage the techniques used by smaller models to become more effiecient?
    Or will they always remain separate?

    • @teleprint-me
      @teleprint-me 5 месяцев назад +2

      There are pros and cons to each approach. Larger models are scaled in a way that makes their capabilities proportional to their parameters. So, larger models are smarter and that will always be the case.
      Both techniques feed off of one another, so improvements in one will lead to improvements in another.
      It's cheaper and easier and faster to iterate over smaller models and any gains made throughout the process are applied to larger models.
      Not sure if this helps. Anyone can feel free to correct me if I misrepresented any information.

  • @markfitz8315
    @markfitz8315 5 месяцев назад +1

    very good - thanks

  • @Murat-hh4hu
    @Murat-hh4hu 5 месяцев назад +67

    For a moment I thought she is AI generated)

  • @imanrezazadeh
    @imanrezazadeh 5 месяцев назад +5

    Excellent explanation! A minor note: the analogy of curtain makes sense, but then you mentioned fine-tuning makes structural changes to the parameters, which is not accurate. It just changes the values of the parameters.

    • @aymerico11
      @aymerico11 4 месяца назад

      How does it change the value ? Is it token change ? Basically it means that once you've tuned your model f(x) no longer equals y but actually z right ?

  • @oieieio741
    @oieieio741 5 месяцев назад +6

    Excellent explanation. A solid understanding of how AI works. Thanks IBM

  • @gamingbeast710
    @gamingbeast710 5 месяцев назад +1

    awsome , 100% focued :D thx for the professionalisme :D

  • @jediTempleGuard
    @jediTempleGuard 5 месяцев назад +8

    I think customized language models will become more important over time. Companies will want artificial intelligence applications specific to their fields of activity, and individuals will want artificial intelligence applications specific to their special interests. Not to sound like I'm telling fortunes, but with improvements in cost, customized smaller models may become more dominant in the market.

    • @Cahangir
      @Cahangir 5 месяцев назад +2

      what types of AI apps would individuals want apart from personal assistants that would need customizing?

    • @Anurag_Hansda
      @Anurag_Hansda 4 месяца назад

      I very much agree with you... Google could be much more efficient by giving specific detail.

  • @bastabey2652
    @bastabey2652 5 месяцев назад

    I once attended a whole day IBM sales presentation in Delhi for telco CRM/Billing system.. it was an educational experience more than sales.. IBM sales is really good

  • @reazulislam8446
    @reazulislam8446 5 месяцев назад

    So precise..

  • @webgpu
    @webgpu 5 месяцев назад +2

    anyone noticed she kept on talking * while * writing ? women are real multitaskers - i swear to God my brain is 100% monotask and i could never Ever: write AND do anything else. The apex of my manly monotaskiness is to be able to talk while i'm driving (but i can only talk about light subjects, if you talk about anything a little more involved, i will just not follow you.

  • @AdamSioud
    @AdamSioud 5 месяцев назад

    Great video

  • @teresafarrer1252
    @teresafarrer1252 4 месяца назад

    Great video: really clear and professional (unlike a couple of the saddos commenting). Thanks!

  • @fasteddylove-muffin6415
    @fasteddylove-muffin6415 5 месяцев назад +2

    You walk into a dealership & ask a salesperson how much a vehicle will cost.
    Answer: This vehicle will cost you whatever you're willing to pay.

  • @ChrisJSnook
    @ChrisJSnook 4 месяца назад

    What software solution powers this mirrored whiteboard in front of you? It’s awesome and I want to use it?

  • @carkawalakhatulistiwa
    @carkawalakhatulistiwa 5 месяцев назад +4

    And PHI-2 with 2,7 B billion parameters. proves that we have spent a lot of time and money on computerization that is wasted because of bad data.
    with better data PHI-2 LLM can be equivalent to gpt 3 175 billion parameters . and there is still the possibility to reduce LLM to 1 billion parameters with the same capabilities

    • @akj3344
      @akj3344 5 месяцев назад

      There are 1B models on huggingface made for RAGs.

  • @Alice8000
    @Alice8000 2 месяца назад

    Daaaaamn woman. Good explanation.

  • @silberlinie
    @silberlinie 5 месяцев назад +1

    They used an interesting technique to record the video.

  • @mohsenghafari7652
    @mohsenghafari7652 3 месяца назад

    hi. please help me. how to create custom model from many pdfs in Persian language? tank you.

  • @renanmonteirobarbosa8129
    @renanmonteirobarbosa8129 5 месяцев назад +1

    There are mistakes with the information provided.
    PEFT and Lora are separate things
    model size is influenced mostly by numerical choice and how you compile the GPU kernel.
    ...

  • @team-m2
    @team-m2 5 месяцев назад

    Great and concise, thanks! But ... is she writing from the right to the left? 🤔

  • @gihan5812
    @gihan5812 5 месяцев назад

    How can i speak to someone at IBM about working together.

  • @johnnyalam7301
    @johnnyalam7301 5 месяцев назад

    Very nicely and intelligently explained 3:49 pm ( Christmas Day 2023)

  • @potatodog7910
    @potatodog7910 5 месяцев назад

    Nice

  • @mrd6869
    @mrd6869 5 месяцев назад

    Small and powerfulmodels will win out.Phi 2 and Orca2 are some good examples.

  • @luciengrondin5802
    @luciengrondin5802 5 месяцев назад

    Stumbled upon this and feel like asking : how did IBM miss the LLM train? Watson was very impressive IMHO. Very much ahead of its time. How could IBM not capitalize on it? Why was it OpenAI that ended up with the language model breakthrough? Which innovation openAI had that IBM could not think of? Was it RLHF?

    • @VoltLover00
      @VoltLover00 5 месяцев назад

      You can easily google the answer to your question

  • @potatodog7910
    @potatodog7910 5 месяцев назад

    How much of this can be done with GPTs?

    • @scottt9382
      @scottt9382 5 месяцев назад +1

      A GPT is just one type of an LLM

  • @cleansebob1
    @cleansebob1 5 месяцев назад

    Looks like it all depends...

  • @SiegelBantuBear
    @SiegelBantuBear Месяц назад

    🙏🏼

  • @wzqdhr
    @wzqdhr 5 месяцев назад

    Does IBM have anything to do with this AI booming?

  • @rursus8354
    @rursus8354 5 месяцев назад

    If you cannot find the best man, take the next best.

  • @markmaurer6370
    @markmaurer6370 5 месяцев назад

    1:19 So IBM does not believe consumers need to have their data protected.

  • @jameshopkins3541
    @jameshopkins3541 4 месяца назад

    THEN A COMMON PERSON CAN'T DO A LLM FROM SCRATCH???

  • @jameshopkins3541
    @jameshopkins3541 4 месяца назад

    She is 36 years old Isn't it?

  • @jameshopkins3541
    @jameshopkins3541 4 месяца назад +1

    LLM IS BLA BLA BLAAAAA??????

  • @joung-joonlee1037
    @joung-joonlee1037 5 месяцев назад +1

    I think, that LLM or GAI Look like Spread-Sheet if concern the facts that this type of engine inject By SELF toward tokens and Spell Out tokens..!! AND This type of tokens look like iterated by LLM or GAI, because that is also programs using Computer Iterations...! AND The LLM or GAI's using cost can be acquired using calculations over Time/Number of Tokens/Weight of Meaning.... But, I know that this calculations is just approximation by User. Thank you for NICE Video! and I'm korean.

  • @NisseOhlsen
    @NisseOhlsen 5 месяцев назад +1

    What makes them so expensive? Simple. Their Architecture is not right.

  • @Free-pp8mr
    @Free-pp8mr 5 месяцев назад

    It is not intelligent to pay for AI! It’s simply marketing!

  • @ciphore
    @ciphore 5 месяцев назад

    Nancy Pi did it first 😤

  • @bobanmilisavljevic7857
    @bobanmilisavljevic7857 5 месяцев назад +1

    🦾🥳

  • @Canadainfo
    @Canadainfo 5 месяцев назад

    amazon bedrock!!

  • @Drunrealer
    @Drunrealer 2 месяца назад

    Drink from de bottle

  • @thierry-le-frippon
    @thierry-le-frippon 4 месяца назад

    People will pay for that 😅😅😅 ???

  • @aprilmeowmeow
    @aprilmeowmeow 2 месяца назад

    so sad that people cant even write a speech anymore.

  • @ashishsehrawat_007
    @ashishsehrawat_007 5 месяцев назад

    Kinda boring explanation.

  • @talatshahgmailcom
    @talatshahgmailcom 4 месяца назад

    Thank you, very informative and easily understandable.