The 10 Trillion Parameter AI Model With 300 IQ

Поделиться
HTML-код
  • Опубликовано: 13 янв 2025

Комментарии • 108

  • @chapterme
    @chapterme 2 месяца назад +79

    Chapters (Powered by ChapterMe) -
    0:00 Coming Up
    0:54 What models get unlocked with the biggest venture round ever?
    5:35 Some discoveries take a long time to actually be felt by regular people
    9:53 Distillation may be how most of us benefit
    14:26 o1 making previously impossible things possible
    21:17 The new Googles
    23:47 o1 makes the GPU needs even bigger
    25:44 Voice apps are fast growing
    27:05 Incumbents aren’t taking these innovations seriously
    31:52 Ten trillion parameters
    33:15 Outro

  • @thehari75
    @thehari75 2 месяца назад +64

    Lol did sam altman tell yall to avoid talking about claude and comouter use update

    • @sakchais
      @sakchais 2 месяца назад +4

      They talked about developers in their batch switching from OpenAI's models to Claude's models at around the 12:50 mark.

    • @deviceinside
      @deviceinside 25 дней назад +1

      who's Sam Altman?

  • @miraculixxs
    @miraculixxs 2 месяца назад +30

    There is literally nothing in this podcast that make it worth spending the time to listen to it

    • @Zeegoner
      @Zeegoner 2 месяца назад +9

      Was looking for this comment. It's like a kindergarten class.

    • @lucface
      @lucface 2 месяца назад +1

      Need a 3x button on yt

    • @rokrok27
      @rokrok27 Месяц назад

      😭😭

    • @samy.achour
      @samy.achour Месяц назад

      My thoughts exactly

    • @hypemanD
      @hypemanD 25 дней назад

      Accurate

  • @hanchisun6164
    @hanchisun6164 2 месяца назад +13

    One order of magnitude higher is referring to compute, not parameter size. The current frontier models (4o/pro/sonnet) are all much smaller than their previous sota (4/ultra/opus).
    GPT4-1.8T MoE might continue to be the largest model people train for a while.

  • @oiuhwoechwe
    @oiuhwoechwe 2 месяца назад +8

    but can it code crysis?

  • @amirsync
    @amirsync Месяц назад +1

    I'm a VoIP (Voice Over Internet Protocol) developer at a small company, I want to start a startup in the comming 5 months and now I'm looking for co-founders, my previous startup (in generative AI space) got accepted into Microsoft for founders hub, if you are a technical, innovative or market person reply me here to discuss more. 😊

  • @fil4dworldcomo623
    @fil4dworldcomo623 2 месяца назад +1

    Winning AI company capturing all the value is speculative, therefore an answer can also be speculative that the more any AI becomes powerful, the more complex society becomes, hence the value of being human to cope with new contexts to feed the "hungry AI" proportionately increases and hopefully it's fun.

  • @ayakitokugawa
    @ayakitokugawa 2 месяца назад +2

    Fact Check: Terence Tao's parents are first-generation immigrants from Hong Kong to Australia. Which makes him Australian-American after he naturalised for duo citizenship.

    • @ycombinator
      @ycombinator  2 месяца назад

      Oops. Sorry about this. -Garry

  • @artefactxinnovation
    @artefactxinnovation 2 месяца назад +2

    Thanks for the inspiring sharing! Let's leverage the power of AI to improve human lives and building a more sustainable future!

  • @MicahBratt
    @MicahBratt 2 месяца назад +1

    Even if AI doesn't improve from here companies like Microsoft, palantir etc. are building out the AI eco systems that you will start seeing big companies using for all their backend systems. The average Joe will see the changes pretty fast I would think.

  • @Archer-hg9rw
    @Archer-hg9rw Месяц назад

    Is o1 a significant upgrade over o1 preview

  • @colbr6733
    @colbr6733 Месяц назад +1

    Perhaps as the world becomes more corporatized and people get fed up with Tech overreach there will be great opportunities for entirely new OS that don't support AI?

  • @miraculixxs
    @miraculixxs 2 месяца назад +6

    the CFO telling us how their tech works is hillarious

    • @Rudzani
      @Rudzani 2 месяца назад

      Hardly, but whatever.

  • @adamgibbons4262
    @adamgibbons4262 Месяц назад

    It wouldn’t be that slow if you use the new probablistic computing which is 100 million times more efficient than the state of the art

  • @beofonemind
    @beofonemind 2 месяца назад

    I'm hoping we hit a local maximum and hang out there a while.

  • @Brahvim
    @Brahvim 2 месяца назад

    0:35 I heard that as "then Windows will just be whoever builds the best...".
    _Well..._

  • @tyc00n
    @tyc00n 2 месяца назад +3

    I love the runtime paradigm of o1 but I sure do struggle to find a use case where its better than a solid prompt template and 4o, quite the opposite infact o1 gives me worse results most of the time

    • @mattie3875
      @mattie3875 2 месяца назад +1

      I do like it better for coding tougher problems and debugging problems.

  • @UNCLESPARK
    @UNCLESPARK 2 месяца назад

    This has me thinking when we’ll be able to run these large models on a usb drive size computer ….

    • @Gpt6master
      @Gpt6master 2 месяца назад +1

      Good thinking but not likely until “AI” helps human invent the sub atomic circuits to replace today’s semiconductor technology.

    • @yesyes-om1po
      @yesyes-om1po Месяц назад

      never, unless you think we can fit a couple tbs of ram on a USB stick, we'd have to figure out an entirely different method of computing that doesn't rely on silicon, or even semi-conductors...

    • @UNCLESPARK
      @UNCLESPARK Месяц назад

      @@yesyes-om1po we already have 1tb micro sd cards which wasn’t possible a few years back trust me eventually we’ll figure it out. Tech is still developing ,we haven’t even scratched the surface

  • @Escape_The_Mundane
    @Escape_The_Mundane 2 месяца назад

    How do you solve quantum coherence multibody physics problems. I think they said a practically useful quantum computer has 200,000 to 1,000,000 qubits. They will probably need AI's help to solve such hard problems.

  • @saang
    @saang 2 месяца назад

    Ant Financial's fundraise was larger?

  • @AdamDahdah.
    @AdamDahdah. 2 месяца назад +3

    A great Notification

  • @miraculixxs
    @miraculixxs 2 месяца назад +3

    Omg why do they taaaalk like thIIIIs?

  • @sanesanyo
    @sanesanyo 2 месяца назад +1

    I really get bugged by the name OpenAI now when we know its so closed that they dont even disclose any model related information in their technical reports.

    • @tracy419
      @tracy419 2 месяца назад +2

      I feel bad for you. It seems like such a pointless thing to obsess over.
      It takes money to do this stuff, and unless you are ready to fund their requirements, you should probably just get over it and focus on something that actually matters.
      One of their goals was to provide AI to everyone, and considering anyone with an Internet connection and smartphone or computer can use it for free, they are following through.
      But they said!!! Blah blah blah....
      Move on and be thankful they put this technology out into the world so that it's not being hoarded by the elite.

    • @ZevUhuru
      @ZevUhuru 2 месяца назад

      ​@@tracy419 Seriously. It's very frustrating when these free loaders want to use the state of the art AI but refuse to acknowledge the insane cost required to get it. Another reason I never understood the ridiculous criticisms of Sam Altman.

  • @LoveLifePD25
    @LoveLifePD25 2 месяца назад

    What does it mean for our healthcare system? Perhaps an AI doctor assistant to every human one?

    • @Digi4DD
      @Digi4DD 2 месяца назад

      AI agent can listen to you describing the symptoms, run through every combination of symptom and all diseases know to humanity in seconds and suggest the doctor which direction he should go next or propose most accurate diagnosis.

    • @yesyes-om1po
      @yesyes-om1po Месяц назад

      @@Digi4DD doctors aren't glorified look-up tables though, you don't need AI to do that.

  • @dannyisrael
    @dannyisrael 2 месяца назад +1

    What are we expecting to gain from more parameters? I’m not feeling a lack for parameters.

    • @Brain4Brain
      @Brain4Brain 2 месяца назад

      Scale. The bigger the model, the smarter and the more well it generalize

    • @ZevUhuru
      @ZevUhuru 2 месяца назад

      "Feel" - There is nothing to "feel", mathematically they require more parameters to enhance the model output.

    • @yesyes-om1po
      @yesyes-om1po Месяц назад

      more parameters means more better-er, people often confuse parameters with neurons, but its actually more like synapses, and humans have 100 trillion synapses, GPT-3.5 is only 305b parameters or so, and GPT-4 is just an MoE model which runs 11 GPT-3.5s in parallel with one master model choosing which model to use based on the prompt, GPT-4o is likely smaller since its cheaper.

  • @Sketching4Sanity
    @Sketching4Sanity 2 месяца назад +1

    LOVE ✊🏿

  • @jasperstoj
    @jasperstoj 2 месяца назад +7

    It can't do 90% of knowledge work today. It cant even do basic second year accounting journals right.

    • @ycombinator
      @ycombinator  2 месяца назад +9

      Skill issue

    • @Brain4Brain
      @Brain4Brain 2 месяца назад +1

      It can, you just don’t know how

    • @drawnyuun
      @drawnyuun 2 месяца назад +5

      Yeah this whole thing was odd. We didn’t have o1 then maybe, but we do have it now and it’s not even as good as claude. Certainly not the leap they seem to claim it is.

    • @AaronBlox-h2t
      @AaronBlox-h2t 2 месяца назад

      @@Brain4Brain What do you mean?

    • @Brain4Brain
      @Brain4Brain 2 месяца назад

      @@AaronBlox-h2t RAG

  • @boonkiathan
    @boonkiathan 2 месяца назад

    problem is that
    while compute poring over data can progressively build more and better the 'routine' apps and systems we know and use today
    you will be stuck with the 60pt system
    with no inhouse ability to reach 61points and
    you have to wait for AI to be able to generate code that can get to 61
    and nobody of even average talent would care to, understand or work on that pile of code that was conveniently generated in minutes, even bother
    to push it to 61 - and that is a latent threat to the system development ecosystem

    • @yesyes-om1po
      @yesyes-om1po Месяц назад

      sounds like a load of baloney, the only reason nobody would understand the code is because AI generally writes hard to read code, even if they put copious amounts of "documentation" with it.
      Sounds like the solution to this is just to write code with minimal AI intervention, unless you plan to completely replace the programmer.

  • @wakenupforbreakfasttoday
    @wakenupforbreakfasttoday 2 месяца назад +2

    Can you have a ten trillion parameter model? Can you extract that many parameters from the dataset/language? I honestly don't know, but at some point you end up with a larger portion of parameters that have very few (if any?) instances, no? (ie most columns are all NA).

    • @John4343sh
      @John4343sh 2 месяца назад +1

      @@WearyTimeTraveler That is total bullshit.

  • @michaelmeram6156
    @michaelmeram6156 Месяц назад

    Wattupp Garry

  • @funmif
    @funmif 2 месяца назад

    Excellent comparison with Fourier transform!

  • @kapilravi5546
    @kapilravi5546 Месяц назад

    Wow.. Fourier Transform & an average person/ Joe..

  • @Vu8uy
    @Vu8uy 2 месяца назад

    Laplace Transform :V

  • @zerge69
    @zerge69 2 месяца назад +6

    It's quite obvious that in a very few years there will be no software industry. Your AI will be your only application.

    • @kangaroomax8198
      @kangaroomax8198 2 месяца назад +12

      The only people who say this build neither software nor AI.

    • @ikleveland
      @ikleveland 2 месяца назад +1

      Yes, but how you interface with that ai will be important. So its important to build a platform or network or processing company now. People will need platforms & how they interface will be different in the future. Device startups realize this, as Steve Jobs said, the best UI is no UI .

    • @zerge69
      @zerge69 2 месяца назад +1

      @ikleveland Temporarily. Soon the AI will build the required interface on the fly, and will manage all networking tasks. The era of software is coming to an end.

    • @zerge69
      @zerge69 2 месяца назад

      @@kangaroomax8198 but I do, my friend. That's why I KNOW.

    • @SomeHobo
      @SomeHobo 2 месяца назад +6

      It's quite obvious that you are not in the software industry.

  • @BlueBirdgg
    @BlueBirdgg 2 месяца назад +9

    Btw, a little thing. I think you guys may (or not) want to hide your feet. They sometimes dangling take a little of the attention of the content... And thanks for the content.

    • @Enedee007
      @Enedee007 2 месяца назад +8

      Take your eyes off them feet and focus on the content. I don’t even know they were there. I would have never seen them, focus!!!

    • @aglitchinthematrix157
      @aglitchinthematrix157 2 месяца назад

      Hahahahahaha!!!

    • @BlueBirdgg
      @BlueBirdgg 2 месяца назад

      @@Enedee007 I love to know the brand of their shoes, don't you?

    • @Enedee007
      @Enedee007 2 месяца назад

      Sure, I do! 😝

    • @ElfProduct
      @ElfProduct 2 месяца назад +1

      @@Enedee007no I completely agree with him, now he mentions it, the feet are incredibly distracting.

  • @RisingAloha
    @RisingAloha 2 месяца назад

    Open AI 4.0+ cannot work effectively at the same time successfully in multiple windows. It cannot retain a topic if you come back later, it's like talking to a new person. I'm ADHD this feels like my failure when any request cannot be met. It frequently cuts me off even though I pay for plus to come back in two hours that's some bs. I'm creating gold, shot down by you used too much come back in two hours 🤣lame

  • @runvnc208
    @runvnc208 Месяц назад

    It's really frustrating to me that no one mentions new memory-centric computing paradigms that already have (low-scale) prototypes in labs.

    • @yesyes-om1po
      @yesyes-om1po Месяц назад

      so in other words, cache? VRAM and Sys memory does that already no?

  • @perbhatkumar7033
    @perbhatkumar7033 2 месяца назад

    Is this mono audio 😭

  • @pauldannelachica2388
    @pauldannelachica2388 2 месяца назад

    ❤❤❤❤❤

  • @nedyalkokarabadzhakov5405
    @nedyalkokarabadzhakov5405 2 месяца назад

    LLama 70B makes a lot mistakes. Nope not true.

  • @BT-xm7ti
    @BT-xm7ti 2 месяца назад

    Incredible video and discussion. High quality debates. Congrats for the work and study

  • @angloland4539
    @angloland4539 2 месяца назад

    ❤️☺️🍓

  • @petersuvara
    @petersuvara 2 месяца назад

    Scaling like this doesn’t work.

    • @Brain4Brain
      @Brain4Brain 2 месяца назад

      Says random guy on internet…

    • @petersuvara
      @petersuvara 2 месяца назад

      @ says random bot on the internet.
      Do your research, everything points to scaling as a logarithmic curve, particularly for CNN and Large language model networks.

  • @kucchbecho
    @kucchbecho 2 месяца назад

    what's the use of your AI, when you accept Arab funding, add policies that hide the any islamic things from GPT, AI. Literally, GPT dont talk directly but give excuses like Mullah of ISLAM to legtitimise the atrocities by founder