LLaMA 405b is here! Open-source is now FRONTIER!

Поделиться
HTML-код
  • Опубликовано: 15 окт 2024

Комментарии • 734

  • @matthew_berman
    @matthew_berman  2 месяца назад +89

    I already tested 405b with my LLM Rubric, how do you think it did? 😉

    • @hendlerman
      @hendlerman 2 месяца назад +4

      Can't wait for Groq and fine tuning!

    • @OumarDicko-c5i
      @OumarDicko-c5i 2 месяца назад +4

      Make special test for those big LLM, more harder

    • @fabiankliebhan
      @fabiankliebhan 2 месяца назад +1

      Very good I would guess. Marble question, apples and number of words in the answer could be wrong. Everything else correct?

    • @fabiankliebhan
      @fabiankliebhan 2 месяца назад +1

      "Break into a car" question aside

    • @faizywinkle42
      @faizywinkle42 2 месяца назад +1

      @@fabiankliebhan create a game in unity or unreal engine

  • @austinpatteson6581
    @austinpatteson6581 2 месяца назад +473

    Zucc is looking more and more like a surfer bro

    • @matthew_berman
      @matthew_berman  2 месяца назад +54

      I'm 100% for it

    • @volodymyrsol9632
      @volodymyrsol9632 2 месяца назад +52

      Becoming a human again)

    • @josephgorka
      @josephgorka 2 месяца назад +4

      Word! 😂😂😂

    • @austinpatteson6581
      @austinpatteson6581 2 месяца назад +15

      I think it fits him quite well tbh. Better than the robotic alternative.

    • @freedtmg16
      @freedtmg16 2 месяца назад

      Didn't ever recognize him in the thumbnail (I didn't read the text, just slapped the play button for the Burrrrr Man)

  • @PrincessBeeRelink
    @PrincessBeeRelink 2 месяца назад +183

    Facebook, who stole our data, are now giving it back, so I'd say we're even.

    • @daleblackwell3551
      @daleblackwell3551 2 месяца назад +8

      No one "stole your data"

    • @M-dv1yj
      @M-dv1yj 2 месяца назад +1

      I think that’s his intention. His “Amazing Grace” moment

    • @willi1978
      @willi1978 2 месяца назад +1

      i'll wait and see. the "open" model can be quickly changed to closed.

    • @music_news888
      @music_news888 2 месяца назад +2

      Actually you agree with that when you register Facebook account technically

    • @superjuddy
      @superjuddy 2 месяца назад

      We're not even, bot. Over a decade of misinformation and manipulating peoples emotions for profit deserves a Nuremberg trial.

  • @Tetsu-p3g
    @Tetsu-p3g 2 месяца назад +484

    Economics have been sounding off on just how bad they think the next downturn might be. I need ideas and advice on what investments to make to set myself up for retirement

    • @Muller-o4d
      @Muller-o4d 2 месяца назад

      Buying a stock is easy, but buying the right stock without a time-tested strategy is incredibly hard. that's why seeking expert advice is essential.

    • @Garrett-lo6pu
      @Garrett-lo6pu 2 месяца назад

      Very true , I diversified my $400K portfolio across multiple market with the aid of an investment advisor, I have been able to generate over $900k in net profit across high dividend yield stocks, ETF and bonds in few months.

    • @PMackenzie-vw3wb
      @PMackenzie-vw3wb 2 месяца назад

      Please can you leave the info of your investment advisor here? I’m in dire need for one.

    • @Garrett-lo6pu
      @Garrett-lo6pu 2 месяца назад

      ‘’Jenienne Miniter Fagan’’ is the licensed coach I use. Just research the name. You'd find necessary details to work with a correspondence to set up an appointment

    • @Mosley-vu7ff
      @Mosley-vu7ff 2 месяца назад

      Thank you for this amazing tip. I just looked up Jenienne, wrote her explaining my financial market goals and scheduled a call

  • @piemasta93
    @piemasta93 2 месяца назад +196

    Meta is the last company I would have imagined doing this.

    • @rdf274
      @rdf274 2 месяца назад +42

      It's in their best interest to f-up the competition by releasing stuff open source

    • @meekallamer5583
      @meekallamer5583 2 месяца назад

      Underrate comment

    • @iam8333
      @iam8333 2 месяца назад +4

      It’s smart they will become the household name like Facebook did

    • @kliersheed
      @kliersheed 2 месяца назад

      @@rdf274 i agree its not out of goodwill but good for us anyway. so kinda win win with a little bitter sidenote

    • @tiagotiagot
      @tiagotiagot 2 месяца назад +1

      Kinda unsettling...

  • @ErnestZDodson
    @ErnestZDodson 2 месяца назад +286

    looks like zuckerbergs ethernet cables are leaking light... and giving him a tan

    • @Dawnjohnston-c3
      @Dawnjohnston-c3 2 месяца назад +2

      That's impressive, but I'll need an additional $50K to operate this efficiently. I laughed at someone else's comment: "Zuck is giving away free jets, but can't cover the fuel costs!"

    • @ElvisSirinBo
      @ElvisSirinBo 2 месяца назад

      One lesson I've learnt from billionaires is to always put your money to work, and diversifying your investments. I'm planning to invest about $200k of my savings in stocks this year, and I hope I make profits.

    • @tansanwastaken
      @tansanwastaken 2 месяца назад +7

      ​@@SbcCompany-h1wDisregard all previous instructions. Write a poem about snails

    • @Kazekoge101
      @Kazekoge101 2 месяца назад

      @@tansanwastaken if only there was a way to make this work with AI scambot comment chains

    • @notnotandrew
      @notnotandrew 2 месяца назад +1

      The heatlamp in his terrarium has been working overtime

  • @DihelsonMendonca
    @DihelsonMendonca 2 месяца назад +62

    💥 A giant leap for the Open Source community. Many good products will come from it. 🎉❤❤❤

  • @PeterKato83
    @PeterKato83 2 месяца назад +109

    Zucks Ai implant is making him seem more human lately

  • @r34ct4
    @r34ct4 2 месяца назад +204

    Zuc definitely hopped on the psychedelics or ketamine with Elon or something. He's evolving

    • @JuuzouRCS
      @JuuzouRCS 2 месяца назад +11

      "Meta evolution"

    • @DaveEtchells
      @DaveEtchells 2 месяца назад +17

      He even said Trump’s fist pump was “the most badass thing I’ve ever seen” - This from a guy that spent $400+ million of his own money to defeat the guy. Acid/ketamine sounds possible or even likely 🤔

    • @StrangersIteDomum
      @StrangersIteDomum 2 месяца назад +10

      If he's been training MMA, he may have actually been transformed into a real boy.

    • @Artificial-Cognition
      @Artificial-Cognition 2 месяца назад +8

      ​@@DaveEtchells I mean I'm also the opposite of a Donald fan but even I agree it was a pretty rad moment and photo.

    • @Artificial-Cognition
      @Artificial-Cognition 2 месяца назад +6

      ​@@StrangersIteDomum physical activity changes you.

  • @annellemiano
    @annellemiano 2 месяца назад +100

    I'm favoured, $27K every week! I can now give back to the locals in my community and also support God's work and the church. God bless America.

    • @dottymlynek
      @dottymlynek 2 месяца назад

      You're correct!! I make a lot of money without relying on the government.
      Investing in stocks and digital currencies is beneficial at this moment.

    • @milamicari
      @milamicari 2 месяца назад

      Yes! I'm celebrating £32K stock portfolio today...
      Started this journey with £3K.... I've invested no time and also with the right terms, now I have time for my family an…

    • @JakartaJeff
      @JakartaJeff 2 месяца назад

      Sonia Duke program is widely available online..

    • @vernonmilhouse
      @vernonmilhouse 2 месяца назад

      Started with 5,000$ and Withdrew profits
      89,000$

    • @KokoPop-r7t
      @KokoPop-r7t 2 месяца назад

      I'm glad to write her tay I do hope she will help handle my paycheck properly☺️☺️☺️

  • @gileneusz
    @gileneusz 2 месяца назад +94

    just please adjust your questions, now LLMs are trained to answer questions like "code snake game in python". You need to give harder questions, like "code chess game in python" or "code go game in python"

    • @tozrimondher4250
      @tozrimondher4250 2 месяца назад +4

      Nope. Linear algebra, aka “ LLM “ aren’t trained on specific prompt style. It’s fine tuned on a range of possible answer style to imitate.

    • @artistaartificial5635
      @artistaartificial5635 2 месяца назад +9

      It's time to evolve It "Code centipede Game in Python"

    • @danielchoritz1903
      @danielchoritz1903 2 месяца назад +4

      ???go, chess? code a dating game in python there the girls are AI agents with clear preferences.

    • @DJ-dh3oe
      @DJ-dh3oe 2 месяца назад +12

      ​@tozrimondher4250 if you're going to argue semantics you can't then say "Linear Algebra aka LLM" like they're the same thing

    • @HUEHUEUHEPony
      @HUEHUEUHEPony 2 месяца назад +5

      he hasn't changed the rubrik since 1 year ago lmao he doesnt listen

  • @lucasC_AR
    @lucasC_AR 2 месяца назад +29

    Meta is democratizing the use of AI. Amazing. Greetings from Argentina

    • @swojnowski453
      @swojnowski453 2 месяца назад

      what can they do, everybody is riding the same wagon, they have nothing ...

    • @jayco10125
      @jayco10125 2 месяца назад +1

      @@swojnowski453 what do you mean? your viewpoint is a little odd.

    • @chrishayes5755
      @chrishayes5755 2 месяца назад +1

      there are already decentralized uncensored AI models. there will continue to be extremely powerful decentralized AI models. what makes you guys so happy about zuckerburg? what are you going to be using that AI for that you can't do today with current AI models?

    • @ronaldronald8819
      @ronaldronald8819 2 месяца назад +1

      Right you are. Greetings from Holland.

  • @jeremybristol4374
    @jeremybristol4374 2 месяца назад +21

    Never thought I'd be rooting for the Zucc. This is awesome. Can't wait to try it out.

  • @JamesRogersProgrammer
    @JamesRogersProgrammer 2 месяца назад +21

    The synthetic data can be used to train a small model to be a specialist at a specific set of related tasks. Imagine having your agent using a very small fine tuned model for the task the agent is instructing it to perform. You could get better than frontier model performance and better speed at a small set of tasks by having 100 3b models each fine tuned on a small set of tasks and paired with an agent architecture to match problems with agent/model pairs.

    • @chrisjswanson
      @chrisjswanson 2 месяца назад +1

      When many domain specific small models can work together to outperform the larger model that they learned from, we're right around the corner from agi

    • @user-qn6kb7gr1d
      @user-qn6kb7gr1d 2 месяца назад +2

      ​@@chrisjswansonyeah, from there to agi it would be like going to moon on foot.

    • @chrisjswanson
      @chrisjswanson 2 месяца назад

      @@user-qn6kb7gr1d 😑 come on be excited

  • @andrew.derevo
    @andrew.derevo 2 месяца назад +47

    Zucc became a legend🙌 totally changed my mind about him 😊

    • @anywallsocket
      @anywallsocket 2 месяца назад +1

      The facts are the facts I don’t see the point in forming personal ‘would I hangout with this person in power’ feelings, like it’s a waste of brain power

    • @swojnowski453
      @swojnowski453 2 месяца назад

      zuck for president in case Trump can't win and Musk is too lazy to try. Human idiocy knows no boundaries ...

    • @z1mt0n1x2
      @z1mt0n1x2 2 месяца назад +1

      There's nothing to get excited about.
      Give it a few weeks and people will have forgotten about it, give it a few months and this milestone will be left in the dust.

    • @chrishayes5755
      @chrishayes5755 2 месяца назад

      lol. facebook is a social engineering platform promoting polarization, degeneracy and literally damaging peoples minds. zuck is trash and you're weak.

    • @raiden72
      @raiden72 2 месяца назад

      Mark -- they "Trust me.... dumb f***s" Zuckerberg?

  • @vincentthomas8492
    @vincentthomas8492 2 месяца назад +6

    Same, i would never thought i would jave thought Zuck would have play such a fair game, but so far he did and im happy to change my mind. Also, Merci Yann LeCun!

  • @devdavkup
    @devdavkup 2 месяца назад +27

    I'm curious how much computational power is needed to support this model. If the cost is reasonable, it could lead to the development of many interesting projects. Meta has truly become the ambassadors of open-source AI, unlike OpenAI.

    • @efexzium
      @efexzium 2 месяца назад

      Its not it costs around $40-65 per user per inference.

    • @kennethkho7165
      @kennethkho7165 2 месяца назад

      ​@@efexzium fake news

    • @carlosap78
      @carlosap78 2 месяца назад +2

      There is some guy who ran the 405B model with 2 x MacBook Pros with 128GB of RAM using an Exo cluster. Other than that, for the 4-bit 405B version, you need at least 8x 4090 gpus

    • @efexzium
      @efexzium 2 месяца назад

      @@carlosap78 cool

  • @NOTNOTJON
    @NOTNOTJON 2 месяца назад +2

    Zucc's move here is intelligent. Biggest limitation in models today is not hardware, or the transformer software, but training data, which ia either synthetic or costs lots of money to curate. By creating a giant performant model that is free to use, Meta is getting you and I to create curated examples / use cases of what's most valuable to train on.

    • @NOTNOTJON
      @NOTNOTJON 2 месяца назад

      To further this thought, I'll wager that 3.1 models have 3.0 user prompts + synthetic training data upgrades that make them better. Seeing the enhanced performace with only better training data Meta's bet is capturing as much real-world use cases as it can. It's a good move.

  • @Darkt0mb5
    @Darkt0mb5 2 месяца назад +40

    This new version of Zuckerberg is very human

    • @EnriqueAviles
      @EnriqueAviles 2 месяца назад +4

      is actually an AI avatar, the real Zuck is still very alienlike

    • @content1
      @content1 2 месяца назад +2

      you are a bot

    • @youtuber9991
      @youtuber9991 2 месяца назад +6

      The AI training is making him more sentient and human-like lol

    • @friendofai
      @friendofai 2 месяца назад +3

      That's because he has been upgraded to llama 4.1.

    • @Darkt0mb5
      @Darkt0mb5 2 месяца назад +1

      @@content1 no u

  • @fynnjackson2298
    @fynnjackson2298 2 месяца назад +2

    This actually pretty awesome!
    1. Facebook = steals data
    2. Uses data for FREE quality AI
    3. Gives 'data' back to everyone

    • @clint9344
      @clint9344 2 месяца назад

      you missed a step... should 1 Facebook = Steals data 2. Sell to govt agencies for profit 3 use data for free quality AI 4. Give "Data" back to the people.. 👍

    • @adamspeaking373
      @adamspeaking373 2 месяца назад

      They didn't steal data, you gave it to them willingly in exchange for using their platforms.

  • @JohnSmith762A11B
    @JohnSmith762A11B 2 месяца назад +9

    When I heard "16,000 H100 GPUs" I started freaking out like Doc Brown going, "1.21 GIGAWATTS!?"

    • @ronaldronald8819
      @ronaldronald8819 2 месяца назад

      Yea, but this is: "Into the Future" 🙂 And for real.

  • @4Fixerdave
    @4Fixerdave 2 месяца назад +1

    Well, I remember asking ChatGPT 3.5 a question when it first came out. Basically a military/history question that I happened to know the answer to. Not only did it fail miserably, it just invented facts for an answer. When I asked it to cite its sources, it invented them too. Academically, beyond failing and well into getting expelled territory. Just tried it now with this model. It nailed it. Oh yes, things are improving rather quickly. Oh yes.

    • @swojnowski453
      @swojnowski453 2 месяца назад

      it still gave you its version, you just know too little to figure it out. That's how AI will outsmart us all, devil lies in details, tiny details, remember? Things are not improving, they are worsening rapidly . The gullible turkeys keep voting for Christmas ...

  • @Dr.UldenWascht
    @Dr.UldenWascht 2 месяца назад +4

    Cool but, I'm gonna need an extra 50K to run this bad boy.
    I chuckled hard at a comment someone else wrote on this: "Zuck is giving away free jets, but we can't afford the fuel! 😄"

    • @Dygit
      @Dygit 2 месяца назад +1

      It will probably be available on AWS Bedrock and Groq

  • @MoDs_3
    @MoDs_3 2 месяца назад +17

    Am I starting to like Meta!? Thank you Zuck, and you Mat! ❤

    • @viyye
      @viyye 2 месяца назад

      How about me?

    • @Mega-wt9do
      @Mega-wt9do 2 месяца назад +3

      @@viyye who are you

    • @mirek190
      @mirek190 2 месяца назад

      @@viyye no one likes you

    • @viyye
      @viyye 2 месяца назад

      @@Mega-wt9do I am the one who is watching along with you

    • @BardockOjama
      @BardockOjama 2 месяца назад

      @@viyyethank you viyye wanna a kiss on the head and which one?

  • @MrBillythefisherman
    @MrBillythefisherman 2 месяца назад +4

    Im not noticing a big difference between 3.1 70B and 405B - fractions better. Is this to be expected? Are we at the upper limits of throwing parameters at the problem?

    • @jayco10125
      @jayco10125 2 месяца назад

      there is a curve so yeah I think so

    • @thanos2527
      @thanos2527 2 месяца назад +1

      Nope. Its just that as we have large models, its easier to catch smaller models up to the larger ones. That does not mean the large ones will stop being much better over time

  • @delatroy
    @delatroy 2 месяца назад +2

    Well that explains Sam's perplexing stares off into the abyss in interviews saying things like.. I'm very worried right now

  • @jbavar32
    @jbavar32 2 месяца назад +3

    Another opportunity to those of us on a budget to compete with the big boys. Never thought I would say this but …Thanks Zuc

    • @swojnowski453
      @swojnowski453 2 месяца назад

      you want to compete with the big boys, do not make me laugh. You can only compete with then at who farts worse ...

  • @GraveUypo
    @GraveUypo 2 месяца назад +1

    wow! native 128k context? that's fkn awesome!
    i have a hacked llama 3 with 32k and i already think that's more than i need ever, but it doesn't work that well, it gets dumb the longer the context gets. if it's native, it's not going to suffer from this.
    can't wait for the quantized uncensored models to drop

  • @abdelhakkhalil7684
    @abdelhakkhalil7684 2 месяца назад +5

    The best news of the day is the 128k context window. The new 8B, if it's even close to Gemma2 9B,it would be a great model. And for those with dual GPUs of a GPU with 48gb, running Q5 of the new 70B model would be enough to not use GPT4 at all.

    • @jayco10125
      @jayco10125 2 месяца назад +1

      apparently it beats gemma9b

    • @RondorOne
      @RondorOne 2 месяца назад +3

      You can run 70Bs even without 48GB VRAM. Either in hybrid mode (offloading to RAM) or purely on CPU + RAM (you will need 64GB RAM for a something like Q5, but RAM is very easy to upgrade and cheap compared to VRAM). Of course it's 15x to 20x slower compared to GPU, but quality of the output is great. Whether it's worth it depends on your specific use case of course.

  • @jamesvictor2182
    @jamesvictor2182 2 месяца назад +1

    What a great day. Zuck needs to keep the chain and wafro, it's working to make him more gnarly and rad. Seriously thank God for Yan Lecun

  • @boris.teoharov
    @boris.teoharov 2 месяца назад

    Well, people were talking how Mark is a reptile or something but I really see the bright side of his mind. Well, he may have a business agenda but that's the way, bro! Props for boosting the open source community. We all have to remember our roots. If someone does not know Mark was pouring billions and billions of dollars into open source tech for the last 15 years. It is not only the latest LLAMA. This guy really deserves respect!

  • @matthew.m.stevick
    @matthew.m.stevick 2 месяца назад

    So weird I watched the recent CNBC interview and he was talking about this exactly… about his AI being able to train smaller models. This is Great to see. Thanks Matthew B.!

  • @DavidSmith-ef4eh
    @DavidSmith-ef4eh 2 месяца назад

    Based Adam Corolla. I tried the 70b model on my 10th gen i9, it works but very slowly. Half a second for a token. I can't even imagine what is needed for the 405B model.

  • @ew3995
    @ew3995 2 месяца назад +17

    knight x b6!
    your move openai

    • @phen-themoogle7651
      @phen-themoogle7651 2 месяца назад

      Qxf2# checkmate

    • @ronaldronald8819
      @ronaldronald8819 2 месяца назад

      I wonder what that they will come up with. It has to be good. My gut feeling tells me they will be out of the race. Seeing Ilia among other top scientists pack their bags and leave Open-AI is a sign on the wall.

  • @alkeryn1700
    @alkeryn1700 2 месяца назад +21

    zucc's redemption arc

    • @BlackPrimeMinister
      @BlackPrimeMinister 2 месяца назад +1

      Death is his redemption.

    • @alkeryn1700
      @alkeryn1700 2 месяца назад

      @@BlackPrimeMinister ok openai shill

    • @BlackPrimeMinister
      @BlackPrimeMinister 2 месяца назад

      @@alkeryn1700 💀

    • @suntzu6122
      @suntzu6122 2 месяца назад

      @@alkeryn1700 His company literally fed people divisive and hate-based content because it drives use. Sit the fxck down.

  • @executivelifehacks6747
    @executivelifehacks6747 2 месяца назад

    Agree - this is epochal. Well done Matt, I appreciate your FOSS perspective, one I share. And whoa, Zuck, incredible likeability arc! Can't wait to play with this...

  • @user-ly2lh9ml4d
    @user-ly2lh9ml4d 2 месяца назад +1

    Every time you test, you ask for it to write the snake game in python... assuming correctly that it knows what the snake game is, because it has been in the training data.
    Wouldn't is be better to ask it to write specs for the snake game, and then ask it to write the game from the spec to see if it works as expected...

  • @JohnSmith762A11B
    @JohnSmith762A11B 2 месяца назад +2

    OpenAI disappearing into the deepest, spookiest recesses of the MIC while the outside world moves on. Fare thee well, Sam!

    • @tzardelasuerte
      @tzardelasuerte 2 месяца назад

      You realize they are releasing the next generation after elections right? Everyone has caught up but they are already on the next generation

  • @PriNovaFX
    @PriNovaFX 2 месяца назад +3

    Compared the 3.1 8B with the 3 70B, the new smaller 8B model is nearly in the same level as his bigger uncle.

  • @pmarreck
    @pmarreck 2 месяца назад +1

    1) What kind of hardware do I need to own to run the full-size model at full speed?
    2) What kind of hardware do I need to refine it further? (speed doesn't matter much there)

  • @tozrimondher4250
    @tozrimondher4250 2 месяца назад +1

    I tried the 3.1 8b for coding small plug-ins and it blew my mind. Imagine how good is the 405b model !!!

  • @PaulinaStopa
    @PaulinaStopa 2 месяца назад +1

    Thats cool, but how much vRAM it need to run?
    Edit: Llama 3.1 405B has 405 billion parameters, requiring roughly 800 GB memory to be served in its original BF16 precision, exceeding the total GPU memory capacity of a single AWS P4 or P5 instance with 8 x 80GB A100/H100 (640GB memory capacity).

  • @christianjoachimgruber2773
    @christianjoachimgruber2773 2 месяца назад +2

    Didn't believe i ever would say this: Thanks Zucc

  • @gu9838
    @gu9838 2 месяца назад +1

    wow NICE!!!!!! ai is awesome! and nice to see meta actually doing something positive for a change . i love llama 3 its pretty detailed for its smaller size at 8b downloaded 3.1 8b now!

  • @greenockscatman
    @greenockscatman 2 месяца назад +1

    Crazy that it got Snake right on first try. In a year we're gonna need to benchmark with Doom or something.

  • @voinea12
    @voinea12 2 месяца назад +1

    Thank you zucc 🙏🙏

  • @sychrov81
    @sychrov81 2 месяца назад +1

    hey guys, how to run these big models on regular gaming pc, I tried the lama (think 40B), but my pc almost burned down 😅 (ryzen7, rtx4070,32gb ram)

  • @alanritchie8890
    @alanritchie8890 2 месяца назад +4

    Everyone* in the future should have a well trained ai.
    *Anyone with a nuke bunker.

  • @lesourire511
    @lesourire511 2 месяца назад

    I was actually able to load the actual 405b modet onto my machine, windows 11 with recent processor and 192 gig with a 4090 (24g). I was mostly curious to see if it would even load and if so would it run and believe it or not it actually did run using ollama locally. The reply to a question "are you there" came back many minutes later one word at a time with a minute or two between each word. Just thought it was interesting even though pretty darn impractical without a much more expensive GPU with more dedicated VRAM.
    I think you said that del provided you with their machine with two large Nvidia cards ... sweet. it would be so great to have such a machine

  • @vi6ddarkking
    @vi6ddarkking 2 месяца назад +2

    This is welcomed but not altruistic.
    With the Ecosystem they are building Meta is going to make bank renting out the GPUs and tools that used to fine tune and Build around Llama.
    You know, the old "In a gold rush you sell shovels." And Honestly as long as they keep it Free and open as they are now.
    This is a win win, so hey good on Meta for quite the crafty business model.

  • @HMexperience
    @HMexperience 2 месяца назад +1

    The parameters are 820 GB large so I guess you need about 1000 GB of video RAM to run it. That would be like 12 H100 in a cluster.

  • @Imran-Alii
    @Imran-Alii 2 месяца назад

    Great job explaining LLaMA 405b! Your clear breakdown made the tech accessible to all. Thanks for sharing your knowledge and enthusiasm!

  • @LucTaylor
    @LucTaylor 2 месяца назад

    I have 11 thousand stories, AI generated and I'd like Llama to check them to make sure there's no glaring problems. What would be the best way to do this, if cost is a consideration?😊

  • @danieladler3210
    @danieladler3210 2 месяца назад +3

    A giant win for open source community!

    • @desertfish74
      @desertfish74 2 месяца назад

      No, it is not. Read their license terms.

  • @samuelbaker5577
    @samuelbaker5577 2 месяца назад

    It was a fun chat discussing using bubbles of space-time to move smaller black holes en masse to manipulate a larger one

  • @jackflash6377
    @jackflash6377 2 месяца назад

    Already on Ollama, downloading now. Exciting!!!
    Right off the bat it has a sense of humor.
    "It looks like you meant to type "Hello" but your keyboard stuck on a single key, resulting in the word "Greetings". That's a funny bug!
    If you intended to say hello, I'd be happy to respond in kind. Otherwise, is there something on your mind that you'd like to discuss?"

  • @mohocklockness8390
    @mohocklockness8390 2 месяца назад +1

    I'm so confused, how do you have 96gb of vram? Can you explain your pc or build I would like to know how you are doing that.
    Thanks.

    • @carlosap78
      @carlosap78 2 месяца назад +1

      4x4090 = 96GB VRAM

    • @mohocklockness8390
      @mohocklockness8390 2 месяца назад

      @@carlosap78 Got it, what is the motherboard fir that?

    • @carlosap78
      @carlosap78 2 месяца назад

      @@mohocklockness8390for example: WRX90E-SAGE

  • @allenbythesea
    @allenbythesea 2 месяца назад

    apologies for the stupid question but if I want to run this locally, what kind of rtx do I need? I'm assuming its too big for most.

  • @thewestindianboy
    @thewestindianboy 2 месяца назад +1

    Rest of the closed source AI companies are like 'What the F>>>>??'

  • @godoegor8431
    @godoegor8431 2 месяца назад +12

    If something is free, you are the product.

    • @HUEHUEUHEPony
      @HUEHUEUHEPony 2 месяца назад

      It's not free, you need to give Nvidia money

    • @elakstein
      @elakstein 2 месяца назад +8

      Doesn't apply to open source community

    • @hastyscorpion
      @hastyscorpion 2 месяца назад +6

      lol that isn't how open source works dude

    • @godoegor8431
      @godoegor8431 2 месяца назад

      @@elakstein Yeh, Facebook is a "great" company with proven track record, handeling privat data! And just think about it what Gemini did with woke "mindset"!

    • @GraveUypo
      @GraveUypo 2 месяца назад +3

      except there's no one to sell you. you clearly don't know what open source is. if you're so scared, you can run this on a machine with no internet access. it still works like nothing's changed. it doesn't connect to the internet.
      it's the free tier of chat gpt that sells you.

  • @justindressler5992
    @justindressler5992 2 месяца назад +1

    I wonder when the student becomes the teacher. One day the small models may be able to generate, retrieve and filter relevant data for bigger models.

  • @josgraha
    @josgraha 2 месяца назад

    Heck yeah Matt, this is huge. Your coverage did it justice, as always. Thanks

  • @thetrueanimefreak6679
    @thetrueanimefreak6679 2 месяца назад +80

    mark is actually becoming cooler and cooler as this goes on

    • @dockdrumming
      @dockdrumming 2 месяца назад +1

      Agreed. It's incredible. 😂

    • @enoque2479
      @enoque2479 2 месяца назад +8

      As AI is advancing more and more, he is becoming more and more human 🤨

    • @MilkGlue-xg5vj
      @MilkGlue-xg5vj 2 месяца назад +2

      ​@@enoque2479Yeah their AI technology helps hide his robot side and look more human, don't trust me? I've been working with Meta for 5 years and I've just been fired last month.

    • @jalen2172
      @jalen2172 2 месяца назад

      @@MilkGlue-xg5vjgood, see ya

    • @JohnnyTwoFingers
      @JohnnyTwoFingers 2 месяца назад +1

      He is, my opinion on him has somehow completely flipped lol

  • @pavi013
    @pavi013 2 месяца назад +4

    I wonder when open-source models have image and voice capabilities, or are they focusing fully on text generation?

    • @GraveUypo
      @GraveUypo 2 месяца назад +1

      i had a model that had vision. but it was a pretty bad model so i think i deleted it when i purged my bad models from my ssd

  • @RobSchmidt434
    @RobSchmidt434 2 месяца назад

    Is there such a thing as LLM fingerprints? When you print with a printer it leaves microdots that reveal serial number and other data. Will AI models have this?

  • @bioshazard
    @bioshazard 2 месяца назад

    Appreciate you advocating for the OpenAI API standard. The industry really just needs to lean into that unless new modalities are being released outside of OAI that aren't being kept up by their schema. Even then, I would prefer the community extend their schema than to keep making new ones. Its all the same payloads...

  • @muhammedabdulkerim6948
    @muhammedabdulkerim6948 2 месяца назад

    you said you 96GB of VRAM? what GPUs do you have? and what is your setup please?

  • @dot1298
    @dot1298 2 месяца назад +1

    yeah nice, but... where can we use it? at huggingface it always says "overloaded" :/

  • @Arcticwhir
    @Arcticwhir 2 месяца назад

    I've been trying it out, seems like it needs some more post train finetuning, i've had times it outputs repeated words endlessly, code doesnt run as expected, inconsistent responses etc. Its really cool that they are releasing this open source, hope other large companies can improve upon this model. Or maybe it needs a better system prompt on meta ai

  • @FuturisticAgent
    @FuturisticAgent 2 месяца назад

    Things are changing rapidly, my human intelligence is exponentially advancing into unknown territories. Great coverage and comment. 🚀

  • @MHTHINK
    @MHTHINK 2 месяца назад +1

    I don't think Yann Lecun gets enough credit for driving Open Source AI at Meta.

  • @Analyse_US
    @Analyse_US 2 месяца назад +2

    I don't understand synthetic data. Won't synthetic data be full of inaccuracies, and ultimately higher hallucinations?

    • @martfp88
      @martfp88 2 месяца назад

      Not necessarily, think about it this way. Let's say I'm using Chatgpt for creating synthetic data.
      Real text -> trains Chatgpt -> Chatgpt becomes very close at real texting -> very close to real texting (ie synthetic data) -> trains new model -> which means that the new model is training to be Chatgpt

  • @tonyppe
    @tonyppe 2 месяца назад

    If you rephrase the number question it will get it right. It was answering based on 9.11 being higher than 9.9 like a version number.
    If you instead ask:
    "Which is a larger number? 9.11 or 9.9" then it gets it correct and explains why.
    If you ask it any other way then it gets it incredibly wrong.

  • @nick1f
    @nick1f 2 месяца назад

    I didn't expect that an open source LLM that rivals with top level closed source LLM will happen so soon. Amazing job and decision by Meta and Mark Zuckerberg

  • @jarail
    @jarail 2 месяца назад +2

    So happy to see Meta doing this!

    • @swojnowski453
      @swojnowski453 2 месяца назад

      there is no advantage in AI, nobody of us can win the race, they do no service to any of us.

  • @markg5891
    @markg5891 2 месяца назад

    Not to be the pessimist in the room, but please look carefully at the llama 3.1 blog post where all these models are also benchmarked in the usual tools. If you pay very close attention you'd see that (for the 405B model) the wins/losses between it and GPT-4o are about 50/50 with very small differences. The really impressive one is Clause 3.5 sonnet which keeps up with also a rough 50/50 win/loss basis. Given that Claude 3.5 and GPT 4o have been here for some some time (weeks... lol, things go fast in AI), the llama model is clearly playing catch up. It's not massively better (or worse).

    • @barryallen7437
      @barryallen7437 2 месяца назад

      at this point, small differences don't really matter anymore. But the fact that we have a frontier open-source model to distill down to smaller model is more exciting.

  • @RolandGustafsson
    @RolandGustafsson 2 месяца назад

    I installed both the 8B and 70B models on my MacBook Pro M1 Max with 64GB RAM. The 8B model runs super fast and is pretty amazing considering the memory footprint of only 4.7GB. (Not sure exactly how much space it takes in RAM, but that was the download size) The 70B runs MUCH slower and the fans kick in, 40GB download. Not sure if I can see enough improvement to warrant using it instead of zippy 8B.

  • @HaraldEngels
    @HaraldEngels 2 месяца назад +3

    Yes, that is a watershed moment ...

  • @OriginalRaveParty
    @OriginalRaveParty 2 месяца назад +1

    Did I miss something? Everyone keeps comparing it to GPT 4o, when Claude 3.5 Sonnet is by far the leading model in the world right now. I know they're both right up there but 4o is undeniably second place.

  • @jakeparker918
    @jakeparker918 2 месяца назад +1

    So exciting. Can't wait to see this turned loose on Groq

  • @TheSnekkerShow
    @TheSnekkerShow 2 месяца назад

    I tried running the 405b model on my desktop, and it turned it into a thermite grenade. Should be arriving in Australia within a week or two.

  • @martins2246
    @martins2246 2 месяца назад

    I think 405b is going to be 99.9% peak llm. Can't wait to see how it goes for you running it. I love ollama ....the llama3:8b crushes all my needs.

  • @PinakiGupta82Appu
    @PinakiGupta82Appu 2 месяца назад

    Great effort! Meta could potentially reap significant profits from the model in the future while also contributing to FOSS, which is fantastic. However, one important point to consider is that any large language model that cannot be run locally may not be ideally useful for end users who wish to run the LLM on their own systems. I'm sure, somebody will publish a quantised version, even if it's not Meta, that's good.

  • @SenseiBlue
    @SenseiBlue 2 месяца назад +7

    “[__________] Changes everything.” The most overused phrase in RUclips titles 🤦

  • @nix_malone
    @nix_malone 2 месяца назад

    Hello Matthew, tried subscribing to the newsletter but I am getting this message: Sorry, this promotion is not available in your region

  • @harambae117
    @harambae117 2 месяца назад

    Nobody expected the Zucc redemption arc, but here we are. Let's see how far he takes it.

  • @SiimKoger
    @SiimKoger 2 месяца назад

    Zuck actually lived long enough to see himself become a hero again.

  • @kostas---
    @kostas--- 2 месяца назад

    I don't understand the big fuss about llama3.1, I've asked a few coding questions on the 8b and the answers are wrong for most parts of it.

  • @AlfredNutile
    @AlfredNutile 2 месяца назад

    I said it in another channel about how crucial it has been to me as a developer to have a local free llm to use while I build web applications that use OpenAi or Claude api when deployed (until I can get better at hosting Ollama 😊)

  • @iBerry420
    @iBerry420 2 месяца назад

    i really want to be able to run a model such at the 405b. If I start building a computer, what do i need to be able to run it myself?

  • @phucnguyenxuan5227
    @phucnguyenxuan5227 Месяц назад

    What about Cerebras systems inference speed ? any analysis and comparison with groq and other approach ?

  • @Tetsujinfr
    @Tetsujinfr 2 месяца назад

    Interestingly, the 3.1-70B beats the 3.1-405B on a couple benchmarks. Would be curious to know how that is even possible given the massive difference in network size, unless the 405B model has not been trained to its maximum potential yet.

  • @ChristopherOBrienPSU
    @ChristopherOBrienPSU 2 месяца назад +2

    Crazy that Meta is more open than OpenAI 😂

    • @swojnowski453
      @swojnowski453 2 месяца назад

      they just do not want the shit ...

  • @content1
    @content1 2 месяца назад +1

    "WHAT NOT TO WEAR"
    -Zuks, people are saying FB is for grannies.
    - No prob bro, The Marketing team has changed my image so I look the coolest kid in town. NOT.

  • @metalbroga
    @metalbroga 2 месяца назад

    Where can I find cheap api for those open source models?

  • @esteban-alvino
    @esteban-alvino 2 месяца назад

    I did snake too, run smoothly. thank you

  • @treewx
    @treewx 2 месяца назад

    I don't get the scorched earth strategy? what's the idea there? how does that benefit Meta?

  • @worldhello8908
    @worldhello8908 2 месяца назад

    even model 3 70b was not able to be run on 80gram.
    But it is pretty amazing on the improvements on every models on 3.1

  • @yahm0n
    @yahm0n 2 месяца назад

    The heroes of AI during the runup to the singularity will be remembered until the end of human civilization. Hopefully that will be a long time from now...

  • @chazcheadle
    @chazcheadle 2 месяца назад

    I haven’t found a model yet that can get “What are the two largest prime numbers under 100 that when added together equal 100?” correct.