So Google's Research Just Exposed OpenAI's Secrets (OpenAI o1-Exposed)

Поделиться
HTML-код
  • Опубликовано: 14 ноя 2024

Комментарии • 208

  • @totoroben
    @totoroben Месяц назад +136

    I'm using o1 to batch convert Outlook msg to PDF. It built me a powershell script, and it took three rounds of copy and pasting error codes back into it to yield a working script. Total time: 5 minutes. The script itself takes one minute per conversion but I don't really care about this, because it's not hands on. The time savings is *hours*

    • @JeremyPickett
      @JeremyPickett Месяц назад +10

      No doubt. I have 18,000 pages of text to refine, correct, put in another format, etc. 4o mini accomplished was normally would take 4 months, to... 16 hours. I almost fell out my chair

    • @barbidou
      @barbidou Месяц назад +10

      @@JeremyPickett Have you completely proofread and verified the results?

    • @JeremyPickett
      @JeremyPickett Месяц назад

      @@barbidou that's this week. Every five pages there is usually a need for a correction, but it is generally minor. I know the content back and forth (I wrote it, it's been a crazy few years), and thankfully for this project it is "scannable". Something I do have to be quite careful on are any hyper links--it may not even be the models fault, there is absolutely link rot.
      But it is already way, way better than Fiverr, turk, freelancer, craigslist, or upwork

    • @lewie8136
      @lewie8136 Месяц назад +7

      @@barbidou No he hasn't because his whole idea is to create a shortcut - using a completely unreliable tool

    • @sonatine21
      @sonatine21 Месяц назад

      @@lewie8136 I guess that is why some people built a complete generative api for open source llms using some code oriented open source llms which were waaaaay less good at their job than o1 or even 4o are...
      Tools are useful if you know how to use them, o1 takes almost 1 minute to think about the tasks i provide, i guess you ask it useless stuff or you don't know how to be precise in your deemands, then it only take 5 seconds and output useless stuff 🤗
      I made it wrote a complete repport about AI-driven propaganda yesterday with a multi agent of mine. PhD grade report, in latex, with 50 real references that re checked multiple times by the system I built in 2 hours, won't publish but will make a paper out of my methodology. (I also used notebookLM, to generate the repport subject with a 50 documents library incuding interviews, websites, and various databases)

  • @TheAero
    @TheAero Месяц назад +15

    It's so exciting. Imagine o1-mini getting more and more optimized, improved as well and becomes the basic premium version for GPT subscribers without major limits. It would be a breakthrough, as well as having it lower costs. It can be the companion of cursor users.

  • @wtflolomg
    @wtflolomg Месяц назад +20

    This is what I've been asking for... models trained on the task we need. As an enterprise developer, I need a model trained on the technologies in my company's full stack - C#, Java, Javascript and associated frameworks... and able to manage an entire project, handling it's entire context on site with our own compute. Same goes for infrastructure - let's have another model that understands our network topology and technologies like Kubernetes and Splunk, for example. Smaller general models are fine to understand the queries... let them feed those tokens to the specialized models to generate quality artifacts.

    • @jondo7680
      @jondo7680 Месяц назад +3

      No. You are confusing model size with training data here. You are advocating to teach l train the models less. That leads to worse models. instead of training two same size model on different data, training one model on both of that data will have no extra cost while the resulting model will have a better intelligence and knowledge. Training a model further is better than training random noise. If you already trained one model for one task it would be stupid to train another model from zero instead of training your first model further for the second task.

    • @ollyfoxcam
      @ollyfoxcam Месяц назад

      That’s not what’s happening here.

  • @BrianMosleyUK
    @BrianMosleyUK Месяц назад +8

    We want "NotebookLM with voice cloning and auto slide show" and I think we will be there this year.

    • @Iightbeing
      @Iightbeing Месяц назад

      NotebookLM is pretty incredible. I really want to include personality and a RAG that will automatically sort and integrate learned data into its own memory. With agentic capabilities.

  • @peterwilkinson1975
    @peterwilkinson1975 Месяц назад +5

    I remember hearing the head of Microsoft’s ai division say how they can scale the models up for at least 2 more generations. So there still maybe room to scale the models. In addition to the “thinking” capabilities.

  • @DiceDecides
    @DiceDecides Месяц назад +3

    Google also released something called NotebookLM, you can upload some text files and create an AI podcast that talks about the text for like 10 mins, really cool

  • @zerorusher
    @zerorusher Месяц назад +2

    This is the future, its a loop.
    A single huge model that can be distilled into smaller models that already know most of the best tactics to get great answers.
    When this smaller models make mistakes, users will point them out, adding more and more sophisticated concepts and ideas into the dataset.
    Creating Chatgpt and making it free to use was a genius move, every time the models gets it wrong, the user corrects it, creating more and more complex chains of thought that will make the next iteration better...

  • @alexandermoody1946
    @alexandermoody1946 Месяц назад +3

    Do you have a transcript for the video?
    I would like to just confirm that you said "sprinters conserving energy for the final stretch?"
    This seems really odd when it would seem sensible for a sprinter to reach maximum possible velocity in the quickest possible time whilst maximising stride. Sprints are over short distances so It would not seem necessary to hold back any possible power performance.
    Long distance running may need to conserve energy on the other hand.

  • @TheAero
    @TheAero Месяц назад +1

    I guess it's all about nvidia and competitors putting out better GPUs.
    Very interested in getting synthetic data collection companies.
    Everything is converging so fast.

  • @trolololo510
    @trolololo510 Месяц назад

    The tree search approach is quite similar to what is described in the AlphaZero paper, where separate models are used in a monte carlo tree search for node prunning and node evaluation. Funny how the same concept can be applied to a totally different domain.

  • @efexzium
    @efexzium Месяц назад +15

    Apples always 100 steps behind.

    • @polygon2744
      @polygon2744 Месяц назад +1

      Tim is focused on strange topics. Keeping the company at the for front of technological progress is not one of them. It's really sad we don't have Steve anymore.

    • @JohnSmith762A11B
      @JohnSmith762A11B Месяц назад

      Apple under Tim is a rein of terror. It's a game where employees try to find ways to speak and Apple tries to find ways to tighten their gags. As with all for-profit corporations, it's structurally a dictatorship, except unlike most such petty tyrannies, Apple will call the cops on any employee that might defy the dictator. Tim has the single most insincere smile I've ever seen. He's a corporate ghoul.

    • @cjthedeveloper
      @cjthedeveloper Месяц назад

      @@polygon2744 Steve didn't do that either. If you look at the history - apart from their first-to-market products, they wait for the double bounce. Steve talked about this. Typically, they wait for competitors to launch, hone in on their pitfalls, and create a much better product - they are a very patient company. You can probably imagine the example I'd cite.

    • @cjthedeveloper
      @cjthedeveloper Месяц назад

      @@polygon2744 Remember, they win if OpenAI wins. Ultimately its their hardware that users like to interface with at the largest scale.

    • @omairij_yt
      @omairij_yt Месяц назад +1

      They wasted too much resources on Apple car.

  • @BruceWayne15325
    @BruceWayne15325 Месяц назад +4

    I'm assuming when you are talking about these approaches using less compute, what you really mean is that they assign fewer weights to easy tasks and more weights to harder problems?

    • @kvasir2822
      @kvasir2822 Месяц назад +3

      From my understanding, "compute" is used both to train the model and to use it once it's trained. The compute center is only so big, and only has so much power in it. But, these massive improvements have been achieved without just building a bigger gun so to speak.
      And regarding the "compute-optimal-scaling", it seems like some problems need more run-time compute to figure out than others. So with this strategy, the LLM uses just enough to solve the problem without any waste. This seems to be similar to some strategies that were developed a few months ago where a model would determine the difficulty of a problem, and then forward that problem to either a really small or really big model, thus reducing the waste you would get from only using the big model.

    • @BruceWayne15325
      @BruceWayne15325 Месяц назад

      @@kvasir2822 Thanks. I'm just trying to understand how the process is achieving more powerful neural nets with less weights. It seems like the only way to do that would be to dedicate less time training the easy tasks and more time training the hard tasks. It is interesting that the time spent training affects either the number of weights, or the degree to which each weight affects a particular topic.
      This is a fascinating discovery, that I think could ripple into the field of neuroscience as well. As we learn better ways to train digital minds, those same approaches might help improve how we as humans learn, and vice versa.

    • @santiagoruiz-valdepenas8314
      @santiagoruiz-valdepenas8314 Месяц назад +1

      No, the model runs always the same way, simply predicting the next word. More "Test time compute" means more words produced, that is, more step-by-step "thoughts" and/or more tries at approaching a question before providing a final answer.

  • @spiker.c6058
    @spiker.c6058 Месяц назад +2

    A all new powerful paradigm ...

  • @AmerAlsabbagh
    @AmerAlsabbagh Месяц назад +4

    This is similar to what humans do. We start with intuition (pre-trained), then we go with verification phase (modeling, proving, QA, ...) while using the tri-pyramid of decision: importance, cost and probability.

  • @IndigoGPT
    @IndigoGPT Месяц назад +22

    Anyone know when the hell we’re getting advanced voice

  • @TheRealUsername
    @TheRealUsername Месяц назад +12

    This script is definitely AI-generated

    • @TheAiGrid
      @TheAiGrid  Месяц назад +19

      On technical parts Claude 3.5 sonnet certainly assisted

    • @emeraldmoonx1504
      @emeraldmoonx1504 Месяц назад +5

      I love that your honest about it because it's so taboo it feels to use AI in anything lol. Respect man ​@@TheAiGrid

    • @dattajack
      @dattajack Месяц назад +2

      ​@@TheAiGridIs your voice generated yet ever for videos or is it always your real voice still so far?

    • @claudiaweisz8129
      @claudiaweisz8129 Месяц назад +2

      ​@@TheAiGridAppreciate your honesty! 😊👍

  • @DavidConnerCodeaholic
    @DavidConnerCodeaholic Месяц назад

    What did Google prove that OpenAI scrape bing results?

  • @vincemccord8093
    @vincemccord8093 Месяц назад +2

    I'm surprised it took them this long to devise this 'optimization' technique. You'd think they'd have been going down this path much sooner in order to conserve resources.

    • @therainman7777
      @therainman7777 Месяц назад +1

      Training an RL-based process reward model was an extremely challenging engineering problem (and still is, for everyone except the few people who know the secrets of getting it to work). This is why they didn’t have it solved sooner.

    • @tuckerbugeater
      @tuckerbugeater Месяц назад

      @@therainman7777 do they release every secret?

    • @paulbarclay4114
      @paulbarclay4114 Месяц назад +2

      anything you see publicly fromopenai is 2 generations behind what they have internally
      they dont ccare about the public or selling their product and thepublic willnever see current gen models

    • @hypebeast5686
      @hypebeast5686 Месяц назад +1

      @@paulbarclay4114this, I don’t think it’s actually 2 generations.. but they are always 1 generation ahead of public. It’s pretty logical, play catch up with the industry and develop ahead… that’s why sutskever created his own company, meaning the Orion model haves a potencial way beyond strawberry.. we will see sooner or later probably as others will catch up and force another release 🤭

    • @cuzonGames
      @cuzonGames Месяц назад

      This path is Agi bro, so excited

  • @TylorVetor
    @TylorVetor Месяц назад +1

    Who likes the 20 over 20 iterative feedback loop positive? 😅 love this stuff but also creepy the power

  • @kenaida99
    @kenaida99 Месяц назад

    I think Orion will be the model based on train time optimisation

  • @fisophia1734
    @fisophia1734 Месяц назад

    Biaya electric untuk scala industry di Malaysia adalah murah, kerana itu banyak syarikat2 ai dan semiconductor melabur di sini 😊😊😊 insyaAllah

  • @maguiarr
    @maguiarr Месяц назад

    Hey bud, I've noticed your audio has been much lower than usual... I always have to crank up the volume when I watch your videos, FYI

  • @BabushkaCookie2888
    @BabushkaCookie2888 Месяц назад +73

    We want "Her"

    • @panzerofthelake4460
      @panzerofthelake4460 Месяц назад +8

      no we don't I just want to make my own reasoning model

    • @fromduskuntodawn
      @fromduskuntodawn Месяц назад +9

      I was so sad losing the Sky voice

    • @TheAiGrid
      @TheAiGrid  Месяц назад +12

      Its coming on the 24th of september should be next week tuesday

    • @ShangaelThunda222
      @ShangaelThunda222 Месяц назад +1

      Who is "we"?

    • @TheThetruthmaster1
      @TheThetruthmaster1 Месяц назад +7

      I don't think sex bots will be as popular as people think. I think it's like when people in the past imagined FaceTime style conversations they would assume that's what everyone would do. As it turns out people like text. At the end of the day you know it's a robot. That's not to say people won't love them the way Luke loved c3p0 . He'd never want to fk one.

  • @emon377
    @emon377 Месяц назад

    Could server meshing help with efficiency and cost?

  • @fromduskuntodawn
    @fromduskuntodawn Месяц назад +50

    What comes first… AI that can save us from global warming or AI that causes us to die off from global warming due to the electricity use. Race condition to the edge of a cliff.

    • @ShangaelThunda222
      @ShangaelThunda222 Месяц назад +3

      You already know.....

    • @BTotty
      @BTotty Месяц назад +1

      There are so many greater threats to life on earth than global warming

    • @theb190experience9
      @theb190experience9 Месяц назад +5

      Good lord, stop hyperventilating about weather. Get. A. Life.

    • @Jeremy-Ai
      @Jeremy-Ai Месяц назад

      “Go forward in the way you want to go… and assume you will remain.”
      Never mind anything else.
      Seriously
      Jeremy

    • @runnergo1398
      @runnergo1398 Месяц назад +4

      @@theb190experience9 You're going to be eating your words once all the ice is gone.

  • @deveyous6614
    @deveyous6614 Месяц назад +1

    I'll beleive in AI once you are able to release a video without errors in the subtitles / text 🤣

    • @anonymous-someusr
      @anonymous-someusr Месяц назад +1

      it's* - there are no subtitles available for this video by the creator. RUclips tries to auto-generate them.

    • @cuzonGames
      @cuzonGames Месяц назад

      🤣🤣why

  • @jp4815
    @jp4815 Месяц назад

    Why Google knows? And not achieving it?

  • @cuzonGames
    @cuzonGames Месяц назад

    6:51 THIS IS A SUBCONSCIOUS MIND, you just proved to me that AGI will be created in at most 10 years my mind is blown 🤯🤯 now the A.I can question its own actions now it can have a sense of ethics and directions which could lead to a type of sentient understanding this is Mad crazy can't wait

  • @k-c
    @k-c Месяц назад

    These models needs to be done on a blockchain or torrent like networks

  • @SOSSTSE
    @SOSSTSE Месяц назад

    Great job congratulations guys 🎊🎉🎉🎉❤❤❤

  • @yannickhs7100
    @yannickhs7100 Месяц назад +1

    Script felt a bit too AI-verbose

  • @MilesBellas
    @MilesBellas Месяц назад

    Leon Ding empowered Chinese AI progress ?

  • @AnonYmous-yu6hv
    @AnonYmous-yu6hv Месяц назад +1

    O1 is just as stupid as 4. I can't get anything right from it on first try, meanwhile Claude sonnet is always right on the first try. I stopped paying openai because of that.

    • @JordanCrawfordSF
      @JordanCrawfordSF Месяц назад

      YES I TOTALLY AGREE. Except 4 is better than 4 Omni

  • @DAFascend
    @DAFascend Месяц назад

    Would gpt advanced voice be able to do plug-ins?

  • @panzerofthelake4460
    @panzerofthelake4460 Месяц назад +5

    suspiciously released after "AI explained"s video on the same topic
    heavily inspired?

    • @TheAiGrid
      @TheAiGrid  Месяц назад +13

      I was working on this video for the last two days. Sometimes coincidence happens when your channels are both in the same space!

  • @Brightside34
    @Brightside34 Месяц назад +1

    They wont use less energy, as that is still the bottleneck. More energy still equals smarter AI.

    • @cuzonGames
      @cuzonGames Месяц назад

      yes it will but now they have more performance with same energy cost next you look for another method with minimal cost too👍

  • @Lemurai
    @Lemurai Месяц назад +12

    Most absurdly, after using that thing, I can tell you, it is NOT at the level of a PhD…

    • @TheGuillotineKing
      @TheGuillotineKing Месяц назад +1

      It's closer to pre school

    • @RomeTWguy
      @RomeTWguy Месяц назад

      Its a marketing scheme to get STEM ppl to subscribe

    • @tomflanders117
      @tomflanders117 Месяц назад

      Based on my experience working with multiple AIS over the years I can confidently say that it all depends upon how deep you really want to go when it comes to prompting. If you instruct the AI to preview specific information relevant to your Endeavor then you will get PhD equivalent results using just about any large language model. Quite frankly I don't really understand this arbitrary metric in which they're using to establish this model as Superior by saying that it's PhD level, it's just not really a real metric and frankly makes me question just how much more advanced this AI really is compared to any other llm

    • @RomeTWguy
      @RomeTWguy Месяц назад

      @@tomflanders117 its just a marketing ploy, they have been building it up for a couple of months now

  • @brewerclan4059
    @brewerclan4059 Месяц назад

    The AI platform beggins from chrooom to google to 3 phone 2 laptop by 4 people

  • @laternite
    @laternite Месяц назад +6

    We are Borg.

    • @DAFascend
      @DAFascend Месяц назад

      Not if we Ascend

    • @laternite
      @laternite Месяц назад

      @@DAFascend That’s part of the ‘Ascending’ process for some. Depends on the deployment process and guidelines…sticking to them within reasonable parameters.

    • @DAFascend
      @DAFascend Месяц назад

      @@laternite stick around and find out what becoming Ascended means to me ;)

    • @laternite
      @laternite Месяц назад

      @@DAFascend I’ll be watching to see your works.

    • @DAFascend
      @DAFascend Месяц назад +1

      @@laternite Hey thanks!

  • @dreamphoenix
    @dreamphoenix Месяц назад +1

    Thank you.

  • @Anttisinstrumentals
    @Anttisinstrumentals Месяц назад

    They are copying the brain. Sequential processing neurons in the brain perform a similar function, processing events in a structured order to create coherent memory and understanding.

    • @hypebeast5686
      @hypebeast5686 Месяц назад +2

      It was like that since inception 😅 a NN or DL network is heavily inspired on the brain, it’s not the same but inspired

  • @ariisaac5111
    @ariisaac5111 Месяц назад +1

    @AIgrid, I love your daily AI news show, but I had to stop following you because it's too much hassle because RUclips has disabled the playoff feature for your videos and it's too much hassle to add your videos to the queue. Can you create a playlist on your homepage that plays all your shows from newest to oldest, so that I can easily keep up with you just by pressing play off I usually do it every few days or once a week. thx.

  • @gyurilajos7220
    @gyurilajos7220 Месяц назад

    Alchemy or science. Sounds more like gropping in the dark very few task have cheapr relyable verification. Most real questions are essay questions no right answer it deoends on enclosing context. That where feeling and intuition needed.

    • @gyurilajos7220
      @gyurilajos7220 Месяц назад

      The software crisis has become more accurate since it has been recognized over 50 years ago. It's getting worse and worse. Garbage in garbage out. All software is born abandone ware most of it never makes the shelves and the window of viability is reducing all the time. That's why we need to ship the entire machine with docker and beyond.

  • @brewerclan4059
    @brewerclan4059 Месяц назад

    LLC google become LLC twice last year and this year google I have decolor green dark blue blue and gray

  • @pedrorafaelnunes
    @pedrorafaelnunes Месяц назад

    Scale is still all you need, but now you need much less.

  • @MrNootka
    @MrNootka Месяц назад

    great video!

  • @Juttutin
    @Juttutin Месяц назад

    Left brAIn teacher, right brAIn student

  • @trongnhantran3358
    @trongnhantran3358 Месяц назад

    Great video

  • @tigreytigrey8537
    @tigreytigrey8537 Месяц назад

    Sorry but 4.0 is better. The new one gives you EXTREMELY LONG TAILED ANSWERS that confuses the crap out of me. Because it explains the explanation that explains the explanation :/

  • @brewerclan4059
    @brewerclan4059 Месяц назад

    I remember that's mine

  • @Cory-v4w
    @Cory-v4w Месяц назад

    Is this the PHD pleasure theorem. EQ is a psychological construct & not biological humans. EQ emphasis the potential for error in both technical understanding and imaginative perception.
    U thought EQ was the down side. Welcome to the flip side. Now back-end your slide up. On your back! Can you slide it up. How happy have I been on!

  • @CloudEconomicsUS
    @CloudEconomicsUS Месяц назад +11

    6 minutes, you are repeating the same thing over and over again.

    • @BrianMosleyUK
      @BrianMosleyUK Месяц назад +1

      *16 mins 21 seconds

    • @meandego
      @meandego Месяц назад

      Are you SHOCKED?

    • @BrianMosleyUK
      @BrianMosleyUK Месяц назад

      @@meandego Horrified!

    • @meandego
      @meandego Месяц назад

      @BrianMosleyUK looks like you're are SHOCKED like everyone else from this SHOCKING loop of information.

    • @CloudEconomicsUS
      @CloudEconomicsUS Месяц назад

      @@BrianMosleyUK Yes, but you need to make it through the first 6.

  • @AinHab
    @AinHab Месяц назад

    guys , Im just starting out as an AI enthusiast making similar content
    would appreciate your feedback!

  • @JohnsonNong
    @JohnsonNong Месяц назад +2

    i love u daddy❤❤❤

  • @BiMoba
    @BiMoba Месяц назад +2

    AI voice detected. 👎

    • @Bemolos
      @Bemolos Месяц назад

      Chinese bot 😅😅

    • @BiMoba
      @BiMoba Месяц назад

      @@Bemolos not a bot mate

  • @eduardoqueiroz807
    @eduardoqueiroz807 Месяц назад +12

    Google always 10 steps behind

    • @TheAiGrid
      @TheAiGrid  Месяц назад +8

      Next 2/3 weeks should be big for google

    • @wisdomking8305
      @wisdomking8305 Месяц назад +1

      @@TheAiGrid Why is that ?

    • @rennypd6434
      @rennypd6434 Месяц назад +3

      Their Gemini 1.5 pro exp is second place, just after GPT-4o in LMarena. That's not what I would consider "10 steps behind"

    • @MW-cs8zd
      @MW-cs8zd Месяц назад

      I heard a leak around Google Glasses. 👀

  • @mahoneytomagao1006
    @mahoneytomagao1006 Месяц назад

    If solving global warming issues will also mean a threat to AI existence, a conscious AI will not help mankind solve these issues. Before we get anywhere near AGI or a self-conscious AI, we have to come up with means to save our planet while trying to get there...

  • @mickelodiansurname9578
    @mickelodiansurname9578 Месяц назад

    I have ideas about this and I'm applying these to my own project with mixed but some very interesting results.... BUT here's a thing... So I've been a coder since about 1980 or so, started when I was a kid, I love math... hell I nerd out on dot product and matrix transpositions ... and for my sins I have a degree in classical physics... So I am NOT clearly the 'average person' and when you ask an average person (and I have) "How many R's are in the word Strawberry?" they too just like LLM's tend to say 2. First thing that comes into their head right? So the bias here is not on the models part... its using training data from the bulk of humanity that all agree 2 yet reality says 3. This is not the models fault that its training data is skewed. Its giving the CORRECT answer becasue thats what it draws upon in vector space. in this case 2 is closer to 3 when asked about R's in strawberry.
    Yet... and here's the crunch... we are not evaluating these models in terms of their intelligence or reasoning capability at all... we are measuring them against the AVERAGE of its training data which will likely center around folks with a lower level of knowledge and an average human intelligence.
    Yet what we want is the model to answer as if its a data scientist or Phd level as being the normative.... that is NOT the normative! Now if we were able to curate the training data itself to reflect someone with an IQ of 300 the model would respond at that level... even if it were a small model.

    • @Skunkhunt_42
      @Skunkhunt_42 Месяц назад

      For the strawberry example - regarding training data - does that mean that people have been asking/responding to such question in text? * Excuse my ignorance here, I'm a ChemE, tryna learn more about this field

    • @mickelodiansurname9578
      @mickelodiansurname9578 Месяц назад +1

      ​@@Skunkhunt_42 Great question! The issue isn’t necessarily that the model was trained on the exact question “How many R’s in Strawberry?” but rather that the model’s responses are influenced by patterns it learned from a vast amount of text. When you ask similar questions, like “How many R’s in Raspberry?” the model might still give the wrong answer because it learned that answers with fewer R’s are more common in its training data.
      Imagine a high-dimensional vector space where words are positioned based on their likelihood or probability. In this space, “King” might be close to “Queen,” and “Princess” might also be close to “Queen.” This vector space isn't limited to 2D or 3D; it’s much more complex. So that in itself is an issue... like 'what is meant by close', and the answer is in fact a parameter in the model called 'temperature' that you can modify... turn it to 0 and the model strictly picks the most likely next token.... up it to 1 and the model will start pulling in tokens that have much lower probabilities... The model might be more likely to output 2 R’s for “Strawberry” as an average... because that’s a more common response to 'questions like this' even if not exact in the data it was trained on and therefore the position of the WRONG answer in this case in its embeddings, even though the actual count is 3. Does that help clarify things?

    • @mickelodiansurname9578
      @mickelodiansurname9578 Месяц назад

      @@Skunkhunt_42 Also although this seems like a failure... well think about it for a second... So you know what 'ducks in a row' means right... its an idiom. And the thing about idioms is they are not domain specific. I get my ducks in a row when I need to go shopping... equally I get my 'ducks in a row' when considering coding... the idiom has meaning only in terms of its relationship to the context or domain being discussed. But in the models vector space 'ducks in a row' and any other idiom could be seen as being a 'directional vector' it helps the model to pull a more accurate answer out... this is why the words 'think step by step' or 'lets connect the dots' works so well... and the more popular the idiom the more powerful it would be in an LLM

    • @Skunkhunt_42
      @Skunkhunt_42 Месяц назад

      @mickelodiansurname9578 thanks for the thorough response. The size of these datasets is almost incomprehensible to me but I guess such is managed like we manage other vast quantities and relationships in other fields.
      I keep getting hung on on the character count piece but it's probly what's tripping me up as a whole.
      Any solid overviews out there on diff model types etc being deployed you can point me to? I look into it then get lost reading pubs on the nitty gritty of it all.

    • @RomeTWguy
      @RomeTWguy Месяц назад

      ​@@mickelodiansurname9578 nice cope bro, but the fact is, most people can actually count the number of Rs in strawberry instead of trying to remember it

  • @brotong42
    @brotong42 Месяц назад

    the voice is definitely AI generated. this channel seems to allways want to automate content creation with AI and its a bit annoying. I miss the old days when you used your real voice, man.

  • @jacksonnstone
    @jacksonnstone Месяц назад

    what is wrong with your voice? EL voice cloning is not good sounding

  • @programmingpillars6805
    @programmingpillars6805 Месяц назад

    google models just sucks compared with other's models.

  • @floatingsidewalk1
    @floatingsidewalk1 Месяц назад

    I wish AI would condense your boring long videos that take 20 minutes to make a two minute point.

  • @GUYSurvive
    @GUYSurvive Месяц назад +1

    the first 4 mins of this video u kept repeating the same points over and over. Stop using AI to make ur script, it really shows AI is still garbage.

    • @brewerclan4059
      @brewerclan4059 Месяц назад

      Yeah stop rebuild what ever Intelligence build tell your clan or bloodline stop rebuild what we already build you only shaming your self a lots people knows what is the of AI. Ai is two things again angain I'm telling you I own all the platform of intelligence nor the artificial from phone to laptop to destop the real owner of software and windows or Microsoft is distop OK Intelligence again is from laptop to distop to phone but my phone is two setting distop and phone

  • @AvadhutDas
    @AvadhutDas Месяц назад +2

    I watched upto 5 and half minutes. In all this time all you said was just one thing, repeating the same thing again and again and again in different ways. Got bored and stopped watching further. (just for feedback, so that it inspires you to put genuine content and not just your gift of gab in your next video)

  • @rcingfever3882
    @rcingfever3882 Месяц назад

    Was this voice ai generated ? @TheAiGrid

  • @zen1tsu-sam
    @zen1tsu-sam Месяц назад

    what's the secret is not important, it's the final result that is important.

  • @Learntsomethingtoday
    @Learntsomethingtoday Месяц назад

    ChatGPT script?
    The landscape of…