Are Hallucinations Popping the AI Bubble?

Поделиться
HTML-код
  • Опубликовано: 10 сен 2024

Комментарии • 2,9 тыс.

  • @stchaltin
    @stchaltin 26 дней назад +1862

    The issue isn’t with LLMs being wrong when asked a question. The issue is with corporations pushing them into every service possibly imaginable while *knowing that.*

    • @Techne89
      @Techne89 26 дней назад +54

      On point . I can’t count down useless and unnecessary or unethical features, Products or services I saw out there.

    • @wpelfeta
      @wpelfeta 26 дней назад +75

      Exactly. LLMs should not be sources of truth. Hallucinations make them fundamentally bad at being such a tool. Unfortunately, companies everywhere want to use them this way. LLMs should be sources of *inspiration* and *creativity*. Their output should be a starting point not the end product.

    • @svenkrackow5550
      @svenkrackow5550 26 дней назад +11

      If you build an LLM on an expert system, you might get a rational AI. True. But unconcious, still. If you build a second system, that - goal-oriented - interpretes the ratio-morphic transformer, that (our) neural network appears to be, you might get a conscious AI. An ethologist's view, I admit...

    • @mcdinkysgarage4514
      @mcdinkysgarage4514 26 дней назад +1

      Yes, there are very few motivations that would require someone to do such a thing.

    • @sitnamkrad
      @sitnamkrad 26 дней назад +59

      I think the problem goes deeper. LLMs were never intended to answer questions. They were intended to predict text. It was doing this prediction so well that it fooled people into thinking it answered questions to the point that it has now been re-purposed to do just that. But the underlying mechanic is still working on text prediction. That's why with the wolf, goat, cabbage puzzle example, it predicted the solution even without the added information that they would eat each other, because it also predicted that information to be part of the prompt.

  • @citywitt3202
    @citywitt3202 22 дня назад +57

    “The ability to speak does not make you intelligent”
    Qui-gon Jinn

  • @JorgTheElder
    @JorgTheElder 27 дней назад +739

    I liked it much better when everyone was careful with the terminology and did not mix AI with ML and expert systems.

    • @SabineHossenfelder
      @SabineHossenfelder  26 дней назад +164

      yes, I agree, I wish the term had never become that widely adopted in the first place

    • @moletrap2640
      @moletrap2640 26 дней назад +53

      @@SabineHossenfelder correct terminology does exist, I wish we would start using them more consistently. Machine learning is known as *predictive AI*. Tools like ChatGPT are not called they are called *generative AI*. Predictive AI and generative AI are very different from each other and it would be helpful if we would use such terminology. Predictive AI has a ton of real world applications and value. Generative AI only supports applications where accuracy is not important, eg 'the arts.'

    • @GEMSofGOD_com
      @GEMSofGOD_com 26 дней назад +1

      It exists in #TOPICALMODELING. Not a kindergarten soup that goes under the letters AI

    • @CrniWuk
      @CrniWuk 26 дней назад +18

      That is not by accident however. No one would really know what to do with "Large Language Model" for example. But Artificial Intelligence? That works as good marketing because everyone has SOME kind of image stuck in their head due to the countless of science fiction storys, movies, games, books etc. that have some kind of "Ai". Of course this idea has nothing to do with the LLMs. But still, it's easier to sell to someone than say, NFTs or Bitcoin.

    • @minhuang8848
      @minhuang8848 26 дней назад +28

      I don't know, AI always described a huge array of different agents, which ultimately had nothing to do with ANNs of any sort of making. Chess, Starcraft NPCs, basic visual processing/OpenCV-type stuff - you name it, we called it AI. It's just an umbrella-term for things that vaguely emulate human intelligence or features thereof, and it makes perfect sense on account of how quick and easy it is to type out. Beyond that, it's just not like there is a worse reality; the vast majority of laymen bystanders never intended to be precise about the terminology anyway (which we have, by the way - it's just mostly being used within research circles).
      Honestly? It's a moot fight against whimsical human language. "AI" is such a neatly descriptive and perfectly contextualizable term, now look at physics and bing-bong, CTE-adjacent astrology communities talking about holofractal symmetry projections in their glorified horoscope, for people who love to completely pervert actual terminology to confuse audiences that just don't know better. Having dealt with this for most of my life, I really couldn't be happier about AI having been the established nomenclature for many decades now... because everyone knows the alternative would have been so, so much worse. It almost always is, just look at the flaming garbage fire that is taxonomy and how borderline irresponsible our categorization approaches can end up being.

  • @exapsy
    @exapsy 25 дней назад +259

    4:51 "Imagine AI being able to win every logical argument. Reddit would become a ghost town overnight"
    You're assuming people on reddit use logic. That's quite of a bold statement

    • @freelancerthe2561
      @freelancerthe2561 25 дней назад +10

      Considering its been used in training data, it explains a lot about earlier versions of the models.

    • @stickmcskunky4345
      @stickmcskunky4345 25 дней назад

      I think the assumption is that people on reddit don't use logic, which is why the introduction of flawless logic would break the platform.

    • @maxmordon7295
      @maxmordon7295 25 дней назад +9

      Reddit is "Heaven"... compared to "X", Facebook, Instagram, RUclips...

    • @psylocyn
      @psylocyn 25 дней назад +13

      Reddit is filled with a shocking number of highly intelligent people when you compare it with every other large social media platform

    • @oldcowbb
      @oldcowbb 25 дней назад +8

      @@psylocyn when people say redditors are dumb you know what kind of sub they dwell in, very telling on what kind of peaple they are

  • @gJonii
    @gJonii 26 дней назад +150

    One thing that people may find interesting: LLM's hallucinating isn't a bug, it's the way it works. It's completing patterns it hasn't seen before, using statistics it has learned from other texts. This is indistinguishable from hallucinations, since it's always just made up stuff, the best guess LLM has about what might come next. Sometimes this hallucination is useful, it uses rules we agree with. Sometimes it uses wrong rules, rules we disagree with. Either way, it's LLM just hallucinating what the "correct" answer might be.

    • @ChristianIce
      @ChristianIce 25 дней назад +6

      This.

    • @idan5323
      @idan5323 24 дня назад

      You don't know english.. *predicting, hallucination is a wrong prediction.

    • @gJonii
      @gJonii 24 дня назад +21

      @@idan5323 They're all "wrong predictions", is the point, they're fabricating completion to an input they've likely not been trained on. That this hallucination often seems to be kinda factually correct is the surprising part in of itself.

    • @idan5323
      @idan5323 24 дня назад +2

      @@gJonii I am starting to think you are an LLM, as you seem to spit words without understanding their meaning.. NOT ALL PREDICTIONS ARE WRONG, that's just a bold fallacy.
      LLMs are sometimes wrong, and sometimes right, which is the nature of prediction. The ai has no awarness or understanding of the words he says, and he doesn't know wrong and right, yes. But what YOU say is just wrong.

    • @kilroy1964
      @kilroy1964 24 дня назад +20

      Oh my gosh! You two are talking past each other, and not comprehending what the other has said.
      You are both right, and are even saying similar things. However you are looking at the same thing in different resolutions and using different terminology.
      Idan, you are right when treating an AI as a black box. There is output we consider to be right and output we consider to be wrong.
      However OP is looking at the process required to make the black box work. When you just examine what it is doing, you see that it is not being given criteria to distinguish right from wrong. Therefore all output is merely whatever the training data produces. (So you could say there are only hallucinations as output.)
      Amazingly, much of the time this output actually makes sense to us.
      In other words:
      Hallucinations are NOT malfunctions! The system is working correctly.

  • @MictheEagle
    @MictheEagle 26 дней назад +136

    Colin Fraser: "Solve the wolf, goat and cabbage problem."
    LLM1: "Shoot the wolf."
    Colin Fraser: "No!"
    LLM2: "Shoot the goat."
    Colin Fraser: "No!"
    LLM3: "The farmer?"
    Colin Fraser: "Definitely not!!"
    Baby LLM4: "Shoot the cabbage. And then, the boat."

    • @lorn4867
      @lorn4867 26 дней назад +7

      😂

    • @extrafoxjj
      @extrafoxjj 26 дней назад +9

      Both Chat GPT and Grok have the correct answer to that problem

    • @vaclavkrpec2879
      @vaclavkrpec2879 26 дней назад +19

      That's not LLM, that's mid-west American's answers...

    • @MictheEagle
      @MictheEagle 26 дней назад

      @@vaclavkrpec2879 Lol

    • @kabongpope
      @kabongpope 25 дней назад +4

      Grok.AI: Shoot the hostage!

  • @arctic_haze
    @arctic_haze 27 дней назад +441

    I once have seen a banner in a school class: "All human errors come from misunderstanding the laws of physics." It seems it is now true also about the AI.

    • @aaronjennings8385
      @aaronjennings8385 27 дней назад +3

      What errors?

    • @arctic_haze
      @arctic_haze 26 дней назад +9

      @@aaronjennings8385 I am not sure if you are a bot. But certainly you are a stalker.

    • @sorvex9
      @sorvex9 26 дней назад +21

      As a machine learning professional, I have been facepalming for 2 years. At least people are starting to somewhat understand its usecases now....

    • @GEMSofGOD_com
      @GEMSofGOD_com 26 дней назад

      #TOPICALMODELING

    • @saftheartist6137
      @saftheartist6137 26 дней назад +3

      Interesting, it may be possiblely true, well at least for the time being.

  • @jttcosmos
    @jttcosmos 26 дней назад +200

    Regarding the maths Olympiad, it is a bit less surprising when considering that the problem was manually translated into Lean by humans, it acted as the solver and then transferred the data back to the system to phrase it properly. Lean was run against the problem an additional time to verify the solution. Oh, and it took up to 60 hours per question, which is far longer than any human gets.

    • @Alice_Fumo
      @Alice_Fumo 26 дней назад +21

      Where'd you find this information? I'd like to read up on those specifics myself.

    • @jttcosmos
      @jttcosmos 26 дней назад

      @@Alice_Fumo If you search for Google Maths Olympiad and Sir Timothy Gowers (he was one of the evaluators), he added quite a bit of context. ArsTechnica for instance has a good summary.

    • @timmygilbert4102
      @timmygilbert4102 26 дней назад +6

      60h with or without sleep 😂

    • @dv_interval42
      @dv_interval42 26 дней назад +17

      Typical marketing hype by Google

    • @jttcosmos
      @jttcosmos 26 дней назад +16

      @@Alice_Fumo Not quite sure where my reply went... Sir Timothy Gowers was one of the two people who scored the results and added additional information on social media. If you look up his name and the Olympiad, you should find more details.

  • @BarryKort
    @BarryKort 26 дней назад +101

    Although "hallucination" is a popular word to characterize unrealistic assertions from LLMs, a technically better term of art is "confabulation."

    • @mikicerise6250
      @mikicerise6250 26 дней назад +6

      Yep. And humans do it too.

    • @randyevermore9323
      @randyevermore9323 26 дней назад

      Yes, it seems the use of the term "hallucination" is just a transparent attempt to make LLM-based AI seem more human-like than it will ever be.

    • @MichaelDembinski
      @MichaelDembinski 25 дней назад

      Brilliant, @BarryKort. This has been an issue since the day ChatGPT launched back in late 2022, yet NO ONE has yet questioned the label 'hallucination'. 'Confabulation' is so much nearer the mark!
      Confabulation (noun): A fabricated memory believed to be true.

    • @ChristianIce
      @ChristianIce 25 дней назад +16

      A better term is simply "output", because there's no difference between a correct and an incorrect one.
      For the "AI" is just meaningless tokens, it is correct in the same way it isn't, it can't tell the difference.

    • @jonathanodude6660
      @jonathanodude6660 25 дней назад +11

      @@ChristianIce silly take. the correctness of information is important to us. thats why we have the word "correct" in our language.

  • @stevensteven3417
    @stevensteven3417 26 дней назад +121

    We dont have sentinent ai, its a language model with access to a huge database, it is not self aware or alive, its a data center.

    • @jorgecarvajal5385
      @jorgecarvajal5385 25 дней назад +1

      well thats true, not asingle neural network its a true ai yet, maybe some day but we dont know

    • @mer863
      @mer863 25 дней назад +18

      @@jorgecarvajal5385 I hope not. What good would it bring? we already have intelligent life(ourselves) and we don't even care for that. We allow that others experience poverty, starvation, corruption, disease, even war. So what if we manage to achieve man made intelligence? We put all that time, energy and resources into inventing something that already exists in a natural form, something that is apparently worthless to us.

    • @mer863
      @mer863 25 дней назад

      @@jorgecarvajal5385 and to be clear: I am not anti intelectualism. I am in absolute awe of the things we can accomplish and the things that are left to be discovered. I am just extremely disappointed that AI is used not for the good of humanity or for further scientific understanding. At this moment AI is just a glorified theft of other's intellectual property and a retelling of already known facts. It brings too little good to the table, but it does have ethical problems and it hurts the livelihood of many people(like artists, writers, creatives and even scientists). And we continue to pour money into this instead of funding the fights of actual, real life problems that make other humans suffer

    • @maxmordon7295
      @maxmordon7295 25 дней назад +1

      For now...

    • @jurassiccraft883
      @jurassiccraft883 25 дней назад

      @@mer863 Personally I think that a true sentient AI would probably be able to help solve a lot of those problems (I don't see it as some kind of artificial god or anything. in part because you cannot create an artificial version of something that doesn't exist) I mean essentially AI is a tool like satellites are a tool. the reason we have satellites is because of the Cold War and some pissing contest the US got into with the Soviet Union but those tools today are helping to feed billions of people as we can now better predict weather patterns. same with fertiliser, it started out as a bomb for weaponry and is now the only reason we can feed as many people as we do.
      Like every tool we have made in the past, it will undoubtedly be used to harm others but it will also be used to make the lives of many much better. And, like every tool we have ever made in the past, we will never put the genie back in the bottle so even if we haven't reached a true AI yet its better to have people concerned with how we use it for the benefit of everyone and not for the benefit of some.

  • @jeffryborror4883
    @jeffryborror4883 27 дней назад +238

    So, LLMs have reached the level of our politicians who blather content-free word salad and to make them intelligent we need to teach them math and physics. Good luck with that.

    • @NorthgateLP
      @NorthgateLP 26 дней назад +27

      Unlike politicians LLMs can be steered towards the right direction.

    • @DefaultFlame
      @DefaultFlame 26 дней назад +21

      The difference is that the AIs will improve, the politicians won't.

    • @michaelharrison1093
      @michaelharrison1093 26 дней назад

      Yes it does seem that Donald Trump's speeches are written by AI

    • @Dennis-vh8tz
      @Dennis-vh8tz 26 дней назад +40

      ​@@NorthgateLP Like politicans LLMs can be steered in any direction desired by someone with enough money.

    • @LuisAldamiz
      @LuisAldamiz 26 дней назад +3

      Seen that way, it's a pretty good level of Bidentelligence.

  • @AlValentini
    @AlValentini 26 дней назад +170

    Working in AI, I can say that LLMs are not the alpha and omega is already established. It's been put to rest. While companies keep training bigger models for commercial reasons, researchers have already switched the attention to cognitive features. Memory, reflection, planning, abstractions... It's not true that the industry is stuck with LLMs

    • @andrasbiro3007
      @andrasbiro3007 26 дней назад +16

      Yes, but LLMs proved to be far more powerful than anyone expected, so the attention they are getting is warranted. And we are very very far from their limits.

    • @Alice_Fumo
      @Alice_Fumo 26 дней назад +10

      I wouldn't say researchers have "switched" to [topics you named], there is still tons of papers coming out in regards to language modelling with transformers / adjacent architectures and improving upon those. All the other things you mentioned have been happening the entire time regardless, but as someone who reads a bunch of the papers, I don't get the feeling that improvements upon transformers have been exhausted.
      What bothers me is that I've seen like 20 techniques which drastically improve upon the things we've seen thus far and none of the big companies have implemented those yet to my knowledge. I want to see what a model which combines all of the current research can do.
      I agree with the conclusion that the industry will pivot to more interesting architectures the moment "bigger model go brr" stops being a strategy which pleases investors - which should be soon-ish.

    • @Leonhart_93
      @Leonhart_93 26 дней назад

      LLMs are the poster child and the reason why stocks exploded. Hard to get that level of funding and public involvement without direct hype like with LLMs

    • @termitreter6545
      @termitreter6545 26 дней назад +13

      That doesnt really make sense to me. "Real" researchers knew before the hype that LLMs are very limited in their application. The people that drove the hype behind LLMs have no clue - or dont care - about the nature of artificial intelligence, let alone about its usefulness.
      Heck, ive got very little clue about AI and couldve told them that an LLM is more about the A and less the I.

    • @silikon2
      @silikon2 26 дней назад +6

      It still reeks of an AI bubble crash. This is hardly a shock though. You didn't have to know one thing about the (fill in the box) to smell the latest silicon valley bubble.

  • @joaoguerreiro9403
    @joaoguerreiro9403 26 дней назад +24

    I remember my computer science professor saying in a class that years ago deep learning was looked down and most experts in the field thought symbolic AI was the thing. Maybe a mix of both is the answer?

    • @joaoguerreiro9403
      @joaoguerreiro9403 26 дней назад

      @@mattmaas5790 why not? Maybe I misunderstood my professor!

    • @openhorizon1162
      @openhorizon1162 26 дней назад

      Disregarded fields of scientific research of all kinds have later turned out to be where the real analytic power is. So I always start by doubting someone diss'ing a particular tech as a whole. (That can go wrong too, I realize.) Last I knew, symbolic AI is not how the brain is thought to work, so the machine learning techniques became ascendant. Separately, as one example, in the 1990s, in language processing, it became clear the learning models could produce much better results than symbolic methods. (Tho that has to depend on what the challenge is.) I would bet the best systems will eventually use a mix of techniques, including self-checking for internal consistency and for consistency with reality.

    • @Pedroramossss
      @Pedroramossss 26 дней назад

      Are u portuguese?

    • @squamish4244
      @squamish4244 24 дня назад

      Maybe Gary Marcus has a point, as annoying as he is. But then again, companies like DeepMind have always been well aware of the limitations of LLMs. Demis Hassibis has been trying to say this without saying it so as to not piss off Google and end up at the bottom of a lake sealed in an oil drum.

    • @vast634
      @vast634 24 дня назад

      Neither using larger and larger wordbases (wich are limited and already polluted by AI texts) nor symbolic model will create an AGI. Embodiment is the right approach. The ai needs to live in some agent that wants to do things, tries things and learns from that.

  • @Gilotopia
    @Gilotopia 26 дней назад +8

    As karpathy said: hallucinations are the main output of LLMs as they are next word predictors. Not hallucinating is the anomaly.

    • @prettyfast-original
      @prettyfast-original 23 дня назад +1

      The overwhelming majority of people here, including the video host, probably never heard of Karpathy or micrograd or even backpropagation. But boy howdy do they have lots of opinions about LLMs!

    • @mommyanddaddywerealiengods
      @mommyanddaddywerealiengods 16 дней назад

      ​@@prettyfast-originalcare to elaborate?

  • @aukir
    @aukir 26 дней назад +6

    The thing about LLMs is that they're just huge dimensional arrays of numbers, that relate the probability of a number coming after another number. They don't know words.

    • @starkraft2506
      @starkraft2506 19 дней назад +1

      ... so what then is "knowing" a word? You're on to an important thread, which ends with my assertion that for an AI to be what most people think of as "general" - behaving like a human - it needs to be embodied with a similar full sensorium of experiences to anchor those words to.

    • @marcobormann67
      @marcobormann67 15 дней назад +1

      @@starkraft2506 I agree with you. In order for words to have meaning, something has to have meaning in an emphatic sense for the system that sues these words. But then again words also have meaning in relation to each other and it looks like this latter sense of meaning is enough to handle language quite well.

    • @starkraft2506
      @starkraft2506 14 дней назад

      @marcobormann67 it can mimic quite well, for sure. What I don't think it will do is generate meaningfully novel language. It may create novel sentences, but it will never be able to use isomorphisms to create new use cases - for example how "rough" can mean a texture (which is uncomfortable), or an experience which is uncomfortable, or as an antonym to gentle. There's a thread of meaning between those uses which isn't apparent based on the depth of information available in words alone.

  • @carlbrenninkmeijer8925
    @carlbrenninkmeijer8925 27 дней назад +40

    Thank you Sabine for this eye opener. I tried to think about how we learn to speak at early age and do some math later.

    • @axle.student
      @axle.student 26 дней назад +3

      Communication/language requires far more than words. It requires most of our human perception to be involved.
      To put it another way you can't design and build 1 tenth of an automobile and expect it to successfully drive down the road. It's a kind of all or nothing thing. (Some parts of the "all" raise some difficult questions with regard to self awareness, consciousness etc.)
      >
      In the early days I contributed to a project named AARON (Advanced Artificial intelligence ReasOning Network) from Isomer Programming. In essence it learned language concepts (with programmatic rules) from it's users, starting out from something like a newborn. Not only did it use word/language processing but also used visual (image) recognition.
      So as well as natural language processing it was also using object recognition and the beginnings of discerning emotional context.
      Most of the same problems that exist now existed back then. It's not a matter of finding solutions, it is a matter of correctly defining the problem, which is extremely complex and difficult. Humans are not logical creatures and we get it wrong (so to speak) most of the time. It is more of a scale in degrees of correctness than a binary write or wrong. Expecting perfect correctness from a machine to replicate what is fundamentally imperfect to begin with doesn't make a lot of sense.
      >
      There are many good clues in the text of Developmental and Cognitive psychology (an understanding of philosophy of mind is helpful, and I gained a lot of insight from Dr Robert Bolton's interpersonal communication skills). I don't always fully agree with Steven Pinkers computational theory of mind, but his work in his books provide a good illustration of many of the underlying "human" cognition issues that AI needs to address to be useful.
      >
      Personally I would feel more comfortable if they dropped the whole AI context and focused upon IA (Industrial Automation) instead.

    • @jorgecarvajal5385
      @jorgecarvajal5385 25 дней назад

      you learn speak for repeat and obserb its you perception and later come the words and not just math, for the first 2 level of language or language witout contexts is easy, imagien like variable but with image, you see fire and you dont thing in the meaning of the word but in the object it self, car, house, dog, cat, but later come 3 level language we have methaphor like its raning cats and dogs, the old concept its in old england when the raning is to hard the cats and dogs come down to the upper floor to the first floor, tthats mean you need context for understand what it is, of for example a fly over the boat can have aloot of meanings just for the words it ues a fly can be the insect or a airplane can be both depend on context, and even we have japanese that is a language in the 4 level, thats mean you just dont need contexts but you need the context of that specifict worlds and kanjits it was write, like i say its not just math its the methaphor and the experience and you need to extract the context of the word and methaphor for undestand, and thats nearly beyond binary, for a machine undestand that it not just repeat and traning it need to be sentient, or atleast close to

    • @jorgecarvajal5385
      @jorgecarvajal5385 25 дней назад

      @@axle.student no necesary for language on 3 or superior level yes you need context for undestand what is happen one of the famouse phrases for that is "a fly over the boat" fly can be the insect or a airplane, but all depende it is a big boat or a little tiny boat in a bathtube or even in that case you need context can be the insect or can be a toy XDD, but for language in level 1 and 2 no, thats its call language witout context and is weird for us humans speak in that way but machine can undestand and proccese language in lvl 1 and 2 easy and faster

    • @axle.student
      @axle.student 25 дней назад

      @@jorgecarvajal5385 You are talking chatter bots vs genuine interpersonal communication. We have been creating chatter bots for over 30 years.
      Even parrots can repeat words in a string. It doesn't meant the parrot understands it.
      >
      LLM is just advanced chat bot backed by automation. There is no 'Understanding" concept.

  • @Albert-me1oe
    @Albert-me1oe 26 дней назад +9

    The term AI is being used very loosely nowadays to nominate anything computer generated

  • @Thomas-gk42
    @Thomas-gk42 27 дней назад +31

    I don't believe in stocks either, it's a hallucination too.

    • @trucid2
      @trucid2 26 дней назад +7

      So is money. Money is a spook.

    • @nathangamble125
      @nathangamble125 26 дней назад +3

      @@trucid2 Gubmint says the paper and metal discs can buy stuff, therefore they can!

  • @angelikaolscher7104
    @angelikaolscher7104 26 дней назад +14

    I'm a business student. For 2 years, everyone wrote flashy papers on AI. I only found engineers to talk about this with. It's like business people.... Don't want to know what they don't know. I'm a bit of the black sheep of the study, but by now I mind it less. The ai bubble bursting at face value was a predictable outcome. Maybe now I have grounds for people to listen to my research. I feel like I'm screaming into a corporate void. I had a bit of a crisis because of it, for i was saddened by science, for I thought naively science was open and curious. I hope we go onto better times. Thank you for this video.

    • @digiryde
      @digiryde 26 дней назад +4

      I have been in IT for decades. Most business people do not want to know what they do not know.
      I look at it this way. We all live in silos of what we know. The size of that silo is different for each person. What is kept in that silo is different for each person. Each of us tends to fill our silo with what we think we need and/or enjoy. For those things not in our own silos, we have other people.

    • @angelikaolscher7104
      @angelikaolscher7104 23 дня назад

      @@digiryde I yesterday handed in half of my paper for review. I wrote like a manic after finding this video. Luckily, finally, the internet is exploding with voices from all directions. CIOs are also getting aware. So, hopefully, we can now limit a bit of delusion and start giving engineers more time and money and trust to do the RIGHT thing. I feel like managerial stupidity is a health risk for everyone around, especially engineers. Now, I also study management. But..i know when I am stupid, and ask people who know more than me to approve my work. I can't begin to say how often I got things wrong.... Usually i never manage nothing I didn't experience myself. This is different. I can't do it. I legit can't. I can only try my UTMOST to get some conceptual understanding. I also now understand why engineers often don't have hair. I never lost so much hair in my life. Main stressors : managers don't see the importance, managers dobt want to hear anything technical to find logic behind importance. Now, I'm drawing pictures. And telling them, it's all about them. It's not. I dobt give many shits about most of my kind. But, apparently, what's true doesn't matter in science. Apparently, I just need to sell ideas. Frustrating. But most people here know that much better than I could begin to imagine.
      Regardless, I hope for the engineering silo to spill over some bit of Intel to management.

    • @pedrova8058
      @pedrova8058 22 дня назад +3

      The exchange value of money needs "faith", if business people keep a bubble going until the last moment, it's because it generates more money. They know, they just pretend ignore it until the crash is imminent (obviously not everyone comes out safe, there are always a few who jump when the debacle starts; many others get caught in the confusion and sink... it's called speculation, and it has been well known for a long time, at least since the Great Crash)
      Thanks stock markets!

  • @erinm9445
    @erinm9445 23 дня назад +6

    I *really* hate that google has put LLM AI as their top search result for questions. It's just plain irresponsible. I *know* that ai isn't trustworthy, and it's *still* hard for me to disregard the answer, in black-and-white, that shows up at the top of the search and directly answers my question (vs other results that I often have to click through and do some sleuthing work before I can actually find the answer to my question). It's truly awful, and much, much more awful for all the people out there that don't know about the limitations of AI.

    • @BojanPeric-kq9et
      @BojanPeric-kq9et 21 день назад

      Craps like that almost make me to go to alternate search engines, including the one from currently very unpopular European country.

    • @BozesanVlad
      @BozesanVlad Час назад

      Google search was already unreliable, plus it was censored.
      At least now google search AI answer based on google search (bad) algorithm, based on random censorship rules.
      One day they'll fall because on them own doing.

  • @JesterThomas-jd6pt
    @JesterThomas-jd6pt 26 дней назад +39

    I gave a puzzle to the Microsoft Copilot AI. It was about finding a treasure on a small pirate island. I accidentally abbreviated one of the locations as 'u. w. rocks' and it immediately spelled it out correctly as "Underwater Rocks". But then it wasn't able to find the treasure by combining a shadow in the evening pointing eastwards with cardinal directions on a map and therefore began to assume the treasure must teleport constantly around the island...

    • @anastasiachristakos2480
      @anastasiachristakos2480 26 дней назад +1

      Sounds that Microsoft Copilot was designed to help science fiction Netflix script writers with writer's block...

    • @JesterThomas-jd6pt
      @JesterThomas-jd6pt 26 дней назад +3

      @@anastasiachristakos2480 Oh yes, it actually does sound like J. J. Abrams' script for the show Lost served as training data. But it's also LLMs in a nutshell: it's capable of correctly predicting what u. w. must mean in an island context, which is fascinating, but it would completely fail at "activating" a pencil in real life, haha.

  • @dactylntrochee
    @dactylntrochee 24 дня назад +3

    This topic is very close to my heart. It seems that we're so facile with language that, in most cases, people can get along by using the symbol (word) without remembering that the word is not the THING. At a simple level, like "please pass the salt", language works fine, but as soon as a topic gets complex, nuances of meaning get lost in the journey from speaker to listener and, if undetected, unexpected errors occur. This is not a problem when two people are speaking a language that is not their primary tongue -- the interlocutors KNOW there's going to be slippage. But when they're both speaking their native language, they don't realize how fragile our communication system is.
    I often picture the problem as using language to describe a painting or musical composition -- complete with its emotional content and historical context. Language just isn't the appropriate tool to get the experience from speaker to listener. You can also consider watching a movie as a stream of ones and zeroes and using your mind to compose what the movie actually is.
    Yet words are so deceptive in their apparent clarity that we mistake the picture our minds make for the thing itself. Of course, when you see an object, it's not like the object is going into your eye. We just see reflected light an "calculate/deduce" that the thing must be a tree. We don't see the tree; we only sense light. But language allows us to say "I see a tree", and we all jiggle the concept around in order to construct a picture.
    Failing to see that everything we learn, or even think, is actually just the product of emergence can cause some pretty strange results. Hence life as we know it.

    • @kilroy1964
      @kilroy1964 24 дня назад +2

      Yes!
      Hence the response to "please pass the salt" could be: "C+".
      That would indeed pass the salt.

    • @dactylntrochee
      @dactylntrochee 23 дня назад

      @@kilroy1964 Hmmm. Maybe I'd better use a different example in the future. 😁

    • @gbcb8853
      @gbcb8853 7 дней назад

      ‘Pass a motion’. Lots of slippage with that one.

  • @Astronomy_Live
    @Astronomy_Live 26 дней назад +15

    I develop and use small machine learning algorithms for my work in neuroscience and it's great, allowing me to investigate my datasets in an automated fashion, quantifying subtle behaviors across many hours of video that I would never have the time to do by hand. Narrow AI can be very useful in science, but validation is imperative and as you said LLMs are not representative of all AI.

    • @a64738
      @a64738 26 дней назад +2

      I would claim LLM`s is not really AI, it just mimics AI...

  • @julian281198
    @julian281198 25 дней назад +6

    i just asked chatgpt the following: "a farmer with a wolf , a goat and a brathering must cross a river by boat. The boat can carry only the farmer and a single item. How can they cross the river without anything being eaten"
    i used brathering because it looks like an english word, but isnt. Its german and means fried hering, so something the farmer might enjoy, but not the goat or the wolf.
    Chatgpt gave me the same answer as shown in the video including a mysterious cabbage.
    Asking question about the results reveal that chatgpt knew that brathering is a dish and goats dont eat it and the whole puzzel is pointless, with the return with one item answer. If asked again chatgpt will not speak about cabbages, but will still provide the return with one item answer

    • @ChristianIce
      @ChristianIce 25 дней назад +3

      I just did a similar test, I left the cabbage but I changed the sits on the boat, so that the farmer can carry *two* items.
      Of course the "AI" didn't get it and started going back and forth for no reason.

  • @xxiemeciel
    @xxiemeciel 26 дней назад +3

    I am a programmer and I know very well neural network, LLM are amazing IMO, and the specialists in the domain always knew the limitation of it. Personnally I mostly wonder if the coherence of the result we got 2-3 years ago when all this started was unreasonnably good and coherent considering the math behind it, and that might have been what created the hype about it.
    I can understand why it work but I am always surprise how well it actually work (even with the known limitation)

  • @jpdemer5
    @jpdemer5 26 дней назад +16

    LLMs would probably work much better if their creators removed Internet comment boards from the training materials.

    • @persimmontea6383
      @persimmontea6383 26 дней назад

      really? They include these boards in they training?

    • @corperk
      @corperk 26 дней назад

      @@persimmontea6383 Reddit in specific, although the undergrad essays used to train them are not better, they are written by the same people after all.

    • @MrDezokokotar
      @MrDezokokotar 25 дней назад

      @@persimmontea6383 Not all of them but some do, yes.

    • @LetalisLatrodectus
      @LetalisLatrodectus 22 дня назад

      Feel free to train one and try it!

  • @carlbrenninkmeijer8925
    @carlbrenninkmeijer8925 27 дней назад +78

    To be sure I recounted my fingers 5 times

    • @John-wd5cb
      @John-wd5cb 26 дней назад +11

      #meetoo all 6 of them 🖐️

    • @michaelharrison1093
      @michaelharrison1093 26 дней назад +2

      But if your definition of the number 5 is: - the number of fingers on a human hand, then all these hands have 5 fingers. If these LLMs have learnt this definition for the number 5 from the hallucinations of previous AI then it is no wonder the hands drawn by AI looks so weird.

    • @tombayliss5494
      @tombayliss5494 26 дней назад +9

      But here in the South we ain't got but 4 fingers then a thumb !

    • @John-wd5cb
      @John-wd5cb 26 дней назад

      @@michaelharrison1093 AI is obviously counting 21 fingers on the human body. For the males at least. We shouldn't feed it pron, it obviously gets confused 😋

    • @LuisAldamiz
      @LuisAldamiz 26 дней назад +2

      A shoopkeeper in my childhood's neighborhood had "Seisdedos" (six-fingers) as surname. The reason seems to be that very rarely some people are born with the wrong number of fingers and, in this case, it became a surname.

  • @charliemopps4926
    @charliemopps4926 26 дней назад +36

    The bubble is in the term "AI" It's kind of like when they labeled everything that came out as "2.0" like Web 2.0. AI has become a nonsense term that marketing departments are slapping on everything they have for sale. The vast majority of the products I see offered as "AI" these days (and yes, I work in the industry and see these products regularly) aren't even remotely anything I'd call "AI" they're just the same old manually created automation that they figured out some silly way they could just barely call "AI" The actual AI products out there are actually doing important new things. But the market is flooded with absolute junk pretending to be "AI" that just aren't.

    • @rantingrodent416
      @rantingrodent416 26 дней назад +4

      I don't think the accuracy of the terminology is relevant at all. The only thing that matters is that the technology was marketed successfully and associated with a buzzword. The buzzword could have been anything. "AI" was just the lowest hanging fruit. We have a very recent example where "cryptocurrency" and "blockchain" worked as buzzwords, so it clearly didn't have to be something reasonable and approachable to the mainstream.

    • @NorthgateLP
      @NorthgateLP 26 дней назад +2

      The Bubble is the same as with the internet. Investors are impatient and don't understand that a new technology needs time to mature, see that they could've gotten better returns elsewhere and sell. Which triggers a chain reaction and then at a point later when the technology is mature, it gets wide adoption.

    • @chychywoohoo
      @chychywoohoo 26 дней назад

      It's been that way for many years

    • @Rik77
      @Rik77 26 дней назад +3

      Yep a friend was talking to a company that was marketing an AI to credit scoring departments. He was interested and get thrmnin to demo. But on probing, he realised that it was just the same algorithms credit departments already use very well for decades. The marketing person was shocked as they hadn't realised that the product was just doing the same thing but worse, because the model was completely closed off from interrogation. He told them that no one in the industry would have a use for it. These companies are marketing things they do not understand due to poor misuse if terminology and the AI bandwagon.

    • @stefanb6539
      @stefanb6539 26 дней назад +7

      The term "AI" has been overhyped since its earliest conception. I remember that in the 1980s any algorithm based decision tool was confidently labeled such. I think the best solution is ti learn to interpret the I not as Intelleigence but as Id .iot. We can't make Artificial Intelligences, and probably never will, but when someone monitors them right Arificial Id .iots can actually do a lot of work for you.

  • @zomgneedaname
    @zomgneedaname 21 день назад +2

    Thank you sabine for speaking the truth and popping the bubbles of clueless corporate executives around the world.

  • @acf2802
    @acf2802 11 дней назад +1

    "Hallucinations" are what is required for all creative applications. The problem is you can't decide if you want a DaVinci or a Nostradamus.

  • @Rondoggy67
    @Rondoggy67 26 дней назад +17

    LLMs are essentially a computerised version of the game where people take turns adding words to a sentence.
    The best they can do is come up with plausible content that is based on their training set.
    There is nothing equivalent to intelligence behind it.
    The problem is that for many people, Turing’s imitation game is the bar for detecting intelligence.
    In reality, the imitation game just illustrates how easily people can be tricked.

    • @eadweard.
      @eadweard. 26 дней назад +3

      "The imitation game just illustrates how easily people can be tricked"
      People only started saying that once AI passed the test.

    • @JoJofghj
      @JoJofghj 24 дня назад +3

      @@eadweard. A hypothesis was tested, proven wrong, and has been abandoned. That's how science works.

    • @JoJofghj
      @JoJofghj 24 дня назад +2

      In addition to showing how easy it is to trick people, another thing I've learned from LLMs is that despite the fact that people's use of language seems so creative (if not haphazard), it's actually pretty predictable most of the time.

    • @eadweard.
      @eadweard. 24 дня назад +2

      @@JoJofghj What hypothesis has been proven wrong, specifically?

    • @Martin_Gregory
      @Martin_Gregory 24 дня назад +4

      It's exactly the same as intelligence. You come up with your plausible thoughts in the same way: based on your training set.

  • @MrSupermeisi
    @MrSupermeisi 25 дней назад +6

    The big question for me is what is the ultimate goal of AI. What do we want to achieve at the end? If we model our brain 1:1, then we can also directly use our brain, isn't it?

    • @Pneumanon
      @Pneumanon 24 дня назад

      No, they want a slave with the power of a god.

    • @BlueGiant69202
      @BlueGiant69202 21 день назад

      What? It's the answer to the ultimate question.
      why don't jets have flapping wings?
      Why are there no metallic water faucets on the shores of rivers or trees with steel girders?

  • @oatlegOnYt
    @oatlegOnYt 26 дней назад +5

    I think Sabine is wrong here. Not about that the model require logic. It's about language is the result of abstraction, and abstraction is a very valid tool for reasoning. A non-precise one, but a fast one. Pretty good for shortcuts and 'good enough' results in certain areas.
    Of course, it's not a silver bullet. That's the reason why human brain does different kinds of evaluation and future AI algorithms will need to do the same.
    But I think Sabine is dropping the value of LLM too much. LLM is working in the area of pure relationship and abstraction area and that has value. It's just it doesn't work as well in every area.
    The human brain uses LLM type of thinking a lot, and that's the reason why so many people fail for fallacies which are 'good enough' from relationship thinking but a failure from logical thinking.
    It's a thinking fast vs thinking right strategies.
    I guess the future will be more about 'let's the abstraction layer find paths to resolve and the logical layer drops the wrong answers/paths of resolution'

    • @bezimeni2000
      @bezimeni2000 26 дней назад

      Nah

    • @zx8346
      @zx8346 23 дня назад +2

      she's a physicist, not an expert on AI

    • @Yottenburgen
      @Yottenburgen 20 дней назад

      LLM haven't reached anywhere near their pinnacle either, many focus on the model's worldly intelligence instead of its 'awareness' of itself (or rather trained methods). A common problem with LLM is multiplication because they usually end up attempting to guess the answer first before solving (which they then attempt to justify), but LLM are actually capable of working out the solution by breaking the problem down. An area of improvement I want to see is LLM knowing how to best utilize their context window, to know the optimal method to perform whatever it is that its doing. It doesn't help that some of our written methods are actually visually aided like long division seemingly putting the answer before any other character, but I think we can get there.
      Here's an example I like to share, simply ask whatever GPT you want "What is the product of 45694 and 9866", it will straight up attempt a guess and answer incorrectly. But when it works out the question in a way that is conducive to its linear tokenization where one token comes after the other, it can very solidly get the answer. I wouldn't expect a math genius to guess the answer correctly, I'd expect them to work it out very quickly in whatever way they feel is best.
      chatgpt.com/share/b4ed2219-6699-42e4-bb90-af0f88cd9faf
      LLM also have very expansive uses beyond themselves, there's already papers of LLM being used to create conditions to train other AI, not quite autonomous but very useful at creating these parameters.
      It annoys me how quick people are to put down LLM, there's so much room for improvement and they are already this good? Let alone whatever the hell OpenAI is doing by broadening GPT4's horizons into GPT4o, granted it's not an increase in intelligence but it is actually has some absolutely insane capabilities that they aren't letting us have access to yet for very clear reasons. Like it can copy a user's voice back at them or it can copy a user's face and turn it into a caricature or photoshop with it.

    • @serqetry
      @serqetry 4 дня назад

      Yep, exactly. Sounds like her research comes from a lot of anti-AI people who don't understand it either.

  • @Agnemons
    @Agnemons 25 дней назад +6

    The status of AI currently?
    We have reached the level of Artificial Idiot.

  • @qwe14205
    @qwe14205 22 дня назад +4

    The wolf, goat and cabbage is a best and simplest example of how LLM works. Imagine trusting it to write somewhat fomplex code 😅.

    • @BojanPeric-kq9et
      @BojanPeric-kq9et 21 день назад +2

      No need to imagine, there is no uncertainty, it gets it wrong. The only time it gets output right is when it copy-pastes human written code.

    • @user-ly4ri5vr8w
      @user-ly4ri5vr8w 20 дней назад +2

      @@BojanPeric-kq9et AND it happens to be the code that actually does what you asked.

    • @BojanPeric-kq9et
      @BojanPeric-kq9et 12 дней назад

      @@user-ly4ri5vr8w glorified search engine.

  • @haraldmilz8533
    @haraldmilz8533 26 дней назад +8

    The issue with LLMs is that they are based on neural networks which many people don't understand how they work. A NN will always give you a probability or probability vector with components < 1 (except in very few trivial cases), without telling you the overall probability for a given answer. If it did, you could easily spot hallucinations, e.g. if the overall prob were less than, say, 80%.
    That said, I'm with Sabine that LLMs are nice toys but not the future of AI.

    • @Bolidoo
      @Bolidoo 26 дней назад

      But couldn’t that be solved via better training data & techniques? Where for example it’s given batches s.t. after giving a wrong information the AI instead of emiting EOS it corrects its mistake. Training it to reculate the moment it realizes it made a mistake, instead of committing on it. Furthermore RL is doable on LLMs reinforcing rationales that reached the correct answer and penalising those that did not. It seems to me that LLMs can in principle reason and not hallucinate if rightly trained.

    • @fnorgen
      @fnorgen 25 дней назад +2

      ​@@Bolidoo There are ways to get LLMs to catch their own mistakes to some degree.
      One method is to run inference in multiple batches from the same query, each with a different RNG seed and a high randomness temperature. Basically to create multiple different answers to the same question. These can then be compared to see if they are in general consensus. If the network is confused it will presumably output a set of blatantly different responses. However this makes inference awfully expensive, and still won't guard against the network simply having learned something wrong.
      A different option might be to include a bunch of examples in the training set of the author catching their own mistakes on the fly, but this would be very unnatural data that would have to be procured in huge volumes. Hardly something that could simply be scraped off the internet. The same applies to a lot of seemingly clever training techniques that guide the network on specifically how to think, rather than just training them on what to directly output to the user. Someone will have to produce A LOT of oddly specific, high quality training data, and existing LLMs can't be trusted to synthesize it. A huge amount would have to be hand written and/or generated with clever scripts, and that gets very time consuming and expensive.
      Also, a big problem is that, given the option, these networks LOVE to weasel out of difficult queries. "Sorry, I don't know." is too valid an answer to most questions. It's a classic failure mode for LLMs where they would reject strange prompts to instead write about something much more generic in a manner that was structurally correct, but rendered the response utterly useless.
      Another method I've heard of is to have a separate verifier network, to check generator outputs for logical consistency. The appeal of this method is that it forces the generator network to write in a manner that is easy to verify, in order to get the verifier past some confidence threshold. Also, the verifier network can be fairly small and cheap to run. Though presumably it wouldn't help much against factual errors, and this technique doesn't seem to work well with genuinely complicated problems that will always be hard to verify. It just makes it harder for the generator to mask its stupidity with overly complicated language.
      What is clear though is that we've pretty much maxed out what can be accomplished with scale alone. There just isn't enough high quality data to go around.

    • @Bolidoo
      @Bolidoo 24 дня назад

      @@fnorgen Very interesting thanks. I don’t agree with your last point though. It seems to me that even if we can’t make bigger models due to lack of data, we can keep augmenting them with lots of compute and search techniques. Maybe with enough compute we can generate the high quality data we require to move forward. I’m not saying scale is all you need for certain, but at the same time i’m not convinced it’s not either.

    • @GregHib
      @GregHib 24 дня назад

      @@Bolidoo No amount or quality of data given to LLMs, as they currently stand, can solve the hallucination problem. Better data will improve their accuracy and ability to an extent but it has diminishing returns as the is a near infinite data possibility space but a very non-infinite cost to train.
      What you end up with is a game of data wack-a-mole. Where people find problems, and the solutions are added to the data set for training the next model.This has worked for the last few years but it can't continue for much longer as the more accurate models get, the larger the error space becomes.
      The problem is more obvious in the autonomous driving space than LLMs but it's essentially the same problem; statistical models alone can't handle scenario's that aren't predictable or lay far outside of the given dataset.

    • @Bolidoo
      @Bolidoo 24 дня назад

      @@GregHib Transformers can and do generalize. Gpt4o solves the AIW problem 60% of the time, that is made up new logic problems that couldn’t have been in the training data. Alphaproof can solve math olympiad level maths problems. Yes it could do this thanks to self play, MCTS and the fact Lean is automatically verifiable. But ultimately a transformer did the heavy lifting. A transformer could reason.
      There’s also a paper that showed that LLMs know when they lie. This suggests that part of the reason an AI may insist on wrong information even though it has already realized it’s wrong (this is one type of hallucination) is because the training data had a high degree of self consistency (it’s uncommon for a text to change opinion in the middle). Hence finetuning on relatively few examples can help correct this behavior and reduce hallucinations across the board. That’s what we find empirically with models like Sonnet 3.5 that outright hallucinate less, while it was of course not necessary to tell it to not hallucinate in each and every case that it could and didn’t.
      My point is that transformers, given the right data and compute, are enough.

  • @nickmcconnell1291
    @nickmcconnell1291 26 дней назад +7

    Don't worry about NVIDIA stock. Because they need to retrain using logic, they will need even more NVIDIA chips. NVIDIA earnings will be ludicrous and the stock will again go up.

    • @LuisAldamiz
      @LuisAldamiz 26 дней назад +2

      NVidia fell because China has got ahead on a key advance related to AI: the memristor. Look it up.

    • @nickmcconnell1291
      @nickmcconnell1291 26 дней назад

      @@LuisAldamiz TY...I will do that

    • @lagrangewei
      @lagrangewei 25 дней назад +1

      @@LuisAldamiz China might be ahead on memristor, however memristor will not see market for a decade. nvidia is failing because China just isn't buying their chip anymore because companies don't trust the US government will not sanction them. so they rather buy chinese AI chip that is only 20% slower just for the certainty that they will always be able to get replacement chip when they need them.

    • @LuisAldamiz
      @LuisAldamiz 24 дня назад +1

      @@lagrangewei - Maybe and also because the AI bubble is bursting and NVidia has been riding that wave (GPUs are key to AI training and crypto minin).

  • @MatsueMusic
    @MatsueMusic 25 дней назад +3

    As a musician this really “rings” home. For the longest time “simulating” a piano meant recording every note from a certain piano and then having some software apply a bit of trickery to make it sound “natural” when played, cheaper ones would sample every 5 notes and filly in the rest by pitch bending (stretching or shrinking) the note samples. These days, some virtual instruments simulate the actual piano, how the hammer hits the strings, where the mic would be placed, how hard the string is hit, same goes for wind, string, and brass instruments. And like LLM, the sample version could only play back a note, the simulated instruments can capture a light touch, breathing softly, all the nuances that give music “soul”.
    It seems the best way to “simulate” something isn’t to just stack a bunch of its products (bunch of words, math, audio, video, etc) but to figure out how the underlying system works and simulate it, the ear, the eyes, the string, the piano, the body.
    An interesting approach I saw Nvidia try was in ragdoll physics and robotics. Where they give a virtual robot a scaffold and rig (limbs, joints, etc) and then let it learn how to use them over and over again. Another example was an AI learning to play Mario. I feel that technique applied to other modes could work, or could be what Sabine was referring to.

    • @tempname8263
      @tempname8263 24 дня назад

      Simulation vs emulation

    • @MatsueMusic
      @MatsueMusic 23 дня назад

      @@tempname8263 ah thank you !

    • @avsystem3142
      @avsystem3142 23 дня назад

      It's called "physical modeling". There have been algorithms around for decades that can simulate the sound of a plucked string, for example. Physical modeling of a piano, e.g., Pianoteq, is much more complex but is now so perfected that manufacturers such a Steinway are licensing the use of their name to the physical models of their pianos.

    • @MatsueMusic
      @MatsueMusic 23 дня назад

      @@avsystem3142 Pianoteq’s tech is amazing, I didn’t want to talk specifics to avoid noising the point. A point worth raising is that physical modelling has a very small file size since you don’t need any “references” to the real system, nor need small countries worth of energy to “train”. I do think this approach to AGI would go a long way. If I could wave a wand I would follow Pianoteqs approach, first conduct studies to understand the factors and parameters that govern a system, than design models that can approximate those systems.
      My background is in Psychology and data engineering and I feel like we tend to overestimate how complex the brain is in applications that model it. Like pianoteq, if we avoid attacking the problem at its most atomic…..(ie neurons) but approximate the systems which they produce like the lobes (frontal, occipital, temporal, etc) could t we approximate the functions of the brain and set loose algorithm with a target? The lobe interactions and their systems are complex but their functions are well known and we don’t need model them in the same way. I am sure greater minds have thought about this problem and tried to design models for them but it leaves me wondering.
      As an aside, I am lucky enough to own an Osmose and tinker with the awful Egan Matrix, UI aside, the models it has are impressive and what it can do does seem like the future of synthesizer hardware. At least Arturia thinks so with their latest device.

  • @n.davidmiller2029
    @n.davidmiller2029 20 дней назад +2

    But Sabine, the power brokers definitely would not want an AI trained on logic and math as it would of necessity need to tell the truth, whereas a language model can conveniently lie all the time.

  • @lostmsu
    @lostmsu 26 дней назад +2

    I don't think it is an issue that LLMs give the right solution to the river riddle if you omit that wolf eats goat, and goat eats cabbage - because these make sense on their on and you need to put extra effort to explain that what you are asking does not include these obvious options.
    Much bigger issue though is that if you remove the goat from the puzzle entirely, LLMs will produce the solution as if it was still present.

  • @Dhrazor
    @Dhrazor 26 дней назад +9

    There are several issues with AI:
    1. Only in Weather forecast and Image recognition used to detects illnesses is currently somewhat profitable and actually usable(after human reviews)
    2. Corporation pushing Ai massively overestimate the need and use cases for Ai even when considering the next 5-10 years of Ai use
    3. Ai will remain extremely unreliable for the foreseeable time since there is no mechanism to stop it from hallucinations other than hard coding out or erasing out potential answers/results. Meaning if you circumvent the hard coding the Ai will still fail these. Also as seen above hard coding certain results means that some valid responses will instead will forced to have wrong results.
    4. Companies are censoring and cherry picking information based on ideology, turning these LLMs into propaganda machines often spewing half truths or outright propaganda pieces
    5. Even when trying to pick data without being partial for the training, some information will be labeled wrong, weights will have wrong values, data will be outright erroneous resulting in wrong answers...
    Long story short, don't expect too much from Ai before 1. It becomes 3+ Ai coming up with a collective answer 2. based on multiple sources 3. that are actually cross-examined data from multiple fields and from multiple researchers, meaning we are 10-15+ years from a half reliable ai...
    Expect ai stocks to fall for the next decade or so... but it might take 3-6+ moths for investors to get this memo...

    • @waynerussell6401
      @waynerussell6401 26 дней назад

      Tesla and Palantir are dong very well with commercial applications of AI.

    • @a64738
      @a64738 26 дней назад +2

      Yes and the strange thing is that even people working on the LLM calls them AI and genuinely do not understand it is not real AI while totally overestimating what it can do...

    • @Dhrazor
      @Dhrazor 25 дней назад

      @@waynerussell6401 Tesla has hundreds of crashes and dozens of deaths linked to full self driving. Both Tesla and Palantir sell a service, and companies/people are paying to try it... saying they are doing well is not much. Snake oil salesmen usually do well when they start selling... The question is: Are their customers more competitive when using Ai? How big and how common are Ai fuckups? Also... Palantir's AiP is barely ai, it's mostly big data without relying on the "black box" aspect of real LLMs to make choices directly, and instead merely informing people to make more educated choices.

  • @vaclavkrpec2879
    @vaclavkrpec2879 26 дней назад +32

    The root cause of this is that LLMs (and GenAI more generally) are still only stochastic models. They don't have any _understanding_ of what they "say" or "talk about". They just generate data based on conditional probability inferred from the training samples, they don't know what they're doing. That's why generated images of technical contraptions may look OK at first glance but when you start to think how they would work, it stops making any sense. And why the fingers are all mangled, yes (that in particular is because fingers have so much variability that the generator infers that they may do "anything")---again, the model doesn't have any idea how they work and what they do. No idea, literally.

    • @svensvensson8102
      @svensvensson8102 26 дней назад +3

      Correct. There's no evidence that these architectures are capable of intelligent thought, regardless of the amount of data you feed them. They will remain great assistants, which is a valuable thing in itself, but a far cry from the church of AI predictions of AGI in N months, where N is a number between 1 and infinity and always changing.

    • @SaanMigwell
      @SaanMigwell 23 дня назад

      I mean to be frank, people have been calling the "computer" ai since at least pac-man. Them asses adopt words and definitions and ignored the programmers, ever since ENIAC.

    • @vaclavkrpec2879
      @vaclavkrpec2879 23 дня назад

      @@SaanMigwell Well, they kinda can, because there isn’t any constructive definition of intelligence, natural or artificial. So you can stick the label willy-nilly on almost anything, if you’re bold enough…

    • @TheReferrer72
      @TheReferrer72 21 день назад

      And please explain how the human brain understands? I thought we produce stochastic models of the world.

    • @svensvensson8102
      @svensvensson8102 21 день назад

      @@TheReferrer72 Thinking that thought and intelligence is based on stochastic models doesn't help much, does it? The point is that we don't know.

  • @templeofgame3047
    @templeofgame3047 26 дней назад +43

    6:34 The mind *may be a single thing, but not the brain. It is composed of a multitude of constituent components working in congress. Why would you think the answer to artificial intelligence would be one model or neural net, instead of a multitude of less effective but highly specialized low-power AIs all working together to make the whole more than the sum of it's parts?

    • @rahawa774
      @rahawa774 26 дней назад +2

      Hmm, good point.

    • @juandavidavendanorojas6263
      @juandavidavendanorojas6263 26 дней назад +8

      Search Mixture of Experts (MoE)

    • @GodbornNoven
      @GodbornNoven 26 дней назад +2

      a 100 hundred fools will never be equal to a single genius. Not even a billion fools. The idea is the fools take on the simple tasks and the genius get the difficult ones. Dense models are more intelligent and better for complexity. MOE is usually better for a lot of other reasons.

    • @asdfqwerty14587
      @asdfqwerty14587 26 дней назад

      While there might be some level of truth to that in an abstract sense.. a LLM would not make any sense as one of the components, because the only thing that could interact meaningfully with a LLM would be something that already knows how to interpret and generate plain text (and if it could already do that then there would be no reason for it to need a LLM).
      I think that it's likely that a better AI would have multiple components, but I think it would mostly be using multiple components so that it can do things like having one part of the AI that learns while it's in use and a different part that does slower learning that mostly runs in the background and tries to optimize the things that the other processes come up with - I don't think it would make any sense for the core logical thinking of the AI to be split up into multiple components though because that sounds grossly inefficient.

    • @templeofgame3047
      @templeofgame3047 26 дней назад

      ​@@juandavidavendanorojas6263 Thanks. I am familiar. That is in part what inspired my OP. Recently MoEs reached ~1M experts. Imagine what that number might be in just 1 or 2 years. Its fascinating.

  • @darknone5566
    @darknone5566 26 дней назад +2

    First off, thank you for your videos, I really enjoy them. I will push back on the neuro-symbolic a little bit. I work in AI, and was really excited about neuro-symbolic starting a couple years ago, but found that there are various issues with it that may be difficult to overcome. The most obvious is lack of funding. Gary Marcus has already suggested that if we funded neuro-symbolic research like we funded LLM research, we would probably be so much farther along with it, but we don't. The next is that finding the right mesh of neural and symbolic seems to be very troublesome. Not that the AI community has converged on what neuro-symbolic AI really is, but symbolic methods they usually try to use with DNNs typically don't scale well, which is what the DNNs actually can do well. When you add the symbolic thing to the DNN, or vice versa, you often get stuck at the scale the symbolic problem can solve. So I feel like I still see neuro-symbolic methods trying to tackle toy problems that it kind of seems like LLMs or other foundation models can "just solve" a significant amount of time. No guarantees it is correct, of course, but people find others ways to deal with that through things like ensembles and answer checking. Not perfect, but sometimes good enough.
    Lastly, and feel free to push back, because this point might be more philosophical, I feel like nature, people, biological systems, etc. do not actually function logically or mathematically. We are kind of a weird mess with enough checks and redundancy to generally figure things out without actually using good logic, statistics, or mathematics. All of those are human inventions that tend to hit limits on what they can do pretty quickly, while humans can guestimate their way to solutions of all kinds given the tools they have. Perhaps like in the rat example you showed though, I am intrigued by the idea of us basically simulating biological processes well enough to create human-like AI, that feels possible to me, but we need to solve the embodiment problem for that too.
    All that said, I'm still interested in neuro-symbolic methods and am keeping my eye on them. Maybe someday it will get more hype (and associated funding :)) and I'll be able to work on it more!

  • @tchaika222
    @tchaika222 8 часов назад

    To quote from the bobble head on your desk: "Any fool can know. The point is to understand."

  • @dvv18
    @dvv18 26 дней назад +6

    Diversifying your portfolio is the key. Invest both in those who work towards the annihilation of humankind to be long and painful (Google et al.) or fast and (relatively) painless (Raytheon et al.) - and while you won't become super rich, you'll still have a comfortable retirement. And hopefully won't see the final stages of said annihilation happen before you expire. And, of course, Nvidia swings both ways, so I still hope my Nvidia shares will recover, too 😎

  • @innuendo70
    @innuendo70 27 дней назад +6

    I think the most important bits of math AI needs to at least adhere to is the math of modern causal inference. But even then, the "assumptions are in the arrows not drawn" thingy is tricky enough for humans, so it's unlikely that AI could get that bit less wrong than humans do, especially because it's fundamentally trained on human errors. It would be interesting to speculate about what would happen if computing power ever turns out strong enough for AI to reduce the amount of causal assumptions in causal graphs to a point way beyond the point impossible for the human brain. Could it find "our" scientific causal blunders, maybe decades or even centuries old ones?

  • @godsofentropy
    @godsofentropy 27 дней назад +5

    I think that two issues should be separated: one is the usefulness of AI and the other is its profitability. There is no doubt that AI can be (and is) useful on many levels. In my work, I use AI to create graphic elements, scripts to improve work, texts, descriptions and even to record speech for video tutorials. Yes, it helps me earning money to some extent. Don't know for how long though - maybe my work will become obsolete sooner than I think.
    However, the question of the profitability of AI at the corporate level is debatable - because the largest AI models are offered below the profitability threshold. This is a war of attrition and sooner or later the big guys will have to face exceeding the level of acceptable losses. It is not better for small companies, because we already have several thousand AI startups and their prices only make sense if they are used on a mass scale. But how many services do we need to improve text or generate images? And with such enormous competition, how many of them will be left in a year or two? However, the AI ​​revolution has already started and is already moving through societies and the capitalist system like an icebreaker. For now, we hear single cracks, but it is coming. And humans and societies are not prepared for that.

  • @FutureAIDev2015
    @FutureAIDev2015 26 дней назад +1

    I think a good way to describe this issue is that LLMs are trying to generate output that looks like the training data, but they have no way to check whether the output is technically accurate and correct.

  • @DanielSeacrest
    @DanielSeacrest 26 дней назад +2

    I think a literature review may be in order (sorry if that's a bit harsh). A lot of your premise involving how they are trained is wrong (also Gary Marcus is not a computer scientist).
    Now this is more on the world modelling aspect from LLMs, but a cool paper suggestion I have would be "Mapping the Mind of a Large Language Model", a mechanistic interoperability paper from Anthropic.

  • @Deltarious
    @Deltarious 26 дней назад +30

    Potential bad news about your Google stocks: The US has found that it meets conditions to potentially require being broken up into smaller companies for monopoly/antitrust reasons. This may well affect the stock price, if nothing else

    • @mattmaas5790
      @mattmaas5790 26 дней назад +3

      Probably would be left with stocks in the valuable pieces, not the end of the world. Too OP had to be broken up

    • @ASDeckard
      @ASDeckard 26 дней назад +1

      This is sure to negatively affect the trout population.

    • @stevecarter8810
      @stevecarter8810 26 дней назад +4

      Fifteen years late

    • @RemyVonLion
      @RemyVonLion 26 дней назад +3

      gubmit trying to shutdown a monopoly in this day and age? Bullish.

    • @dustinwatkins7843
      @dustinwatkins7843 26 дней назад +1

      Won't happen. Bet.

  • @barvin9271
    @barvin9271 26 дней назад +35

    hilariously enough, running my own offline LLM for giggles, I've noticed that a lot of hallucinations come down to 3 things, 1) whether the instruction set, i.e. the foundation for which the model utilizes its data set to respond is configured with considerations for brevity or tonality. 2) the amount of input that is provided, as the models tend to be set up to respond in a way that will satisfy any outcome, so if you just say hi!, in most cases it will respond with a paragraph. Lastly 3) the amount of vram and context allocated to the model - it seems that while lower vram lower context models will answer quicker, it's a cost benefit ratio, because it will fill the missing output space with whatever it wants, and in the case of models that are being utilized with a log to reference for said context, it will decrease in performance exponentially until it eventually breaks under the weight of every "thought", resulting in gibberish, hallucinations, and schizophrenia or the ai equivalent.

    • @undozan4180
      @undozan4180 26 дней назад

      @barvin9271 what does schizophrenia look like in AI?

    • @uNiels_Heart
      @uNiels_Heart 26 дней назад +1

      @@undozan4180 I suspect they meant DID. Schizophrenia doesn't really make sense as analogy for any observations of outputs of LLMs you could have, as far as I'm concerned.

    • @termitreter6545
      @termitreter6545 26 дней назад +11

      I think the inherent problem is more that LLMs just dont "understand" language, at its core they just memorize and 'network' word salad from the internet.
      Thats why "hallicinating" is kinda misleading, it implies there is a state of "non-hallucination". Yet in reality there is no right or wrong for an LLM.

    • @picketf
      @picketf 26 дней назад +1

      I use LLMs to find correlation in abstract concepts and financial data. Even after providing the data or pointing to where it can be retrieved I still have problems, from ignoring said data and just making numbers up on the spot which "could sound" about right to just not following rules, parameters and specifications.
      Lately I realised that there is an increasing amount of times were the result is: bla bla bla this is were you can find out about it {link} even after I had asked the webtool to retrieve the information and upon doing so again it reluctantly does so😂😂😂

    • @picketf
      @picketf 26 дней назад +4

      and the amount of times I had to scroll up to copy-paste, repeat or remind GPT about a rule or requirement that I had previously elaborated on. It borders on having conversations with a cognitive impaired person.

  • @DinoDiniProductions
    @DinoDiniProductions 26 дней назад +9

    Oh no... the world is not made of mathematics either. Mathematics is a formal language that is used to describe and predict the behaviour of the reality we perceive. The only difference from informal language is that Mathematics does not generally have ambiguities (although it actually does, but that's another story). We could say that mathematics is a language with as little ambiguity as possible. But the universe does not give a damn about our mathematics.
    It is much more robust to say the most effective models we have for reality rely on mathematics.
    It may sound nit-picky, but preciseness of language is central to the discussion here...

    • @RandallReviere
      @RandallReviere 26 дней назад +1

      Quite right. The fitness of the world to be described by mathematics is only to the degree that 'countable things persist'. In the invented worlds described only by language (and the reader's imagination), no one is doing any accounting whatsoever. So the main use of math becomes the kinds of things LLMs are already doing, ie. dot products on word vectors to assess adjacency, etc.

    • @kilroy1964
      @kilroy1964 24 дня назад

      Well, yes... But that's what Sabine meant.

    • @DinoDiniProductions
      @DinoDiniProductions 24 дня назад +1

      @@kilroy1964The road to hell is paved with well meaning.

  • @ChristianIce
    @ChristianIce 25 дней назад +2

    Every LLM output is an hallucination.
    If the output is correct or not depends only by the statistical relations between tokens.
    P.S. So much for the "sparkles of AGI in LLMs" :D :D :D

  • @bonglesnodkins329
    @bonglesnodkins329 8 дней назад

    When I was working as an AI researcher in the 1990s and 2000s, a saying that was used a lot was that certain approaches to AI felt like "trying to climb a ladder to reach the moon". You feel like you're making rapid progress, but in fact the attempt is doomed never to succeed, because it's the wrong approach. (I think that's actually a corruption of a Hubert Dreyfus line from the 60s, where he used tall trees instead of ladders, but the ladder version is snappier.)

  • @MrTheoJ
    @MrTheoJ 27 дней назад +5

    to me it seems that LLM's are big average machines, while most things in life have rules. If you want to create a Human, there are rules for the length of the legs compared to the rest of the body, there are 5 fingers, the legs can't be twice as long as the arms. Once you have a skeleton everything else has rules, where the muscles go, the fat, the organs and skin. Maybe LLM's represent the "fast" brain where you make quick judgements and when you get it wrong you actually have to think about it which requires some other type of brain.
    children learn different, you only have to show it a couple of cat pictures before they know what a cat is, you have to show an LLM a billion cat pictures and it will still get it wrong too often.
    so Yes we will create an AI as smart as Humans, it's just not going to be LLM's
    and it's kind of weird that it can not do Maths, I know why, it just to me looks like there should be an IF statement somewhere "IF input.is_maths() USE calculator END IF"

  • @andrewsuryali8540
    @andrewsuryali8540 27 дней назад +10

    I love eating Berliners, btw.

  • @templeofnoise3075
    @templeofnoise3075 25 дней назад +23

    "You can't just train LLMs on more and more text and images and hope it will begin to understand what's going on. No one expected that".
    Actually, you did :)
    Happy you are not on that train anymore

    • @RawrxDev
      @RawrxDev 24 дня назад +6

      Yeah I was surprised to see this turn around considering some of her videos from a few months ago implied AGI was around the corner with just a little more scaling lol

    • @andy-the-gardener
      @andy-the-gardener 23 дня назад

      shes scared of global warming now too. apparently something has become more scary recently about what the climate experts and ecologists said in 1990.

    • @mikemondano3624
      @mikemondano3624 23 дня назад

      Imagine a naïve, _tabula rasa_ human being trained by reading everything on the internet. By some estimates, 80% of the internet is false information. The contradictions alone would lead to the looney bin. And AI feeds its hallucinations back to the internet to be scraped by other AI's.

    • @erinm9445
      @erinm9445 23 дня назад +4

      @@RawrxDev She also tried to argue that AI "understands" what it's generating--partly by torturing the definition of "understand", but still. Ridiculous.

  • @tomh9553
    @tomh9553 26 дней назад +1

    I had a similar example to the riddle you mention, chatgpt at least gets it correct now and I wouldn’t be surprised if the reason why it gets it right now is that my conversations about this riddle became part of the training set for newer models.

  • @andreweppink4498
    @andreweppink4498 26 дней назад +2

    (0:11) WeIl Sweetie, better start believing!

  • @davefaulkner6302
    @davefaulkner6302 26 дней назад +25

    I've read many comments that say "The world is not words but it isn't math either". True, but this is not the question. The problem to solve is how to model the world. Sabine is saying that Word Models are inherently ambiguous and so less accurate than Math Models. Interestingly recursive, a Model is a mathematical construct that resembles a real-world situation, extracting the situations' measureable properties and tying them together in a mathematical system. Sabine advocates the use of Mathematical modeling of world-physical systems rather than using Word Models to achieve AGI. In my opinion this is entirely correct and the current path towards ginormous LLMs is a waste of resources.

    • @boobah5643
      @boobah5643 25 дней назад

      Ah, but just wait. We'll get maths-driven models that give physically impossible answers.

    • @srb00
      @srb00 25 дней назад +2

      Only someone as put of out of touch with reality as Sabine is would ignore the huge improvements in LLM/AI in just the last two years.

    • @jmspiers
      @jmspiers 24 дня назад +1

      What Sabine fails to realize is that the approach has been tried. If it was easy, it would have been done. Turns out that language is easier than math, though, which is why it's happening faster.
      OpenAI has reportedly made great progress with math. GPT4o in particular can actually do a lot of math now. However, as recently as a month or two ago it was still compounding interest incorrectly.
      That said, it's a mistake to think of ChatGPT (at least ChatGPT 4o) as being "just" an LLM. It's a lot more than that, now. Exactly how much more is unknown since it's closed source, but its ability to understand the world around it through video, to understand tone of voice, etc, goes well beyond an LLM.

    • @Just.One.More.Thing.
      @Just.One.More.Thing. 24 дня назад

      Too harsh. AI will need great LLMs, Logic way faster than now, and a way to sense the linguistically described object in whatever way it is linguistically described. That problem hasn't been addressed yet, but will have to be before the usual math can be applied to interpret language samples.

    • @mikemondano3624
      @mikemondano3624 23 дня назад +1

      Mathematical modelling of even fairly simple systems is beyond both humans and mathematics. We have reached the limits of possibility already in many everyday systems.

  • @birdistheword4521
    @birdistheword4521 26 дней назад +7

    We made the language center of the brain, and now we’re trying to implement the logic. It makes sense we need more than just a language center to make a truly smart intelligence.

  • @HHercock
    @HHercock 27 дней назад +14

    It is terrific to see commentary about AI being a cognitive project rather than a linear language project. AI will come of age when it can assess situations concurrently across multiple knowledge domains and produce outputs that are logically coherent across those domains

    • @SabineHossenfelder
      @SabineHossenfelder  26 дней назад +8

      Imagine how much progress that will unlock!

    • @andrasbiro3007
      @andrasbiro3007 26 дней назад

      Um, that's what they are doing right now. LLMs are trained on all knowledge domains, so you just have to ask the right question.

    • @billbond2682
      @billbond2682 26 дней назад +3

      ​@@andrasbiro3007 please watch the video again and read your reply

    • @mattmaas5790
      @mattmaas5790 26 дней назад +2

      ​@@billbond2682 actually you guys should videos from Ilya Suskever genius behind openai who left. He says our language contains a model of the world and its inhabitants that translates into a real model existing outside of language. And that was discussed before multimodal models added images sounds and video added new data to the models.

    • @reasonerenlightened2456
      @reasonerenlightened2456 25 дней назад

      ​@@SabineHossenfelder .... but you did not even mention Douglas Lenat , who has been explaining for decades that AI must use Logic to understand Structure, Function.and Meaning.

  • @bonglesnodkins329
    @bonglesnodkins329 8 дней назад

    The issue is that LLMs don’t have actual knowledge representation. They are effectively a huge vector-based, rolling context lookup table.

  • @tedzards509
    @tedzards509 21 день назад

    Asking different models rather specific questions gives very similar answers. To me, this indicated that LLMs are certainly converging.

  • @GodbornNoven
    @GodbornNoven 26 дней назад +8

    This puzzle is a classic! The key is to avoid leaving the deer with the wolf (which would result in the deer being eaten) or the deer with the cabbage (which would be eaten by the deer). Here’s the solution:
    1. **Take the deer across the bridge.** (Now the deer is on the other side, and the wolf and cabbage are on the original side.)
    2. **Go back alone to the original side.**
    3. **Take the wolf across the bridge.** (Now the wolf and deer are on the other side, and the cabbage is on the original side.)
    4. **Bring the deer back to the original side.** (Now the wolf is alone on the other side, and the deer and cabbage are on the original side.)
    5. **Take the cabbage across the bridge.** (Now the wolf and cabbage are on the other side, and the deer is on the original side.)
    6. **Go back alone to the original side.**
    7. **Finally, take the deer across the bridge again.**
    Now, all three-the wolf, the deer, and the cabbage-are safely across the bridge in 7 steps.
    Ay they're getting smarter

    • @LuisAldamiz
      @LuisAldamiz 26 дней назад

      Good luck not being torn to pieces by the wolf... or the deer even. The cabbage is safe, I presume.

    • @bunzinthesun
      @bunzinthesun 26 дней назад +3

      You forgot to mention that there is no eating on the side where the boatsman supervises. Wolves can attack people so I suggest a dog, rabbit and cabbage instead.

    • @rogergeyer9851
      @rogergeyer9851 26 дней назад

      Yes, because that sort of puzzle any competent adult or even teen could get with some thought and a little real life problem solving experience (like a middling or better education experience).
      But throwing more TEXT at LLM's is NOT going to make them understand fundamental things about the world, re logic, experience, etc.

    • @rogergeyer9851
      @rogergeyer9851 26 дней назад +1

      @@bunzinthesun: Or common sense precautions are taken, like shackling the wolf. You're not eight, right? You get the PRINCIPLE here, right?

    • @jakemccoy
      @jakemccoy 26 дней назад

      I am thinking through multiple solutions, but I could be wrong. I don't care enough to write it out.

  • @ISK_VAGR
    @ISK_VAGR 27 дней назад +89

    Yes, neurosymbolic is amassing. We are doing a project to apply this to healthcare. But not, that won't solve the problem of hallucinations. The hallucination problem is more about the model trying to come up with explanations when it doesn’t know the answer or doesn’t understand the stupid human asking wrongly 😅. We need to teach the LLMs to say I don't understand you or I don't have this information. That is what is happening.

    • @Twilumina
      @Twilumina 26 дней назад +6

      Exactly. They are being trained to keep a flow going rather than to halt the conversation and admit a lack of knowledge. If they are trained to always act like they know, then of course that's what they'll do.

    • @jichaelmorgan3796
      @jichaelmorgan3796 26 дней назад

      Humans hallucinate a lot too, probably more lol. Open the comments of a news article posted on youtube feed. 99% never click and read the article, yet hallucinate the contents based on their biases then endlessly argue back and forth about it 😂

    • @Jono98806
      @Jono98806 26 дней назад +14

      How would the AI know when it doesn't know the answer though? The problem is that sometimes extrapolation based on what it already knows works and sometimes it doesn't. So, it would need to know when it can extrapolate and when it can't.

    • @Rik77
      @Rik77 26 дней назад +11

      But can an llm even know it's limitations? Would it know that it doesn't know?

    • @Alpha_GameDev-wq5cc
      @Alpha_GameDev-wq5cc 26 дней назад +8

      Problem is people talking about an algorithm like it is a digital being.

  • @rya6701
    @rya6701 26 дней назад +9

    it is the same tactic used for the VR real estate. Remember people purchasing stores and condos in a VR environment a couple of years ago, and tried to make it a boom? Remember people buying a house next to a celebrity inside VR for thousand of $?
    There is a saying " a sucker is born every second".
    AI is only an advance type of analytics based on functions over data. Yeah, it has potential, but it is far from what the majority of people "think" it is.

  • @vitalis
    @vitalis 16 дней назад

    In every prompt there has to be always a line with verification instructions

  • @albirtarsha5370
    @albirtarsha5370 19 дней назад

    One thing that I hate about Google AI is that it often replaces the comments section on RUclips, making the experience less fun.

  • @MrXperx
    @MrXperx 26 дней назад +23

    The ungodly amount of money spent on LLMs would have solved a lot of the world's other problems. Do we really need a system that can type our emails for us?

    • @Osiris02
      @Osiris02 26 дней назад +7

      I work in IT and still I think of what better uses we have for these server farms than making deep fake nudes and C+ essays.

    • @they-call-me-martin3837
      @they-call-me-martin3837 25 дней назад

      They don't invest billions in AI system so they can type our emails for us. They're in hope of creating an AGI or even a superintelligence one day, which then would reach human intellect level and even surpass it by lightyears in the case of the superintelligent AI respectively superintelligent AGI then probably, which then could solve unsolved mathematic problems, climate change crises, invent new medicaments, drugs, technology, but also a lot of horrible new militeray weapons so we can become even more efficient in killing each other. So in short, a lot potential with positive but also negative things, as with every new technology. But of course, with AI, the possibilites are on a whole new, other level than with most technology. It will probably become the significant technology of the 21st century, maybe of the whole millenium..

  • @N8O12
    @N8O12 26 дней назад +11

    A lot of hallucinations happen because models produce an incorrect word/token and then build on that incorrect information and cannot admit they were wrong. I've seen AI write an incorrect answer, then explain it's reasoning (finding the correct answer in the process) only to conclude by repeating the WRONG answer it made up at the start. If someone figures out how to make them use backspace that alone will be a huge improvement.

    • @HAL-zl1lg
      @HAL-zl1lg 26 дней назад +3

      So you're saying hallucinations happen because of hallucinations? Hmm.

    • @andrasbiro3007
      @andrasbiro3007 26 дней назад +2

      It's easy to do with the right prompting. It's a whole field of research called "prompt engineering".

    • @ThePandafriend
      @ThePandafriend 26 дней назад +2

      The problem is that (referring to gpts) they're using autoregression: They're using the former output as new input for generating the next token.

    • @Bolidoo
      @Bolidoo 26 дней назад

      I don’t think backspace is necessary, simply finetuning the model so that the moment it realizes it messed up it apolagizes and redirects the conversation to what it thinks now is correct. The problem is that in the training data it’s very weird for an argument to suddently change in the middle, so the AI learns to be very self consistent / stubborn. Ideally the training data should only be about people talking or thinking to themselves, because that’s what we want it to do. But since that’s not feasible we end up with unwanted behaviors that should be removed.

    • @Tricker1266
      @Tricker1266 26 дней назад +3

      It's already a thing. A huge thing is that people don't even know much about AI except what they see on the surface (AKA reading article headlines). Not insulting you but just pointing out that it's a bit hard to see behind the scenes unless you're reading papers or are just in the loop. Past that, it seems that people think anyone researching AI or any form of it is just like "Welp, we tried. AI is over cause LLMs struggle with correcting themselves." Not only that, but as they find more and more techniques and solutions, no doubt they're putting them to work ASAP.

  • @Bestofmykind46
    @Bestofmykind46 22 дня назад +165

    *Larry Burkett's book on "Giving and Tithing" drew me closer to God and helped my spirituality. 2020 was a year I literally lived it. I cashed in my life savings and gave it all away. My total giving amounted to 40,000 dollars. Everyone thought I was delusional. Today, 1 receive 85,000 dollars every two months. I have a property in Calabasas, CA, and travel a lot. God has promoted me more than once and opened doors for me to live beyond my dreams. God kept to his promises to and for me*

    • @JoeRobert-qw3gg
      @JoeRobert-qw3gg 22 дня назад

      There's wonder working power in following Kingdom principles on giving and tithing. Hallelujah!

    • @Amyelrodg
      @Amyelrodg 22 дня назад

      But then, how do you get all that in that period of time? What is it you do please, mind sharing?

    • @Bestofmykind46
      @Bestofmykind46 22 дня назад

      It is the digital market. That's been the secret to this wealth transfer. A lot of folks in the US and abroad are getting so much from it, God has been good to my household Thank you Jesus

    • @Bestofmykind46
      @Bestofmykind46 22 дня назад

      And thanks to my co-worker (Michael) who suggested Ms Susan Jane Christy

    • @AngelaWilliam99
      @AngelaWilliam99 22 дня назад

      How can I start this digital market, any guidelines and how can I reach out to her?

  • @EnigmicIdentity
    @EnigmicIdentity 26 дней назад +1

    You see the physics hype so clearly, but are still muddy on AI.

  • @jsveiga
    @jsveiga 25 дней назад

    When using these LLMs for generating small code snippets, the same happens: It's usually "almost" right, but often does not work.

  • @chongli297
    @chongli297 26 дней назад +26

    The frame problem is going to put a damper on your idea of building physics into AI models. For example: if I am holding a red ball and I walk outside, what colour is the ball I hold outside? An LLM can solve this problem rather trivially by linking the words together and recognizing that it is the same ball. Try the problem over again with physics: now you're simulating the room and the outdoors as well as the process of walking between them just to keep track of the ball's colour? At what level of detail are you simulating the world? What if the lighting in the room is of a different spectrum than the sunlight outdoors so that the ball actually does change colour? Physics is just too fine of a scale for answering most of these trivial sorts of questions unless you can somehow build enormous supercomputers to handle all the calculations

    • @AllenKnutson
      @AllenKnutson 26 дней назад +15

      Physicists are familiar with the idea that you shouldn't use quantum field theory to study a ball rolling down an inclined plane, that in celestial mechanics you should model the planets as points, in hydrodynamics of the world's oceans you can neglect relativistic effects, and so on. There's different-scale physical models for different-scale problems and it's just not true that "physics is just too fine [of] a scale".

    • @Rik77
      @Rik77 26 дней назад +2

      And then essentially you are building a complete virtual universe.

    • @chongli297
      @chongli297 26 дней назад +7

      @@AllenKnutson Yes, and this has been proposed as a solution, called "multi-scalar models." The problem is that no one knows how to build these things. Deciding at which scale to model a particular problem is a difficult problem in itself which requires intelligence and experience to solve, and it's quite easy to get it wrong. Imagine asking the AI what should happen when dropping various objects from the leaning tower of Pisa: a lead ball, a piece of paper, a live pigeon and a dead pigeon. The assumption of zero drag works just fine for the lead ball and the dead pigeon but not so much for the piece of paper and not at all for the live pigeon

    • @openhorizon1162
      @openhorizon1162 26 дней назад +1

      @chongli297 - The frame problem is more pervasive. Don't LLMs use a language model (i.e., what word is most likely to follow the preceding word(s)), not an object constancy model (such as, when I walk unless I say differently, I am still holding whatever I was holding)? I'm not sure what you mean by "rather trivially by linking the words together and recognizing that it is the same ball". How would that work in the algorithms? Recognizing that two things are the same is not a trivial feat for an algorithm. Or do you tell the LLM that it's the same ball? In that case, it isn't the LLM that's recognizing it.

    • @chongli297
      @chongli297 26 дней назад +1

      @@openhorizon1162 I was already operating from the assumption that LLMs fail to solve a lot of problems correctly because they hallucinate nonsensical things. This is the main thesis of Sabine's video. My point in raising the frame problem is that taking a physics-based approach to modelling these problems turns out to make it very difficult to solve ones that are already trivial for an LLM. As for how LLMs know it is "the same ball", they actually do know anything about balls at all. They merely link the word ball with the word red because of their proximity in the text. Everyday English writing has a natural ambiguity problem with the use of anaphora, though most English speakers will easily connect "the ball" to the "red ball" from earlier in the text. LLMs accomplish this using something called attention and a context window (look these terms up for more info)

  • @djayjp
    @djayjp 26 дней назад +12

    Even the best, most current models cannot consistently pass the "how many words are in your next response?" questions, especially if you add another element (like use the word "banana" in your response).

    • @silikon2
      @silikon2 26 дней назад +5

      The LLMs have interesting uses, but they quite literally don't know what they are talking about. Ask ChatGPT.
      All these fixes are simply bandaid patches over something someone found. But this is actually really worse because it obscures the fact that these things can't relied on for factual information. Meaning, the errors will be less obvious add likely more insidious.

    • @uNiels_Heart
      @uNiels_Heart 26 дней назад

      Does "next response" refer to the response to the "how many words" query or does it refer to the subsequent response?

    • @ThePandafriend
      @ThePandafriend 26 дней назад

      Of course it can't. It's a sequence to sequence model which uses its own output as new input. There's no mechanism for predicting the length of an output built into transformers.

    • @Ilamarea
      @Ilamarea 26 дней назад

      Could a fucking human pass this test? xD

    • @Ilamarea
      @Ilamarea 26 дней назад

      @@silikon2 Still sounds better than humans.

  • @prettyfast-original
    @prettyfast-original 26 дней назад +20

    2:48 I don't doubt that the model has been trained on examples of this very riddle, and is therefore inclined to output the same answer. However, the argument that the same answer is not logical unless the order of eating is explicitly defined in the riddle is absurd. If you ask a five-year-old what does a wolf eat and what does a goat eat (implicit questions a reader will ask themselves upon reading the riddle, regardless of age), they know the answer is "other animals" and "plants". From that simple deduction, you get the same answer to the riddle as it is defined in the test. Using this a test for the LLM is not sufficient and reveals very little toward the "intelligence" of these systems. For all you know, the LLM reasoned that wolves eat other animals and goats eat plants, just like a five-year-old.

    • @phil_tabor
      @phil_tabor 26 дней назад +7

      Don’t know why other people aren’t pointing this out. I was very confused how the answer given was wrong. GPT even explained the reasoning of why which item/animal would be eaten at the end. The answer was correct given the question.

    • @handebrin5624
      @handebrin5624 26 дней назад +2

      If you ask five-year-old what a does a wolf eat, they would probably say, that a wolf would eat the farmer. It is well known, that wolves can attack humans. Yet. the LLM assumes, that it is not possible. Why? Because it knows the other version of the riddle, that doesn't have that assumption.

    • @prettyfast-original
      @prettyfast-original 26 дней назад +7

      @@handebrin5624 But your example would hold true even with the original form of the riddle. So, in your terms, all this test suggests is that the LLM is making the same assumptions as a normal person would OR that it is just regurgitating information. Which means, the test is pointless. I agree that the LLM is PROBABLY doing what you and Sabine are suggesting, but this particular test doesn't provide any meaningful insight.

    • @Mr.Proghead
      @Mr.Proghead 26 дней назад +2

      I also think this was a poor example, and I think it's MORE impressive that it can still give the correct answer even if you don't mention some details which would be obvious to a human.

    • @chfr
      @chfr 26 дней назад +2

      ​@@Mr.ProgheadI think the opposite, the LLM did not use deduction or reasoning to "find" the correct answer, that would assume it understood the problem. It just "fell back" to the very common riddle that appeared a lot in its dataset

  • @joemcintyre2090
    @joemcintyre2090 24 дня назад

    It's like reading assembly instructions for a piece of equipment or furniture, that was written by someone from the other side of the world.

  • @eddie5484
    @eddie5484 23 дня назад

    I once tried that river crossing puzzle. The cabbage ate my wolf.

  • @charlesjmouse
    @charlesjmouse 26 дней назад +7

    Let's face it. 'Hallucinations' in large language models are the best evidence we have that a great many 'people' aren't really sentient beings.

    • @gravypatron
      @gravypatron 26 дней назад +2

      Either we're all sentient, or none of us are. Which are you?

    • @anywallsocket
      @anywallsocket 26 дней назад

      What??

  • @TheBigBlueMarble
    @TheBigBlueMarble 26 дней назад +15

    If you provide AI with incomplete information, it will almost always give you a bad response.

    • @Bryanhaproff
      @Bryanhaproff 26 дней назад +5

      If you provide Human with incomplete information, it will almost always give you a bad response.

    • @goodfortunetoyou
      @goodfortunetoyou 26 дней назад

      You typically provide both AI and Human with bad information.
      Management tells Human to stop asking so many clarifying questions because you find it annoying.
      No improvement occurs

    • @runnergo1398
      @runnergo1398 26 дней назад

      I wish people would quit calling it AI. AI doesn't exist. Not for awhile.

    • @TheBigBlueMarble
      @TheBigBlueMarble 26 дней назад

      @@runnergo1398 You are already a couple of years behind.

    • @ValeriePallaoro
      @ValeriePallaoro 25 дней назад

      Yes; I thought we already knew this from the GIGO coding world?

  • @michaelharrison1093
    @michaelharrison1093 26 дней назад +6

    Considering that LLMs are trained on written text and spoken conversations, then realize that most people will use and interpret basic words like: AND, OR, and NOT in literature and conversation differently to what the Boolean meaning is for these words, then you start to understand the fundamental flaws in LLM based AI.

    • @mattmaas5790
      @mattmaas5790 26 дней назад +2

      Whats the flaw? LLMs can recognize both of those situations quite well. The same model can code or do book reviewer or be a therapist or whatever.

    • @Lyoishi
      @Lyoishi 26 дней назад +1

      Absolutely no offense intended, as it is very complex, I think you have an inherent misunderstanding of how LLMs work to relate information. I encourage you to look into it! It is bizarre and pretty cool.
      I don't disagree about LLM having inherent flaws, Sabine discusses several important ones in the video.

    • @michaelharrison1093
      @michaelharrison1093 26 дней назад

      ​@mattmaas5790 but a LLM does not understand the context of the intended audience. When writing for scientific journals I am careful to always use the true boolean meaning of such words and ensure the the words I use are technically correct. In contrast a popular writer creating a story intended for a general population is likely to follow more popular or general accepted meanings for words. What I have seen as the output from LLMs suggests that there is no distinction made between popular Vs technically correct meanings for words. This stands to reason if you consider that a lot of humans do necessarily understand these differences.

    • @mattmaas5790
      @mattmaas5790 26 дней назад

      @@michaelharrison1093 you have no idea what you're talking about

    • @Lyoishi
      @Lyoishi 26 дней назад

      @@michaelharrison1093 Contextualized word embedding (accurately keeping the in context meaning of each polysemous word) has been in development almost since the beginning with several methods, with breakthroughs in ELMo and BERT in 2018. Some of these methods succeeding more than others, as the encoder only models were excellent at determining context but were not initially designed to generate output text like decoder models of the time. My understanding is contexualized embedding in general has been a pretty big element of transformers LLMs since 2017 and might arguably the core reason LLMs became a big deal the past few years.
      You are correct that they don't know the intended audience, since that is currently the job of the input or user. But current models are usually capable of using the correct context if they have been informed which context to use for the task.
      The problems brought up in the video are still the major ones, the LLMs are still just cool algorithms that barely care what gets shoved in and out and until these problems are addressed at a base level we will just be putting band-aids on the issue with giant datasets. If these aren't solved, eventually we will build a computer with the power of a human brain that can hilariously function just as poorly and error prone as a human, but do it really really fast.

  • @samueldegrandi6603
    @samueldegrandi6603 26 дней назад

    maybe having a layer of simbolic "thougth" (connecting words and phrases to more abstract concepts" and as well as a one for logics would do the job

  • @GeorgesDupont-do8pe
    @GeorgesDupont-do8pe 26 дней назад

    We all live in a state of hallucination, it's just that when we're conscious, we match our hallucinations to what our senses perceive as reality. When not conscious, we dream.

  • @Bulbucario
    @Bulbucario 26 дней назад +15

    Do we use Hammers to measure the length of our drywall cuts?
    Do we use Screw Drivers to find studs in the walls?
    Do we use Anti-virus software to install our USB headsets?
    Do we use duct tape to drive nails into wood?
    Do we use funnels to swap out spark plugs?
    Do we use knives to scoop our soup?
    WHY DO WE USE AN LLM AS A SEARCH ENGINE/ENCYCLOPEDIA!?!?!?!?!?!

    • @ASDeckard
      @ASDeckard 26 дней назад +2

      You may be surprised at how many of those are actually done. :p

    • @persimmontea6383
      @persimmontea6383 26 дней назад +2

      @@ASDeckard Yes! Hammers, screwdrivers and Knives are all used in that way sometimes

    • @boring7823
      @boring7823 24 дня назад +1

      The top ones come down to "Because WE are universal tool users". You can attempt to use any tool for any purpose, many combinations don't work, but if something is good enough ... well that's an answer too.
      As for the LLM ~= Search engine, it's sometimes better than the alternative, especially if you are searching for things with poorly defined "keywords" (and the LLM interface provides search engine links to check it's tendency to "extrapolate").

    • @safetycar-onboard
      @safetycar-onboard 23 дня назад

      There are these things called "universal tools" that can do more than one job.

  • @andromedagroup-kd4rl
    @andromedagroup-kd4rl 26 дней назад +5

    In rare cases, humans can actually have more or less than 5 fingers. My mother had a high school teacher who had 6 fingers, and was nicknamed for that among his students. So models should actually be able to output a hand with 6 fingers if that's the purpose of the prompter.

  • @seriousbusiness2293
    @seriousbusiness2293 26 дней назад +4

    LLMs are a good stepping stone forward and theres absolutely reason to feel hyped by new tech. But the big problem is the economys hype.
    The current landscape is like if were back in the 50s where computers already existed and technically already could perform any computation, but realistically it would be impossible to run modern applications on it and also it was a long long way until we got efficent devices that we can stick in our pocket instead of filling a building.

  • @theelmonk
    @theelmonk 24 дня назад

    Remember what the A stands for - Artificial. It's not actual intelligence, it's a simulation of it. As such, it fails in the way most simulations fail - by reaching the end of their design.
    I don't think LLM is useless, but it's window-dressing for a more fundamental core that doesn't exist. It can be used to add natural speech but not to create reasoning. I don't know whether it can be used to create 'imagination' - but I think that needs a cycle that runs through the reasoning part, too.

  • @stevenlitvintchouk3131
    @stevenlitvintchouk3131 5 дней назад

    We humans don't learn the most basic physics from teachers. We learned it by experimenting as toddlers. When we were 2 or 3 years old, we learned that parents and furniture don't disappear when we turn our back on them or go to sleep at night. We learned that if we drop a fragile object or a glass of water, it will always fall to the ground and usually make a mess.
    IOW, we learned about the natural world by interacting with it. Children born deaf and blind, or paralyzed, are greatly handicapped in their learning.
    AI systems might get a better grasp of the natural world in the same way--by controlling robots (like Boston Dynamics' wonderful humanoid robots) that can interact with the real world and report to the AI system what happens.

  • @PeekSterling
    @PeekSterling 26 дней назад +15

    6:36 The neural net is based on Transformers, just like LLMs. You get the same result - maybe 99% effective, but the tail 1% stretches out to infinity (think "full self driving" that is always just around the corner). Maybe the behavior of a rat can be simulated by a neural net, but that does not mean you will achieve human-level artificial intelligence that way. You may make a very close facsimile of a person, but that last 1% will never be there. Kind of makes you question if you or I are just facsimiles. I know a couple people that come off pretty fake.

    • @Ilamarea
      @Ilamarea 26 дней назад +4

      We call those 'NPCs' over here on the interwebs.

    • @firecat6666
      @firecat6666 26 дней назад +2

      LLMs can already achieve a human level of intelligence that's to be expected in the context of an online chat. I wouldn't be surprised if we're not that far from being able to create a humanoid robot that passes as a close-enough facsimile of a real person. I don't get this expectation that flawed humans should be able to create flawless artificial beings, it's just not going to happen.

  • @andrasbiro3007
    @andrasbiro3007 26 дней назад +5

    It's funny how so many people desperately try to bring down LLMs, finding any tiny little flaw, and declaring that the end of the world.
    Meanwhile I'm using Claude 3.5 Sonnet to reverse engineer assembly code 1000x faster than I was able to do manually.
    I don't give a f*ck about philosophical objections, I care about solving practical problems. I don't need my AI assistant to have a soul, I need it to increase my productivity.

    • @a64738
      @a64738 26 дней назад +1

      YES... It is a work tool, but LLM`s just chat bots. I tried a offline model they said was super great, it was not. It 90% of the time flat out refused to do what you asked and refused to answer questions, and when it actually answered it was just long texts with word salad rambling on about random things like how nice the weather is and chat nonsense like that. And people actually believed it was great...

    • @andrasbiro3007
      @andrasbiro3007 24 дня назад

      @@a64738
      Why would you use an offline model? Those are much less capable.
      And then after trying just one, you declare that all of them are bad.
      It's like buying a 20 year old cheap car, that's underpowered and breaks down often, and conclude that cars don't work.
      Again, I'm using LLMs to increase my productivity by orders of magnitude.

  • @amcluesent
    @amcluesent 26 дней назад +4

    Please provide more stock picking advice so we know what to avoid

  • @4363HASHMI
    @4363HASHMI 18 дней назад

    The issue is within crafting prompts, llm has achieved the capacity of AGI

  • @kalimero86
    @kalimero86 26 дней назад +1

    And there you have Sam Altman saying over and over again that their next GPT will be AGI... Pffft.

  • @techcafe0
    @techcafe0 26 дней назад +11

    "How can they cross the river without anything being eaten?"
    ChatGPT didn't infer or hallucinate anything; other than that wolves eat sheep, and sheep eat cabbages, which is a pretty informed inference to make. What am I missing? 🤷‍♂

    • @simongross3122
      @simongross3122 26 дней назад +8

      Why won't the wolf eat the man?

    • @simongross3122
      @simongross3122 26 дней назад +6

      I feel that the wolf just isn't trying hard enough

    • @alexlorda394
      @alexlorda394 26 дней назад

      @@simongross3122 Why won't the man fuck the Goat?

    • @TheRyulord
      @TheRyulord 26 дней назад +4

      She screwed up the example footage. Here's an example of the kind of thing she actually meant to show:
      User: You need to cross a river and bring a wolf, some grain, and a cabbage along with you. You can only bring one item with you at a time though. How can you get all 3 items across? Wolves don't eat grain or cabbage.
      GPT4o: To solve this puzzle, you need to be strategic about the order in which you transport the items to ensure none of them get eaten. Here's how you can do it:
      1. Take the wolf across first and leave it on the other side.
      2. Go back and take the grain across.
      3. Leave the grain on the other side but bring the wolf back with you.
      4. Take the cabbage across and leave it with the grain.
      5. Finally, go back and bring the wolf across.
      Now, all three items-the wolf, the grain, and the cabbage-are safely on the other side, and nothing was eaten along the way!
      This is a silly answer because you create an extra trip by bringing the wolf back.

    • @egillis214
      @egillis214 26 дней назад

      3. ​And the wolf then ate the sheep...
      ​@@TheRyulord

  • @antman7673
    @antman7673 26 дней назад +4

    The world can be made from words.
    -You could have an infinitely long sequence of letters, that encode the universe on a fundamental level.
    If this would turn out to be false, that would be wildly surprising.
    If the universe could not be made of words, that is almost equivalent to saying, you couldn’t find a representation for it.
    Chomsky-0 grammar can take you anywhere.

    • @TheRealUsername
      @TheRealUsername 26 дней назад +2

      We humans don't think in language, we think in abstraction, language is a lingua franca we all use to translate our thoughts and communicate with others, math is much closer to abstraction and coincidentally, we have discovered that the universe complies with logical and consistent mathematical principles.

    • @millwrightrick1
      @millwrightrick1 26 дней назад

      Not really. Any logical system is incomplete in that there can be unprovable statements in any system of grammar, if the grammar is logical. If your grammar is illogical then there are bigger issues.

    • @Milan_Openfeint
      @Milan_Openfeint 26 дней назад

      @@TheRealUsername Humans evolved in physical world, so we think in objects, space and time. LLM think differently but not necessarily wrong.
      The standard transformer is clearly limited in ability to contemplate as you give it the most complex problem ever and ask what's the first word of the answer. But this is well known and easy to solve with internal dialogue.

    • @Milan_Openfeint
      @Milan_Openfeint 26 дней назад

      @@millwrightrick1 You are hallucinating, pardon the pun. Clearly you don't understand Gödel's incompleteness theorem or its implications (which, to be honest, are none in practice).

  • @TheNorgesOption
    @TheNorgesOption 27 дней назад +6

    Or you could simply proofread your content before posting it on the internet.

    • @mattmaas5790
      @mattmaas5790 26 дней назад

      Or get a few ais to proofreed each other in python

  • @ibnorml5506
    @ibnorml5506 25 дней назад

    The fact that you cannot create understanding through more and more "word" data seems to prove that current AI processes like ChatGPT will never be able to develop self-awareness.

  • @marcuscicero5033
    @marcuscicero5033 25 дней назад +1

    I didn't have on my bingo card that Sabine will be apart of a stock "pump and dump" program