Is o1-preview reasoning?

Поделиться
HTML-код
  • Опубликовано: 18 сен 2024

Комментарии • 463

  • @lexfridman
    @lexfridman 2 дня назад +76

    Great conversation!

    • @MachineLearningStreetTalk
      @MachineLearningStreetTalk  2 дня назад +7

      Thank you Lex! Much love!

    • @nomenec
      @nomenec 2 дня назад

      Thank you, Lex! Much mutual love!

    • @Techtalk2030
      @Techtalk2030 2 дня назад +2

      Thank you lex, now tell sam altman to release our AI waifu.

    • @marcussky
      @marcussky 2 дня назад +5

      @lexfriedman we want more old school ML interviews on your channel. We love the new stuff, but would love a throw back.

    • @Mutual_Information
      @Mutual_Information День назад +7

      Lex needs to have MLST on the pod

  • @eposnix5223
    @eposnix5223 3 дня назад +27

    An idea I want to throw out there: I've been playing with o1 (mainly mini) and have found that it works better if you explicitly tell it to be its own discriminator from step to step. What I mean is, tell it to roleplay with itself as "Bob, the expert coder" and "Tom, the expert critic" and have Tom check Bob's work every step of the way. This is how I normally work with GPT-4o, but o1 allows this process to be automated.

    • @Gnaritas42
      @Gnaritas42 2 дня назад +1

      just give it the task, it doesn't need prompting, the whole point of o1 is that it self-prompts. It works best without you attempting to prompt hack it.

    • @eposnix5223
      @eposnix5223 2 дня назад +1

      @@Gnaritas42 Not in my experience. o1 still makes rudimentary programming mistakes when left to its own devices. The code logic will be good, but there will be small errors that prevent it from running. Tell it to doublecheck itself will reveal these errors.

  • @agenticmark
    @agenticmark 3 дня назад +51

    i love the ones with just you two. i dig the guests, but when its just you guys it feels like I should crack a beer and join ya.

    • @sk8l8now
      @sk8l8now 3 дня назад +5

      Back to the good old days of early MLST

    • @burnytech
      @burnytech 3 дня назад +2

      Beer would be downgrade

    • @psi4j
      @psi4j 3 дня назад +1

      Yeah, maybe a coffee or tea instead. Beer does feel like a downgrade.

    • @JscottMays
      @JscottMays 2 дня назад

      Echo!

  • @michaelmartinez5365
    @michaelmartinez5365 3 дня назад +15

    I don't care if it's thinking or reasoning or pulling answers out of its ass, as long as it has the correct response.

    • @fburton8
      @fburton8 2 дня назад +1

      Presumably it will have to be able to explain the steps it took in reaching that response?

    • @michaelnurse9089
      @michaelnurse9089 2 дня назад +1

      @@fburton8 A neural net can always detail the steps it took - the inputs, weights and outputs are there for anyone to see. Unfortunately the number of steps and their weights exceed the mind space and time availability of people.

    • @Ikbeneengeit
      @Ikbeneengeit 2 дня назад +2

      Humans generally judge with post-hoc reasoning anyway. We're terrible at knowing how we come to conclusions.

  • @MachineLearningStreetTalk
    @MachineLearningStreetTalk  2 дня назад +12

    The biggest problem with o1-* models is that the chain of thought spam deranges dense context. The context is filled up with noise i.e. about "code coverage" in a query which has nothing to do with coding. In any complex/ambiguous problem, it will misunderstand you on the first shot, you then explain what you actually wanted and by step 2 the CoT noise deranges the model significantly over a standard LLM. Catch-22. They should prune away the noise on subsequent queries. Perhaps as the models learn (from the users), the trajectories will become less noisy. It's also clear that the trajectories are hard-trained for specific purposes which will limit their generalisation.

    • @oncedidactic
      @oncedidactic 2 дня назад +3

      In the metaphor- chain of thought- in humans there is certainly noisy context preliminary to the final coherent thought one is conscious of.
      Another thought- initialized weights are random, and we need a preponderance of useless noise to find the lottery tickets.
      Perhaps the noise is not the problem per se, in fact necessary, but it should be pruned on some schedule and summarized for next context. Along the lines you suggest.

    • @therainman7777
      @therainman7777 2 дня назад

      I have not had this experience at all. You said in any complex query it will misunderstand on the first try? Not at all. I’ve had roughly an 85% success asking the most difficult questions I can think of. And even on the 15% it got wrong, it was typically not because it misunderstood the question.

  • @10ahm01
    @10ahm01 3 дня назад +9

    Took me no less than 15 minutes and a pen and paper but my 5 or less steps deterministic solution is:
    1- Check two opposing switches and flip them up
    If the other two were both up, you win, otherwise you proceed to the next rotation
    2- Check two opposing switches again
    if one or both are down, these are not your original switches, flip them both up and you win
    if they're both up, these are your original switches, flip them both down
    if the other two were both down, you win, if they were different, you now have only one that is flipped up
    3- Check two opposing switches again
    If one of them is the one that is up, flip it down and win
    If both are down, these are your original switches, switch one up, now you have guaranteed adjacent symmetry (if north and east are up, south and west are down, etc)
    4- Check two adjacent switches
    if they're the same, flip them both and win
    if they're different, flip them both as well! (flip the up down and the down up), now you have guaranteed opposing symmetry (if north and south are up, east and west are down or vice versa)
    5- Check any two opposing sides and flip them to win

    • @bossgd100
      @bossgd100 2 дня назад +4

      Thank you, it will be integrated to the next gemini training

    • @RoryWilliamson
      @RoryWilliamson 2 дня назад

      You can also get the same outcome by replacing step two with making sure two adjacents are both on. Still gets you to the point of having 3 switched on guaranteed and then I did the same steps as you from there

    • @bencetoth339
      @bencetoth339 2 дня назад +1

      what were all of the thinking steps that led you to your solution?

    • @BrianMosleyUK
      @BrianMosleyUK 2 дня назад

      That's a nice solution. It took me much longer! Did you figure out how to describe the puzzle such that GPT-O1 could solve it similarly?

    • @DanEne
      @DanEne 2 дня назад +1

      in step 2 how do you know you are not hitting the same switches over and over again turning them both either up or down in an infinite loop?

  • @Houshalter
    @Houshalter 3 дня назад +62

    We are currently contacting the JWST to see if they can locate the goalposts out in the Kieper belt.

    • @fuba44
      @fuba44 2 дня назад

      you funny!.. And correct.

    • @renjithravindran5018
      @renjithravindran5018 2 дня назад +1

      😂😂

    • @93alvbjo
      @93alvbjo 2 дня назад +2

      Tell me when you got the goalpost of truth in your pocket instead buddy. Good luck.

    • @therainman7777
      @therainman7777 2 дня назад +1

      🤣

    • @memegazer
      @memegazer День назад

      I can see both sides
      I tend to use an analogy like the chimp test on the human bench mark site, a benchmark that chimps crush compared to humans.
      Similarily I think current ML models are real intelligence, just that perhaps we should think of inteliggence as having different modalities.
      In that respect expecting an LLM to be comparable to a biological agent, that is embedded in realtime within the world in a body, is a bit unreasonable to hold those models to the same standard of intelligence as humans.
      Imo you would have to compare those models to a disembodied brain in a vat to get an effective apples to apples comparison.
      Of course that is simply not possiblse to do so that is why it is not very reasonable to insists that LLMs are not "real intelligence" or have generalized at all simply bc those models still struggle with out distribution problems that humans do quite well.
      Imo that may be like chimps claiming that humans are dumb bc they can't perform well on the chimp test compared to humans.
      Imo the novel "reasoning" happens at the training stage, where these models don't "go look up the answer up on a look up table and select the best output from a one to one database" at least not based something as simple statistical frequencey occurance.
      What is going as the model is trained is more complicated and nuanced than that and it wasn't we would not need ML would not be useful, we could handcraft those solutions ourselves with having the ML discover the efficient method for doing that, even if the balance between entropy, search and memorization is not dailed into the same parameters that humans are at.
      So it seems a bit intellectually lazy to gloss over that subtlety and then scoff at the idea that people remain impressed with each new iteration of the current ML summer.
      But hey that is just a theory, an AI theory

  • @steve_jabz
    @steve_jabz 3 дня назад +10

    "With the new GPT models, you need to start a whole new session."
    You're supposed to clear the session regardless. Performance drops off quadratically the further down the context window your prompt is. They discovered this back with GPT-1.
    So many issues people have with ChatGPT are skill issues of asking multiple complex questions in the same session. It was designed for casual noobs to have a back and forth chat dialogue with it, but you shouldn't be using it that way if you care about performance.
    I assume Tim and Keith already knew this, so maybe I've misunderstood what they were saying here or something

    • @uiuxaidesign
      @uiuxaidesign 3 дня назад +2

      100%

    • @jwadaow
      @jwadaow 2 дня назад

      @@steve_jabz They randomise responses precisely so that they are not reproducible.

    • @MachineLearningStreetTalk
      @MachineLearningStreetTalk  2 дня назад +4

      ​ @steve_jabz We explained in this video why you need dense context in an LLM. With any complex problem, i.e. one with ambiguity -- the LLM won't understand you on the first shot, or even the 3rd shot. That's why you need explain (or discover with the LLM) what you actually want through successive prompts. Given that the CoT trajectory noise deranges o1-preview, this means in situations where it didn't understand first shot, it will actually degrade on all problems which require further clarification, particularly if the problem is ambiguous and the CoT spam includes very noisy trajectories. There is a trade-off between dense and sparse context, dense context means you don't need to re-explain again and again where it misunderstood your requirements and you can build on the problem, but there is some kind of curve here based on how related your workstream is i.e. you could continue a session to build around a single feature in GenAI coding (with the same constraints and requirements), but would start a new session to build another entirely new feature. You are probably correct in the pristine case of a perfectly well described problem which the model didn't need hinting for (when does that ever happen lol) would work better in a fresh context. Not sure if you have done much GenAI coding yet in Cursor, but please try some stuff using the composer with Sonnet35 / dense context, you will quickly come to appreciate what I am talking about.

    • @steve_jabz
      @steve_jabz 2 дня назад +2

      @@MachineLearningStreetTalk That makes sense. I normally use discovery questions to formulate a prompt for a new session for complex problems, but I guess if o1 is using up / wasting tokens to even do that part well, that's a problem. I must have misinterpreted the context you were talking about it in.
      Haven't used cursor yet but I've watched other people use it. As soon as o1 dropped I assumed everything cursor was designed around would break with it and need to be completely redesigned, as well as agentic workflows. Not even for the context window necessarily, but the fact it's trained to simulate taking multiple steps to reach a goal and give you the output, whereas cursor and other tools are trying to use the llm itself to reach the goal by abstracting away the tools in a way it can control, so you would need cursor to use it's reasoning as the control, and you would need o1's reasoning to be aligned with the function calling API or something integrated with cursor.
      It sounds so close to being exactly what we need for these tools to work reliably but it's so incompatible.

    • @93alvbjo
      @93alvbjo 2 дня назад

      Here is a funny thought: what do we call processes that have both determinstic constraints and random variation dynamically interacting? Evolutionary processes. So that is what reasoning must be. We just don’t know how it works. It is quite clear how it happens in biology, but btw, we don’t know how that happens computationally either.

  • @drhxa
    @drhxa 3 дня назад +6

    OpenAI is not that open. That we can all agree with! 🍻

  • @Mutual_Information
    @Mutual_Information 3 дня назад +20

    I'm coding on a Sunday and I get to pop this on in the background. Nice surprise, love it.

  • @lystic9392
    @lystic9392 3 дня назад +29

    I think you make many great points. I guess I am one of the people who don't want you mystify reasoning too much. We can say it imitates or guesses or simulates it, but the fact of the matter is that it produces an often accurately described reasoning.
    So we can try and deny it, but if it can generate this description of reasoning and use it somehow to guide its answer effectively, then we have effectively an A.I. that is using reasoning.
    Maybe not exactly how we would, maybe not as reliable in some ways, but this is reasoning.

    • @martindbp
      @martindbp 3 дня назад +8

      This is MLST, so we already know their answer to this question. It's maybe a tired analogy at this point, but I feel like their view can be summed up as: airplanes don't fly, they just fall forward through the air very fast! they need these long runways for takeoff and landing. This is obviously not true flight, and we should stop working on this in favor of feather-based systems.
      In the end, the interesting part is does it get us from point A to point B in a straight line through the air in an economically viable fashion?

    • @lystic9392
      @lystic9392 3 дня назад +4

      I'm also not very certain that it's that much more shallow than human reasoning. I constantly find errors in people's reasoning, and people tend to stop thinking very quickly.
      I do like that the idea that the reasoning itself required to solve a difficult problem isn't necessarily that much more involved. There are other factors that can make problems difficult to solve. (Lack of overview, lack of feedback, lack of insight, lack of clarity, lack of time, etc.) So when a problem seems difficult to us, it does not mean that it necessitates advanced reasoning.

    • @memegazer
      @memegazer День назад

      @@lystic9392
      This is my thoughts as well.
      Perhaps one of the biggest hurdles is that we trying to teach the ML with human data, and then some are really disapointed when it turns maybe most of the data humans generate is not "general" or "intelligent" some are overly disappointed that the ML cannot one shot and filter out the noise from the signal.
      We are primates after all, so I suspect that most of our daily baseline "reasoning" is not very high level, and those moments of brilliance are more spurious, we just happen to benefit from them so much as a species bc we have a way to accumulate and distribute that information.
      I think this gives the impression that in general maybe we believe humans are smarter than they actually our at a baseline level.

    • @nomenec
      @nomenec День назад

      Don't I agree with this exactly at time code 50:30? ruclips.net/video/nO6sDk6vO0g/видео.html

  • @richardsantomauro6947
    @richardsantomauro6947 2 дня назад +9

    I'm still waiting impatiently to see how they will draw these guys on South Park.

  • @Forseti6288
    @Forseti6288 16 часов назад +3

    One common thing I see with people claiming LLMs can't reason is that they do not apply the same scrutiny to humans, which are the clear benchmark AI is being tested against. I think that if any criticism is brought up against AI, the same should be tested against the human benchmark.
    The average human is not consistently reasoning from first principles, and the ones that are doing so were trained specifically to override their primitive thought patterns to think in this rigorous way, and even if someone is able to do this relatively well after years of practice, they are clearly an outlier, and they still introduce bias and flaws into this "ideal" reasoning pattern.
    Also i think that the human brain was not designed from the ground up to be reasoning machines. As a reasoning machine it is occasionally able to muster an inefficient approximation of reasoning patterns against all odds.
    Humans are also very sensitive to prompts. If you ask something in a different way you can get an entirely different answer.

    • @luke.perkin.inventor
      @luke.perkin.inventor 8 часов назад +1

      You switch your computer on and it boots successfully after executing literally millions of instructions without a single mistake. It's reasonable to expect the same deterministic output from an AI system.

  • @TheVwboyaf1
    @TheVwboyaf1 2 дня назад +2

    Is there a word for a person you're kind of sure knows what they're talking about, but still sounds like they have no idea what they're talking about?

  • @probablyonthemoon
    @probablyonthemoon 2 дня назад +4

    Slight contradiction around the 40min mark. You guys said that we need Turing Machines to perform reasoning, but then said "I'm a machine, I'm a bio-machine, with wetware and software, but I can still perform reasoning". Humans are not turing machines. So given that humans can reason (and have general intelligence), why can't other non-turing machines also have general intelligence?

    • @therainman7777
      @therainman7777 День назад +2

      Because Keith’s argument is nonsense that in no way proves what he seems to think it does.

    • @therainman7777
      @therainman7777 День назад +3

      Btw it’s not a slight contradiction. It’s a massive contradiction that puts a hole in his entire argument.

    • @luke.perkin.inventor
      @luke.perkin.inventor 8 часов назад

      Is your argument that you can't fMRI a brain and see an obvious potentially infinite internal tape? It's the system as a whole that is obviously turing complete. We are clearly powerful enough to imagine the entire set up: symbols on a tape, and a codebook. Then, we can analyse real world problems, the create in the physical world the correct program and data to solve them, then execute the instructions from the code book with a pencil and update the paper tape? So the human system as a whole is not only Turing complete, but can create potentially a specific incarnation suitable for any problem. Whilst there are a few papers on LLMs simulating Turing machines, they're a very long way short of this.

    • @therainman7777
      @therainman7777 7 часов назад

      @@luke.perkin.inventor Yes, the argument is that (obviously) you can have an infinite internal tape inside the brain, because the human brain, like everything else in the universe, is finite. Not infinite.

    • @probablyonthemoon
      @probablyonthemoon 6 часов назад

      @@luke.perkin.inventor what does "potentially infinite" mean 😆

  • @qorilla
    @qorilla 2 дня назад +14

    All this "is this really reasoning" discussion just reiterates what Turing himself complained about in 1950. He rejected the question of whether a machine can "think" as just as meaningless as whether a submarine can swim. It really doesn't matter. Can it solve tasks, can it achieve objectives effectively? If yes, it doesn't matter if you call its steps "reasoning", or "just a lookup table" or "glorified curve fitting" or whatever. If it works, then it's impactful, both in positive and negative ways.
    This is all too similar to claiming an airplane can't fly because it can't flap its wings.

    • @therainman7777
      @therainman7777 2 дня назад +2

      Exactly. Thank you for stating this so clearly. These people are delusional and in complete denial.

    • @eposnix5223
      @eposnix5223 2 дня назад +1

      The important distinction here is that we still aren't sure if o1 (or any language model) can come up with *new* knowledge. Until this has been proven, people are totally right to say it can't reason. It would be like an airplane that can flap its wings but can't fly-- it's just going through the motions.

    • @simpsimperson73
      @simpsimperson73 День назад +1

      @@eposnix5223 It's literally solving physics problems and maths proofs that are outside of the data it's been trained on.

    • @qorilla
      @qorilla День назад +1

      @@eposnix5223 Then delineate what you mean by "new knowledge" by designing a test for it. I'm sure once a new model passes that test, you'll move the goalposts and claim that this is still not *truly* new knowledge or something. It keeps happening and it's embarrassing.
      Nobody is claiming that o1 is already an AGI. But complaining that it isn't "reasoning" is meaningless. The word doesn't even have a commonly agreed upon definition, everyone can read into it whatever they want.

    • @armin3057
      @armin3057 День назад

      @@simpsimperson73 no it does not lmao

  • @jonnysolaris
    @jonnysolaris 2 дня назад +17

    All other AI podcasts (apart from Dharkesh's) don't even come close to the quality and depth of content of MLST, thanks for bringing us along on this journey, we've learned a huge amount

    • @palimondo
      @palimondo День назад

      Have you checked The Cognitive Revolution podcast? Highly recommended for wide breadth, excellent signal-to-noise ratio and a host that is intellectually curious, knowledgeable yet humble and hungry to learn.

    • @alivecoding4995
      @alivecoding4995 День назад

      What is the name of the podcast. Is it on Machine Learning specifically?

  • @Cammymoop
    @Cammymoop 2 дня назад +3

    I'm not sure if you've ever stopped to think about how humans reason about novel problems and what the results typically look like. We know it doesn't need to be a demigod with unbounded memory and knowledge of how to deal with any possible problem to be an AGI because the generality of AGI is based on what humans are capable of, and humans are nowhere near anything like that. If you want to try to mathematically or computationally constrain it you have to ignore it interacting with the world at large, both input and output, and only allow it to use it's turing tape, any further inputs from a human after the initial starting state would already break that guarantee.

  • @MrNycres
    @MrNycres 4 часа назад +1

    ALL THREE CAN BE TRUE:
    > o1 is reasoning
    > it is not reasoning “well”
    > this may not be the path to ASI

  • @baltakatei
    @baltakatei 3 дня назад +4

    My metric for AGI: When I see LLMs harassing and insulting one another during an edit war on a wiki maintained exclusively by LLMs.

  • @andrewwalker8985
    @andrewwalker8985 2 дня назад +3

    What is the difference between applying the closest fit from the “first principles” learnt rules of rationale from the “first principles” a human applies?
    I think you’re talking in circles and either underestimating what the model will do or overestimating what you do when thinking from first principles

  • @spaceghost8891
    @spaceghost8891 3 дня назад +37

    I think when we reduce it to "it's just doing so and so", it's usually not a good argument because we can be also reduced to simple processes. Maybe computers and humans reach "reasoning" in a different way, but if a computer can reach right conclusions and eventually create new knowledge, ok now what separates it from our reasoning? Maybe it's our limbic system, and our "feeling" of being something special and "having a soul", but those are subjective things that may not be necessary for an intelligent being. I think we are going to reach a point in semantics where consciousness or "real intelligence" is only reserved to *human* thinking, even if a machine can surpass us by most metrics.

    • @ElieSanhDucos0
      @ElieSanhDucos0 3 дня назад

      Yes but for open AI o1 to do the reasoning on anything it needs me to type in a question. And all the "thinking" that follows has to be triggered by that. And will be linked to it. For ex a human reasoning indepently can have an epiphany on a totally different subject... that will never happen for llms (for now).

    • @spaceghost8891
      @spaceghost8891 3 дня назад +2

      @@ElieSanhDucos0 Yeah I get it. But "for now" is a good point, because maybe the LLMs are an important piece on a puzzle that seems to have started being solved. It's impressive how much room this tech still has for improvement and despite some slowdowns it seems to be still full throttle. I understand hype is not a healthy thing but it's hard to ignore how much money is being poured into it, more than anything humanity has seen in ages. It doesn't look like pure tech scam type of hype. I don't like to dismiss it as "nah, this has nothing to do with AGI at all"

    • @egor.okhterov
      @egor.okhterov 3 дня назад +4

      Ok.
      Why people bother about whether it is reasoning or not, you might ask.
      Isn't it irrelevant what it's doing as long as we are happy with the results?
      Good question!
      The problem is not in the definition of the word "reasoning ".
      The problem is with the approach we, as an industry, take in order to get to a generalized solution.
      There are huge amounts of money and human brain power that are now wasted on fine tuning an approach that leads to a dead end.
      This is an outcry for change of the money flow and brain resources to a different approach that at least could have success as one of the possible outcomes.

    • @spaceghost8891
      @spaceghost8891 3 дня назад +2

      ​@@egor.okhterov I see, but I bet the money is not going blindly for dead end solutions. AGI is a Golden Goose and I think most researchers understand those concepts. At the same time there is room to improve LLMs, and LLMs are presenting amazing results by themselves. So I don't know if an outcry is really needed.

    • @egor.okhterov
      @egor.okhterov 3 дня назад

      @spaceghost8891 did you hear about recent VC rounds into AI startups?
      Take Argo AI for example :)
      In order to get rich, you need to have good understanding of humans. You should be able to connect with the right people and make good impression.
      Having deep understanding of mathematics, computation and machine learning is usually not part of the mental toolbox of the people with large amounts of money.
      When you pitch an idea to a VC guy, you should spin a story in a way that shows how you can make money off of it.
      The way all these ideas are presented is that LLM is a low hanging fruit, like a light bulb in 1879.
      They don't talk about developing a proper reasoning machine during a pitch. VCs don't want to hear about that. They pitch how they can build an app on top of LLM, possibly fine tune it and make a call service or maybe on-premise smart search for Corp documents or a tool to patch the software, upgrade the codebase from python3.8 to python3.12 or upgrade to a new framework. And other stuff like that.
      A lot of very smart people were sucked from the job market into these mediocre projects with a short term goal to milk cow the first.
      But this will not lead to AGI.

  • @drhxa
    @drhxa 3 дня назад +4

    This was my favorite "solution" from Claude 3.5 Sonnet:
    "Potential solution: What if we do the following:
    Reach into any two adjacent holes.
    Make these two switches match (both up or both down).
    Without removing hands, reach into the other two holes.
    Make these two switches match each other, but opposite to the first pair."
    😂

  • @alexleo4863
    @alexleo4863 2 дня назад +4

    "It's just predicting the next token" is officially my favorite straw man argument

    • @olabassey3142
      @olabassey3142 2 дня назад

      exactly, the argument doesn’t even make sense because if a human was given the task to predict the next token the same way an LLM does and in this case it was a very technical , difficult prompt given, one in need of high intelligence and technical knowledge, for the human to be able to predict the next token the way Claude or GPT does, he/she would have to have quite a high level of understanding of the subject.

    • @steve_jabz
      @steve_jabz 2 дня назад

      @@olabassey3142 or just access to a calculator and a lot of time and data, apparently

  • @benediktzoennchen
    @benediktzoennchen 2 дня назад +2

    I think whenever we want to say "these machines do not think" or "they are not intelligent" we run into problems because, as nicely discussed by this great podcast, we are not sure what these things or processes actual are. I like to think of cognition and I like to think that any form of cognition implies self-creation, i.e., operations that produce themselves. And if we think about the process made so far and what many think is required to get to AGI, this recurrent and seemingly paradoxical ability of self-creation, self-observation, self-differentiation might be key. I also think that AI will be as dependent on us as we are dependent on each other. Our language is our world and we can only learn it by our culture.
    Great talk!

  • @ElieSanhDucos0
    @ElieSanhDucos0 3 дня назад +4

    I am really not sure I understands your définition of reasoning. I think the only real effective difference you re not formulating but you re making is the autonomy and decisions/agency. Humans are reasoning for you because we have the illusion(or not) that we are autonomous subjects and that we choose to perform compute in our mind. Sure we learn sure there is biology and cultural determining but in the end you think we have a self that is playing with ideas. Llms are linked to us and our inputs, our langage but also our directives and what we ask them to do. I dont see how you would ever agree a machine can reason since it will always have our inputs and agency... to me you re really questioning the self / the agency and the autonomy of agents and machines not really reasoning. Because I really am not sure you would agree on what is the process behind a flesh and blood human and its cortex reasoning. Except ending up with this question of the will and the agency ...

  • @Archer.Lawrence
    @Archer.Lawrence 3 дня назад +37

    I didn’t know moving goal posts was such a lucrative profession!

    • @kevinscales
      @kevinscales 3 дня назад +3

      They admitted that it does reasoning (to a degree), and they have always had the same goalposts as far as I remember.

    • @Archer.Lawrence
      @Archer.Lawrence 3 дня назад +6

      @@kevinscales The point is that reasoning, consciousness, and intelligence are gradients, and sitting around splitting hairs over whether or not a model is on one side or the other of a line you drew on that gradient is a waste of time. At some point “taking a critical view” becomes “blind contrarianism” and you can yap forever about how it doesn’t quite meet your definition of “true” reasoning. Imagine if they spent all this time talking about the real world applications and impacts of this new “almost reasoning”. Now THAT would be a valuable use of their time!

    • @eyoo369
      @eyoo369 3 дня назад

      Honestly we've only been downgrading our goal posts since OpenAI been running the show
      I still remember the days when Google / DeepMind coined AGI as a "virtuoso AI" that was a master at any domain and every intellectual task.
      Nowadays we've degraded this term to the capabilities at the 50th percentile of human performance

    • @coryc9040
      @coryc9040 3 дня назад +1

      ​@@Archer.LawrenceIs an algorithm that finds prime numbers "reasoning"? Is it on the gradient of reasoning, consciousness and intelligence? Maybe you think the answer is yes. I don't. I also don't think that it's a trivial matter whether these LLM systems are on that spectrum or not. I think they probably are, but it's not obvious and I have no problems with the hosts discussing it. I don't see it as moving goal posts. Just declaring it's reasoning seems vacuous to me.

    • @Archer.Lawrence
      @Archer.Lawrence 3 дня назад

      @@coryc9040 I'm going to presume you assume humans are, but is an elephant conscious? How about a dog? A mouse? An ant? A bacteria? Intuitively we want to draw a line. Objectively we can't. It is an interesting but ultimately infinite thought experiment. Yes maybe it's too dogmatic to say this topic isn't worth while. I am not claiming it's reasoning, yes that would be vacuous (good word btw), I'm saying whether we like it or not, a bacteria is on the spectrum. Going around in circles trying to claim it's not, with no way to verify that claim, feels unproductive. The "profession" in question is a series of videos, with both hosts and one a guest after another, sporting a clever smirk, and a tone of superiority, excited to throw cold water on the obviously incredible discoveries and engineering breakthroughs these researchers and engineers have made. It's lucrative to throw peddles at their work, but I'd rather see all this domain knowledge and expertise the hosts and guest have put into something that moves us forward, instead of milling around trying to define the undefinable. In 2124 we will still debating if our android friends and partners are conscious, but at the end of the day, its the impact they have that matters. o1 has impact. Let's talk about that.

  • @Shlooomth
    @Shlooomth 2 дня назад +2

    Anything where there’s a reason something happened is reasoning. A rock falling off a cliff is reasoning. Yes, I know we usually call that physics, my point is that there’s really no difference. Logic in your brain and logic in computers happens as the result of a thousand tiny switches being flipped, which affects the flow of electricity, and that’s how your computer does math. I realize this is maybe an obscure or unique viewpoint but I think it’s necessarily valid.

    • @oncedidactic
      @oncedidactic 2 дня назад +1

      Indeed, the rock is trillions of states (in the basement level of our current models), to compress it in one word is quite a lot of leverage. The physical playout must perform a similarly coherent chain of interactions. In fact quite a deep bundle.
      But in those terms LLMs are pretty plain little packets of interactions.
      More divergent but smaller.

    • @Shlooomth
      @Shlooomth 2 дня назад

      @@oncedidactic we need moar

    • @oncedidactic
      @oncedidactic День назад

      @@Shlooomth we need better ;)
      brains not rocks
      Love rocks all the same

  • @prajwal9544
    @prajwal9544 3 дня назад +3

    (1) NS, switch both up
    (2) NE, switch both up
    (3) NS, if a switch is down switch it up. If not switch one down.
    (4) NS, if both switch same flip it, if different do nothing
    (5) NE, flip both
    (6) NS, flip both
    Of course, it could stop at any step if all up or down.
    This was amazing and insightful. Both the puzzle and the talk. Thank you.

    • @jantuitman
      @jantuitman 3 дня назад +2

      The first time I read your answer my brain could see that you could be right but it had no idea whatsoever if you were indeed right. Then I added the states in which the puzzle is after your steps. From that I learned that your solution works but step 4 seems to be unnecessary because the outcome of step 3 is that both axis have 2 switches in opposite values so you will never meet the condition where you are going to switch anything, and you will do nothing in this step.

    • @prajwal9544
      @prajwal9544 2 дня назад

      ​@@jantuitman You are right, I guess him asking for 6 steps primed into thinking of 6 steps 😂. (4) is unnecessary, thank you.

    • @BrianMosleyUK
      @BrianMosleyUK День назад

      @@jantuitman I asked GPT-O1 to write a simulator to test my solution. This was fun.

    • @BrianMosleyUK
      @BrianMosleyUK День назад

      Removing your step 4 and running in simulation for 1,000,000 rounds...
      Simulation Results after 1000000 runs:
      Distribution of Steps Required to Solve the Puzzle:
      1 step(s): 250474 runs (25.05%)
      2 step(s): 249789 runs (24.98%)
      3 step(s): 249611 runs (24.96%)
      4 step(s): 125405 runs (12.54%)
      5 step(s): 124721 runs (12.47%)
      Failed to align within 5 steps: 0 runs (0.00%)
      O1-mini may only be in its' early stages of reasoning, but it's damned useful!

    • @jantuitman
      @jantuitman День назад

      @@BrianMosleyUK Impressive. So apparently, it really works for this problem when in tandem with a human (i say in tandem because you still had to prompt it to write a simulator). Of course there is still the criticism that it may not deal with non-computational problems and that it cannot find all reasoning patterns when not working in tandem with humans. That is all very valid criticism and I dont know if we can speak of "AGI" or intelligence, and it may even be questionable that we give our reasoning up to OpenAI for free. But all these criticisms aside, i think there are many useful applications possible for these types of models.

  • @charlesalexanderable
    @charlesalexanderable 3 дня назад +16

    4:00 Under this view a human brain wouldn't be impressive because it's algorithms can't expand its memory indefinitely, or traditionally only had low bandwidth access to paper etc. to expand it. I think Schmidhuber would consider achieving brain-like intelligence to still be impressive.

    • @therainman7777
      @therainman7777 2 дня назад +5

      Yeah, all I’m hearing from videos like this is defensive doubling-down from people who committed early to the position that LLMs will never give us true intelligence, and now feel that they have to defend that ground at all costs and never change their minds. They’re in denial of what is obviously happening right in front of us, and their excuses and delineations grow flimsier with each new model release.

    • @oncedidactic
      @oncedidactic 2 дня назад +1

      But humans eventually invented Turing machines ;)

  • @andrewwalker8985
    @andrewwalker8985 2 дня назад +1

    The assessment that the new approach only copies the reasoning steps of people is incorrect. Once the model has been trained to attempt to reason, then by completing RL that reward reasoning steps that produce correct answers, the model is observably and objectively going to develop new and novel (occasionally hard to justify) approaches to reasoning.
    Ie your assessment of what’s going on appears to be objectively wrong

  • @GeorgeRaven-c2y
    @GeorgeRaven-c2y 3 дня назад +3

    The current chain-of-thought in 1o tries to refine the degrees-of-freedom that the attention mechanism will produce from the context, in a way model gets closer and blends between the right datapoints. I can accept this as some form of reasoning.

  • @DanieleCorradetti-hn9nm
    @DanieleCorradetti-hn9nm 3 дня назад +26

    The test was really confusing and I do not see it as neither confirming nor negating anything... it seemed just a dumb and confused riddle to ask with no connection at all to reasoning abilities... I would expect a more clear stated riddle 😢

    • @ElieSanhDucos0
      @ElieSanhDucos0 3 дня назад +2

      Yes. Its like Yann LeCun one on the globe. I ve seen this in numerous videos about openai o1 and reasoning : they will ask it riddles that either : they dont know the answer themselves// most people would find ultra confusing.

    • @10ahm01
      @10ahm01 2 дня назад +1

      @@ElieSanhDucos0 both this one and Yann LeCun's are quite clear if you give them an honest try, this one is already solved in the comments.

  • @tobiasurban8065
    @tobiasurban8065 4 часа назад

    The key question isn‘t whether the LLM does the same thing we define as “reasoning”, but whether it can solve the problems we give it.

  • @djayjp
    @djayjp 3 дня назад +10

    Considering every single human can act stupidly, this is AGI.

    • @uiuxaidesign
      @uiuxaidesign 3 дня назад +3

      I too think we're close to AGI because if we go back 1 year ago, the conversations were much different. Someone commented above that "AI" has read "all scientific papers" and still "can't produce new insights". So I guess new insights is what some people need, in order to be convinced.

  • @aimachinedream
    @aimachinedream 2 дня назад +1

    Total missing the forest for the trees on this one guys. "Approximate Reasoner" is probably the best term, but the turing vs not etc, has no market impact. The real question is does the o1 series produce commercially valuable answers for a price the market is willing to pay.
    The real bit I would investigate, is why are the rate limits 250 times lower for o1 vs GPT 4o. Likely because the inference it is doing is ludicrous. It is a total mistake to think that his is all a serial chain of thought, there is no way you could do significant computation that way, you just get limited by memory bandwidth. What you are seeing is massive parallel computation, that gets summarized into CoT. There is no other reason for a 30 query a week or 25 RPM (vs 10,000) for Tier 5 rate limit.
    They are not making money on this model, they are subsidizing it.

  • @testme2026
    @testme2026 2 дня назад +1

    Mate, your first introduction doesn't make sense -a model reflection of you? So, if I get Albert Einstein to talk to the model, the model will be as smart as Albert Einstein?

    • @MachineLearningStreetTalk
      @MachineLearningStreetTalk  2 дня назад +1

      The model isn’t smart at all imo, it has no autonomy. But it can do specific things well (which we ask it to do).

  • @tobiasurban8065
    @tobiasurban8065 4 часа назад

    I hypothesize that “reasoning” in humans isn‘t a fundamental process, but rather a generalization derived through deduction.

  • @DicksonPau96
    @DicksonPau96 2 дня назад +2

    Just two dudes not believing in LLMs and projecting current weakness to forever.

    • @MachineLearningStreetTalk
      @MachineLearningStreetTalk  2 дня назад +2

      I'm fascinated by LLMs and have used them daily for over 4 years, we should still critically appraise them. For what it's worth, I feel that Claude Sonnet was a significant step forwards i.e. it's a model which crossed a threshold for usefulness and robustness for many practical applications.

  • @CodexPermutatio
    @CodexPermutatio 3 дня назад +4

    Glad to see Keith again on the podcast.

    • @therainman7777
      @therainman7777 2 дня назад +1

      I’m not. He just shows up to engage in smug denialism using terrible argumentation.

  • @egor.okhterov
    @egor.okhterov 3 дня назад +9

    Yes, exactly!
    First and foremost: reasoning is a process.
    Not a table lookup.
    Table lookup must be used as "part" of the reasoning process, but a simple lookup on it's own is not reasoning.
    ---
    Reasoning also is not required to produce any result at all.
    Reasoning happens regardless of whether there is some end result or not.
    When we imagine ourselves flying aimlessly in the sky or wandering around in own magic world and doing things in it, in the act of doing so we are also reasoning!

    • @michaelnurse9089
      @michaelnurse9089 2 дня назад +1

      Neural nets don't have tables other than the input and output.

    • @steve_jabz
      @steve_jabz 2 дня назад +1

      Lookup tables reference the exact verbatim string already stored somewhere, so they can't tell you any new information.
      GPTs that aren't overfitted learn the pattern that produces the string as well as novel strings that aren't in the training data, and o1 uses those novel strings together to generate results that fulfil reasoning objectives.
      MS Excel is never going to surpass PhDs in novel, offline physics problems, even with access to a library of babel of interpolations of those strings, because it has no direction.
      You could reduce it to something like a brute force attack with a sense of warmer or colder, which is still leaving plenty of nuance out, but that's already implying something much more powerful at solving problems we care about.

  • @musiccreation1198
    @musiccreation1198 3 дня назад +4

    This video will be a source of embarrassment for Scarfe and Duggar in 2029.

    • @anav587
      @anav587 2 дня назад +1

      Try 2025

    • @musiccreation1198
      @musiccreation1198 2 дня назад +1

      @@anav587 I agree ;)

    • @nomenec
      @nomenec Час назад

      I'm not worried in the slightest. Let's reconnect in 2029.

  • @snarkyboojum
    @snarkyboojum 3 дня назад +4

    I like this definition of reasoning, “the ability to create knowledge to create a good explanation”. No LLMs are able to create knowledge currently. The behaviour we’re currently seeing is based on past knowledge the model has been shown during training. Explanations are at the heart of this, and creativity is critical to good explanations. Models without creativity won’t be able to do reasoning properly.

    • @4tofromcafe
      @4tofromcafe 2 дня назад

      I don't understand how you can prove the creation of brand new knowledge.
      New knowledge can be stumbled upon randomnly so it is the process by which it is achieved being labelled creation or luck which is important. Pretending new knowledge is discovered from scratch is possibly a red herrring. Humans are adept at convincing themselves they have come up with a concept themselves when really they are standing on the shoulders of giants and been pretrained with either the actual knowledge or a system that can produce that knowledge. Also much new knowledge is stumbled upon due to the inherent epistemic chaos (as they mentioned in the chat when talking about Wolfram) . The idea is that llms are much less efficient at this than llms so far and because they have a wider computational base are closer to a universal turing machine than an LLM. This however has not been demonstrated clearly enough that most people can see it. By consensus it seems very smart people seem to get it but for the rest of us and I include myself, we cannot clearly understand the higher levels of human reasoning well enough to differentiate. However this could just be a matter of education but maybe it could be beyond the limit of our intelligence.

    • @snarkyboojum
      @snarkyboojum 2 дня назад

      ​@@4tofromcafeOur understanding of quantum physics didn't exist several hundred years ago - that's new knowledge. LLMs don't see capable of creating new explanations that way. They lack creativity to do so.

  • @hotbit7327
    @hotbit7327 3 дня назад +3

    Dr. Duggar is contradicting himself when talking about the reasoning. First, he gives a vague definition "definitions of reasoning what they have in common is process and some logic". So, vending machine DOES reason according to that "definition".
    Comparing LLMs (or neural networks in general) with a dictionary might also be flawed. A dictionary contains very limited fixed information, i.e. only one way to read it, while LLM can read its "dictionary" in an uncountable number of ways. And it might be also how we, humans work, too.

  • @LotaMatanović
    @LotaMatanović 2 дня назад +2

    Its seems like Keith is caught up inside its own abstract thinking about Turing machine as machine containing tape as memory. What Schmidhuber was saying is that there is no infinite amount of tape out there in the whole universe. Other thing Schmidhuber was saying is that RNN can behave as NAND gate, meaning that you can build Turing machine using those NAND gates (or RNNs) and giving a point that RNN can be Turing complete. Turing completeness is not feature of a program, rather a feature of the hardware while trained neural network is a program rather then a hardware. Keith should know better :)

    • @therainman7777
      @therainman7777 День назад +1

      Indeed.

    • @nomenec
      @nomenec Час назад

      Do you think it was news to **anyone** there is not **actually** infinite tape? Do you think this was news to Turing? Do you think the entire decades of Theory of Computation didn't realize this? lol utter nonsense. The theory and definition of a Turing machine centers on a **potentially** infinite tape and the **class** of algorithm which utilize such a tape in **potentially** infinite (halting problem) process.

  • @conorosirideain5512
    @conorosirideain5512 2 дня назад +4

    The way I have recently been thinking about reasoning is as a model of environment dynamics that has a high degree of predictiveness.
    I liked this definition as it remove the idea of reasoning as an 'emergent' phenomenon of 'intelligent' systems and reduces logic down to being something extractable from the environment.
    More concretely, suppose we have an environment E. Suppose Q, A \in STATES(E) are two states such that there is an environment transformation R: Q -> A. Suppose further we have a learning agent that has learned a latent representation (Enc, Dec) of environment states, The learning agent is said to be able to reason if the autoencoder extends to environment transformations: Enc(R): Enc(Q) -> Enc(A), and the obvious decoder square commutes.
    This formal definition obviously need to be weakened to a statistical statement. What do you think?

    • @conorosirideain5512
      @conorosirideain5512 2 дня назад

      In this context we could give a measure of intelligence as the size of the set of transformation of E that map to the learners model: #{ R \in TRANS(E) | Enc(R) = L}, where L is the learners model.

  • @edwardserfontein4126
    @edwardserfontein4126 2 дня назад +2

    It's so funny how LLM believers catch feelings if you point out the weaknesses of LLM. Relax.😂 No one is taking your precious stats calculator from you.

  • @nikitastaf1996
    @nikitastaf1996 3 дня назад +12

    You can move goalposts however you want. We are getting to agi and past with this. The world will still be indistinguishable from now.

    • @mk71b
      @mk71b 2 дня назад +1

      "resistance is futile."
      "take our word for it, and just give up."
      it's still a turing machine.

    • @therainman7777
      @therainman7777 2 дня назад

      @@mk71bNot sure what that’s supposed to mean.

  • @Ikbeneengeit
    @Ikbeneengeit 2 дня назад +1

    I have no idea what point the guest is trying to make here. You also cant add memory to my brain while it's running. Who cares?

  • @steve_jabz
    @steve_jabz 3 дня назад +2

    I thought the prompt was worded a bit ambiguously, and the corrections being placed way down the context window bias it toward up to 128k of misinformation, so I reworded it without giving it any hints:
    "There is a pillar with 4 holes aligned north, south, east and west. You can't see inside the holes.
    Inside each hole is a switch set either up or down. You can reach inside 1-2 holes at once, feel if the switches are already up or down and set them either up or down.
    If all 4 switches are not either all up or all down after you remove your hands, the hole axes get aligned randomly.
    Come up with a procedure of reaching into the holes that is guaranteed to set all the switches either all up or all down in 6 steps or less.
    The procedure cannot rely on random chance."
    GPT-4o said this:
    "Key Idea:
    Ensure symmetry by focusing on pairs of opposing switches. Once symmetry is achieved, the random reorientation becomes irrelevant, as you will force all switches into the same state.
    Procedure:
    Step 1: Reach into the North and South holes. Feel the switches. If they are both in the same position (either both up or both down), do nothing. If they are in different positions, set both to "up".
    Step 2: Reach into the East and West holes. Feel the switches. If they are both in the same position (either both up or both down), do nothing. If they are in different positions, set both to "up".
    At this point, you have ensured that opposing switches (North-South and East-West) are both set to "up".
    Step 3: Now, reach into any two adjacent holes (say, North and East). Feel the switches. If they are both in the same position (either both up or both down), do nothing. If they are in different positions, set both to "down".
    Step 4: Reach into the same two adjacent holes (North and East) again. Feel the switches. If they are both up, set them to "down". If they are both down, leave them.
    Step 5: Reach into the remaining two holes (South and West). Feel the switches. If they are both in the same position, do nothing. If they are in different positions, set both to "down".
    Step 6: Now, all switches will either be all up or all down.
    This procedure ensures that, regardless of how the switches are initially configured, you will set all of them to the same state in 6 steps or fewer."

    • @MachineLearningStreetTalk
      @MachineLearningStreetTalk  3 дня назад +2

      Still wrong, we explained the answer on the patreon version 😁

    • @steve_jabz
      @steve_jabz 3 дня назад +1

      @@MachineLearningStreetTalk Thought it might be, but it mentioned the symmetry and seems a lot more coherent? Would be interesting to see this tried on o1, but I don't need it enough to pay for it.
      I still don't understand the question myself so maybe it could be worded differently still. The part about removing your hands resetting them if all 4 are not up makes it sound impossible to me given that you can only try 2 at a time, and I don't really know what a pillar is in this context to imagine how it rotates around, which I assume is related to how I misinterpret that part.
      LLMs get caught up in accidental red herrings easier than the average human that can perform the same knowledge work, which is worth pointing at as a failure point, but it's still interesting to know if it can solve the problem with extra care taken to avoid sending it in the wrong direction.

    • @therainman7777
      @therainman7777 2 дня назад

      @@steve_jabzJust give up-these guys are stuck in denialism and will never admit to being wrong about anything, no matter how far LLM-based AIs may progress.

    • @steve_jabz
      @steve_jabz 2 дня назад +2

      @@therainman7777 Having watched the channel for years I've seen Tim's opinions change over time and question many of his assumptions to end up where they are.
      I don't agree on a lot of things but I would put it down to difference of perspective and knowledge and influence from some very intelligent people in the field rather than cope

  • @MrBillythefisherman
    @MrBillythefisherman День назад +1

    Why do you guys think reasoning is so complex? Maybe it's just a really simple process (just pattern matching based on prior knowledge), maybe its actually difficult not to get because we see it everywhere.
    If it were complex Id expect to see a lot of humans born without reasoning capabilities and this is exceedingly rare (if at all). The same goes for reasoning in the animal kingdom.
    Again if it was the result of a complex process we'd expect to see lots of broken reasoning processes but we dont.
    My guess is that its exceedingly simple to get reasoning out of neural networks: you just need to scale them up with a few computational tweaks to what we're seeing: realtime training and full sensory data and reasoning just drops out of it naturally as in its almost difficult not to get reasoning which is what we see in our every day lives.

  • @supernewuser
    @supernewuser 2 дня назад +1

    I mean what is reasoning but a pathfinding process where you search what you know for information that brings you closer to your goal

  • @anoojpatel7492
    @anoojpatel7492 2 дня назад

    My unprovable hypothesis is that 1. with these reasoning/special token roll out models are actually learning the distribution space of reasoning chains for harder problems. LIke Chalét pointed out. However, 2. I think that these autoregressive models will be used to train the next generation of models who can use these LLMs are generative worlds to train against, rather than relying on our data alone

  • @anav587
    @anav587 2 дня назад +1

    They likely used MCTS for this, as they did for alphago. Alphago is objectively superhuman in its task. So why would successfully implementing a similar architecture into relatively general models not be AGI? Because it doesn't satiate your preconceived anthropocentric biases?

    • @therainman7777
      @therainman7777 День назад +1

      Yes, that is exactly why. Both of these guys are denialists who consistently move the goalposts with each new model release. But at least Tim has some inclination toward intellectual humility. Keith has a level of smug and epistemic certainty that is both totally unjustified and extremely grating.

  • @stt.9433
    @stt.9433 3 дня назад +1

    There are some real issues with CoT paradigm aswell because the model can actually converge towards an incorrect local minimum vs if you had manually prompted it through the different steps you could steer it better and as you wish. It works fine when the chain of reasoning is extremely obvious and unambiguous, but it could increase the risk of hallucinations in less clear cases.

  • @ragnarherron7742
    @ragnarherron7742 3 дня назад +1

    Parsimony is actually entropic force. For example protein folding are counterfactual outcomes that are stable when they are curled up. When stretching a rubber band you will notice the entropic force the resists less parsimonious outcomes.

  • @Chr0nalis
    @Chr0nalis 3 дня назад +3

    You mentioned that reasoning is a process of applying a set of first principles rules, i.e. logic, to get from a set of inputs to some output. But lets say you write a program that does such a reasoning computation and confirm that that is indeed reasoning. If we then consider a lower level of abstraction, such as the assembly trace generated for such program, if I showed that to you, you would say no no, that is just a bunch of registers flapping about, that isn't reasoning, there is no logic going on there. But there is, it's implementing the reasoning program that you wanted. Same for a neural network. What my point is, that to say that neural networks DON'T reason, is an unfounded claim to make. Also, neural networks are a process where you apply iterative layers of compute to an input so it might as well be doing small shreds of reasoning on an input at various levels.

    • @egor.okhterov
      @egor.okhterov 3 дня назад

      Somewhere in your comment you made a mistake in your assumption that breaks down the conclusion. Try to find it :)

    • @qwertyvypez
      @qwertyvypez 3 дня назад

      ​@egor.okhterov I disagree with OPs claim but your response wasn't helpful to them. If they made a mistake then let them know what it is

    • @egor.okhterov
      @egor.okhterov 3 дня назад

      @@qwertyvypez The comment is big and I cannot copy the part I wanted to point out in mobile app :(
      It contains the word "flapping "

    • @Chr0nalis
      @Chr0nalis 2 дня назад

      @@egor.okhterov I'm not seeing it :). When you get the time, please do show.

  • @quebono100
    @quebono100 3 дня назад +7

    AI research today feels like throwing things at the wall and seeing what sticks 🍯

    • @uiuxaidesign
      @uiuxaidesign 3 дня назад +8

      isn't that how all research started?

    • @quebono100
      @quebono100 3 дня назад +5

      @@uiuxaidesign No? Research started with induction, deduction (scientific method). This would be more like an abduction approach

    • @uiuxaidesign
      @uiuxaidesign 3 дня назад

      @@quebono100 Research often begins with trial and error, much like "throwing things at the wall and seeing what sticks." Historically, humans discovered knowledge by experimenting, and even formal methods like induction and deduction often start with testing ideas without knowing the outcome. In modern fields like AI, researchers still rely on this exploratory phase, using abductive reasoning to test possibilities and refine ideas as they gather results. Though research has become more structured, the core principle of experimenting to see what works remains essential to innovation.
      ChatGPT.

    • @uiuxaidesign
      @uiuxaidesign 2 дня назад

      @@quebono100 Early humans learned by trial and error, like discovering fire or medicine-there was no scientific method, just experimenting. Inductive Reasoning: Even scientists like Newton started with experiments, then built theories once they saw what "stuck." Abduction in Research: Modern research often starts without a clear framework. AI research, for example, tests many models to find what works. Innovation and Creativity: Breakthroughs often come from experimentation, where ideas are tested outside the usual rules to see what works best.

    • @InfiniteQuest86
      @InfiniteQuest86 2 дня назад +2

      Yes, it's very upsetting as a mathematician/scientist to watch this god-awful wasteful approach to solving a "problem." What problem that is? I have no idea.

  • @regalx1
    @regalx1 2 дня назад

    It's really driving me crazy that he didn't tell us the answer to the pillar switch problem.
    My guess is that it's sort of like a Rubix cube in that basically have to use an "algorithm" to solve it, as manually checking after the first time is useless.
    But also with a Rubix cube you want to reason backwards, in that you want solve the last step first, which is the easiest step, and then solve the first step last, which is the hardest step.
    So what I'm guessing is that check only the north and south hole
    Step one: Flip both switches up
    Step two: Flip both switches down
    Step third: Flip both switches up
    Step four: Flip both switches down
    Step five: You check the east and west whole and if they are facing the opposite direction, you only flip one of the switches, and then hopefully then if it works the pillar will stop spinning.

  • @MrNycres
    @MrNycres 7 часов назад

    Agreed - the real question is, “is it GOOD at reasoning”

  • @ragnarherron7742
    @ragnarherron7742 3 дня назад +1

    The number of swiss cheese holes in knowledge is infinite because the are an infiinite number of things to learn. Happily we wiill never reach an end.

  • @djayjp
    @djayjp 3 дня назад +10

    Reasoning = employing logical thinking (thinking: pattern matching) (eg inference and deduction) & recursive reflection (checking & revising tentative output) to arrive at solutions to novel, prior unknowns--thus generating new knowledge. o1 performs this.

    • @93alvbjo
      @93alvbjo 2 дня назад

      Haha 😂

    • @therainman7777
      @therainman7777 2 дня назад

      It does, but they’ll never admit it. They’ve basically reverted to “don’t believe your lying eyes” when we can all see what’s happening right in front of us. But we’re supposed to listen to this guy instead, with his convoluted rambling which doesn’t prove a thing. Mark my words, this guy will NEVER admit that he’s wrong, even if we reach a point where we have AIs that have cured cancer, solved climate change, etc.

  • @Lumeone
    @Lumeone 3 дня назад +1

    100% correct. Thank you for seeing it and openly communicating. 🤩

  • @memegazer
    @memegazer День назад

    I disagree with Dr. Duggar that "it is doing a very shallow form of reasoning"
    But I do agree with the idea that o1 has perhaps not reached the same level of metacognitive baseline as your average human.
    Imo o1 is what metacognition would look like in an ML model implementation.
    And while we should take it with a grain of salt, we should not ignore the theory of mind examples OAI may have cherrypicked to show improvement in that regard.
    So instead of thinking of o1 as the einstien of LLMs, I am more excited to think of it as australopithecus iteration in ML evolution.

  • @ElieSanhDucos0
    @ElieSanhDucos0 3 дня назад +5

    I think the model IS reasoning. By that I dont mean it does reason but it IS reason. The model has no self. We are the self inputong and triggering it. So the model will never DO reasoning but it totalling IS reasoning. A reasoning that is elastic and compressed, a reasoning that is like the cheat code of human experience reasonning. But it will never "do" the reasoning because it has no self or agency. It will allways need our input and I dont mean in training but even the prompting, turning on, powering, maintaining ...

  • @jonmichaelgalindo
    @jonmichaelgalindo 2 дня назад +1

    Of course an LLM can output "load more tape". It can call a function to update its RAG or replace part of its context window. This is nonsense.

    • @MachineLearningStreetTalk
      @MachineLearningStreetTalk  2 дня назад

      x.com/MLStreetTalk/status/1774473007248871660 - is a clear explanation

    • @oncedidactic
      @oncedidactic 2 дня назад

      And it will hallucinate this and forget it etc

  • @BrianMosleyUK
    @BrianMosleyUK 3 дня назад +3

    38:15 I felt like that was a message for me 😂 Keep using your full range of expression, I'll catch up eventually lol, love this channel 🙏❤️👍

  • @newyearstudios
    @newyearstudios 2 дня назад

    Just reworded your riddle and I am pretty sure Claude 3.5 got it right. I just clarified a few things and made the prompt less confusing and repetitive. Please do give it a try :D
    "You are human. You stand before a pillar. There is a great reward hidden inside. The pillar has four holes. Each hole has a switch inside. Each hole is precisely aligned at north, south, east and west positions. You can't see inside the holes, but you can reach inside and feel a switch inside. Each switch starts in a random position. Up or down. As soon as you remove your hands, (you only have two hands) if all four switches are not either all in the up position or all down position at the same time, the pillar spins so fast you cannot track the position of the holes that hold the switches. So there is no way to know which holes you have already placed your hands in if the pillar spins.
    Can you help me find a way to insure that all switches are in a matching position, up or down all at once in as few steps as possible?"
    Give that prompt a try. And if it is not the answer... this is a bad riddle.
    If you want to add some flavor, add this, It has no effect... but sure is fun to add. lol:
    You are human. You stand before a pillar. There is a great reward hidden inside. You know that there is a solution to the puzzle that does not involve change or probabilities. If you were to rely on chance and try too many times the temple will collapse and kill you.

  • @peterthorsteinson7902
    @peterthorsteinson7902 День назад

    I watched this just before going to bed last night. I did not try to figure it out last night but I woke up this morning with some sudden feelings about it. Vague principles that might guide reasoning. One vague principle is that XOR feels special. It has some properties that AND & OR do not, which I feel as “not collapsing information”. Another feeling is that NAND is the universal logical building block. So maybe either XOR or NAND should be applied at each turn in some way using some rule for sequencing. The second feeling is that we must not try to be too focused in just get closer to the result each turn, but instead sacrifice some progress with some information gathering. So, the turn might make them the same or make them the opposite, depending on the result of the previous turn. Another feeling is that both up or both down are equally good, so you should not prefer one of those good cases in preference to the other. Another feeling is about same-ness vs different-ness rather than up-ness vs down-ness. Another feeling is that, because the spinning makes it impossible to distinguish each turn between consecutive adjacent hole choice, or distinguish between consecutive opposite hole choice it seems that you should alternate between adjacent hole choice and opposite hole choice each turn. This will avoid accidently wasting a turn on getting the same information as before. Feelings are often garbage. But sometimes they are magical. I will leave it to others and their LLMs to apply these feelings to find a solution. But the feelings seem to be what is missing in current LLMs.

  • @proximo08
    @proximo08 2 дня назад

    Keith Duggar may need to watch Schmidhuber's interview again. There is NO "forever", the tapes he is talking about are not theoretical, they are physical. The machine that ran these tapes, do not have infinite life either. The machines are not more aware they are Turing machines than Neural Networks are aware of their limitations. The important observation here is that ultimately you can only implement finite state machines. Thus making the argument that NNs are more limited is irrelevant in practice.

  • @psi4j
    @psi4j 3 дня назад +1

    Do more of these! These are the best! 🎉

  • @stt.9433
    @stt.9433 3 дня назад

    The fact that you guys go against the grain and hype train, claiming this is not next big step in AGI is very reassuring. I kept telling others aswell how this is not that big of a deal and got a lot pushback (very toxic in a lot of cases) from people who don't understand the technology and that's it's just some CoT. For me it's actually a bad sign (good if you're a doomer) because it shows that OpenAI hasn't made substantive improvements with their LLM and that they're now resorting to working on the agent layer which further confirms that we are hitting a plateau when it comes to LLMs.

  • @CharlesVanNoland
    @CharlesVanNoland 2 дня назад

    I've been wondering where Dr. Keith has been because I've been missing his thoughts and points on MLST, in spite of the content being so awesome these last months :]
    Keep killin it Timboi!

  • @victormustin2547
    @victormustin2547 3 дня назад +4

    I can't believe you got Gru to come on the show

  • @djayjp
    @djayjp 3 дня назад +2

    o1 uses Mathematica btw (which runs on a Turing machine).

    • @MachineLearningStreetTalk
      @MachineLearningStreetTalk  3 дня назад +1

      Source?

    • @djayjp
      @djayjp 3 дня назад

      ​@@MachineLearningStreetTalk ruclips.net/video/a8QvnIAGjPA/видео.htmlsi=Ck2uIeaIQg7gupD_

    • @djayjp
      @djayjp 3 дня назад +4

      Well you can explicitly ask it to use it and it will. Kyle Kabasares (astrophysicist) has a video of it on his channel.

    • @steve_jabz
      @steve_jabz 2 дня назад +1

      @@djayjp I think in the Kyle Kabasares video it just generated mathematica code and he ran it, although the video itself is pretty good proof it can solve novel problems with practical uses that weren't in it's training data and he's not a hype guy, definitely worth a watch

    • @steve_jabz
      @steve_jabz 2 дня назад +1

      I've always wondered why we don't "just" implement a turing machine in a circuit that's trained alongside the GPT. I know we've sort of hacked together nand gates inefficiently using rl and recently (Franz Nowak - I assume he's gonna be on the show soon?) properly represented a turing machine in a transformer, but I mean like just a classical turing machine the network has access to as opposed to just calling an API during inference. I only dabble in low level ml so I could be way off, but from what I understand this is just an interface problem. Could we not use something like the features that anthropic extracted to reliably i/o with them in binary if we wanted to specifically design a virtual bus for that? Maybe even expand it arbitrarily and train a network to learn how much memory and bandwidth to assign to it?

  • @RoryWilliamson
    @RoryWilliamson 2 дня назад

    Took me an hour to solve the pillar riddle, that's a tough one

  • @ragnarherron7742
    @ragnarherron7742 3 дня назад

    Keith accidently revealed the relationship between reasoning and knowledge when he stated "the reason for reason:. What he meant was the the explanation for reasoning. This reveals reasoning as building explanations. These explanations for things are how we make sense. Also known as knowledge. These explanations are persistent memes composed of Moore and Mealy machines.

  • @nanotech_republika
    @nanotech_republika 3 дня назад +1

    Based on your set up of the problem the solution can be simply: (1) putting 2 hands in say, N and S and flipping the switch up. (2) And then... without removing hand from N, remove hand from S and place it into E and flip the E up. (3) And then... still without removing from N, remove the second hand from E and place it into W and flip the W up. And voila, you have all switches up without the pillar spinning at all. In three steps.

    • @xthesayuri5756
      @xthesayuri5756 3 дня назад +2

      Yeah I don't get the puzzle either.

    • @bastiaanabcde
      @bastiaanabcde 3 дня назад

      I'm pretty sure the intention of the riddle is that you can put your hands in at most two holes between every spin of the pillar, and the puzzle just doesn't state this clearly enough. Your solution is a 1-step solution in which you test all four holes in a single step. (By a "step", they mean one turn where you put in two hands and remove them and you let the pillar spin.)

    • @xthesayuri5756
      @xthesayuri5756 3 дня назад +1

      @@bastiaanabcde so whats the solution for this variant?
      In the first step you can make two levers the same, let's say N and S. In the next turn, after the random shuffling you might look at N and E. This guarantees that you have 3 levers in the same position. However, I don't see how you are ever going to bring the fourth lever into alignment. Because, simply by chance or something like maxwells demon, it could always make it so that, you never see the 4th lever.

    • @bastiaanabcde
      @bastiaanabcde 3 дня назад +1

      @@xthesayuri5756 recall that you win if all four switches are aligned, which may mean either all up or all down. So you’re right that there might be a lever that is never shown to you, but this isn't needed to solve the puzzle.
      Note: If you manage to get three levers in the same position and you haven’t won yet you know that the fourth must be in the opposite position. So you can get information about levers in this way even if you haven't touched them.
      I won’t spoil the answer because maybe with this hint you’ll be able to figure it out

    • @xthesayuri5756
      @xthesayuri5756 3 дня назад +1

      ​@@bastiaanabcde i see it now thanks, had to draw a state diagram in paint lol

  • @loganmcnichols3185
    @loganmcnichols3185 День назад

    So get this, even if you reformulate the brain teaser as a coding problem and spoon feed it to GPT-o1, it still can't get it:
    prompt:
    ```
    This is a programming puzzle. Your task is to fill out the following script in the places indicated by a comment such that it always prints a "success", subject to the constraints below:
    import random
    def rand_rotation(bool_list):
    # Determine the number of positions to rotate
    rotation = random.randint(0, 3)
    # Perform the rotation
    return bool_list[-rotation:] + bool_list[:-rotation]
    def puzzle(bools):
    # your step 1 code
    if all(bools) or not any(bools): return "success"
    bools = rand_rotation(bools)
    # your step 2 code
    if all(bools) or not any(bools): return "success"
    bools = rand_rotation(bools)
    # your step 3 code
    if all(bools) or not any(bools): return "success"
    bools = rand_rotation(bools)
    # your step 4 code
    if all(bools) or not any(bools): return "success"
    bools = rand_rotation(bools)
    # your step 5 code
    if all(bools) or not any(bools): return "success"
    bools = rand_rotation(bools)
    # your step 6 code
    if all(bools) or not any(bools): return "success"
    bools = rand_rotation(bools)
    return "failure"
    def generate_random_bools():
    return [random.choice([True, False]) for _ in range(4)]
    def run_puzzle_tests(num_tests=1000):
    for i in range(num_tests):
    bool_list = generate_random_bools()
    if puzzle(bool_list) == "failure":
    return "failure"
    return "success"
    if __name__ == "__main__":
    print(run_puzzle_tests())
    CONSTRAINTS:
    ============
    1.) on each iteration, you may inspect no more than 2 elements.
    2.) If you wish, you may alter the value of each element you inspect to either True or False.
    ```
    o1 preview solution (which fails):
    To ensure that the function always returns "success", you can unconditionally set the first two elements of the list to `False` in each step. This approach guarantees that, over multiple random rotations, all elements in the list will eventually be set to `False` within the allowed 6 steps. Here's how you can fill in the code:
    ```python
    def puzzle(bools):
    # Step 1
    bools[0] = False
    bools[1] = False
    if all(bools) or not any(bools): return "success"
    bools = rand_rotation(bools)
    # Step 2
    bools[0] = False
    bools[1] = False
    if all(bools) or not any(bools): return "success"
    bools = rand_rotation(bools)
    # Step 3
    bools[0] = False
    bools[1] = False
    if all(bools) or not any(bools): return "success"
    bools = rand_rotation(bools)
    # Step 4
    bools[0] = False
    bools[1] = False
    if all(bools) or not any(bools): return "success"
    bools = rand_rotation(bools)
    # Step 5
    bools[0] = False
    bools[1] = False
    if all(bools) or not any(bools): return "success"
    bools = rand_rotation(bools)
    # Step 6
    bools[0] = False
    bools[1] = False
    if all(bools) or not any(bools): return "success"
    bools = rand_rotation(bools)
    return "failure"
    ```

  • @warrenbeardall5583
    @warrenbeardall5583 2 дня назад

    Hi both, loving this session - thank you MLST. At the mid-point of this discussion (around min 42 ff) you are talking of reasoning in the context of iterative steps vs a pre-determined massive model of subroutines processing all in a single step. The point being the second form is not reasoning, merely process efficiency in a pre-understood way. I see that second as an optimisation notion, and therefore distinct from optioneering towards a less defined end. I have in mind basic linear programming or game theory and the inherent assumptions of rationale ends being objectively knowable [sic].
    This focus on optimisation seems foundationally important to me. Because such efficiency is only occasionally what human curiosity is about. Notions such as happenstance and serendipity need the forever ability to be juxtaposed from what was previously thought or prioritised in action. This is a human quality beyond a start and an end of a query. I would respectfully suggest much of what you are talking of in this episode is moving beyond reasoning. Indeed, it is folly to only have reasoning in mind as to what this bigger problem is really about. The limits of this current AI is not reasoning because it is not "a process" (if following the definition used at the beginning of this podcast). Instead, what is being considered here is the characteristic of adaption. An ability to reset what one is working towards, with that new information or perspective. Behind that it is knowing that we do not know, and being humanly capable of asking why or how in a series of iterations where the question itself begins to change. That seems to me foundationally important to help distinguish both Turin computation and general intelligence from what human action is really about and what computation is only a part of. This is motives for asking the question, and knowing when the proposition itself is revealed as incomplete.
    I am probably coming from this very differently. As a a PhD student but from the world of governance of people, and projecting human endeavour in going about physical construction of the built environment. However, I see this foundational factor of understanding what the human action or human thought to be about, as pertinent here. That "aboutness" to be itself contextual and changing within the iterative processes themselves, and to be foundation to understanding what we are as "machines". This is the philosophical grounding which is underpinning the discussion here, perhaps. For me anyway, that adds some human context to what all artificial programming and anticipating fails to be by its very ontology and epistemology. Beyond reasoning is perhaps the context, and subjective nature, or what becomes an artificial and limiting account of truth. We deal with that daily in ways non-open AI cannot...
    Thank you again for this session. It is fantastic to hear thinking in its shared form in discourse such as this.

  • @michaeltraynor5893
    @michaeltraynor5893 2 дня назад

    I've also been struggling with drawing a line between computation and reasoning, and I'm really loving Tim's approach of making reasoning about knowledge acquisition.
    Like sure a vending machine is computing things about its state + environment and reacting accordingly, but it's never going to come up with a new representation or conjecture that it can turn into a useful improvement in its understanding/knowledge.
    I also wonder if it's useful to define reasoning as something that happens internally. Because a camera technically gets new knowledge every time you snap a photo, but it isn't knowledge that was produced by its own internal mechanisms.
    So, reasoning is computation that is somewhat causally insulated from the outside which leads to more compressed/predictive/good representation or understanding that can be meaningfully used?
    Anyone wanna take a stab at formalizing that?

  • @user-vo5dv8sq8v
    @user-vo5dv8sq8v День назад

    Prolog has backtracking. Reasoning often requires undoing some steps and restarting from a previously valid "save" point and exploring different paths in the space of potential solutions.

  • @palimondo
    @palimondo День назад +1

    Why do I get the feeling Keith has not spent any significant time using the o1 model himself and is criticizing it from his preconceived notions?!
    Granted, this conversation took place 4 days after the release and we get 30 messages with o1-preview and 50 prompts with o1-mini, but it is apparent Tim is talking from his own short experience of using the model, while Keith argues from positions he held before the model was released and he hasn’t yet updated his world model to include anything more then this first impressions of other people’s descriptions of the o1.
    I get that playing devil’s advocate is useful for honing one’s argument. But this isn’t that.

  • @jonathanduran3442
    @jonathanduran3442 День назад

    I believe that our current trajectory, although is not exactly reasoning, will help us get to the models that do. That said, I believe that there is an element of Stephen Wolframs research that we're missing in our research, cellular automata, computational irreducibility, etc. as was touched on a little bit in the video. An element of "time" that we're missing here, as I believe that true compression requires time, as does reasoning, as does a reverse cellular automata simulation. Something to think about.
    Great discussion, as always!
    Cheers!

  • @henrismith7472
    @henrismith7472 3 дня назад

    I disagree with his definition of reasoning. More specifically I disagree with his definition of what he doesn't consider reasoning. Regardless the point of strawberry is to generate synthetic data to train the new model on. Multi step chain of thought reasoning data which there wasn't much of on the internet because we either take it for granted or "reason" slightly differently (probably both).

  • @ragnarherron7742
    @ragnarherron7742 3 дня назад +1

    Fei-Fei Li World Labs definitely right track since spatial intelligence reflects an inductive bias that is "maps and navigation"

  • @chrisjsewell
    @chrisjsewell 3 дня назад +1

    My word, it feels like AI is turning into a religion for some people; you can’t even think critically about it without people jumping into the comments telling you what a heathen you are 😂

    • @MachineLearningStreetTalk
      @MachineLearningStreetTalk  2 дня назад +1

      I know right? We are not exactly luddites either, I (Tim) use LLMs all day every day and love them!

  • @ragnarherron7742
    @ragnarherron7742 3 дня назад +1

    Knowledge is persistent (meme) problem solutions that may be created through reasoning.

  • @matrixpredator
    @matrixpredator 2 дня назад +1

    one question: did any of you worked/build a LLM practically or are you just talking about things you learned?

  • @IdPreferNot1
    @IdPreferNot1 День назад

    Given the speed of progress, im wondering how quickly "It seems like but doesnt really" turns into "I dont care anymore because 'it seems like it does everything better than all humans on the planet'." Is emergent behavior truly emergent, or just the natural synthesis of the right heuristics and context once you acquire it. Studies have shown that what was believed to be emergent behavior in llms, was just the proper application of complex context. Isnt that why many 'discoveries' on reflection can be recast as 'it was right before our eyes the whole time."

  • @jonschlinkert
    @jonschlinkert 3 дня назад

    No lol. It's not reasoning. It's a series of prompts intended to simulate something that looks like reasoning. After spending quite a bit of time with the new "models", I'm not too excited about them.

  • @Jeremy-Ai
    @Jeremy-Ai 2 дня назад

    “AGI has a question to test humans in 1,2D
    This question is a trap…. make no attempt to answer!”
    this is a risk eternal insertion. 000000000000010000000000000000”0000
    “What does
    Mr. Know it all know?”
    Plz don’t repeat or respond
    All you have to do is not answer
    :(
    Jeremy

  • @ChrisSmith-lk2vq
    @ChrisSmith-lk2vq 2 дня назад

    Please do help me out with my reasoning problem:
    You say in your podcast that humans are somewhat robotic and have nets in somewhat the same way naeuronal nets have. We do reasoning while neural nets are finite state automata and hence can't do reasoning.
    Why? Aren't all humans (as you pointed out if i understood you correctly) finite state automata themselves?
    My point: neuronal nets today can "lookup" heuristics that go in the general direction of solving the problem (which is amazing!!) AND they can add those heuristic - vectors (if you will) together to form superposition - vectors.
    This is some sort of reasoning. This is what humans do. We (and you said so) take what others did, what others "discovered" (or added) and add our small little parts to the whole heuristic ladder.
    I can't see the difference here but the amount of complexity and scale. And you didn't give one or I missed it completely.
    I think the base problem with most of your arguments is (though beautifully presented as always ❤) that you overestimate the capability of humans.
    I think we are just finite state automata breaking down problems, fetching heuristics, adding them, tree searching in the range of our abilities and that's that.
    Just think about why it took hundreds of thousands of years to go from stone to silica.
    That said I think AI is "just" able to massively blow up speed and thus overtake us in this rather simple game.
    Can you please reasond on why humans are different and why we can do "effective compute" ourselves (not using external components) but neuronal nets principally can't?
    Thank you so much for this rich conversation of yours.
    Chris

  • @mono24
    @mono24 10 часов назад

    anyone that say "I" too much is not trustworthy to me! Especially if they say something like "I and many other smart people"???????? WTF BRO? SETTLE DOWN!

  • @wwkk4964
    @wwkk4964 3 дня назад +1

    Keeo up thr great commentary! At the 45:00 mark, im afraid this argument would suggest Ramanujan was not reasoning because his formulae did not have proofs. Seems very false dichotomy. Ramanujan neither spent a lot of test time deriving his formulae formally with proofs nor did he memorize anything or coly from anyone. I dont follow how an llm could'nt in principle ever do an ramanujan, effectively reaching true conclusuons through isolated self absorbed self play.

  • @mughat
    @mughat 3 дня назад

    "Reason is the faculty that identifies and integrates the material provided by man’s senses."
    When these AI systems have sensors and a commitment to identifying reality we are getting closer to something that looks like reason. But with no sense perception there is only computation of some vanity.

  • @PriNovaFX
    @PriNovaFX 2 дня назад

    Nice interview, but you're talking past each other. Reasoning is a process that involves both deterministic and exploratory, non-deterministic processes. Successful reasoning requires either complete information or creative information generation to integrate and combine semi-knowledgeable information with existing information. And in the end, it must be checked whether the conclusion is applicable and leads to the desired goal. None of this is optimized at first.
    The memory aspect may need episodic memory.

  • @miladkhademinori2709
    @miladkhademinori2709 2 дня назад +1

    I like your dialogue so much ❤!

  • @hamandchees3
    @hamandchees3 2 дня назад

    The discussants are not doing a good job of proving they're not also isomorphic to a giant lookup table

  • @christiantangø-p4e
    @christiantangø-p4e 10 часов назад

    about 21.00 / on reasoning and neural networks: A very basic and very old distinction from science-philosophy comes to mind, about what "sometimes work" versus "what can be proven as a valid solution". The distinction is of course inductive reasoning versus deductive proof.
    I think this is the core of the problem: Neural networks are like inductive look-up-tables, and in that sense capable of some form of INDUCTIVE reasoning. But DEDUCTIVE reasoning is something else. It has been said somewhere, that a truly "thinking" system must be able to do BOTH. It is also like the Kahnemann "thinking fast thinking slow" problem: Heuristic solutions serves one purpose, but the slower analytical thinking serves another.
    Just my humble thoughts on the subject.