Tree of Thoughts: Deliberate Problem Solving with Large Language Models (Full Paper Review)

Поделиться
HTML-код
  • Опубликовано: 5 окт 2024

Комментарии • 161

  • @YannicKilcher
    @YannicKilcher  Год назад +24

    OUTLINE:
    0:00 - Introduction
    1:20 - From Chain-of-Thought to Tree-of-Thought
    11:10 - Formalizing the algorithm
    16:00 - Game of 24 & Creative writing
    18:30 - Crosswords
    23:30 - Is this a general problem solver?
    26:50 - Ablation studies
    28:55 - Conclusion
    Paper: arxiv.org/abs/2305.10601

    • @ozordiprince9405
      @ozordiprince9405 Год назад

      I was literally about to go through this paper myself. Thanks Yannic

    • @EdFormer
      @EdFormer Год назад +1

      So glad you're back on with regular content. The hype train since ChatGPT's release has led to an intolerable rise of wishy washy AI content by people who clearly don't care about fundamentally understanding machine learning and disseminating that knowledge but are instead motivated by views, and are happy to peddle the poorly thought out arguments for how close to AGI we are and what the consequences will inevitably be, in order to rake those views in. So I'm so grateful for you continuing to present your sobering and realistic perspective on AI in these entertaining videos that actually cover the detail of methods (where you can - so called "technical report"s can die), as it really augments my experience of researching the subject with much more enjoyment and insight than I would otherwise have had. And I'm sure I'm not alone in this. Thank you!

  • @JurekOK
    @JurekOK Год назад +69

    For multi-step agents, it is exponentially important that each "step" has as high success rate, as the compound success rate decreases very quickly with both count of steps and unit success rate: overallSuccessRate=(stepSuccessRate^Nsteps). Foing from e.g. 90% to 95% is actually a lot, as it enables the chain length to go from 7 steps to 14 steps and still have a ~50% compound success rate, so enables vastly more complicated problems to be solved. Hence, it will often be very valuable to review and iterate on each sub-step to maximize the chance that it doesn't block the entire chain.

    • @avatarcybertronics2584
      @avatarcybertronics2584 Год назад

      You are right, we call this phenomena catastrophic error compound (similar to neural network effect to forget previous style when tuned - take a look at FractalGPT - self evolving true multi-agent system in core no LLMs so no this problem)

  • @ixion2001kx76
    @ixion2001kx76 Год назад +57

    I don’t mind at all that you didn’t cut out the “um”s. It probably saves you a heap of time that is better spent on reading papers, and it makes your videos feel more personable.

    • @1000niggawatt
      @1000niggawatt Год назад +6

      Yes, videos that focus on production are just a lot of clickbait, hype, fast movement for adhd zoomers, and there is no substance - they just put a paper into chatgpt and read the result.
      Yannic please do not bother with production, we come here to hear the critique, anyone who's just clicking on le funny ai vids will go to one of the many fastfood ai channels instead, anyway.

  • @1000niggawatt
    @1000niggawatt Год назад +163

    Yannic is the one man who's actually giving intelligent critique of new papers, instead of just throwing the paper into chatpdf and making a video.

    • @N.i.a.m.a.t.u.l.l.a.h
      @N.i.a.m.a.t.u.l.l.a.h Год назад

      @TigerGold 59 www.chatpdf.com/

    • @Candyapplebone
      @Candyapplebone Год назад

      Ooof

    • @TheManyMan
      @TheManyMan Год назад +4

      @TigerGold 59 joke but there are pdf reader plugins for GPT; useful for running through summaries of papers you don't have time / don't want to read fully ngl

    • @1000niggawatt
      @1000niggawatt Год назад

      @TigerGold 59 there's literally "chatpdf" and pdfgpt.

    • @mamadsc
      @mamadsc Год назад +5

      Yes, I also really like www.youtube.com/@hu-po for his deep analysis of papers (and always reexplaining basic concepts along the paper so it's easier to follow for everyone)
      And www.youtube.com/@ai-explained- for what I think is a more global understanding of AI trends

  • @jit_rs
    @jit_rs Год назад +13

    One application of this "AI-guided tree search" is in automated theorem proving. There was a reasearch project termed GPT-f, where they took the Lean proof assistant which can precisely check if a proof up to a certain point is correct and designed a plugin that constructs a proof step-by-step with backtracking using a language model (GPT-f itself) as the decision maker and it was able to prove about 60% of common geometry/algebra theorems with zero user intervention. As a type theory nerd myself I am excited to see what this branch of research brings next 🎉

    • @luck3949
      @luck3949 Год назад

      Do you work at Yale-NUS Singapore?

    • @jit_rs
      @jit_rs Год назад

      @@luck3949 no, I am a systems programmer

  • @dribrahimel-nahhal2477
    @dribrahimel-nahhal2477 Год назад +12

    Yannic, thank you for this excellent video on the 'Tree of Thoughts' research paper. Your explanation was very clear and concise, making it easy for even a layman like me to understand. I appreciate your efforts in breaking down the decoding technique used in large language models and highlighting its usefulness in investigative problem-solving patterns. Keep up the great work!

  • @ixion2001kx76
    @ixion2001kx76 Год назад +48

    A very nice addition to the new field of computational philosophy.

    • @television9233
      @television9233 Год назад +17

      Computational philosophy is the use of computation for philosophical research (as per Stanford's encyclopedia).
      You are probably thinking of the philosophy of computation, in which even Alan Turing, the father of the modern computation, engaged in back in the 50's in his paper "Computing Machinery and Intelligence". The same paper is also why he is partly remembered as a philosopher.
      So I'm not sure what you mean by "new field".

    • @trulyUnAssuming
      @trulyUnAssuming Год назад

      ​@@television9233 it feels like a joke along the lines of "this isn't computer science anymore - people are just throwing shit against the wall"

    • @television9233
      @television9233 Год назад +3

      @@trulyUnAssumingdon't think so. using a value based tree search algorithm is literally one of the most CS things you can do.

    • @davidw8668
      @davidw8668 Год назад

      ​@unAssuming hilarious interpretation, and indeed it's getting harder to determine where the jokes end or begin. However I don't think the paper is shit even though for sure some people already figured how this is the new agi mechanism to erase humanity.

  • @ilianos
    @ilianos Год назад +5

    When I saw this paper, I was hoping someone like you would cover it. Thanks a lot!

  • @marshallmcluhan33
    @marshallmcluhan33 Год назад +7

    Awesome I saw this and wondered if it was profound. Thanks for explaining it.

  • @amalzubidat1897
    @amalzubidat1897 Год назад +8

    Thank you for reviewing this! Yannic is always on top of things :)

  • @clray123
    @clray123 Год назад +6

    I'm pretty sure in the picture at 10:46 the authors meant to descend into the left branch first and backtrack to later descend through the solid green branch, not like Yannic explained.

    • @Zankras
      @Zankras Год назад

      That’s how I read it too.

  • @lucastononrodrigues1069
    @lucastononrodrigues1069 Год назад

    Awesome, I was reading it last night! Very glad you posted it right on time :)

  • @titastotas1416
    @titastotas1416 Год назад +3

    I like your content so much that I felt it necessary to express my gratitude in the comment section, simply pressing the like button does not cut it for me in this case.

  • @tiagotiagot
    @tiagotiagot Год назад +8

    Could this pattern of thinking be trained on, so that models may spontaneously chose this approach when suitable and produce better results straight out of the box?

  • @nangld
    @nangld Год назад +2

    LLMs are N-gram Markov models, in that they output a single token, based on the last N tokens of chat history. So outputting intermediate steps helps the follow up calls to the model to organize its reasoning. Just like a human being has more chances solving an equation with a piece of paper, instead of relaying solely on their brain. In other words, some problems inherently requires N tokens of memory to be solved by a given model. Guess in the end scientists will expand the big-O space and computation complexity to LLMs. Obviously you can also ask the model to introduce different personalities, like the engineers from the relevant fields or simply different psychological models, which will explicitly reference associated knowledge during the solving the problem, and you will get a several totally different answers, and all of them could be worthy of considering.

    • @ThetaPhiPsi
      @ThetaPhiPsi Год назад +1

      LLMs are, if anything, N-gram Markov++ models. Try to replicate some of the results of LLMs (eg. Llama-7b) with a N-gram HMM. Its an overly simplistic view of LLMs I only would use as an intro to NLProc. On the other hand, I would be interested to know if one could replicate an LLM by an N-gram HMM. If that works, I'll take everything back.

  • @sgramstrup
    @sgramstrup Год назад +7

    Hey Yannick. You mentioned something important. You said that we shouldn't 'pick' control questions but let the AI suggest what to do. The idea oc is to remove hooman rigid thinking and find a more 'fluid' approach. Here's the question then. Why are people not using NEAT or one of the novelty seeking algorithms, to optimize their cognitive architecture ? This problem isn't much different from a genetic algorithm car trying to pass a maze. Just because we hoomans didn't design the 'maze' this time, doesn't mean that we should try to lead the car through step by step. For fucks sake - this is what genetic algorithms excel at ?? Let a GA develop general architectures, optimize methods like CoT/ToT and what not.
    We simply don't have to try out all this shit by hand. Lets use the great tools we have to the maximum, and let the genetic control network develop 'it self' ! How long would it have taken a genetic algorithm to go from 'AutoGpt', to CoT, to 'sampling' decisions to a 'Tree of Thoughts decision ? Not long, is my bet, and then we have to ask: What other cognitive architectures could such an adaptive algorithm discover ?

    • @ankitaharwal5886
      @ankitaharwal5886 Год назад

      Yannic should pin this comment, as someone might just look into it and implement it

  • @nicktasios1862
    @nicktasios1862 Год назад +2

    Another possibility why chain of thought prompting works could be that data in the training set that has this form is more likely to be correct?

  • @mono_onamoto
    @mono_onamoto Год назад

    Very informative and good voice for radio. Cheers Yannic!

  • @jabowery
    @jabowery Год назад +14

    Sounds like a Stack-RNN may be the next step for DeepMind given the prominent mention in the recent Princeton/DeepMind paper "Neural Networks and the Chomsky Hierarchy". However, since there are no authors in common between the two papers, it may require overcoming some of the Big Org problems that have plagued Alphabet's ability to execute on its in-house talent.

    • @Rotbeam99
      @Rotbeam99 Год назад

      what is a stack-rnn? thanks

    • @jabowery
      @jabowery Год назад

      @@Rotbeam99 See "Neural Networks and the Chomsky Hierarchy"

  • @killermike1209
    @killermike1209 Год назад +1

    Yannic, Your sunglasses are strikingly stunning.. Much thanks for keeping me informed on AI goings on.. Also thanks for being anti-boring, funny and or highlarious.. - Cheers!!

  • @florianbehrens690
    @florianbehrens690 Год назад

    Thank you for making it much easier to consume these papers!

  • @cutebabyseal621
    @cutebabyseal621 Год назад +1

    Watching Yannic try to come up with a crossword cue for "ape" was hilarious.

  • @washedtoohot
    @washedtoohot Год назад +2

    Can’t wait to see this in Langchain 😮

  • @Rockyzach88
    @Rockyzach88 Год назад

    This is cool. Sort of my first video I've watched about prompt engineering. The idea of creating sort of virtual neurons comes to mind. And yeah right as this was coming out, I was thinking the exact same thing, like they would replace parts of algorithms or "functions".

  • @DaKingof
    @DaKingof Год назад +1

    I'd think this could improve coding with LLM's tramendously. One huge problem I see is that they don't seem to know what versions they are using to write code. It would be wonderful to have the LLM look back and find the code it used for a snippet to see what version it is, then review for the latest or needed version and update it's response to include the latest or selected versions. This way it always knows what codebase it's using and can compare live rather than attempting to use it's training data. As of now it seems to get really confused when I try to ask it to do any of this.

  • @eruiluvatar236
    @eruiluvatar236 Год назад +3

    I wonder what would happen if chain of thought or this technique or refinement or the other techniques that increase the quality of the output were used to produce a dataset that contains only the initial question and the final answer and that was used to fine tune the model.
    If the thinking needs to happen explicitly in the context window, that might not help much or it may still help. Then if it helps, it would be even more interesting to try those techniques again to see if they still provide a benefit. If they do, continue the fine tuning loop and see where it leads.

    • @jonnicholasiii2719
      @jonnicholasiii2719 Год назад

      It eventually leads to God-level cheat codes.

    • @eruiluvatar236
      @eruiluvatar236 Год назад

      @@jonnicholasiii2719 Lol, I doubt it. I don't think transformers can reach consciousness or AGI without some serious architectural changes so no God mode yet.
      But there is plenty of evidence that more training and better quality data help a lot and that you can squeeze way more intelligence in the weights than we are currently able to. Ie you can quantize and prune most of the weights with minimal performance loss.
      So I wonder if this can be an improvement, much like you can fine tune smaller models on larger/better models output and get some improvements in some benchmarks.

  • @florianhonicke5448
    @florianhonicke5448 Год назад +3

    Thanks for the summary!
    Can we also have an interview with the authors? :)

    • @Sven_Dongle
      @Sven_Dongle Год назад

      Maybe next spring.

    • @joech1065
      @joech1065 Год назад

      ​@@Sven_Dongle So 10 years in AI time

  • @joepike1972
    @joepike1972 Год назад +1

    7:44 I would be careful about asking it a question like "do you think that is a good thought?" My experience has been the model might make up reasons why it thinks you think it is wrong instead of taking the time to understand what it is talking about, especially in a subject it is not that versed in. So I would watch out for over corrections from leading questions representing the model seeking only to appease and not reflecting understanding. Particularly if you are taking an unconventional stance or exploring a subject the model is not particularly strong in.

  • @ericadar
    @ericadar Год назад +3

    Do you think one train a new LLM with maybe 50% more parameters than the original LLM producing the input-output pairs of the final tree-of-thought prompting so that the new (larger) LLM already encapsulates the entire tree-of-thought tree expansion/pruning process on a single feedforward run and thus save on inference compute?

    • @drdca8263
      @drdca8263 Год назад

      Does it need to be larger? Compare AlphaZero.

  • @Candyapplebone
    @Candyapplebone Год назад +2

    Nice to see an actual pro do a video on this xD

  • @Ernest_Viger-Beaulieu
    @Ernest_Viger-Beaulieu Год назад

    Thank you so much. Best explanation I found about this paper. 🎉

  • @falklumo
    @falklumo Год назад +2

    It should not go without saying that the original paper seems to be "arXiv:2305.08291v1 [cs.AI] 15 May 2023" which is NOT DeepMind and was published 2 days prior to the work you cite here ...

  • @Veptis
    @Veptis 6 месяцев назад

    This could be really useful for coding problems/debugging. And you use someone like LSP to recursively walk more information into the prompt for the model to solve it.

  • @sabofx
    @sabofx Год назад

    Really helpful explanation of TOT! Thanx bud! 🤓

  • @joepike1972
    @joepike1972 Год назад

    5:13 I have noticed that is seems to be related to the models text limitations as well, or just general capabilities possibly related to their number of tokens. In that I have seen large language models give more efficient usage of such though process capabilities were as older models might just use such space to continuously insist the same points and not make much progress with it.
    But the other aspect it will try to do several things at once in the limited space and not take the time needed to fully expand each thought to the degree it needs to deal with matters efficiently.

  • @jonbbbb
    @jonbbbb Год назад +5

    Could this technique (or even just chain-of-thought) be used in the training process itself, or as a separate step like RLHF? This would be RLAIF I guess heh.

    • @skyebrows
      @skyebrows Год назад +2

      My feeling is we could train a new neural net using inputs and outputs using gpt4 and this method to create a much more efficient/intelligent base model before these techniques are applied.

    • @simonrouse9461
      @simonrouse9461 Год назад +2

      In RLHF, they already use a language model as a critic model. Although it’s called “human feedback”, they actually only use a small amount of human feedbacks to train that critic model. It’s the critic model that actually gives the feedback.

  • @joe_limon
    @joe_limon Год назад +1

    One can prompt an ai like bing, asking it to implement a tree of thought process to solve your problem. It can look up the paper and construct/execute the process by itself.

    • @sgramstrup
      @sgramstrup Год назад

      Because it uses something like it already. Gpt4 and all other tooling LLM's are already connected in these cognitive architectures CA, when you chat. We are no longer chatting with the model directly, but with their CA on top.

    • @joe_limon
      @joe_limon Год назад +1

      @@sgramstrup bing can and has in my testing looked up the paper, interpreted the strategy, and then applied it in its response.

    • @clray123
      @clray123 Год назад

      @@sgramstrup Do you have any proof of that? It may just be that we're talking to larger and better trained models.

  • @aa-xn5hc
    @aa-xn5hc Год назад +1

    Really brilliant analysis

  • @ChocolateMilkCultLeader
    @ChocolateMilkCultLeader Год назад +1

    A whatsapp group I'm in, consisting of non-AI people shared this. Yannic is making it

  • @guillemgarcia3630
    @guillemgarcia3630 Год назад

    Really well explained! Thanks!!

  • @anishbhanushali
    @anishbhanushali Год назад

    So this is basically learnable Beam search where we're using the same GPT ( or anyother LLM ) for learning the best tree-path. Also here the beam is formed by 'thoughts' as oppose to 'tokens' in general .. !!

  • @jackderrida
    @jackderrida Год назад +2

    I just have one criticism of the video. Otherwise, I love it and am glad you explained Tree of Thoughts. My one criticism is that it would be easier to follow you explaining the diagram if you didn't make up your own paths that contradict the graph at first.
    I get that you wanted to give an example where it goes back up the tree, which the graph doesn't. But I'd rather you just run through the displayed graph and then do your own to illustrate how it can go back up the tree, separately. It's just confusing for me, when learning, when the example contradicts the explanation I'm listening to.

  • @FranAbenza
    @FranAbenza Год назад

    Flood Fill algorithm could be an interesting way to benchmark the probability of a branch to solve our maze

  • @billxu9799
    @billxu9799 Год назад

    Good name to catch the hotspot, but kinda trivial work considering extra token usage / computation

  • @luciengrondin5802
    @luciengrondin5802 Год назад +1

    This seems to be good to eek better results out of a language model, but if I understand correctly, it can't be used to improve the language model itself, can it? Basically that tree search procedure is not differentiable, thus it can't be back-propagated, right?

    • @drdca8263
      @drdca8263 Год назад

      training sets aren't differentiable either though?

  • @television9233
    @television9233 Год назад

    The idea of a value based tree search on LLM outputs sounds interesting in theory, but the results of this specific implementation are lackluster, especially when taking into account that their method was given task-specialized prompts.

  • @XorAlex
    @XorAlex Год назад

    Thanks for explaining!

  • @PaganPegasus
    @PaganPegasus Год назад

    Not to sound like an ass or anything... but I feel like ToT is just CoT with beam search but pruning happens after self evaluating each thought rather than pruning on the conditional probability of each token. Because quite literally the beam search score evaluation is gonna be:
    `P("good"|{step N}) - P("bad"|{step N})` for each possible thought at step N.

    • @CoolIcingcake3467
      @CoolIcingcake3467 21 день назад

      Claiming ToT is "just CoT with beam search" is like saying a Formula 1 race car is just a bicycle with a motor. It fundamentally misses the point.
      ToT isn't just an upgrade on CoT - it's a completely different approach to how AI reasons.
      Here's where the comparison falls apart:
      Think Parallel, Not Linear: ToT explores multiple reasoning avenues at the same time, not one after the other like CoT with beam search. This lets it explore the solution space far more effectively.
      Smarter Evaluation: ToT doesn't rely on simplistic scores. It uses expert knowledge and analyzes the entire chain of thought, not just individual steps. This helps it understand the quality of reasoning in a more nuanced way.
      Adaptability & Second Chances: ToT has a secret weapon; adaptive search with backtracking. It can adjust its approach on the fly and even revisit past decisions if needed. This is something standard beam search just can't do, especially for complex problems.
      That simple formula you mentioned? It doesn't even scratch the surface of how ToT evaluates its reasoning. ToT uses far more complex methods to assess the quality and logic of its thought processes.
      Frankly, comparing ToT to CoT with beam search is like comparing a world-class chef to someone microwaving a frozen dinner. The underlying principles and skills are just worlds apart. ToT's structured exploration, ability to adapt and backtrack, smart evaluation, and knack for balancing broad and focused thinking are simply not present in standard beam search approaches.
      To dismiss ToT as a minor tweak is to ignore a major breakthrough in AI. ToT pushes the boundaries of what AI can achieve and opens doors to solving problems that were once impossible for traditional methods.

  • @piotr780
    @piotr780 Год назад

    space of solutions in Game of 24 is really small, so mayby they simply induce brute force search inside the network

  • @aamir122a
    @aamir122a Год назад

    So they have taken Tree search , which was applied to the game GO, to drive policy and value network and applied to NLP tasks.

    • @television9233
      @television9233 Год назад

      Tree search has been a thing since the early days of computing and has been used extensively.
      And NLP research (as well as any other subfield within computer science) has constantly used tree search algorithms as they are one of the fundamentals of CS.

  • @dennisestenson7820
    @dennisestenson7820 8 месяцев назад

    It doesn't seems like anyone realizes that when you "instruct" an LLM to do something, it's not being instructed to do anything but continue writing text that could syntactically correctly follow the given instruction.

  • @MrSuntask
    @MrSuntask Год назад

    Great vid. Why do you wear sunglasses?

  • @rikvermeer1325
    @rikvermeer1325 Год назад

    Polluting the ToT critic (prompt) makes way for some intricate goals.
    Looks like me this is the part where the AI gets to plot.

  • @Amipotsophspond
    @Amipotsophspond Год назад +1

    This reminds me of Psychology Therapists just blindly repeating "...And How Does That Make You Feel..."

  • @unclecode
    @unclecode Год назад

    This topic is interesting, but I have reservations about these papers. They often use large language models for classic algorithms, resulting in high costs, redundant tokens, and increased environmental impact. Rather than substituting parts of existing algorithms, it would be better to introduce solutions that leverage the language model's capabilities.
    If this research were valuable, one could publish a paper on bubble sort using LLM instead of comparison operators-a worthless endeavor. Using LLMs for traditional algorithms like quick sort may not be groundbreaking. Efforts like LORA or QLORA, which introduce new paradigms, are more significant.
    In my view, this research doesn't bring substantial innovation. For instance, replacing trees with graphs, naming it 'Graph of Thought' or GOT (Not Game of Thrones ;) ), could yield similar results using classic graph algorithms. This doesn't add much value when compared to zero-shot methods.
    While reputable institutions are involved, I'm curious if they see something I don't.
    Your content is consistently excellent. Keep up the good work!

  • @RedCloudServices
    @RedCloudServices Год назад

    Yannic do you predict this capability be integrated soon with openai GPT, Llama or the other LLMs public or private?

  • @karlitucha
    @karlitucha Год назад

    What tools and platforms do you use to stay up to date with the latest papers?

  • @hanskraut2018
    @hanskraut2018 Год назад

    At 5:30 you said 2 hypothesis:
    1) working memory (can calculate something then write it down and use all calculation to go in other branches/directions)
    2) more compute time
    I think you you are spot on and u even used the framework of a „hypothesis“
    Very nice. Nice 🏅 i dont know im just midlessly intuition commenting here like ur suposed to on youtube right

  • @lingred975
    @lingred975 Год назад

    Feynman technique? Is the AI trying to explain, discovers its mistakes and corrects? :)

  • @SLAM2977
    @SLAM2977 Год назад

    Yannic at his best: with glasses!:)

  • @-mwolf
    @-mwolf Год назад

    yesssss, thanks yannik!

  • @Timotheeee1
    @Timotheeee1 Год назад +1

    can you review rwkv?

  • @ajit60w
    @ajit60w Год назад +1

    This is back to writing programs. Whoever said that advent of LLMs will make teaching programming unnecessary.

  • @ericadar
    @ericadar Год назад

    @YannicKilcher can you do a review of Meta's Megabyte?

  • @aitools24
    @aitools24 11 месяцев назад

    00:05 Deliberate problem solving with large language models
    04:12 Chain of Thought prompting helps in better problem solving.
    07:57 Using a tree search algorithm with pruning for model self-critique and improvement
    11:39 Implement Chain of Thought in two ways: one approach is to explicitly sample the next thought, while another is to input all thoughts at once and generate a linear sequence.
    15:28 Language models can be integrated into programming by handling specific parts, resulting in more evaluations.
    19:14 Backtracking is useful in language models for solving crossword puzzles.
    22:47 The algorithm implemented a crossword-solving algorithm using language models.
    26:16 The paper introduces a technique for improving performance in language models
    Crafted by Merlin AI.

  • @petevenuti7355
    @petevenuti7355 Год назад

    How can this be integrated into the conceptual structure of the network itself?

  • @questmarq7901
    @questmarq7901 Год назад

    This kind of thing can help me with the world building in my book

  • @rootthree9436
    @rootthree9436 Год назад

    how's it different from beam search?

  • @wizix9877
    @wizix9877 Год назад

    being thoughtful before acting is even true for AI :)

  • @JTMoustache
    @JTMoustache Год назад

    Kilcher is back

  • @jacktherater3533
    @jacktherater3533 9 месяцев назад

    This is what my social anxiety looks like in white paper.

  • @Amipotsophspond
    @Amipotsophspond Год назад

    19:52 this is really good point because all the models are forward only and this reevaluation allows some pseudo-backwards from a forward only system.

  • @zyxwvutsrqponmlkh
    @zyxwvutsrqponmlkh Год назад

    I thought you didn't do papers anymore. ☼

  • @Adventure1844
    @Adventure1844 Год назад +1

    Tree of thoughts looks like a neural network

  • @mikelewis1166
    @mikelewis1166 Год назад

    I’ll be dropping the transcript of this video into an llm to generate some interesting python scripts and system commands…

  • @clray123
    @clray123 Год назад

    They should have called their methodology Clever Hans.

  • @kurort5099
    @kurort5099 Год назад

    yoooo nice sunglasses bro ;P good good

  •  4 месяца назад

    Nice sunglasses 😎

  • @vighneshsablok5851
    @vighneshsablok5851 Год назад

    Would have appreciated this video more with a live demo of solving a small crossword with ToT method. It is a tough one to visualize.
    Otherwise great work!

  • @tommornini2470
    @tommornini2470 Год назад

    Neural network + algorithm = science

  • @charlesd774
    @charlesd774 Год назад

    I wonder how long until someone makes a GAN where the discriminator is a trained LLM and the generator trains itself.

  • @drdca8263
    @drdca8263 Год назад

    oh heck,
    uhhh...
    I hope the idea I have for how this could be extended, doesn't work?
    Edit: not to say that I think I had any rare insight or anything. I think the idea I had is probably obvious, especially if it works.
    I’m just refraining from mentioning it in case it works just in case, on the very unlikely off chance that me saying what it is makes it be done sooner
    I mean 90% of the idea is already described in the video,
    and the other 10% idk if it is likely to work at all...
    and maybe if it does work, it still might not be *that* effective, but like,....
    it seems *conceivable* that the idea might work quite well, with enough training data,
    and idk that seems like it could be bad...
    edit3: Ok, yeah, no, many other people definitely thought of the idea before I did.

  • @rothn2
    @rothn2 Год назад

    If the authors had restricted their scope to _planning models_ I think this could have been a much more sound paper, with the opportunity to dominate a class of problems.

    • @rothn2
      @rothn2 Год назад

      You know, the control systems formerly done by RL.

  • @aleksanteri_r
    @aleksanteri_r Год назад

    Why is no one talking about how ToT is also a funny emoticon??

  • @markopancic6060
    @markopancic6060 Год назад

    Ant poe eta would be a way to finish that crossword 😂

  • @testboga5991
    @testboga5991 Год назад

    Interesting

  • @danberm1755
    @danberm1755 Год назад

    Sounds like we need an AI assistant to prompt the AI assistant for complex problems 😁

    • @danberm1755
      @danberm1755 Год назад

      Actually this gave me a longer run time (which you mention as the reason tree of thought might have high success rates). I'm pretty sure this can be expanded apon for tree of thought like results that are actually correct.
      ----
      Act like you have the ability to prompt yourself as a human would to figure out complete answers.
      If you made change with $20 how many ways could you split the bill into 8 bills?

    • @danberm1755
      @danberm1755 Год назад

      The big question is how long OpenAI would allow the inference to continue.
      In other words you'd probably have better results when you pay by the token (not ChatGPT).

  • @SimonJackson13
    @SimonJackson13 Год назад

    Morpholigical application complexity.

    • @SimonJackson13
      @SimonJackson13 Год назад +1

      It's like an alpha beta pruned minmax search?

    • @SimonJackson13
      @SimonJackson13 Год назад

      So pattern matchers, rule expanders, rule factorisers and similarity scorers. GANminmax? The score list does appear like a genetic algorithm cross pollination for something. Would the later thoughts have more than one node as parent then?

  • @hurktang
    @hurktang Год назад

    This is basically the algorithm of the annoying kid in the car.
    1. Ask "why".
    2. Do noise.
    3 Ask "how long before we get home?"
    4. Goto 1.
    But with a twist! If the kid gets bored, the parenting driver gets disintegrated and replaced by a new one.

  • @stacksmasherninja7266
    @stacksmasherninja7266 Год назад

    Figure 1 has subfigures (a) (c) (c) (d) lol

  • @Chillingworth
    @Chillingworth Год назад

    We need adversarial promoting that makes any model output an exact phrase or extremely similar thing from innocuous input prompts. There must be a way to get ChatGPT to be redpilled

  • @tacticalgold
    @tacticalgold Год назад

    This isn't new isn't this what I've always done? Now I'm just more qualified… lol

  • @milos_radovanovic
    @milos_radovanovic Год назад +1

    All of these sound like we are teaching AI to do the philosophy by combining expert intuition with formal reasoning!
    I'm waiting for AI that can at least do its own science through numerical model experiments. :)

  • @jacktherater3533
    @jacktherater3533 Год назад +1

    chatgpt API cost go brrrrrrrr..

  • @GNARGNARHEAD
    @GNARGNARHEAD Год назад +1

    baby steps to an Auto argument mapper 🤯

    • @GNARGNARHEAD
      @GNARGNARHEAD Год назад +1

      have a search for Argument Mapping by Tim Van Gelder, I promise you will have an epiphany

  • @arandomguy46
    @arandomguy46 Год назад

    This was actually made a while ago by another youtuber before this report came out but i forgot the name of the youtuber.

  • @manslaughterinc.9135
    @manslaughterinc.9135 Год назад

    Aft, Poe, Era makes Ape, For, Tea

  • @killermike1209
    @killermike1209 Год назад +2

    Oh yeah I forgot to ask will humanity perish in the grey goo or at the hands of liquid metal lawnmower men?

    • @hanskraut2018
      @hanskraut2018 Год назад

      How about u investigarte / care about real problems? Listen to russel barkley lecture gor parents adhd ever 10.-20. person does qualify so ur well of selfishly to listen to the most cited adhd guy
      Or watch documentarys sbout the medical times or ways people can be fucked over in bussness and all the way in history and recent history how systems dont protect the productive/well meaning always.
      You find far more useful worry’s than terminator AI……

    • @hanskraut2018
      @hanskraut2018 Год назад

      „Killermike“ ok fits the comment

    • @Amipotsophspond
      @Amipotsophspond Год назад

      If you make a Ai that works as expected, that "work as expected" many would say that means it follows orders as given. but Totalitarian humans can give orders too, so you are only shifting to the devil you know this time with better tools and even more compliance from their subordinates. don't let your fear of a machine overlord blind you to a human overlord in control of a machine.

  • @BrutalStrike2
    @BrutalStrike2 Год назад

    26:26

  • @wizix9877
    @wizix9877 Год назад

    ToT ...looks like the AI crying from being over tasked...