AI Can Do Maths Now, and it's Wild

Поделиться
HTML-код
  • Опубликовано: 24 дек 2024

Комментарии • 782

  • @AnotherRoof
    @AnotherRoof  9 месяцев назад +72

    Check out Sabine Hossenfelder's video here -- it is a great overview of AlphaGeometry, I just respectfully disagreed with that one part
    ruclips.net/video/NrNjvIrCqII/видео.html

    • @twiddle7125
      @twiddle7125 9 месяцев назад +2

      Yeah, no thanks. Don't need to watch a transphobe's videos.

    • @Tacom4ster
      @Tacom4ster 9 месяцев назад +1

      Sabine is just an alt right troll that probably got her degree from a cereal box

    • @diribigal
      @diribigal 9 месяцев назад +40

      I won't be supporting Sabine but I enjoyed your video.

    • @Tacom4ster
      @Tacom4ster 9 месяцев назад +30

      Screw Terfs@@diribigal

    • @Nzargnalphabet
      @Nzargnalphabet 9 месяцев назад

      Video 6h ago but comment one day ago how?

  •  9 месяцев назад +550

    This brings back some memories of chasing the angles endlessly during a competition and wondering when to stop. Just to realize later that the solution is simple using a lesser known theorem. I hated geometry problems.

    • @AnotherRoof
      @AnotherRoof  9 месяцев назад +121

      Yeah, I saw a one-line proof of this using the nine-point circle and Apollonius's Theorem! Never would have thought of that!

    • @Osirion16
      @Osirion16 9 месяцев назад +53

      this is what makes me hate and love geometry problems at the same time. kind of a toxic relationship. I suck at them, but they're always so easy, it's just that they always require this one theorem that makes the problem trivial but you don't know it until you do... I feel that algebraic problems are more fun because there are often many ways to solve them not necessarily by brute force but by brilliant bit of ideas that lead to a solution. Where, in geometry, I don't feel like I can just invent a theorem out of thin air like that, but oh well, i'm not so good at geometry so anyways |:

    • @remotepinecone
      @remotepinecone 9 месяцев назад +4

      What I want to know is does it know the theorems and identities involved in the proof or did it make them all up as it went along?

    • @AnotherRoof
      @AnotherRoof  9 месяцев назад +16

      @@remotepinecone It knows a foundational set of results, which it then used to deduce valid theorems (known in the paper as "synthetic theorems"). Collectively these are the results it uses to make inferences in problems.

    • @stargazer137
      @stargazer137 9 месяцев назад +3

      @@Osirion16 I relate to this so much omg.

  • @wiadroman
    @wiadroman 9 месяцев назад +132

    27:00 The painting of Yellow Alex is the most beautiful thing I have ever seen. My eyes are welling up with the sudden surge of emotion and awe.

  • @blacklistnr1
    @blacklistnr1 9 месяцев назад +347

    I mean.. it's a general search algorithm:
    - the "logic" part has the edges of the graph and
    - the "creative" part picks less random directions than brute-forcing the whole graph.
    So while it does generalize to basically all other fields, as you said its utility and insights are questionable for harder/novel problems.
    The creative part would be really interesting if it offers a good true positives rate, to maybe offer in the future novel ideas to problems we are stuck on

    • @johnsherfey3675
      @johnsherfey3675 9 месяцев назад +9

      I think a step minimization step would be great.

    • @Benw8888
      @Benw8888 9 месяцев назад +28

      More training can help it replicate "beautiful" proofs, even if it's hard for us to explain what makes something beautiful- as long as we can label which proofs are beautiful, we can theoretically train an AI on those labels. (technical details: train a reward model on a set of human "beautiful proof" labels, and use that reward model to finetune a LLM using RLHF)

    • @Dogo.R
      @Dogo.R 9 месяцев назад +11

      I think you misunderstand just how vastly more competent human brains are.
      Human brains cooperate, adapt, and compute over months to years to solve a hard problem.
      An AI that runs in even a few minutes is just not even logically close to the computational resources and inherent efficency of what humans do.
      We cooperate, mentally adapt, and do it over months to years.
      That is so ungodly powerful.
      The idea that AI will help solve novel problems is a drastic misunderstanding of how insanely small its inherent computational ability and efficency is compared to a group of humans.

    • @Benw8888
      @Benw8888 9 месяцев назад +18

      @@Dogo.R AI scaling laws and biological anchors disagree. Even the most generous estimates place AI as using as much compute as human brains within decades at most. Already, AI LLMs have >1/1000 the number of parameters as we have synapses, and that gap can be closed quickly with exponential growth, larger monetary investments, etc.

    • @OmateYayami
      @OmateYayami 9 месяцев назад +4

      I think that your focus on the time scales is a proof of how deeply you are mistaken by putting value in how long it takes to solve the problem. There is abundance of historical evidence that the logic is inverse.
      Or, take a look how could apple and ms pass such a great well established company as IBM.
      That's a sunken cost fallacy.
      Human civilizations wouldn't collapse if your logic was true. If you see something that is able to compete and is way younger and uses less resources it means it's more effective and closing the gap.
      Also if you think computers are unable to cooperate in problem solving... Well, yeah...
      Not to mention there groups of people behind those computers. They are still tools. Just as with industrial revolution humans elevated physical strength using tools. This has potential to elevate intelligence.

  • @bloom945
    @bloom945 9 месяцев назад +37

    Pausing at 12:45 and just LOOKING at that diagram... I can't even begin to imagine the pain of animating this

    • @KookyPiranha
      @KookyPiranha 17 дней назад

      it's not an animation, they probably used something like geogebra
      and no, it probably wasnt pain to move around points or highlight regions

  • @EebstertheGreat
    @EebstertheGreat 9 месяцев назад +101

    I think the British pronunciation of "upsilon" and "epsilon" highlights the etymology of the names. Long ago, the Greeks forgot the names for ε, υ, ο, ω, and some other letters. So they just called them by their sound, the way we do vowels in English, like we just call E "ee." But as some of these vowels started to sound like each other, they added words to the names to distinguish them, like "little o" for omicron and "big o" for omega, similar to the way some Spanish-speakers call b and v "be larga" and "ve corta" or similar. In the case of ε and υ, "psilon" meant "bare," on its own, so an epsilon is a bare e. This distinguished it from αι, which had the same sound. Similarly, upsilon is a bare u, as opposed to ου.

    • @eth3792
      @eth3792 9 месяцев назад +30

      I...can't believe I've never put together that "o-micron" means small o and "o-mega" means large o. 🤯

    • @Anhilare
      @Anhilare 9 месяцев назад +3

      not quite-it's bare υ as opposed to οι, which by then had come to sound the same. thus, at that time, ὗς "pig" and οἶς "sheep" were homophones pronounced "üs", because of which both being replaced by χοῖρος and πρόβατον, respectively, in the bible

    • @JivanPal
      @JivanPal 8 месяцев назад

      > I think the British pronunciation of "upsilon" and "epsilon" highlights the etymology of the names.
      It doesn't do this any more than the non-RP / American pronunciation does. Indeed, the Greek pronunciation of "psilon" has the stress on the second sylalble: "psi-LON", not "PSI-lon", and regardless, it's definitely not "SIGH-lon"/"ep-SIGH-lon". The RP pronunciation is merely a result of the application of British/English phonics rules to the English spelling of "psilon"/"epsilon". See also "pi", "phi", "chi", which in Greek are pronounced "pee", "fee", "kee" (well, /xi:/, but close enough), respectively.

    • @EebstertheGreat
      @EebstertheGreat 8 месяцев назад

      @@JivanPal It's not eps + ilon but e + psilon. Neither the British nor the Americans pronounce these letters the same as the modern Greeks or the ancient Greeks, but the way the British pronounce them emphasizes the two morphemes, whereas the American pronunciation does not.

    • @JivanPal
      @JivanPal 8 месяцев назад

      @@EebstertheGreat I disagree that either English pronunciation highlights that it's e+psilon.

  • @ElaineWalker
    @ElaineWalker 9 месяцев назад +7

    @26:45 I'm so relieved you stressed the point of beauty! And also fun! My parents were mathematicians who did research together so I grew up hearing all the language at the dinner table. I had no patience for algebra, and still don't, but in my 50s my mathematician-husband helped me realize I can visualize math in my head (geometry, topology.. an if it isn't a visual math problem I'll turn it into one!). It helped that I already knew how to "talk math." Now we have several published papers. It's sooooo fun. I literally only do it for the joy, and also the hope that some day physics will find a use for our math - we already found a deep connection with light polarization. Anyway, now if I'm not busy with anything else I'm doing math research. So yeah, I'm a little weary of the AI. We're hoping it can eventually take the place of what Bruce does in Maple, cuz that is usually a pain.. but not take the fun away. 🤔

  • @GearHeadedHamster
    @GearHeadedHamster 9 месяцев назад +40

    The animation visually showing the ai's step by step solution must have been a nightmare to animate. But it made fallowing along so much easer. Props!

  • @QuargCooper
    @QuargCooper 9 месяцев назад +75

    I really appreciate you highlighting the point "the AI wasnt designed to make beautiful proofs". Too often in the public discussion around AI, people leave out the fact that the AI has been designed in the first place. It has a reward function - for example, LLMs are designed to mimic human conversations. But this mimicry doesn't require a deeper understanding in order to succeed; it can just cleverly explore the space of possible conversations, in order to see what sounds most like a human. Similarly, here, it is trying to cleverly explore the space of known problem solutions, in order to create a new result. Which is why it looks that way - it always seemed to me to have a lot in common with the "press the middle suggested word on your phone keyboard to write a sentence" game. Or Clippy.
    At least for this flavour of artificial intelligence (which I guess is a name that is sticking now), it seems difficult to see how it can develop novel solutions. It might be useful in order to fill in the gaps of existing fields - e.g. find all possible consequences of a set of theorems. But it seems like making the leap to inventing a new field will be beyond it.
    Not to mention that the usefulmess/validity of any theorems it does come up with are suspect - as mathematicians, we often _need_ to appeal to authority, because there is too much mathematics for us all to hold in our heads. The way we square this is by knowing that with enough effort, we could reproduce/have explained to us the results we need. But because AI is trying to maximise its reward function, it might not actually be generating valid proofs - each proof would need verified. And if its proofs are thousands of steps long, and not guaranteed to be correct, then how can we trust its results? This is a huge problem that needs to be resolved before it can be useful - this is where the application differs from that of e.g Deep Blue, or AlphaGo. We might be surprised at the way an alien plays chess, but we can at least verify whether or not it has beaten us.

    • @gabitheancient7664
      @gabitheancient7664 9 месяцев назад +3

      I think (and that's just a guess) that the AI is designed to be perfectly logical, like, straight up use the theorems it knows' statements

    • @NNOTM
      @NNOTM 9 месяцев назад +11

      As​ @gabitheancient7664 says, the proofs can be automatically verified if the "left brain" part only uses deduction steps that are known to be valid. It's fairly straightforward to guarantee that if it can find a proof, it's a valid proof.

    • @gabitheancient7664
      @gabitheancient7664 9 месяцев назад +3

      @@NNOTM yeah having difficulty verifying proofs or making up one that's valid is a very human thing we aren't very "logical" by nature, but computers tho

    • @Jackson_Zheng
      @Jackson_Zheng 9 месяцев назад +11

      Dude, you literally have no idea what you're talking about. Appeal to authority? What a joke.
      Proofs in mathematics are verified through formal logic. LLMs might not be efficient at finding proofs yet, but the process of verifying mathematical proofs is very well defined. All it takes is a large amount compute.
      As for your point about mimicry, it is literally proven that ANY sufficiently large neural network is able to approximate any function to any degree of accuracy.
      Since any deterministic process (ie. same inputs = same outputs) can be represented as a function, so can your brain and the brains of every single human on earth.
      The only way it can be false is if you deny the assumption that the laws of physics work the same no matter where you are in the universe.

    • @vicktorioalhakim3666
      @vicktorioalhakim3666 9 месяцев назад +8

      @@Jackson_ZhengAll the points that you mentioned are pretty much in agreement with what the OP said. Not sure why you act so triggered by it. Do note that modern theorem provers are not fully automated, and still require a lot of human intervention (e.g. coq, Lean, Isabelle HOL, etc..), which in large part is what an LLM is supposed to substitute (and it is far away from it, just ask LeCun). And just because something can be approximated arbitrarily close, does not necessarily make the approximation useful, feasible, practical and/or good. That's something a lot of people miss when they say that LLMs just need to be scaled up to achieve AGI. No, we are missing something very fundamental besides mimicry that would enable an AI model to accurately represent a reasoning machine.

  • @gabitheancient7664
    @gabitheancient7664 9 месяцев назад +37

    tbh the a fun part of a math olympiad is the fact that you were able to solve it, a human person with a physical brain, other body parts, and limited time, and that needed you to develop some abilities (I mean my main interest in math olympiads is developing problem solving skills, almost a pedagogical objective)
    that seems almost as impressive as saying we could make a machine that beats people in long jumping, like yeah amazing technology but that's not the cool part of seeing someone jumping really long

    • @ram527
      @ram527 9 месяцев назад +4

      This has far reaching implications right, someone who was previously unable to solve such problems can now use AG and boom he can work at the same level as a person experienced with math olympiads. After working on your skills for years you get beaten by an AI, that is bad to me.

    • @ram527
      @ram527 9 месяцев назад

      Basically you dont have an incentive to learn geometry anymore sadly if ag is perfected. you will never be as good as the AI.

    • @mironhunia300
      @mironhunia300 9 месяцев назад +7

      @ram527 We already have (non-AI, non-heuristic) algorithms for solving olympiad and olympiad-adjecent problems like inequalities (many inequalities can be bruteforced by multiplying everything out and applying Muirhead) and integrals (Risch algorithm). Geometry also had these sort of algorithms: analytic geometry is popular brute force approach that is commonly used by olympic contestants. Alpha Geometry is only unique in a sense that it is a form of machine learning algorithms, but it is neither the only nor the first thing that can algorithmically solve math olympiad problems. It'd be surpriced if it affected math olympiads in any way.

    • @gabitheancient7664
      @gabitheancient7664 9 месяцев назад +2

      @@ram527 the fun isn't seeing the solution tho, it's solving it

    • @gabitheancient7664
      @gabitheancient7664 9 месяцев назад +3

      @@mironhunia300 there's also combinatorics and soome nt problems where you could theoretically read the question and just put it in your computer and test all cases, the fun part is having the creativity to make the cases very small and solve it by hand

  • @G.Aaron.Fisher
    @G.Aaron.Fisher 9 месяцев назад +13

    I'm not surprised that this method has results, as I've always thought that one of the best applications of LLMs is providing leads to subject experts that can follow up on them. That subject expert could either be a human such as a lawyer or a bespoke program as is the case here. LLMs are essentially playing the role of general-purpose heuristics to prioritize what nodes are explored by the subject expert in what order. AlphaGeometry is just the first notable instance where a computer program played the role of the subject expert.

    • @iankrasnow5383
      @iankrasnow5383 9 месяцев назад +2

      Well put. It's important people understand that this "subject expert", what they call a deduction engine, is not an AI model; or at least not a machine learning model. It's a more traditional sort of computer algorithm.

  • @jellewijckmans4836
    @jellewijckmans4836 9 месяцев назад +64

    I do think it's important to note that for a lot of mathematical problems just knowing what the answer is might give a human mathematician a better starting point to figure out more elegant and more meaningful proofs even if the AI generated one was a over complicated mess.

    • @jeffrey1025
      @jeffrey1025 9 месяцев назад +8

      Yes, proofs are like code. You can refactor the core ideas into something neater

    • @eragon78
      @eragon78 9 месяцев назад +8

      Exactly. Brute force calculations are used to prove stuff in math all the time. Computers revolutionized many fields of math just because calculations could be run to far faster by computer than by humans, even if the methods were exactly the same. Computers also dont make mistakes unlike humans.
      That doesnt mean computers are always useful for every situation in math, but even the relatively "dumb" computer brute force methods are insanely useful. This type of AI makes brute forcing stuff like geometric problems way more doable as well, and that can be very useful for certain projects. Is it going to be useful for everything geometry related? no. But itll be useful for some stuff, and thats what matters.

    • @mrtthepianoman
      @mrtthepianoman 9 месяцев назад +5

      I agree. Even if the brute force proof cannot be refined into a more beautiful proof, knowing a result is true will allow a mathematician to focus their efforts in that direction. If we knew the Riemann hypothesis were true, we wouldn’t have to waste effort looking for counter examples or trying to disprove it.

    • @mrtthepianoman
      @mrtthepianoman 9 месяцев назад +2

      Also, for important results (like the Pythagorean theorem) mathematicians will often come up with many different proofs. This is not because we need reassurance that the result is true, but because proving the result in new ways may provide new insights. A brute force approach may not be insightful, but it does tell us something.

    • @eragon78
      @eragon78 9 месяцев назад +2

      @@mrtthepianoman another thing to consider as well is, if a brute force solution ISNT possible, then it means there a more novel approach is required. That can also help narrow down where to look or what kind of ideas to be thinking of.

  • @murmol444
    @murmol444 9 месяцев назад +26

    "Or it's neither like my phd thesis"
    I am feeling you

  • @dj67084
    @dj67084 9 месяцев назад +9

    I love this. I, for one, don't feel like AI has to be elegant, perfect, and utterly mind-blowing at first.
    I think a lot of self-proclaimed "nerds" (for the lack of a better word) reward being amazingly good as a requisite for sharing anything on the internet, as opposed to consistent, often imperfect, growth (which I think is the more realistic progression toward anything good)... just like how people will nitpick on every youtube video for not being perfect, we'll nitpick AI and find a million reasons that "it's really not that impressive if you think about ..."
    I'm more excited that we even arrived at this point. It's a step in the right direction toward having an amazing and creative logical supplement to human intelligence. What a time to be alive! Love these videos, I love that you took the time to do a deep dive and I'm super interested in more breakdowns. Subscribed!

  • @huwbelling9901
    @huwbelling9901 9 месяцев назад +6

    There's an epistemological problem here as to the "beauty" of a proof. Turing's supposition, in respect of what we now call LLMs, was that perfect mimicry is indistinguishable from some other immeasurable sign of agency, to the extent we know when LLMs fall down, in the strictest sense they are not passing the Turing test. I suspect that (rather quickly) AI will acquire enough learning to meet our indefinable sense of 'beauty' in a proof - if educated mathematicians are able to 'know it when [they] see it', there must be some set of insights, available to any organic or inorganic reasoning machine, that would include 'beauty' in respect of proofs (or for that matter general reasoning or 'creativity')

    • @afterthesmash
      @afterthesmash 9 месяцев назад +1

      It's not clear to me that Turing believed this. What he certainly believed is that many people naively believe they know the contrary when they certainly don't (that there is for certain some elegant test). If you have such thing in hand, you should be able to distinguish mimicry from the real thing. If we can't distinguish the "real" thing from mimicry, then we are all wet in thinking we know what the "real" thing looks like. When the computer passes the Turing test, we don't actually know it only uses mimicry. It might actually be doing the real thing the same way we do the real thing. Unless you think we mimic ourselves to prove our own intelligence.

  • @proxxyBean
    @proxxyBean 9 месяцев назад +27

    I want to throw problems proven impossible at it and see what hallucinations it creates.

    • @freddupont3597
      @freddupont3597 9 месяцев назад +1

      haha, you read my mind!

    • @iankrasnow5383
      @iankrasnow5383 9 месяцев назад +14

      I don't think it will give you a solution at all. This type of algorithm should be able to throw out any hallucinations generated by the language model portion of it. It shouldn't ever give a wrong answer, but it won't always find an answer. Worst case is it ends up in some endless loop.

    • @k-c
      @k-c 9 месяцев назад +3

      The results wont be any different to most human answers

    • @Kaurenjalalifarahani
      @Kaurenjalalifarahani 2 месяца назад

      @@iankrasnow5383everything makes mistakes , top computers , and even the best chess algorithms which have been more developed than math algorithms sometimes make mistakes , no matter how narrow the probability is its never zero

  • @mathcat4
    @mathcat4 9 месяцев назад +15

    Hype! As a contest maths enthusiast, I'm curious what AI can do in this field.

    • @G.Aaron.Fisher
      @G.Aaron.Fisher 9 месяцев назад +7

      I think our ability to solve contest mathematics with computers has always been bottlenecked by our ability to formalize the mathematics involved, rather than our ability to develop new solving algorithms.
      Areas that have been formalized (e.g. algebra, calculus, and probability) have been child's play for WolframAlpha and the like for decades.
      There have always been a countable number of statements that can be proven from any given setup, and only a finite number of them that can be proven in a fixed amount of steps. Computers' abilities to explore several more orders of magnitude of that space is easily going to overpower humans' abilities to draw on experience and pattern recognition to explore the right paths. This is especially true given our ability to codify more and more of our intuition into these programs.

  • @An-ht8so
    @An-ht8so 9 месяцев назад +14

    I've always thought that for a computer to be good at math, it would need to be able to gauge how close it is to finding a proof. A chess computer is strong because because it knows wether or not the position is good/winning even if it hasn't found a winning sequence. In other words, the math computer would need to know how easy/hard something is to prove with respect with what it knows, even if it doesn't know the proof. Now that would be impressive.

    • @An-ht8so
      @An-ht8so 9 месяцев назад +1

      Now, I also that there is some hope for AI to discover meaningful things and not just prove give propositions. Given a dataset of results with their proofs we could try to identify results that would significantly simplify the existing proofs. However I don't really see how an AI would be able to define useful new objects, which is th richest type of math innovation. This amounts to the AI enriching the language it is working with.

  • @saiphrivas1437
    @saiphrivas1437 9 месяцев назад +20

    The first thing I'll do will be looking at the description and crying in pain upon discovering the empty investigator section

    • @AnotherRoof
      @AnotherRoof  9 месяцев назад +4

      I apologise for the lack of developments recently!

    • @yours-truely-sir
      @yours-truely-sir 9 месяцев назад

      same here

    • @saiphrivas1437
      @saiphrivas1437 9 месяцев назад

      ​@@yours-truely-sir YOU. YOU ARE LITERALLY THE ONLY OTHER PERSON I'VE SEEN IN THE COMMENTS TRYING TO CRACK THIS. we should be friends XD

    • @saiphrivas1437
      @saiphrivas1437 9 месяцев назад +1

      ​@@AnotherRoofat least the videos are great, keep it up!

    • @yours-truely-sir
      @yours-truely-sir 9 месяцев назад

      how far in the puzzle have you gotten so far, and have you solved the countdown one?@@saiphrivas1437

  • @gejyspa
    @gejyspa 9 месяцев назад +2

    When I was taking AI in college back in the 1980s (yes, I'm Old (TM)), one of the neat programs they talked about was Automated Mathematicians (see wikipedia), which starting from real rock bottom definitions about sets and used heuristics to generate interesting lines of inquiry, eventually recapitulating Goldbach's Conjecture (although reading through these things now, it may be that the author overstated the case)

  • @jaafars.mahdawi6911
    @jaafars.mahdawi6911 9 месяцев назад +1

    Mathematics as logical poetry, love this, being a math major and poet myself. Also the video itself is very well done. Keep it up!

  • @afterthesmash
    @afterthesmash 9 месяцев назад +3

    The only difference between this and brute forcing Sudoku with all the codified inferences is when you have to add lines to the diagram, and there's an inexhaustible supply of novel lines to mess around with, and you could end up chasing angles forever, and never arrive at your desired conclusion.
    If it was training on past Olympiad problems, then it has a good statistical model of what kinds of lines you might need to add for the kinds of problems that test tends to pose. And then you're pretty much back to brute force again, because the space of lines you might need to play with is relatively closed relative to your statistical model.

  • @Tuned_Rockets
    @Tuned_Rockets 9 месяцев назад +39

    the thing i notice is that it doesn't seem like the AI part of this is very important. i'm pretty sure it would be easy to code up a conventional program to play the part of left/right brain. it feels more like another thing trying to capture the AI craze than a real advancement.

    • @biner01
      @biner01 9 месяцев назад +4

      Is it not half of the algorithm? I'm not sure AI is difficult to code (in general). You do need resources to train it though,

    • @electra_
      @electra_ 9 месяцев назад +4

      Yeah. I imagine the AI helps somewhat mainly in directing the results toward the actual answer. Effectively, all you need is some measure of distance from the current position to the proof and it's just like, a basic pathfinding algorithm.
      I imagine the actual practical use of this system though would be to just integrate the part which is actually consistent into programs mathematicians use, so they can choose what new things to add while the computer automatically chases angles and does the "dirty work". Which, means that you're effectively throwing out the AI part in favor of a real intelligence and just using the already existing proof checking techniques with a bit of automated search applied.

    • @iankrasnow5383
      @iankrasnow5383 9 месяцев назад

      @@electra_ I just browsed the paper, and you pretty much hit the nail on the head. The actual deduction engine is not based on neural networks. It's based on methods that were designed decades ago and are human-understandable. This kind of program just outcompetes humans in the way computers always have: by doing a lot of logical operations very fast. The major advance made was figuring out a way to train the language model, since there isn't enough training data. They avoided this problem by algorithmically generating millions of problems and their proofs, and then training the language model on that.
      AlphaGeometry isn't better at the actual problem solving part, it just has a better language engine for converting a human-readable problem into a usable form, and then judging each attempt based on heuristics. The language processing part isn't doing anything that would be difficult for a human. The deductive engine is, but it's just a traditional algorithm, not an "intelligent" algorithm
      A trained human with the benefit of such a deduction algorithm on their computer would vastly and easily outcompete AlphaGeometry. A general LLM like GPT without any deduction engine would not be able to solve these problems at the current time.

    • @hedgehog3180
      @hedgehog3180 9 месяцев назад +3

      The AI part just seems like an interpreter that gives instructions to what is just a maths program. That potentially has uses and is sorta cool in the general sense that LLMs enable computers to more directly tackle problems you'd need a human to input before but also like all LLMs it isn't doing anything a human couldn't do and it does it at a fairly mediocre skill level. It's really just the general issue that current AI seems like a solution in search of a problem that isn't just producing spam.

    • @electra_
      @electra_ 9 месяцев назад +2

      I think what this does sort of show is that in order for an AI to be actually useful in a specific context, especially one where you need to be objective, you sort of need to heavily restrict the AI by placing it in the middle of a larger system, and have components that can validate its input.

  • @giacomohermesferraro6673
    @giacomohermesferraro6673 9 месяцев назад +1

    5:22 The simplified version is sufficient because if the three circumferences were distinct, their pairwise radical axes should intersect, but they are the three sides of the triangle, which have no point in common

  • @СергейМакеев-ж2н
    @СергейМакеев-ж2н 9 месяцев назад +2

    To be fair, "International Olympiad with a calculator" is a different category of competition from just "International Olympiad". And a symbolic deduction engine is basically a very powerful calculator.
    Still, even in this other category, it's an impressive achievement.

  • @crazygoatemonky
    @crazygoatemonky 9 месяцев назад +2

    When I did math comps we called this approach "geometry bashing". When you have no idea what else to do, draw some lines and work out the angles, see if you get anything useful. The most generous interpretation is that AI has mastered the worst technique high-schoolers use. And that's the only tool in their toolbox. Obviously it's very cool that somebody taught an AI how to geometry bash, and it suits computers' strengths fairly well, but this is leagues away from anything generalizable.

  • @TheCommuted
    @TheCommuted 9 месяцев назад +2

    I realized AI was doing original research when a smaller LLM told me butterflies had no legs. I'm sure it looked at a lot of butterflies.

  • @TheRealJeffDiamond
    @TheRealJeffDiamond 9 месяцев назад

    I also love the optimistic view of the video - that this could be a worthy foundation. This is the "messing around in the dark" part in coming up with a solution. If you could build on the solution - simplify it - connect it to higher level concepts in other proofs - you might start raising its beautification score in the answer. Then you'll get papers saying "our new solution averaged 0.26% more beautiful solutions than other leading AI mathematicians over a course of 20 standard proofs. ;)
    Kudos for this video shouting out that we SHOULD build on this solution and we SHOULD have beauty as a metric.
    As an aside, "I can't describe why it's beautiful, but I know it when I see it", is absolutely the hallmark of a neural net, so you must be using one in your head. :)

  • @RichardGrigonis
    @RichardGrigonis 9 месяцев назад +2

    It makes perfect sense if you imagine an evolutionary algorithm behind human problem-solving and learning. Something generates lots of mutated procedures, tests them, mates them, reproduces them, and then through Darwinian (or even Lamarckian) evolution, comes up with ways of proving and/or learning stuff. As the Wiki for the No Free Lunch Theorem says, "The 'No Free Lunch' Theorem argues that, without having substantive information about the modeling problem, there is no single model that will always do better than any other model. Because of this, a strong case can be made to try a wide variety of techniques, then determine which model to focus on."
    Through optimization (again, using something like genetic algorithms) the system learns better and better ways how to prove things and how to learn.

  • @GhostyOcean
    @GhostyOcean 9 месяцев назад +8

    Can't wait to watch this with everyone later!

  • @optozorax
    @optozorax 9 месяцев назад +290

    Rename this video to "AI can solve geometry, buuuuut...", and do thumbnail like "left: human - 12 steps vs right: AI 49 steps, and draw simple proof on left and complex proof on right" and you will get a ton of views. I thought your video will be just paper explanation, and not your deep look into the theme. Please search a new title and thumbnail, that express that core of your video. It's really good, but current thumbnail is not that catchy.

    • @TheALLVEGAS
      @TheALLVEGAS 9 месяцев назад +3

      Spot on

    • @MrAdamo
      @MrAdamo 9 месяцев назад +18

      This video is his highest viewed in a few months, and its larger than his subscriber count. It is objectively catchy.

    • @herobrine1847
      @herobrine1847 9 месяцев назад +38

      Maybe he’s satisfied with his work and doesn’t want to play the algorithm game

    • @optozorax
      @optozorax 9 месяцев назад +8

      @@herobrine1847 it's not only about playing algorithm, it's about reaching certain people. I read every article and wathched every video about AlphaGeometry, because I obsessed with new ML advancements. At first I thought that I will learn nothing new from this video, but I started to watch it by mistake and pure chance.
      So, there are may be a lot of people who will not watch this video, because they already watched Yannic or Sabrine and thought like me, but they want information from this video. But my variant of a thumbnail can boost their interest.

    • @Oricorio-Pom-Pom
      @Oricorio-Pom-Pom 9 месяцев назад +17

      @@optozoraxwhile true, this thumbnail appeals to a different audience of people, your suggestions just shift the audience, not necessarily expanding it.

  • @knnob9
    @knnob9 9 месяцев назад +1

    This video is a wonderful review of the state of the art in the development of assisted proofs, taking alphageometry as a case study!

  • @firefist3684
    @firefist3684 9 месяцев назад +19

    What happened to the first roof?

  • @varunahlawat9013
    @varunahlawat9013 9 месяцев назад +1

    The amount of work put into this video is crazy

  • @goodtothinkwith
    @goodtothinkwith 9 месяцев назад +3

    Really excellent video! I now understand why others haven’t taken a deep dive into how Alpha Geometry works 😂

    • @AnotherRoof
      @AnotherRoof  9 месяцев назад

      I had this exact thought. "Why isn't anyone covering this in detail??" Then when I made lots of incredibly messy Geogebra diagrams: "Ah. This is why."

  • @u.v.s.5583
    @u.v.s.5583 9 месяцев назад +1

    Honestly I find angle chasing done well extremely beautiful and creative. In fact, more creative than the shortest solution You gave. Just my honest opinion. Beauty lies in the eye of beholder.
    Does it enlighten? Of course it does. It introduces all those angles, giving a nontrivial parametrization of the layout of the problem in question.

  • @logician1234
    @logician1234 9 месяцев назад +244

    That was a really cool video! (don't mind me I'm just trying to confuse your future viewers)

  • @HerbertLandei
    @HerbertLandei 9 месяцев назад +52

    As a chess player, I can understand your obsession with "beauty", but you have to get over it. In chess, sometimes your "beautiful" sacrifice or opening line is not correct, because a chess engine like Stockfish finds an "ugly" refutation. But sometimes, the "ugly" computer move reveals a brilliant and beautiful idea. Computer chess changed not only the game profoundly, but also its aesthetic. And it won't be any different in geometry, or generally in mathematics, even if the first steps look crude and ugly. It will open a whole new world, whether you like it or not.

    • @bennettzug
      @bennettzug 9 месяцев назад +10

      i think beauty in math is more than superficial, as it serves to more effectively enlighten the proof reader

    • @EebstertheGreat
      @EebstertheGreat 9 месяцев назад +6

      But did it change chess for the better? I mean, play is now higher in quality, but is it as exciting, or beautiful, or fun? A lot of GMs seem pretty unsatisfied with the current state of chess at the top level, what with it becoming so drawish and dependent on specific preparation.

    • @mouldyvinegar5665
      @mouldyvinegar5665 9 месяцев назад +11

      Proofs and chess are two very different things; you're comparing apples to oranges. There are objective, mathematically "good" moves in chess, so we have some measure of what moves are "beautiful" - that is what moves win the game. They don't think it's ugly in that it is a bad move, but that it goes against what players are used to - in fact these are very insightful moves, and you appreciate the beauty in the move once you understand the reasoning. However, there is no such thing as an objectively "good" proof. There are only correct and incorrect proofs - otherwise you are given pretty much entirely free reign as what tools you want to use. He is not refusing the idea that AI will not get better (and very much states that it could improve in the future), but that in the current state, it's not giving elegant, insightful proofs (and it is subjective as to what makes a proof insightful or elegant or whatever). Rather, AlphaGeometry is just angle bashing, which is something that pretty much anyone and anything can do.
      A more apt comparison would be asking a GM what makes a move good versus asking a brute force algorithm. One could explain to you what moves to look out for, what lines are threats, etc. The other would just throw millions of possible futures in your face and say it just works. What explanation do you think is more beautiful?
      Finally, why do you think 3b1b has such a wide audience? Everything and more I can learn from textbooks, but it's almost as if there's a beauty as to how he makes his videos...

    • @niwoh
      @niwoh 9 месяцев назад +1

      @@EebstertheGreat Maybe not, but is that relevant? Chess is a game, math is about understanding reality

    • @ChaoticNeutralMatt
      @ChaoticNeutralMatt 9 месяцев назад

      I think it's like cleaning. They tend to be emotional responses (too much, ignore or stress, or compulsory). Not bad or good, just there.

  • @cyb3r._.
    @cyb3r._. 9 месяцев назад +1

    21:58 well, deep blue actually lost to Gary Kasparov the first time they played, but its search algorithm was improved and it beat Kasparov on their rematch a little while later

  • @copiryte9535
    @copiryte9535 9 месяцев назад +4

    Great video, I really loved the insights into why we do proofs, I definitely agree with what you were saying about the beauty of proofs. For me Is if I could have thought of it myself? As in could I guess the next step of the proof or have an idea of the result you're trying to get to without already reading through the proof. AI just randomly works out enough information to try to guess of what the inner workings of the result which it does with enough trial and error, maybe this is the way a human could get to the answer but it isn't satisfying.
    Seeing the underlying structure of a beautiful proof allows us to abstract those ideas used in the specific proof to other results in the same beautiful ways. To me that's what mathematics is all about seeing those connections between not necessarily intuitive topics but being able to construct a framework where those connections do become intuitive, that's why results like Fermat's last theorem are so satisfying...

    • @blumoogle2901
      @blumoogle2901 6 месяцев назад

      This is all true, and yet I feel that an AI proof, even a long, ugly, unintuitive, meandering proof - as long as it is probably correct - would serve as a very solid foundation for a human mathematition to know that a result is achievable and thus refine-able into a beautiful, concise, intuitive proof that does everything you could want.
      I feel this way about AI in every field - it replaces brick-makers so that everyone who used to make bricks can focus on making walls, Wall-makers can make rooms, room builders can make buildings and those used to making buildings can start to design neighbourhoods. It frees every thought worker to stand on a higher foundation and increases the scope of what they can do.
      So what if AIs start replacing the need for mathematitions to find proofs to simple problems; all this will do is free up those same mathematitions to find proofs for entire systems of related proofs instead of doing them one by one - which I think further improves math in gulps and bounds instead of sips and steps.
      This is my view on why I don't mind AI making static images; I expect the artists who used to make static images to instead shift their focus to coordinating/creating collages, videos and interactive museums of static images; AI that makes music - great. I expect musicians to be freed from making sounds to conducting orchestras, running virtual symphonies and theatres with many pieces of music.
      As an analogy: If farmers no longer have to plow because they have tractors, I expect the farming world to improve and the role of farmers to become coordinating different tractors working together and that this is actually an improvement in the role of farmers and their scope rather than their replacement.

  • @RealKingOfTheLab
    @RealKingOfTheLab 9 месяцев назад +2

    The brute force attack of any problem presented to AI does have the simplicity of use very base, core postulates. Where a human would use multiple time-saving theorems that each contain their own elaborate proofs, the lack of using or identifying these upper level theorems to solve a problem doesn't make them less real.
    Does brute force like this take longer than using proved theorems to short-cut the system? yes.
    Does such a longer process matter in a time scale that computers operate on? Not always, but sometimes it does.
    To make this AI Math algorithm faster, and more intuitive, programmers WILL be able to program future versions to use more and more proven theorems... HOWEVER...
    Is AI able to invent, or even know to invent, new theorems for future stream-lining? Not at this time... but that's the goal isn't it.

  • @Zicrus
    @Zicrus 9 месяцев назад +2

    It can also be solved fairly trivially with radical centres, without knowing anything about cyclic quadrilaterals

  • @_Blazing_Inferno_
    @_Blazing_Inferno_ 9 месяцев назад +3

    18:45 this was the way I solved self-posed problems when I was young - playing around and brute forcing things because I didn’t have the tools or experience yet. I still do this to an extent today as an undergraduate student (CS and Math double major), where I often have to solve the problems I mess with using some concrete examples to catch most patterns

  • @Benw8888
    @Benw8888 9 месяцев назад +15

    Some comments on beauty: As an AI researcher with a math degree, it's 100% possible for us to develop AI that generates "beautiful" proofs. As long as you can label whether a proof is beautiful or not, we can train an AI to optimize based on those labels. (Even if you don't understand what makes it beautiful, as long as you can provide a training signal, you can still optimize for it- that's one of the very powerful facts about AI)

    • @pedroteran5885
      @pedroteran5885 9 месяцев назад +4

      You can optimize for proofs similar to those, but they wouldn't necessarily be similar in the aspects that make them beautiful.

    • @JGLambourne
      @JGLambourne 9 месяцев назад +4

      Just what I was thinking. We need a big "beautiful proofs" dataset. I think the way to do this will come via automatic formalization of hunan generated results. It feels repetitively feasible to consume maths papers and spit out formalization in something like lean, at least within the 10 year timeframe

    • @BobJones-rs1sd
      @BobJones-rs1sd 9 месяцев назад +3

      I'm curious to know more about your optimism in this regard, because my immediate impression is that this problem will be a LOT harder than simply feeding in a training set of "beautiful" proofs. I'm actually very optimistic about AI overall (after being a skeptic for many years), and I'm amazed at the progress of LLMs in the past couple years.
      But there are things I think AIs and this type of learning model are quite poor at. Specifically, they're good at generalizations from large numbers of examples (which is I think where you're coming from), but they're also often quite poor about being able to pick out details in some cases. This comes out of the fact that training requires specific goals (again as you note) and to an AI, it may sometimes have no idea why some seemingly minor details are important.
      We saw this in image generation for a long time, though recent models are starting to get better. I'm talking about things like how the AI images often would create models with 4 or 6 fingers. To a human, this is an obvious problem and flaw, and if you don't understand how fingers work, you don't understand how we touch things, hold hands, etc., hence making a lot of awkward weirdness in AI imagines. To an AI, it's often a seemingly minor detail in an image, some smaller cluster of pixels that it has no reason to realize the importance of or optimize for.
      Now, if we see that specific flaw, we can optimize of course and provide better training data, with better goals for the algorithms. But even in language learning, we see this fundamental lack of "understanding" from LLMs about basic concepts.
      For a long time, it was difficult even with the language fluidity of ChatGPT for example to understand that a haiku contains a pattern of syllables and to emulate that. It was even harder to get it to understand how to emulate the meter, specific number of syllables per line, and rhyme scheme of something like a sonnet. Even though GPT models could easily tell you what a sonnet was in great detail, even interpret what a lot of those definitions about meter and poetic feet mean, it had no "understanding" of how to apply that to generate a poem in that form, even after having been fed probably tens of thousands of sonnets in its training base.
      It just didn't know it should optimize in generating a sonnet around concepts like accent, meter, and rhyme, because even though it can quote definitions, it doesn't "understand" them in the way a human does.
      I'm not at all saying AI won't eventually develop such types of understanding, but to me while I'm hugely impressed by what it can do now, it still often lacks the ability to generalize a "concept" as a human would, and explaining such a concept to it often doesn't quite work either (though surprisingly it sometimes does, which gives me hope). While even a pretty dumb kid can understand how to create a haiku at least in terms of syllable restrictions very quickly, as they know what a "syllable" actually is and understand that concept.
      Still, one could still train the model to have some understanding of poetic meter, and it seems GPT models have improved over the past year in that regard. But AI wasn't able to extract this knowledge, I don't think -- but you can correct me, simply from having thousands of examples of sonnets in its training set.
      So now, going back to geometric proofs, the problem with "beauty" or "elegance" in proofs is that each new proof is its own optimization problem, often specific to that particular proof (or at least a small class of proofs). What makes something elegant in math is that it demonstrates true "understanding" of a concept, which AI models really seem to struggle with. And worse, each proof often has different concepts that need to be highlighted or optimized for.
      In the video example, for example, an intuitive deep understanding as noted can come from a theorem about secant lines. I'm sure this AI model can be given a long list of geometrical theorems, but how will it realize that in this particular case, such a route is not only more efficient, but "beautiful" because it highlights an intuitive understanding?
      To me, as someone who taught basic proof-based geometry to human students for quite a few years, I realize how humans gradually build up a set of tools and theorems which create shortcuts. But which shortcut is helpful for a particular proof in order to make it "elegant" is often a very non-intuitive process. It comes from doing lots of proofs, knowing lots of theorems, knowing which ones may be applicable to particular situations, and then also knowing how humans learn and understand, so stating the proof in that "beautiful" way makes a human go, "Oh... yeah, that makes so much sense now!"
      That feels to me like every single complex geometry proof is kind of its own little "figure out how a sonnet works" type problem for an AI. Not a general class of things that can easily be grouped under one category of "beautiful." And just like it's harder to highlight to an AI how to decide non-obvious features are important for generating a "sonnet," I think each geometrical problem beyond really basic proofs will have their own internal non-obvious features that need to be highlighted and optimized for before generating a "beautiful" proof.
      So, to me, this feels like something that is several orders of magnitude degrees harder than you make it sound. Simple labels of "beautiful or not" aren't going to be enough here, at least with how LLMs seem to work now. Maybe with lots more computing power, eventually a more general AI with actual reasoning and conceptual "understanding" can be developed that approaches the level of, say, a 5-year-old human child. Once we're there, I imagine grasping the concept of a "beautiful proof" won't be that much farther away, as understanding can grow exponentially (theoretically) with an AI.
      But for right now, I'm skeptical about your approach in this case. Maybe in a few years things will advance further, but "elegance" in math feels like a much harder problem to tackle.

    • @BobJones-rs1sd
      @BobJones-rs1sd 9 месяцев назад +1

      @@pedroteran5885 I agree. AI algorithms currently are fantastic at optimizing toward a specific quantifiable goal. "Beauty" in the case of proofs like these isn't easily quantifiable, and if it eventually will be, it will have to involve some very complex multistage set of concepts (having to do with how human intuition works, as that's often what makes proofs "beautiful"). For now, it feels like except for the most trivial classes of proofs, what makes something "beautiful" in a specific proof is often quite particular to that situation, not an easily generalizable class of similar things. It's not even just a search for a "shorter" proof (even if that can be meaningfully quantified), as often rather terse proofs can be very non-intuitive and not insightful at all. What you're really optimzing for with "beauty" is something like, "Find me a proof that feels really intuitive to a human." And AIs are good so far at a lot of things, but they definitely don't understand the nuances that connect concepts in what humans often feel to be "intuitive" ways.

    • @Benw8888
      @Benw8888 9 месяцев назад +3

      ​@@BobJones-rs1sd ​ There is a whole field of AI research called "AI Alignment" that is designed around making AIs do what we want them to, even if we are unable to describe what that means. Basically, through RLHF, you can sorta train AI to follow vague concepts like "be moral" and "do a backflip". It's not perfect and it's an actual worry whether we're aligning AIs correctly, but for "beautiful math proofs" I think it'll be enough.
      TLDR as long as you can label proofs as "beautiful" or not, an AI can learn what you mean by it, even if you don't understand it yourself and can't explain what makes it beautiful. This is highly nontrivial and is part of the magic of AI.

  • @transponderings
    @transponderings 9 месяцев назад +1

    I think you’ve hit the nail on the head with your observation about the aesthetics of mathematical proofs. I think current AI tools are a long way from ‘understanding’ beauty, and I suspect that in mathematics, as in other domains, they will show a lack of a creative spark, producing work that is derivative and soulless, with proofs that fail to capture our imagination.

  • @PeridotFacet-FLCut-XG-og1xx
    @PeridotFacet-FLCut-XG-og1xx 9 месяцев назад +1

    For me this really highlights the role of serendipity, i.e. being LUCKY, in discovery. There could be problems that are deductively or analytically solvable with our current information, but in reality will never be discovered because the universe never produces the right conditions necessary for someone to generate that idea in their brain (genetic, upbringing, education, that cup of coffee in the morning etc.)... Or in this case, that electric surge in the transistor and lava lamp configuration in San Fransisco that would've pushed the language model to generate a useful step.

  • @MrSonny6155
    @MrSonny6155 9 месяцев назад +1

    It's hard to comprehend the true difficulty of phrases like "aligning AI with human values" until you are faced with real situations like this. Any human value I could brainstorm would have been superficial compared to so human as the *soul* of mathematics or the cultural necessity of beauty.

  • @dliessmgg
    @dliessmgg 9 месяцев назад +5

    My question for the creators of this AI is the same as for those that generate text or images: why do you create AIs that do the jobs that people want to do?

    • @moussaadem7933
      @moussaadem7933 9 месяцев назад

      Making impressive stuff is fun

    • @VirusTree1000.
      @VirusTree1000. 8 месяцев назад

      The absolute stupidity of your comment.

    • @CeuAzulll
      @CeuAzulll 8 месяцев назад

      Progress. Why we discovered the fire? Why we created/discovered math? Why we create physics? Why we start organizing in society? Why we look up at the sky and started questioning the reality? Why we created the wifi, internet, medicine and all other things? AI has already been use to decodify and represent in 3d models for the first time a molecule, and this help and save so many people, AI are being used in nuclear's fusion reactors to produce more efficient and clean energy . The progress is beautiful if used correctly.

    • @carultch
      @carultch 4 месяца назад

      @@CeuAzulll Progress to causing mass unemployment, IS NOT PROGRESS.

  • @ethanlabelle737
    @ethanlabelle737 9 месяцев назад

    I like the perspective of framining proof of a theorem as a search problem. All thats left to do is intelligently search the space.

  • @vib80
    @vib80 9 месяцев назад +5

    Yes. Deep Blue couldn't be easily turned around to play Go. But AlphaGo was quickly turned around into AlphaZero, which didn't even need the examples of expert play. Just given the rules of various games (Go, Chess, Shogi) and a day to study, it could play them at expert levels. That's more typical of what to expect for the development of this system.

    • @realGBx64
      @realGBx64 9 месяцев назад +1

      To be fair, deep blue had to basically be “tuned” to play against Kasparov while Kasparov was not allowed to see how Deep Blue plays to develop his strategy.

    • @vib80
      @vib80 9 месяцев назад +1

      @@realGBx64 No need for the "to be fair"... that's the point. Because of tech limitations (hard and soft), Deep Blue was very specialized to the task (special chess chips and programmed knowledge). The special hardware in modern systems is for general deep learning calculations, and the programming for AlphaGo itself was not an expert system (the programmers weren't great Go players, or relying on such knowledge to tune things). You don't have to work at the low levels to do powerful stuff anymore.

    • @afterthesmash
      @afterthesmash 9 месяцев назад +1

      It was an ego war. IBM had a huge ego about being the first computer to crush the world champion, and Kasparov had a huge ego in being the last human to withstand this incursion. When they negotiated the terms of the match, all this was taken into account.
      Kasparov had a calculation that computers wouldn't be able to solve the problem of long tactical horizons in that time frame, and he felt he could steer the program into positions where this blindness would prove fatal. He got the computer into precisely such a position, and then it the chose the move he was convinced it couldn't find. This crushed his confidence about winning, and he made some blunders over the board subsequent to this.
      Later he sued IBM to justify how the computer had been able to make this move, and they came back with the rather unsatisfying explanation that the computer had crashed, and that a safety system had randomly picked that picked that move so that the computer didn't default on the whole game. This makes no sense to me as a safety system, since computer chess algorithms have long had a structure where you solve to a certain depth, order the available moves accordingly, then solve again to a higher depth, until you judge that your improvement won't justify a further time investment. This tends to be the best way to do it because the transposition table makes the repeated analysis rather inexpensive, and the successively improved order of evaluation makes the alpha-beta pruning more effective.
      When you do that, you always checkpoint the best known response from the last level of analysis, so there should always be a good checkpoint available, and you should never have to resort to a random move selection (unless you crash on the first ply at the first depth, in microseconds, suggesting you should start again and hope for better).
      I didn't actually read the details of this imbroglio, but I collected a few points from much peripheral discussion. I've also viewed an entire DVD documentary where Kasparov is allowed to litigate whether IBM cheated. It seems unlikely IBM needed to cheat, the program was incredibly strong, but the final explanation from IBM was not a good look.
      In any case, Kasparov was doomed in a year or two at the rate of progression, and it was just a matter of time in any case.

    • @iankrasnow5383
      @iankrasnow5383 9 месяцев назад

      AlphaGeometry doesn't work much like AlphaGo/AlphaZero. Those are "simple games" in that they have simple rules and a limited set of moves. AlphaGeometry is a transformer based language model (like GPT) connected to a hand-coded "deduction algorithm" which doesn't use machine learning. The language model was trained on millions of computer generated geometry problems/proofs. This type of program is only as good as its human-coded engine allows it to be.
      AlphaGo is much simpler and more elegant, which is why it's so powerful. It's a machine learning algorithm that learns through play rather than through extensive tree-search methods like older chess engines. AlphaGo has limitations too. It tends to be strong against the world's best players, but it can be overwhelmed by rudimentary strategies that would never fool a good player, but haven't come up during its training. For instance, a year or two ago, someone discovered that a simple technique of surrounding AlphaGo's tiles, letting them win every time. (i forget the details since I don't play Go). Exploits like this for fooling adversarial neural networks will always be a problem.

    • @hedgehog3180
      @hedgehog3180 9 месяцев назад

      @@vib80 But that's exactly what this is though, the actual AI portion of this is not the one doing the math and actually figuring out the proof. The AI here is just an interpreter that also tries to make suggestions for what to do next and then lets a totally normal math program just sort through everything until it hits a dead end. So they can't actually generalize this to other problems until someone figures out how to make an AI that itself can do math, as in not an AI with a CAS module bolted onto it. Until then this is basically how a lot of AIs in strategy games have worked where there are two seperate systems that pass ideas back and forth to each other in order to limit how long the AI spends looking for the next move.

  • @HebaruSan
    @HebaruSan 3 месяца назад

    A very rudimentary theorem prover is an accessible project for a computer science undergraduate without any LLM or neural net tech. Any axiom or theorem can be translated to disjunctive normal form, and any set of such statements can be trivially brute-force churned until a proof-by-contradiction is reached. The hard part is making it _good_ and pruning the junk out of the proofs.

  • @toasterlogan
    @toasterlogan 9 месяцев назад +2

    AI makes me worry about being a math major sometimes. If AI performs math better than most mathematician, the people who are passionate about math and want to pursue a career in math are less likely to get hired in the future. It's really conflicting feeling like I'm committing myself to a profession that could start to die out in my lifetime.

    • @Am33304
      @Am33304 8 месяцев назад +1

      If it’s comforting to you, which it is to me, consider seriously that “AI” is not a threat to maths nor to working mathematicians. I think it’s a technological dead end. Some people will cling to it with clawed hands but it’s too expensive to build and maintain, and it’ll be too expensive to manage and administrate. The economy will take care of AI on its own.

  • @AntonioLasoGonzalez
    @AntonioLasoGonzalez 9 месяцев назад

    AH is perpendicular to BC which is parallel to EF (midsegment), therefore AH is perpendicular to the line through the centers of the two circles, and it passses through H, a point the we know is on the radical axis of the two circles. We conclude that AH is the radical axis of the two circles and so AB2*AB1=AO*AH=AC1*AC2 implying that B1B2C2C1 is cyclic and we are done.

  • @rupertmillard
    @rupertmillard 9 месяцев назад

    Great video
    Thank you for going into this depth. I agree with your conclusion.
    At 25:33, I would argue that your PhD programme / qualification fulfilled both useful criteria, even if the thesis itself did not:
    (1) It was of immediate practical benefit because it qualified you to take the next step in your career
    (2) It equipped you with the personal methods (resilience, perseverance, wider imagination) required to work on hard problems where the results are not already known - I could never do this

  • @aship-shippingshipshipsshippin
    @aship-shippingshipshipsshippin 5 месяцев назад +1

    so have you seen how new ai from deepmind got silver medal in imo? its amazing

  • @tunafllsh
    @tunafllsh 9 месяцев назад +1

    Thanks for covering this paper elaborately. Seeing it solve the geometry problem like a middle schooler make it clear AGI is still very far from reality.

  • @Dogo.R
    @Dogo.R 9 месяцев назад +4

    Always gotta be carefull about almost all "tests" as they go through human weaknesses that reality doesnt.
    I'll explain.
    For example humans can remember insane amounts of things... but remembering a few things quickly and recalling them a bit after is hard.
    And its WAY harder if you ask them to use their brain for other stuff in between.
    And even harder if those things dont just need to be remembered but need to be used in the proccessing. Geez they are going to have an extremely hard time. Their familiarty with the concepts has to be insanely good to efficently encode and hold that information in such an on demand way and while asking them to leave brain power to do computation on it.
    Computers can save data, use it for logic, then rewrite it like its litterally nothing.
    This results in almost all "tests" being vastly favored towards computers... because most tests are lots of small problems that you need to load into your brain, do work with, then compute and answer, and repeat.
    Which is just not like normal real life problems that humans are good at.
    Real live problems you spend hours and hours and hours and months and months and months working on.
    And along with other humans and other stimuli.
    And with tons of hours of sleep to restructure the brain for being better at solving the problem.
    Almost all "Tests" are very much not like reality AND and very much favored towarda computers and their memory model.
    Only tests that are very long term act like reality. Reality is humans solving problems over long periods of time.

    • @Dogo.R
      @Dogo.R 9 месяцев назад

      Also for the math topic. Think about mathamatica. Is mathamatica "a big step towards super intelegence"?... I mean its an ungodly powerful tool for math.
      Anyways just a seed for thought.

    • @hedgehog3180
      @hedgehog3180 9 месяцев назад

      Hence why LLMs aren't writing papers or doing research.

  • @elshadshirinov1633
    @elshadshirinov1633 9 месяцев назад +2

    AI is still missing this "big picture" part. When this is somehow added, then we do have AGI. Not sure, if this might just happen from training the next monster model.

  • @antoine2571
    @antoine2571 9 месяцев назад

    The video is really incredible. I learnt so much, wether about geometry or AlphaGeometry.
    Thank you for explaining it, I was very intrigued about how it works but never took the time to get my hands into it.

  • @empireempire3545
    @empireempire3545 9 месяцев назад +1

    I would argue that there is a very pragmatic and material value to beauty - as a subtype of things which brings us pleasure. Pleasure is necessary to manage our moods. People in bad moods work less efficiently, and moreover are less cooperative which further impairs the performance of the whole system they participate in. Conversely people with strong positive charge usually produce much more results of better quality.

  • @dard1515
    @dard1515 9 месяцев назад +1

    Reminds me of a younger me, doing a dozen fractions because I refused to learn long division.

  • @nyanbrox5418
    @nyanbrox5418 9 месяцев назад

    22:13
    This is a good point, though I think there is no reason to only report on 1st place, considering 10 years back it was more a consensus among regular people that computers could never even attempt the kind of tasks that it is doing at the moment, while everyone knew computers could play chess before it beat Kasparov, because that had already been reported on earlier, it's just that nobody remembers or cares about "the first computer to ever be able to play chess", because who cares
    The first AGI, will be historic, something people likely remember for years to come, just like Deep blue, up until then we will see incremental improvements followed by randomly gaining unexpected abilities thought never before possible, until eventually an AGI is as good or better as the average human at everything
    AGI doesn't need to be better than every expert, it just needs to be average to break the worlds economic system,
    Like, if it takes governments seeing a headline in the paper, new AI virus maker can generate computer viruses (or genetic material for real viruses) that can break the worlds most secure everything, then it is already too late for them to do anything about it
    In the case of Deepmind, they could take as long as they liked, and the reporters never had to report anything, they literally could have ignored the story completely because it's "just some silly game"
    Chess getting botted just affects chess players, but imagine 90% of all traffic on all social media being bots with human-level intelligence or even slightly below average
    It may not be important to mathematicians because they have already all been made obsolete by an AGI that is better and cheaper and more scalable than all mathematical effort on earth, that's also more creative and more reliable, but, I think it's definitely worth the discussion if nothing else, because something definitely is happening, even if the media machine is overhyping it for the sake of ad revenue, something they didn't need to rely on in the past due to subscription service news outlets

  • @ruilopes6638
    @ruilopes6638 9 месяцев назад +1

    Thanks for making the point about the beauty of it.
    I often feel quite frustrated because my natural talent is in mathematics or things related to it, because this isn’t something that is looked upon as artistic. It might be looked as something that will benefit you, but never truly admired.
    And there is so much beauty in it as in a great song, or a beautiful picture or any other art. But we are never thought to seek it or appreciate it.

  • @stanleydodds9
    @stanleydodds9 9 месяцев назад +8

    Yeah, I think these sorts of AI papers are meant to impress average people who don't really know the exact sort of problems it's working on.
    I remember back when I was in school doing the olympiad round 1 and round 2 questions, the geometry ones were not very appealing to me because it was just a tedious process of angle chasing and constructing new points and lines until you get the result you were looking for. It's very much on the easy-for-computers side of problems, where there are just a few different options to consider at each step, but you might need to try them a lot of times to get to the result.
    That's unlike e.g. the number theory side of things, where there are seemingly endless things you could try, and it takes a bit more of an informed or refined approach to get to the desired result. You have to have more of an idea of what methods will get you closer to proving the result, and that's the sort of thing I think AI should be training to understand.
    Maybe that's just me, but I'd be much more impressed if it could solve any of the other olympiad problem types. And even then, that's school level maths, so it's still got a way to go.

    • @empathogen75
      @empathogen75 9 месяцев назад +3

      They’re just doing the ordinary engineering practice of tackling tractable problems. They could figure out a way to do this with the tools they had, so they did it. In a few months someone will have have another breakthrough which people will also dismiss as trivial in retrospect and so on..

    • @Jackson_Zheng
      @Jackson_Zheng 9 месяцев назад

      It's literally no different. If we take all the mathematical proofs we know as a set, then the set naturally increases over time. So with or without AI, if we assume that this set will increase between now and the next 1000 years as an example (including proofs in number theory), and if we assume that all the proofs are derived from existing axioms and built up over time from our current understanding, then this space of possibilities is finite. It doesn't matter if it's an AI or us humans doing it. If we assume that new information is derived from existing information via some transformation, then a neural network of a large enough size will be able to find it given long enough time.

    • @hedgehog3180
      @hedgehog3180 9 месяцев назад +1

      @@Jackson_Zheng Someone hasn't heard of Gödel's incompleteness theorem. Also this is literally just a “monkeys on typewriters” argument so it doesn't really have any weight. Sure a Neural Network probably is capable of discovering something if we give it an arbitrary ammount of time and computing power but that's true of random chance as well. The thing that actually matters is whether it can do so using a reasonable ammount of time and a practical ammount of computing power. So far neural networks have only been able to approximate the average skill of humans in some very limited tasks like summerizing information, and come up with novel approaches in some very limited spaces like board games. I'm not saying that isn't impressive or useful but it is very far away from actually discovering useful information.

    • @hedgehog3180
      @hedgehog3180 9 месяцев назад

      @@empathogen75 Right and Dogecoin is also supposed to go to the moon in just a month! For all the hype of AIs they have yet to actually find a practical application and haven't actually revolutionized anything other than spam so far.

  • @VforVictorYT
    @VforVictorYT 9 месяцев назад +1

    AI learns so fast, you can see the evolution in front of our eyes, it may be far of today but who knows what next version will bring, its still a baby.

  • @onradioactivewaves
    @onradioactivewaves 9 месяцев назад

    The way you drilled through tbis video, you seem very passionate about mathematics.

  • @FireyDeath4
    @FireyDeath4 9 месяцев назад +2

    We need AI that is able to train itself and store symbolic propositional memories for it to be better, because pre-trained LLMs don't get to learn; they are literally just static functions running on a context window of token parameters. One of the major differences I noticed between humans and LLMs is that humans have a continuous thought stream that differs from their speech output, whereas LLMs don't think without saying every thought they have, and since they're pressured to do so, they don't get to refine their speech before spitting it out improvisationally and without forethought. That makes them unable to learn dynamically, and unable to be creative. Also, the AIs here are not particularly multimodal, so they can't look at the images like we can to get a feel for them, and they also weren't challenged to optimise their proofs once they came up with them, which I find to be a bit unfair in this case. It's interesting how symbolic AI is already a thing that exists though. I myself came up with an idea for an AI that formally generates propositions based on syllogisms and deductions to create and strongly substantiate knowledge before I heard about this today

    • @asdfqwerty14587
      @asdfqwerty14587 9 месяцев назад +1

      The biggest problem LLMs have by far is that they see "mimicking human speech" as the end goal. Before I could ever consider it to have any level of intelligence it would need to be using speech as a means to an end rather than being the end goal, ie. instead of learning to speak by copying humans it needs to learn why it needs to be able to speak first. If an AI is built to accomplish some task completely unrelated to speech and then learns how to interpret plain text as a way to improve how it performs at that task, even if it did it at a the level of a child, it would be 1000x more impressive to me than anything that LLMs are doing.

    • @gvi341984
      @gvi341984 8 месяцев назад +1

      For mathematics? They have thousands of papers to read from and the combination of the Wolfram AI with Chatgpt is changing how math is learned

    • @FireyDeath4
      @FireyDeath4 8 месяцев назад +1

      @@gvi341984 Generally, I mean. It should be able to do it for mathematics by learning the frameworks, and it should even be able to discover entirely new branches of mathematics all by itself if it runs for long enough. I probably haven't put nearly as much thought into it as some people, but still it seems way better for making something intelligent than your run-of-the-mill LLM. I wrote a bit about the idea on my site

    • @gvi341984
      @gvi341984 8 месяцев назад

      @@FireyDeath4 It already can the issue is that it can do anything. I mean you can go through the API and try to cycle itself into learning its own math.

  • @SquintyGears
    @SquintyGears 9 месяцев назад +1

    I totally agree with your understanding of what this AI achieved. But I think you'd also be very surprised by the maths that prove Machine learning is possible and the equations that they use in the process.
    Edit: for your last point, I'm not really scared of AI entering the creative space. It'll replace the bargain bin tier stuff. The shovel ware books and bad content already exists all over the place. It's actually the most common stuff (which is also why it's the easiest for AI to produce). But there will always be those of us that care to put the work and make things exactly as we imagined them. That's beauty, art, philosophy, or what ever else you might call it. AI is a tool, not a replacement. And those who aren't seeing that and are going to try and use it as a replacement are going to fail.

  • @jaideairblade
    @jaideairblade 9 месяцев назад

    The fascinating aspect is that we often can't predict what will become a key driver for future innovation. It might be an issue that currently seems ridiculous, but once it's resolved, along with a few others, it could lay the groundwork for groundbreaking developments.

  • @nerd5865
    @nerd5865 8 месяцев назад

    for the first problem
    i might be wrong here but cant you split the large triangle to 3 triangles, OAC, OAB, OBC, OAB is 90-x, name angle OAC a and angle OCB b, since its a circle OC=OB=OA so OBC=b, OAC=a,
    triangles must sum to 180
    CAB+ABC+BCA = 180
    (a+90-x)+(b+90-x)+(a+b) = 180
    2(a+b)+180-2x=180
    2(a+b)=2x
    a+b=x
    since BCA = a+b,
    BCA=x

  • @johnchessant3012
    @johnchessant3012 9 месяцев назад +1

    I broadly agree with your take, but I wonder how this applies to how modern math research is going, one specific case I have in mind being Mochizuki's attempted proof of the abc conjecture. With proofs of major results nowadays being easily hundreds of pages long, and take the leading experts in the field years to digest, do you think it'd be worth having the AI so it can check our work? Perhaps if we could run Mochizuki's manuscript through the checker then Scholze and others might've decided to spend their time elsewhere.

  • @xiutecuhtli15
    @xiutecuhtli15 9 месяцев назад +1

    I bet they can cut out the LLM and replace it with a random chooser. LLMs are expensive to run and I feel like the random chooser would have done just as well.

  • @Inspirator_AG112
    @Inspirator_AG112 9 месяцев назад +1

    *[**03:16**]:*
    Symmetry of circles.

  •  9 месяцев назад +1

    Does it generate a formal proof in some theorem proving framework or do we need to check the output by hand?

    • @AnotherRoof
      @AnotherRoof  9 месяцев назад +4

      I should have said -- the steps I display on screen are the output of AlphaGeometry so it is able to output a stream of (usually) followable deductions

    •  9 месяцев назад

      @@AnotherRoof Sorry, I probably should have been more exact: is correctness guaranteed by any means? In contrast to chatgpt on maths problems, where it just makes things up.

    • @AnotherRoof
      @AnotherRoof  9 месяцев назад +3

      @ I think so by the nature of the deduction engine basically operating through classical "premise 1, premise 2, conclusion." I know I came off as quite critical in the video but I think this is the cool thing about it, in that it has the wild ChatGPT-esque "ideas" but the inferences are correct

    • @yonatanbeer3475
      @yonatanbeer3475 9 месяцев назад

      You have to check the output, but the actual deductions are made by the angle chaser which does not make mistakes and proofs aren't that long, so in practice it's not that hard to check and it will probably never make an error

    •  9 месяцев назад

      Ok, so my conclusion is that it is designed to be correct, but it doesn't produce any self-validating proof (such as verified steps in a formal theorem proving system) for it. @@yonatanbeer3475

  • @BooleanDisorder
    @BooleanDisorder 9 месяцев назад +1

    I would say that Sabine is correct to the point where you can say that it's a promising direction that doesn't require all that much innovation going forward. AKA, it's a good framework for future AI. Hell, maybe maths will generalize with enough parameters similar to how language tasks seem to with large enough models?
    Anyways, I find AI like AlphaGeometry to be beautiful in themselves like math.

  • @Bruh-bk6yo
    @Bruh-bk6yo 9 месяцев назад +1

    I remember using AI for a topology task to prove something with mappings. God, never using it again...

  • @lukasm5254
    @lukasm5254 9 месяцев назад

    I feel like there are a lot of interesting problems in math that were first solved by flopping around. And later they are made compact and beautiful. So this might become a very useful tool in a few papers down the line.

  • @Galahad54
    @Galahad54 9 месяцев назад

    FYI: The engine that played chess, AlphaZero, and played go, AlphaGo, only needed the rules of the games as a difference between the two. This was a completely different approach to the previous alpha-beta pruning approach to chess only, and appropriate to all small games. Until this approach was developed for Go, and got beyond top human level in one day, it was thought that Go was a large problem, not a small one, due to the astronomical number of multi-move paths, dozens of orders of magnitude more than chess.
    Now, as a combinatorial matter, I have no idea of the complexity level of the various mathematical fields, just that I do not even know the names of the majority of current branches of mathematics. Just in the matter of physics, there are dozens of competing versions of string theory, and no way at present to determine if any of them has additional properties that might be reflected in observed reality, requiring modification of the standard model. Similarly, it's quite clear that humans have not yet exhausted number theory, groups theory, topology, etc. Most humans would not distinguish between Howard Thurston, the magician, and William Thurston, mathematician. So far, baby steps, yet exceeding the realized skills of all but the top 0.1% of humans in mathematics.

  • @silvanomicele1373
    @silvanomicele1373 9 месяцев назад

    I really liked the last part of your video ❤

  • @hydra147147
    @hydra147147 9 месяцев назад

    There is actually a really interesting reason why one just needs to prove that B1B2C1C2 are concyclic (that doesn't even leverage the fact about midpoints of A1A2, B1B2 and C1C2 coincide with that of BC, CA, AB - just the fact that they lie on BC, CA and AB). By symmetry we will have C1C2A1A2 and A1A2B1B2 are too. Now suppose that by contradiction they are not the same circle.Then, they must all be different (if any two are then all three are) and by the radical axis theorem their radical axes should meet at a point. But their radical axes are A1A2=BC, B1B2=CA and C1C2=AB clearly do not. Obviously this particular step is beyond what AlphaGeometry can reason about (it's not even capable of reasoning by contradiction in the first place).

  • @albertmagician8613
    @albertmagician8613 7 месяцев назад

    There is a simple addition to the program that probably engenders proofs that you would consider beautiful. That is reconsider all know theorems that are related to this problem and try to apply them. That is a technique that I applied to problems at university, resulting in tens. (A ten in an European university 1968 is worth more than an A, more over these are pen and paper exams, not multiple choice ones. A minor error in an otherwise correct proof somewhere results in at most nine.

  • @abrvalg321
    @abrvalg321 9 месяцев назад +1

    0:10 would have been nice of you to mention that later an exploit was found that allowed even amateurs to beat alpha go.

  • @MathNotationsVids
    @MathNotationsVids 6 месяцев назад

    My limited understanding of AI suggests that we can set a step limit and employ a backtracking mechanism. It's possible that Alpha Geometry can explore multiple solution paths concurrently and remain within the desired complexity bounds.

  • @CTCTraining1
    @CTCTraining1 9 месяцев назад

    Fascinating video. I seem to remember that Sabine had plenty to say about the search for beauty (in physics) and how it guided (damaged?) thinking and the process of forming new theories. Worth a read, even if in another domain. 😀👍

  • @funktorial
    @funktorial 9 месяцев назад

    re: why we solve problems: beyond utility or beauty, we also solve problems simply because we desire knowledge and understanding. that's why it's an important feature of proofs that they be explanatory, and that's exactly the feature that is missing in these AI-generated proofs.
    another feature that doesn't get talked about often is how "close to the ground" plane geometric proofs are. they're more like the formal proofs in formal logic than the natural language proof "sketches" that mathematicians typically make. I haven't seen any AI-generated natural language proof sketches that I would deem successful or all the way correct (but maybe that'll be around eventually).
    like you, I see the actual uses of AI in mathematics to be as part of something like a formal proof assistant. often when you formalize a proof in Lean or Coq, you'll run into lots of tedious details which are routine but cumbersome to write out in full. AI could be used to shorten that; essentially, it might give an easier a way of translating the proof sketches into formal proofs.

  • @tunafllsh
    @tunafllsh 9 месяцев назад +1

    Actually I think beautiful proofs are very useful as you already told. It gives insight into the understanding the problem better and it builds our intuition.

    • @hedgehog3180
      @hedgehog3180 9 месяцев назад

      I mean it's the only way we can really learn.

  • @Tom-vk2rv
    @Tom-vk2rv 9 месяцев назад

    My goodness im so happy i never ever have to do eucalyptic gneometry ever again. Its so fascinating but so hard.

  • @Notthatkindofdr
    @Notthatkindofdr 9 месяцев назад

    First time I have come across your channel (thanks, YT algorithm!), and I just subscribed! Fascinating video, amazing amount of animation work, and thank you for saving me from having to watch Sabine's video on the topic. 🙂

    • @AnotherRoof
      @AnotherRoof  9 месяцев назад +1

      Welcome! Praise be to the almighty algorithm.

  • @RobotProctor
    @RobotProctor 9 месяцев назад +1

    Really interesting take. I think you're right. Otoh, getting more mathematical results in an automated way will not decrease the amount of interest in mathematical aptitude, similarly to how AlphaGo actually increased interest in Go. Let's get more people into math!

    • @hedgehog3180
      @hedgehog3180 9 месяцев назад

      I think AI in a lot of cases really just lowers the burden of entry and that's ultimately a good thing. Like LLMs are basically just doing the same thing to the humanities that calculators and computers did to science and maths a long time ago, it's taking away a lot of the drudgery to enable people to focus on learning what actually matters. Like calculators didn't result in us forgetting how to do basic arithmatic but it meant that students could move onto more interesting stuff like algebra much quicker and usually in doing so you end up getting better at arithmatic. LLMs do the same thing for writing, they take care of all the basic stuff like grammar and decent sentence structure and in doing so allows a student to move on from struggling with those aspects and instead lets them think more critically about what they're writing. If in this case the only thing the AI accomplishes is making geometry more accessible by being able to understand natural human language then that's a job well done in my book.

  • @3Max
    @3Max 9 месяцев назад

    Great take - I agree that the aesthetic aspect is very important, I wonder if the algorithm will eventually start to incorporate this (trying to get fewer steps, quantifying "more powerful" statements, etc). One thing I'll add is that even if we have an ugly proof of say RH, I think that humans will be able to use that to identify areas where our understanding was lacking and build new theories from there.

  • @maxv7323
    @maxv7323 9 месяцев назад +2

    I find it unclear where the ai even is in alphageometry. Obviously there is an llm making suggestions but how is that different from brute forcing suggestions? If the answer is a higher rate of useful suggestions, then is there any reason to believe there is no simpler, conventional (as in not ai) method that does even better than that?

    • @noel.friedrich
      @noel.friedrich 9 месяцев назад +1

      Yes, it's unlikely that there isn't. But nobody has come up with one yet.

  • @LordMarcus
    @LordMarcus 9 месяцев назад +1

    AlphaGeo isn't solving geometry problems, it's trying to communicate in Gallifreyan.

  • @dominiquelaurain6427
    @dominiquelaurain6427 9 месяцев назад

    I am a computer science engineer and a maths fan. I can give you this geometry problem (Euclidean geometry) "Given an ellipse, find a triangle with minimum perimeter where the ellipse is inscribed in" and try AI on it. I bet you get no answer. Why ? the answer is hidden deep in maths. My POV is AI will be an help only if it is used for developing creativity. Usually for maths Olympiads, the idea is not to repeat the off-beaten track, but to get more beautiful tracks. If you get more and more AI techs, which dont tell how they are working...they are no more than crystall balls.

  • @tomatoso27
    @tomatoso27 9 месяцев назад +1

    I think ia could easily create beautiful and fresh hypothesis and conjectures if it had a third agent that actually construct measures and compares, lo fi with error bars, changing the variables, like a physicist playing with a toy model. Then the the other ones would try to prove them

  • @wompastompa3692
    @wompastompa3692 9 месяцев назад +2

    Someone should ask it to square the circle and see what it spits out.

    • @MichaelDarrow-tr1mn
      @MichaelDarrow-tr1mn 6 месяцев назад

      That's not a proof

    • @wompastompa3692
      @wompastompa3692 6 месяцев назад

      @@MichaelDarrow-tr1mn
      I know, I just want to see what gobblygook it tries to pawn off on us.

  • @MrAmalasan
    @MrAmalasan 9 месяцев назад +1

    Font on diagrams too small for mobile viewing

  • @palfers1
    @palfers1 9 месяцев назад +3

    I just subscribed because I like your general take on beauty in maths. Recently I found something beautiful in number theory and wonder if you have any insights.
    Start with Euler's quadratic E(n)=n^2-n+p, p=41. The first 40 n's generate only primes which is well-known. Thereafter there's a curious mixture of primes and composites, and the patterns formed by the composites caught my eye. Define a block k s.t. kp

  • @MathNotationsVids
    @MathNotationsVids 6 месяцев назад

    The more it is trained on multiple solution paths including beautiful creative human thinking the more refined the output will be. The fact that it was even able to generate this cumbersome circuitous route is still inspiring to me because I see its potential. And I’m speaking as a pure mathematician focused on algebraic number theory.

    • @MathNotationsVids
      @MathNotationsVids 6 месяцев назад

      and I was remiss in not starting out by saying this video lecture is extraordinary.

  • @fleefie
    @fleefie 4 месяца назад

    People just really need to understand what AI is used for best IMO. We should see it as a tool to fill gaps rather than a tool to create new things, because it can't really create new things like we do just yet. I use some AI to learn math, but when finding a proof I use it to find one specific self-contained step that I know what to do with if I can prove it, but asking it for a general proof is pretty pointless and won't lead anywhere. And of course it falls appart when trying to generate new ways to reason.
    Basically, AI is to logic what a calculator is to arithmetic. It doesn't replace everything, sometimes it'll be inherently slower or worse, and you're not going to get anything outside of brute force. But if that fits in with what you need, it's a neat tool.

  • @TomJones-tx7pb
    @TomJones-tx7pb 9 месяцев назад

    The first problem is more simply solved by drawing the lines OA, OB, and OC and noting that you have 3 isosceles triangles and so OAB = OBA, etc. I won't go much into why this is a better solution, as well as simpler, but it only uses the facts that lines from the centre of a circle to the edge are of equal length, and that a tangent to a circle forms a right angle is shown in the video. oh and that the angles of a triangle always add up to 2 right angles.

    • @AnotherRoof
      @AnotherRoof  9 месяцев назад +2

      The whole point of that section was to demonstrate that AlphaGeometry doesn't always spot the simplest solution. See: rest of the video :P