AlphaZero v Stockfish (2018): Lecture by GM Ben Finegold

Поделиться
HTML-код
  • Опубликовано: 4 окт 2024
  • This lecture was recorded on December 20, 2023 in Roswell, Georgia. Thank you to the Zatloukal Family for sponsoring!
    Games/Positions:
    09:07 AlphaZero vs Stockfish
    24:56 Stockfish vs AlphaZero
    37:01 Stockfish vs AlphaZero
    Check out Ben's Chessable courses here! www.chessable....
    If you're interested in sponsoring a lecture of your choice, email Karen at karen@atlchessclub.com
    Signup or gift a chess.com Premium membership to get access to their lessons, remove ads, improve your chess, AND help Ben at the same time!! Upgrade your chess.com membership or join! - go.chess.com/fi....
    Buy Merch today! ccscatlmerch.com/
    Watch live at / itsbenandkaren
    Donations are appreciated at: streamlabs.com...
    Follow me on Twitter: / ben_finegold
    #benfinegold #chess #AlphaZero #Stockfish

Комментарии • 231

  • @zvonimirtosic6171
    @zvonimirtosic6171 9 месяцев назад +38

    V. Kramnik: 'Alpha 0 beats Stockfish 8? With only 9 hours of self-learning? And without openings book? How interesting ...'

    • @Corteum
      @Corteum 9 месяцев назад +6

      It has a chip in its butt lol

  • @askthepizzaguy
    @askthepizzaguy 9 месяцев назад +14

    Congrats on using the word "suffice" twice in a row in a grammatically correct sentence. Achievement unlocked.

  • @ernietollar407
    @ernietollar407 9 месяцев назад +57

    GM Finegold at his funniest while talking about computer chess.. Way better than the stockfish channels cheesy scripts

    • @许玄清
      @许玄清 9 месяцев назад +7

      That channel is just loads of AI generated crap. The games they show are probably just Stockfish vs Stockfish itself.

  • @AdamCHowell
    @AdamCHowell 9 месяцев назад +93

    Beating stockfish is easy. It’s cheating to use an engine. Stockfish is an engine. I automatically win as a result. Checkmate stockfish.

    • @paulgoogol2652
      @paulgoogol2652 9 месяцев назад +3

      But can you prove that Stonkfish ls cheating?

    • @paulgoogol2652
      @paulgoogol2652 9 месяцев назад

      Also: Cheating basically means getting help from a third party? Like when I as a 500 elo fish get hints from a master while I play some ofishel chessgame like a real nerd and stuff.

    • @buuythbuuyth1412
      @buuythbuuyth1412 9 месяцев назад +6

      Stockfish doesn't cheat. Its chess speaks for itself.

    • @vibovitold
      @vibovitold 4 месяца назад

      It's also illegal to get human assistance during a game, so wouldn't you get disqualified too? : )
      And if the assistance provided by your own brain doesn't count... you see where it is going then; does Stockfish use itself?

  • @raskalnekov
    @raskalnekov 9 месяцев назад +28

    Not only did I stay awake long enough to see the Stockfish win, I also stood awake long enough to write this comme

  • @f.d.3289
    @f.d.3289 9 месяцев назад +41

    come on, millionaires, sponsor GM Finegold analyzing all 1000 games in full

    • @jayyy5270
      @jayyy5270 9 месяцев назад +3

      best comment 2023

    • @-zelda-
      @-zelda- 9 месяцев назад +1

      I don't even think they made all 1000 games public, I think just around 100 or so are available

  • @JM-tj5qm
    @JM-tj5qm 9 месяцев назад +134

    AlphaZero was great for it's time. But I don't think most people realize Stockfish is way stronger than AlphaZero ever was now.

    • @keithoma7265
      @keithoma7265 9 месяцев назад +16

      I think StockFish would roflstomp AlphaZero even from the starting position. We live in crazy times.

    • @josephmathes
      @josephmathes 9 месяцев назад +30

      For real. It incorporated techniques from AlphaZero, just didn't replace itself with them entirely. Current stockfish is like old stockfish and AlphaZero combined.

    • @kingsolo6241
      @kingsolo6241 9 месяцев назад +4

      It just takes time for Alpha Zero. It has a much stronger computer that Stockfish will never see. It a matter of minutes Alpha Zero will always surpass Stockfish.

    • @MadMetalMacho
      @MadMetalMacho 9 месяцев назад +14

      Wasn't Alphazero trained for just 5 hours? It's just that they stopped after it became the best in the world and moved on to other things, but had they continued, with 5 years of development/training it's hard to see how Alphazero wouldn't be on top today...

    • @JM-tj5qm
      @JM-tj5qm 9 месяцев назад

      @@kingsolo6241
      Alphazero is not a computer and neither is Stockfish. Both are computer programs.
      Alphazero advantage at the time was the fact that it was a NNUE (Efficiently updatable neural network) Once Stockfish updated it's evaluation function too it surpassed AlphaZero. And Since then Stockfish has gotten way stronger still.
      AlphaZero doesn't have any advantage now, while Stockfish has the advantage of being open source and is constantly tested by it's community.

  • @sage5296
    @sage5296 9 месяцев назад +16

    The main reason A0 is more willing to sac material is that since it's using AI/NN to evaluate the position, it can more or less see the compensation, where as SF can't due to the horizon effect. Very similar things happen in some of the SF v Leela0 games. One of the main things at play I think is long term piece mobility.

    • @AG-ld6rv
      @AG-ld6rv 2 месяца назад +1

      Alpha Zero beat Stockfish ON PARTICULAR HARDWARE. More or less, Alpha Zero was using custom hardware designed to execute the code of the engine whereas Stockfish was running on a regular CPU. AI engines can execute their logic on a CPU. The team just didn't do that, because they wanted the sensational reporting of "beating the #1 engine with AI."
      There is actually a post by a Stockfish developer around the time of that tournament. Here were his main points:
      (1) Extraordinarily powerful custom hardware versus regular CPU. Their tournament is like saying Stockfish on a US$700 CPU beats Stockfish on a US$133 CPU.
      (2) They weirdly chose to use a hash size of 1 GB, which hinders the performance of a chess engine. For comparison, I looked into what hash rates are allowed in TCEC, a respected chess engine tournament, and it was 100x the amount!
      (3) They oddly chose to use Stockfish in 1-minute per move mode rather than let Stockfish manage its own time. According to him, Stockfish has a lot of code dedicated to deciding what positions to invest more time into, and that contributes to its performance heavily.
      (4) He said Stockfish would have done a lot better in a fairer situation, that the results mean very little, and that it's like comparing "apples to orangutangs."
      Alpha Zero contributed a ton with its neural network engine e.g. showing h pawn pushes that many have adopted, giving numbers to rate various openings, and being a new technology/technique for chess engines (now implemented in the latest Stockfish but not in as large a degree as in Alpha Zero). Still, with fair hardware, Stockfish likely would have won. Stockfish is no joke. Leela Zero, an open source version of Alpha Zero, runs in CPU mode in chess tournaments, and it doesn't beat Stockfish at all. It's extremely good + often plays unique moves compared to traditional engines, but it shows you how misleading the team was in their private testing.

  • @xwngdrvr
    @xwngdrvr 9 месяцев назад +6

    Show ALL the games! Five THOUSAND minutes with Ben!

  • @xianglong2871
    @xianglong2871 9 месяцев назад +11

    Has anybody consulted Vladimir Kramnik and his mathematicians about these games?

    • @Corteum
      @Corteum 9 месяцев назад

      Kramnik is several hundred points lower rated. His opinions dont apply here. lol

  • @wreckim
    @wreckim 6 месяцев назад +1

    Thanks GM!

  • @jdubrule
    @jdubrule 9 месяцев назад +4

    IIRC, the resignation rule was that both engines had to agree that one side was up by at least 10 pawn-equivalents. So, one engine could be forced to play on, even though the other engine was the only one that had hope for its position

    • @archsys307
      @archsys307 9 месяцев назад +3

      Just like human chess 😂😂😂

  • @miroljubbelic2007
    @miroljubbelic2007 9 месяцев назад +1

    Thanks , very instructive . Your comments are so funny. Keep going

  • @art58wong
    @art58wong 4 месяца назад

    Great lecture. I remember the AZ headlines and even saw a video of your first relentless example. Your explanation filled my swiss cheese understanding of why AZ isn't mopping up the chess world and leaving the game in tatters. Again my thanks for giving much needed perspective. Your series is quite refreshing.

  • @ek5688
    @ek5688 9 месяцев назад +4

    Go Ben!

    • @johnolah8853
      @johnolah8853 9 месяцев назад +3

      But stay right there!

  • @kmarasin
    @kmarasin 9 месяцев назад +5

    Those endgames by AlphaZero remind me of Capablanca, but on steroids. Capa loved to push his pawns and gain space.

  • @curiousnerd3444
    @curiousnerd3444 9 месяцев назад +28

    There is a very good reason for not using opening book.
    chess engines at that time played at high depths and evaluated positions based on materials (Queen 9 points, rook 5, pawn 1) or simple rules which were coded into the engine by humans.
    This made chess engines great at tactics but chess grandmasters would still outplay them because the engines didn’t understand strategy.
    This is the reason stockfish uses opening book to compensate for lack of understanding chess
    People at that time used to believe that learning chess was therefore a sign of human intelligence.
    The point AlphaZero was trying to make was an AI can learn to play chess at a grandmasters level with zero human training just by getting the rules of the game.
    Pitting Stockfish against AplhaZero without opening book is a test of “Artificial Intelligence” learning strategies that beat calculations

    • @ClarkPotter
      @ClarkPotter 9 месяцев назад +2

      Even in 2018 the eval functions were more sophisticated than what you outline.
      I would guess they used a genetic algorithm with parameters as genes to converge upon optimum values, and to determine which to jettison to improve speed.

    • @mohammadhaider8946
      @mohammadhaider8946 9 месяцев назад +7

      Your claim that engines were great at tactics but they would still be outplayed by grandmasters due to a lack of strategical understanding is just plainly wrong. Even years before this match, the best engines had reached a point far beyond human ability to match. There might have been deficiencies, but they were not going to be outplayed by humans.

    • @curiousnerd3444
      @curiousnerd3444 9 месяцев назад +1

      I guess I could frame it better, the main point is after finishing AlphaGo, they made AplhaZero which didn’t use human games and trained entirely on self play aka it learned long term strategies on its own. They could quickly train it to play other games than go (including chess)
      They were not trying to compete with stockfish and that’s why they quickly moved to other things instead of improving AplhaZero unlike stockfish

    • @许玄清
      @许玄清 9 месяцев назад

      Evaluating based only on material will give you a significantly inferior engine than Stockfish 8. HCE was the result of three decades of development and this is just disrespectful

    • @vibovitold
      @vibovitold 4 месяца назад +1

      You've oversimplified it to the point where it's blatantly untrue.
      The rules were anything but trivial.
      If you ever looked into (old at least) Stockfish source code, it contained a huge pile of "manual" corrections - surprisingly specific, like "in an endgame with just Queens and minor pieces value the mobility of minor pieces 16% less" etc. Not literally ofc, but this sort of stuff.
      Of course it's still heuristics (albeit very complex and multi-layered), and it doesn't quite convey the notion of long-term strategical sense (this part is correct).
      But modern engines were very far from evaluating positions only based on the material value of pieces.
      This is really unfair to their creators (who btw often were strong players themselves, or strong players assisted with the development).
      If you built an engine based on such simple criteria, it would never be stronger than an amateur.
      It's true that engines relied on opening books because they weren't very good at grasping the strategical implications of opening choices (the "fog" was too dense from the starting position).
      But still, even if you disabled an opening book, an engine would still play a pretty strong opening based on general principles (mobility, control of the center etc.).
      The main reason for including an opening book is just optimization.
      Similarly, (serious) engines use endgame tables, even though endgame is "pure" calculation, and relatively easy at that (as you can calculate much farther when there are only a few pieces left on the board). But the reason is optimization. In an engine vs. engine match time matters: just like in human chess, they both have only got so much time allocated for the game.
      So why waste time on computations when you can replace them with the equivalent of "home preparation".

  • @paulgoogol2652
    @paulgoogol2652 9 месяцев назад +3

    Great video. Love the Berlin Defense. They are never as drawish as top players make it look. Especially this endgame is very tricky. I actually like much less the lines that don't head to this early queen trade. The old mainline. Then the structure is just symmetrical. The dull exchange french basically.

  • @baoboumusic
    @baoboumusic 9 месяцев назад +18

    Stockfish 16 really likes what AlphaZero did and predicts pretty much all moves. I expect it's about twice as good as Stockfish 8.

    • @addisonmigash8227
      @addisonmigash8227 9 месяцев назад +16

      8 * 2 = 16

    • @nilsp9426
      @nilsp9426 9 месяцев назад +3

      If you can now tell me what "twice as good means" :'D

    • @baoboumusic
      @baoboumusic 9 месяцев назад +1

      @@nilsp9426 the answer is fries.

    • @DarthVader-wk9sd
      @DarthVader-wk9sd 9 месяцев назад

      Way more than twice as good but ok

    • @archsys307
      @archsys307 9 месяцев назад +3

      @@DarthVader-wk9sdHow do you quantify twice as good
      Obviously elo only went up some hundreds of points. Not to 6000
      What does it mean to be twice as good? A 2 to 1 win probability? Then you could use the elo formula to compute the exact elo gap, should be about +100 points.

  • @christofferore6285
    @christofferore6285 19 дней назад

    Honnestly such a great person to Watch and learn Chess from and always tries to joke and be funny. Tysm for making alpha zero video.

  • @-_Nuke_-
    @-_Nuke_- 9 месяцев назад

    Oh yes! I was waiting for these!

  • @talhakonjic3197
    @talhakonjic3197 9 месяцев назад +2

    hi Ben, this game is awesome, its almost as awesome as the games you play are!

  • @rhiehn
    @rhiehn 9 месяцев назад +1

    Most versions of stockfish will accept the berlin draw, so magnus(or anyone else) can draw it with white, but I don't know that i'd really count that, but if a GM said that I'd assume that's what they meant, since it's obviously absurd otherwise.

  • @pinofshame6253
    @pinofshame6253 9 месяцев назад +1

    46:44 my guess AZ saw the idea of promoting the H pawn and didn't want to open the a1-h8 diagonal for the dark square bishop, so that he can put his rook to h8.
    Notice how the Bishop on f2 can't get to f6 or g7 any way

  • @dannygjk
    @dannygjk 9 месяцев назад +1

    Incorrect off the bat. Alphazero is/has an engine. It's just that at the time AZ used neural net tech whereas SF didn't use NN tech, (SF started using NN tech starting with SF 12). Technically any machine that plays chess has an engine built into it.

  • @sachatostevin6435
    @sachatostevin6435 9 месяцев назад +2

    Thanks, Ben! Great video as always.
    You might be interested to know that I've just downloaded a free version of LeelaZero, and my first test was to setup board position: 1. d4 d5 2. Nc3 Nf6 3. Bg4...
    Onwards from here I let LeelaZero play both sides. Strangley enough it accepted a 3-fold-repetition at move 29. Maybe I'm supposed to tweak some settings, I don't know.

    • @sage5296
      @sage5296 9 месяцев назад +1

      the setting to look for is "Contempt", which basically just tells the engine how much it should avoid draws, or in other terms, how much worse of a position it's willing to accept to avoid a draw (on the assumption that it can outplay the opponent in the long term and win anyways)

    • @sachatostevin6435
      @sachatostevin6435 9 месяцев назад

      @@sage5296 oh yeah. I think i remember ben using that word now. Thanx!

  • @CARPB147
    @CARPB147 2 месяца назад

    I think Stockfish 8 didn't have neural network (NN) support (like Alphazero). NN was added to Stockfish in later versions (10 and above, if I'm not mistaken). NN is what gives it the game-changer "AI" capability. Yes, they are all "engines" of one type or another.

  • @DavidYoo-m7z
    @DavidYoo-m7z 9 месяцев назад +2

    AlphaZero "cheated". Stockfish 2018 would have drawn almost all the lost games had it been given a proper computer to run on. The deep mind team put Stockfish on the crappiest computer they could find, the worst time controls Stockfish played at, and only published the results where AlphaZero won. Oh and they removed Stockfish's opening book and endgame database. AlphaZero on the other ran on a custom gpu-driven machine. The match was a complete farce.

    • @许玄清
      @许玄清 9 месяцев назад

      Idk it’s pretty hard to come up with fair hardware for GPU/CPU engine matches. Even modern engine tournaments get criticized over them.

    • @DavidYoo-m7z
      @DavidYoo-m7z 8 месяцев назад

      ​@@许玄清 I agree it is hard to compare apples to apples CPU vs GPU, but look at what Stockfish was given vs. Google's machine. It was totally lopsided.
      If you Google "Hardware used in AlphaZero vs Stockfish match", and read this line from the StackExchange discussion:
      "[AlphaZero] used 4 TPUs for the games, so a processing power of 180 TFLOPS. Note TFLOPS = 1000 billion floating point operations per second."
      The GPU network AlphaZero ran on was roughly 10,000 times more powerful than the hardware SF was allowed, and AlphaZero was *designed* to work without an opening book, whereas Stockfish was designed to work with an opening book an endgame database, and at classical time controls. The game structure was obviously designed handicap SF to the maximum.

  • @unclvinny
    @unclvinny 9 месяцев назад +2

    My nasty comment is that Janet Jackson had a song years ago called "Nasty Boys".

    • @Corteum
      @Corteum 9 месяцев назад +1

      Lol. I dont think anyone even knows who that is now!

  • @CTJ2619
    @CTJ2619 8 месяцев назад

    i like BF discussion about compensation and judgement

  • @-zelda-
    @-zelda- 9 месяцев назад

    8:31 Some versions of Stockfish will gladly play the Berlin draw most of the time with black so you can get lots of easy draws that way

  • @grahamsmith2753
    @grahamsmith2753 4 месяца назад

    As I understand it, AlphaZero was playing on massive hardware, and Google was still so worried, it had the Stockfish opening book disabled. Lc0 has just lost the TCEC season 26 Superfinal to Stockfish by 57 points to 43. The Superfinal consisted of 100 games, played at a time control of 120' + 12". The headlines that Stockfish was crushed by AlphaZero were conveniently always taken out of context...but that's Google for you.

  • @jefftaylor1186
    @jefftaylor1186 9 месяцев назад +1

    If I remember correctly, when they didn’t have the opening book, both “engines” decided that the Evans gambit was optimal.
    Which is awesome because I only play 1.e4 with the hopes of being able to go 4.b4

    • @ethangilworth7891
      @ethangilworth7891 9 месяцев назад +1

      Extremely incorrect. Maybe at very low depth, but certainly not at the massive depths they are typically run.

    • @jefftaylor1186
      @jefftaylor1186 9 месяцев назад

      @@ethangilworth7891
      You’re right.
      It was just alphazero that did it. Not stockfish.

    • @ethangilworth7891
      @ethangilworth7891 9 месяцев назад

      @@jefftaylor1186 I’m an Evan’s player myself, so I would love to see your source for this.

  • @ocudagledam
    @ocudagledam 8 месяцев назад

    So, I was curious about the last game and it turns out that Ben was right about a couple of things. I let SF 14 take a look at the position where, at 47:18, Ben says that it looks to him like it should still be a draw and the newer SF engine agrees, and when Ben says that, after white pushes h5, he would go g:h5, well, he's absolutely right. The way to draw was to exchange pawns on h5 and then go Kc4, with the intention of bringing it to d3. Instead, Kb3 that the A0 played instead of g:h5 straight up loses, the eval jumps from something like 0.03 to around 3.5.

  • @Uerdue
    @Uerdue 9 месяцев назад +3

    Does anyone know a game of this match (or the one from 2017) where AlphaZero sacrificed and Stockfish won by defending and holding on to the extra material? Like, did it ever misevaluate / overestimate its attacking chances?

    • @timharris559
      @timharris559 7 месяцев назад

      Precisely this, it lost most of its games by attacking in a drawn position, but A0 believed it had an overwhelming attack but stockfish had a strong defence to counter.

  • @RoxxorzYourBoxxorz
    @RoxxorzYourBoxxorz 9 месяцев назад +1

    AlphaZero actually flipped the board before Stockfish was able to play Be7

  • @mrpocock
    @mrpocock 9 месяцев назад

    One thing i love about computer chess is when the engines radically disagree about the evaluation.

  • @askthepizzaguy
    @askthepizzaguy 9 месяцев назад +1

    Didn't stockfish "resign" if its own evaluation of its position was like, -5 whole points or something like that? I may be getting the number wrong, but I thought they called the game lost if either program felt it was losing by a sufficient amount. I could be confusing this with a different computer chess matchup event, but I know I didn't invent this out of whole cloth, I definitely read or heard about it somewhere.
    edit- literally one second after I wrote this comment you mentioned white resigned. Great timing on my part. /sarcasm

  • @Rubrickety
    @Rubrickety 9 месяцев назад +6

    "It seems like AlphaZero chooses the move that's most likely to win" is an almost spot-on description of how these deep-learning game AIs actually work. They don't really "evaluate" any position; they just simulate the entire game to the end and see what happened. My understanding is that getting AIs to give a traditional "-2.3" type of evaluation pretty much required a hack; what it's actually "thinking" is "from this position, Black wins 74% of games" or whatever.

    • @sage5296
      @sage5296 9 месяцев назад

      my understanding is that it's basically using the standard alpha-beta pruning, but being much more selective/aggressive with the pruning. The AI part is the algorithm that evaluates a position, which is what basically assigns that %win or +- score. The tree is then pruned accordingly in the same fashion as SF would, except it checks less positions overall (often 100x or higher less positions), generally to a comparable depth tho
      In other words, it plays moves towards a position that's most likely to win, but it still evaluates many moves deep

    • @许玄清
      @许玄清 9 месяцев назад +2

      @@sage5296No, Alphazero does not use alpha-beta pruning. Evaluation is too computationally expensive for that

    • @vibovitold
      @vibovitold 4 месяца назад

      No, chess AIs (or engines based on neural networks) don't work that way.
      They don't simulate the entire game to the end (that's known as a Monte Carlo simulation, this technique predates the AI as we know it, and it can only be executed with a rather low accuracy / randomness, due to computational complexity).
      Quite the contrary - they've got a highly developed (via deep learning) evaluation function, which is essentially a black-box algorithm that "wrote itself" (by trial and error, in the training phase).
      Noone really knows how it works, and what patterns trigger the AI into thinking a position is "promising" or "seeds of defeat". It's the same as with eg. image recognition.
      Noone knows what exactly it was that made the AI recognize a dog in the picture.
      AZ, in a way, played more like a human (human brains are biological neural networks, after all).
      It uses highly sophisticated pattern recognition insteadf of a huge tree search with a slimmer (manually encoded) evaluation.
      That's why it analyzes much fewer nodes per second than the likes of Stockfish. But its analysis is of higher quality, and it compensates for that.

    • @Rubrickety
      @Rubrickety 4 месяца назад

      @@vibovitold In fact, AlphaZero (and all similarly-architected game-playing AIs) use Monte Carlo Tree Search (MCTS) as a fundamental part of their training (and, it seems, during actual play, but see below). MCTS is exactly how AZ learns "by trial and error, in the training phase", and is indeed based on playing a large number of games to the end to determine win/loss efficacy for a given move, which it uses to update the network. It's surprisingly difficult to find clear info on whether current AIs use MCTS during play to choose candidate moves. DeepMind's paper on AlphaGo Zero (the Go-playing AI which preceded generalized AlphaZero) states specifically that it does not, but the AlphaZero description is unclear on the matter, and Leela Zero definitely does use runtime MCTS (i.e., "playouts" or "rollouts"). Apparently without them its Elo was only ~2500. In any case, though, the "evaluation function" of deep learning AIs does literally boil down to "Will this move lead to a win, on average?"

    • @vibovitold
      @vibovitold 4 месяца назад

      @@Rubrickety as part of the training, sure, i meant the "over the board" phase.
      According to LZ website
      "Many people call this MCTS (Monte-Carlo Tree Search), because it is very similar to the search algorithm the Go programs started using in 2006. But the PUCT [Predictor + Upper Confidence Bound tree search] used in AGZ and Lc0 replaces rollouts (sampling playouts to a terminal game state) with a neural network that estimates what a rollout would do"
      so if it doesn't "sample playouts to a terminal game state", only "estimates what a rollout would do", it directly contradicts the claim that "they just simulate the entire game to the end and see what happened". no?
      i do admit that while i'm a programmer with some interest in chess (and conversely chess engines), i am not working on chess engines, let alone an expert in the field

  • @PeterDarren-e4d
    @PeterDarren-e4d 9 месяцев назад +1

    Remember, stockfish, at that time, was calculating around 4 million positions a second to alpha zeros 80 thousand... If AO were to have the same calculating power as SF then no stockfish would be able beat Alpha Zero😮

    • @ocudagledam
      @ocudagledam 8 месяцев назад +1

      They probably had the same computational power on hand. The thing is that A0's evaluation function is much more complex and much more computationally demanding than that of a traditional engine, so if A0 is to match SF8 in term of positions per second, A0 has to have many, many times more computational power at its disposal

    • @许玄清
      @许玄清 8 месяцев назад

      They do have the same calculating power

  • @askthepizzaguy
    @askthepizzaguy 9 месяцев назад

    At a certain point black's best idea is to give away all his pieces every move in order to stop white's attack. That's why being single minded about gaining a material advantage was the flaw in the armor of Stockfish 8. At a certain point, it was willing to trade all that material back to stop getting its ass beat, and considered giving away material like that to be its best move. Alphazero was definitely on the right track with its judgment call that the attack was strong enough even at a 4 pawn deficit to be worthwhile.

  • @jessealtum2627
    @jessealtum2627 9 месяцев назад +1

    I wonder why at 30:30 stock fish chose not to capture the dark squared bishop with the knight eliminating the bishop pair. Is the light squared bishop that much stronger in this position, maybe because of the A2 target?

    • @ennerz-hq8pq
      @ennerz-hq8pq 9 месяцев назад +2

      It wouldnt have used human reasoning like you and I. The answer lies in huge branches of calculations

  • @McWerp
    @McWerp 9 месяцев назад +2

    I always wonder at what point, what move, the engine realizes it was wrong and its losing. What moves did one engine miss that the other saw.

    • @ethangilworth7891
      @ethangilworth7891 9 месяцев назад +1

      With Stockfish, it uses Alpha Beta pruning for its tree search, so it saw every move. The question is more “what move did it evaluate incorrectly.”

    • @sage5296
      @sage5296 9 месяцев назад +1

      @@ethangilworth7891 not exactly, it can easily be the case that it prunes away the key line too early. Ig that would be an incorrect evaluation to some degree but non-NN engines rely more on seeing far ahead than evaluating accurately in any given position

    • @vibovitold
      @vibovitold 4 месяца назад +1

      @@ethangilworth7891 that's pretty much the whole definition (and the purpose) of pruning - to NOT see every move.

  • @nilsp9426
    @nilsp9426 9 месяцев назад

    From my look at AlphaZero vs. Stockfish it seems A0 is increadibly good at punishing any setup with b7, Bc8 and e6. Black not getting their bishop out is such a common theme in these games.

    • @JohSmith
      @JohSmith 9 месяцев назад +1

      Interesting

  • @MrBonified66
    @MrBonified66 9 месяцев назад +2

    Kasparov's claim was that GM's on the Deep Blue team were inserting specific recommendations for certain positions, and the key position in Game 6 was one of them. GK knew his move wasn't sound but he knew DB wouldn't respond correctly - which it wouldn't have without the specific human input.
    In any case, my favourite bit is that a few days after Kasparov bumps into Charles Bronson, they have a chat, GK says something like "I think I can beat it in a rematch". Bronson says "they aren't going to let you have a rematch". And he was right.

    • @archsys307
      @archsys307 9 месяцев назад

      Y they didn’t let him rematch

    • @zackarysemancik5491
      @zackarysemancik5491 6 месяцев назад +1

      ​@@archsys307because they won

    • @vibovitold
      @vibovitold 4 месяца назад

      They had never been obliged to keep on granting him rematches until he's satisfied with the outcome.
      And why would they grant him a rematch anyway after he publicly accused them of cheating without providing a shred of evidence? "I just know that computers can't play that way" is an allegation, nothing else.

  • @movatrova3539
    @movatrova3539 9 месяцев назад +1

    Nasty comments! Rawr!

  • @grahamdugan
    @grahamdugan 4 месяца назад

    I would have been more interested in the opening sequences

  • @alfiecollins5617
    @alfiecollins5617 9 месяцев назад

    37:03

  • @AG-ld6rv
    @AG-ld6rv 2 месяца назад +1

    For nerdier people out there, there were other reasons the Alpha Zero (A0) versus Stockfish (SF) match was... designed to give A0 a favor rather than see which engine was best. The company wanted publicity rather than a fair match ("A0 better than #1 engine!!!!") -- sucks I know. I bet the techies are not to blame but rather the "business" side (CEOs etc.). Before I go into that, I want to say that A0 DID innovate, and in fact, some tech inspired by A0 is even in the newest version of SF (the NNUE mode). However, that doesn't mean the match didn't have dubious circumstances to create sensationalist news. Journalists really did a disservice to the people when they wrote their articles... all without asking the SF team about the results. I'll give you a quote from one SF programmer at the bottom (the last two paragraphs). If you don't want to read that, my three points summarize his points.
    (1) A0 used special, REALLY expensive hardware designed to execute A0's algorithm with a HUGE amount of calculations per second. You can think of this as someone saying "SF with a US$700 CPU is better than SF with a US$133 CPU." Well, no duh! Leela Zero, which is pretty much a copy of A0, does not dethrone current SF with both engines ran on the same hardware. They run L0 in "CPU mode" in engine tournaments for fairness although it might crush SF if it were running on specialized hardware in those tournaments by having more calculations / second, which it isn't allowed to do so that each engine has the same amount of calculations per second. They setup the tournaments this way to be fair, using common sense, and judge the engines in a fair setting... something the A0 team clearly was not trying to do.
    (2) They configured SF to use a tiny hash size. They gave it a puny 1 GB. In the TCEC (a well-respected chess engine competition), engines are given waaay more hash size. A maximum of 100x more to be exact. A0 vs SF is kinda sounding fishier now.
    (3) They gave SF a command so that it used exactly 1 minute per move. Instead, engine tournaments tend to let engines manage their own time. With management of time, they can use advanced algorithms to decide when to go into a deep think. More or less, they decide whether a position is a critical moment and invest their clock at that time. One of SF's advantages in tournaments comes from its sophisticated detection of critical moments so that it doesn't waste its precious clock time, all thrown away with the settings they gave SF in that misleading competition.
    Below is a quote from one of the authors of SF, saying much of the same things:
    "The match results by themselves are not particularly meaningful because of the rather strange choice of time controls and Stockfish parameter settings: The games were played at a fixed time of 1 minute/move, which means that Stockfish has no use of its time management heuristics (lot of effort has been put into making Stockfish identify critical points in the game and decide when to spend some extra time on a move; at a fixed time per move, the strength will suffer significantly). The version of Stockfish used is one year old, was playing with far more search threads than has ever received any significant amount of testing, and had way too small hash tables for the number of threads. I believe the percentage of draws would have been much higher in a match with more normal conditions.
    On the other hand, there is no doubt that AlphaZero could have played better if more work had been put into the project (although the "4 hours of learning" mentioned in the paper is highly misleading when you take into account the massive hardware resources used during those 4 hours). But in any case, Stockfish vs AlphaZero is very much a comparison of apples to orangutans. One is a conventional chess program running on ordinary computers, the other uses fundamentally different techniques and is running on custom designed hardware that is not available for purchase (and would be way out of the budget of ordinary users if it were)."
    Overall, I'm impressed SF drew so many games. In a fair setting, it would have both drew and won even more games since it would have been SF playing on equal footing. (1) Run A0 in CPU mode, so it has the same calculations / second as SF. (2) Use a reasonable hash size. (3) Let the engine manage its own time.

    • @MusingsOAM
      @MusingsOAM 4 дня назад

      The whole point was A0 used machine learning and not just calculations (not underplaying calculations btw, as it's very smart for optimisations).
      People didn't expect that that would provide any advantage as chess was seen as a game that can use brute force calculations instead of seeing patterns and strategic ideas.
      This means, at a human level, that you can actually rely on intuition and get very far in chess, even against someone who's good at calculations. That's a big deal, and has the potential to bring about a renaissance in chess, except for opening theory being just memorised.
      So the computational power wasn't the focus, which many people here seem to misunderstand

    • @AG-ld6rv
      @AG-ld6rv 4 дня назад +1

      @@MusingsOAM Just so you know, "neural networks" are just elaborate, nonlinear functions. You know, "calculations." I don't think you have any training in machine learning. Someone looking up something in Google or, these days, in chatGPT gives them a false sense of knowledge. Take a step back, admit you aren't sure, and move on with your life, realizing you aren't an expert in this domain. Your reply has a ton of inaccurate information in it. (And yes, I've actually studied machine learning for 100s of hours both in an academic setting as well as in my free time to keep up with market trends [investing better].)
      To say no one imagined machine learning could be applied to a game with great success is asinine. That is what everyone who studies machine learning actively thinks. In fact, applying machine learning to games is a tried and true tradition going back to the seeds of the field. As an example, reinforcement learning got its first big result when a researcher made a backgamon bot using the content of that machine learning strategy.
      The actual thing people didn't know would be possible is representing the state of the game due to chess having so many states. The state is simply the current position on the board. A lot of the research in applying machine learning to a board game deals with overcoming this state representation problem. A person can trivially train a bot to play tic tac toe. There aren't that many states. Backgammon, on the other hand, has a huge number of states, which is why it was monumental when a person crafted a way to represent those states while the algorithm actually played intelligently. Chess was just a step up in that type of problem since chess has many more states than backgammon.
      So absolutely not: Everyone in machine learning thought machine learning algorithms could outdo hand-coded algorithms IF a person could represent the state of the game well enough for these decade-old algorithms to make decisions based on past "experience" and "learning."
      Recognize you haven't put 100s or 1000s of hours into a field and stop walking around, thinking you can learn this stuff by reading 20 pages you found online on chatGPT, through Google searches, from a few Stack Exchange conversations, and whatever else imbued you with odd confidence for a subject you haven't studied. No one would need to put in 1000s of hours if all is needed is a 20-page summary.
      Reread what I wrote in my original post. Take it as something to learn from rather than something for your incorrect assumptions and shaky foundation to try to poke holes into. You'll better your shaky understandings by approaching my points as something to memorize and something to try to make sense of rather than you trying to be the big bad wolf in the conversation based on a few chatGPT searches. And by the way, chatGPT has about a high school level of comprehension dealing with advanced topics. You should trust expert humans on basically any topic rather than chatGPT trying to mimic that knowledge. There are some exceptions for particular questions where it can wow people, but overall, anything that takes a textbook in college to learn, a human will understand it better.

    • @AG-ld6rv
      @AG-ld6rv 2 дня назад +1

      @@MusingsOAM No. Yet again, you are completely incorrect. People DID expect machine learning to outperform handcrafted algorithms. Reread my original post for more information. You might be an arrogant know-it-all with people around you who dislike it or perhaps just trust your confidence, but in this case, you are talking to someone who has studied machine learning for 1000s of hours.

  • @askthepizzaguy
    @askthepizzaguy 9 месяцев назад

    22:00 black decides that the white bishop is worth more than the black rook, and wants to trade those two pieces. White also decides that the white bishop is worth more than the black rook, and decides to take the black rook with his own rook instead of with his bishop. Both computer programs agree that white's bishop is worth more than a rook. It's not three points versus five points at all, each piece has a fluid value depending on where it is placed in the current position. Sometimes it's basically worthless, like when it is trapped, or when its only value is in stopping checkmate by blocking an attack and putting itself in an absolute pin to stop the attack. Sometimes, like this bishop, the piece is so strongly placed that it is worth more than a supposedly much stronger piece that is having to play defense to stop the assault of the bishops. Being passive and forced into the role of being a meat shield for the king truly lowers the value of a piece that normally gets its value from the range of spaces it can move to and the speed at which it can attack an opponent's position. Lowering that mobility and putting it on passive guard-only duty really wrecks the value of the piece, to the point you'd gladly trade it away for the minor piece that is utterly dominating your king.

  • @MrHeroicDemon
    @MrHeroicDemon 9 месяцев назад +1

    I will allow the ads to run only for some FineGold.

    • @Corteum
      @Corteum 9 месяцев назад

      You dont use ad blockers?

  • @vibovitold
    @vibovitold 4 месяца назад

    I'd really like to show one of these games to someone like Capablanca or Alekhine (without telling them where it came from), i wonder what they'd think of it.: ) would they "get it"? would they think it was played by genius players? or dismiss it as nonsense

  • @juliandiazromero4101
    @juliandiazromero4101 9 месяцев назад

    OG EBN!

  • @mikeweaver8495
    @mikeweaver8495 9 месяцев назад +1

    Everyone is ignoring the elephant in the room. Alpha zero surpassed stockfish in a MUCH shorter development period. Stockfish had years of development at the time alpha zero was better, but alpha zero had mere months of development. So yes BECAUSE alpha zero was shut down in 2018 and stockfish continued developing for years after, stockfish is better now than alpha zero was then. This is like comparing Magnus Carlson to Paul Morphy or even Bobby Fischer. Yes, Magnus Carlson of today could beat them then, but they were so much farther ahead of their competitors than Magnus is from his competitors. Now, if we look at leela chess zero, an AI that has been in development for a few years instead of months, we see that it CAN beat the best modern stockfish. Now think about how much more advanced alpha zero was than leela and apply the same or faster advancement for alpha zero. I don’t understand how anyone would think that if alpha zero had continued development, it wouldn’t be the best chess entity in the world. It just doesn’t make sense.

    • @vibovitold
      @vibovitold 4 месяца назад

      Carlsen, not Carlson.
      And the fact that every strong language (including Stockfish) includes a neural network now (NNUE) pretty much answers the question of whether the approach prototyped by AZ is superior or not.
      It is even more obvious where you look at Go. Due to (much) greater computational complexity, there have NEVER been any "traditional" Go engines capable of beating the best humans (without handicap). No "Go-Stockfish" existed.
      Which is why AlphaGo was tested against the best human players (beating them).
      This has only become possible once deep learning / neural networks were used, as opposed to traditional algorithms.

  • @bobi5202
    @bobi5202 9 месяцев назад +3

    Rawrrrr

  • @DekarNL
    @DekarNL 9 месяцев назад +2

    A human brain works on roughly 20 Watts of power. I wonder how strong an engine is when limiting it's computing power to a cell phone's vs a GM. @Ben: thoughts?

    • @许玄清
      @许玄清 9 месяцев назад +1

      Stockfish on your phone is already stronger than all GMs out there

    • @DekarNL
      @DekarNL 8 месяцев назад

      Noted, but what if the brain power required to play chess is between 1% and 10% the computing power of a phone. Limiting computing power to that: Who would win? It'd be a cool experiment 😇

  • @ra1u
    @ra1u 9 месяцев назад

    In final position of last game, after Bf8 Rxf8 . h8Q Rxh8 Rxh8 is still hard to break the fortress as there are no pawns available for attack.

  • @perakojot6524
    @perakojot6524 9 месяцев назад

    That black pushing pawn to e4 (move 28.) is an instant loss. It's like a ??? type of a blunder for a strong engine. There is really no reason to analyse the game any more since after that move it's practically a forced loss (even though engines of that era - 5 years ago didn't really have a clue).

  • @ThunderChickenBucket
    @ThunderChickenBucket 9 месяцев назад

    nice

  • @ReyBasilisko
    @ReyBasilisko 9 месяцев назад

    You are not cool Ben. There you go! 1:58

  • @Aaron-i4k
    @Aaron-i4k 9 месяцев назад

    My favorite game in this match is when Alpha zero plays as white in the Queen's indian

  • @ernietollar407
    @ernietollar407 9 месяцев назад

    has Alpha Zero run away hiding from new Stockfishes

  • @andsviat
    @andsviat 9 месяцев назад +1

    Leaving a nasty comment even before you started explaining anything and everything.

  • @EliasMheart
    @EliasMheart 8 месяцев назад

    So is this "Great Engines of the Past"?
    Or ... "Great Systems of the Past"... No, I don't think we have a good word that specifically includes AI(ML) and also engines/expert systems, but is still narrow enough...
    Ah well, I like the first idea, anyway

    • @vibovitold
      @vibovitold 4 месяца назад

      Chess engine is basically any program that plays chess. Pretty much all strong chess engines (including Stockfish) incorporate NNUE these days. So whether it's based on neural networks or manually encoded heuristics, it's still a chess engine.

  • @goodlaugh1094
    @goodlaugh1094 9 месяцев назад +1

    Stockfish 13 was the last version of stockfish to be weaker than Alphazero. Stockfish 14 was already considered stronger. Stockfish 13 and Alphazero do not maximize their advantages from the opening, while stockfish 14, 14.1 have near to complete drawable foresight to the game. Stockfish 15 does better sacrifices slightly. 15.1 and 16 reached a new super level by world record sacrifices and depth to the game. Leelazero, Dragon, and even Torch are a bit behind since they are still only around the level of brilliancy as Stockfish 15.1.

    • @Corteum
      @Corteum 9 месяцев назад

      I'd agree with your assessment. When I tested SF13 by having it analyze some of AZ's wins, it couldnt find some of the moves that AZ played in critical positions. But SF15 and 16 are most definitely stronger than AZ. I wish someone would hack DeepMInd, steal the code for AZ, and make a a new "Frankenstein" version of AZ. 😂 I reckon it'll be even stronger!

    • @许玄清
      @许玄清 8 месяцев назад

      @@CorteumGoogle Leela Chess Zero

  • @atwarwithdust
    @atwarwithdust 9 месяцев назад

    And AlphaZero says no kibitzing.

  • @ocudagledam
    @ocudagledam 8 месяцев назад

    Ginger GM was pushing Harry up the board before A0!

  • @FranciscoTriemstra
    @FranciscoTriemstra 9 месяцев назад

    1000 games? Tyler1: Hold by beer...

  • @TymexComputing
    @TymexComputing 9 месяцев назад

    Can A0 play F3 - only as white?

  • @victorfinberg8595
    @victorfinberg8595 9 месяцев назад

    never play f3, never start a land war in asia, and never trust a headline
    (and 7 times never kill man)

  • @Writerscabin
    @Writerscabin 9 месяцев назад

    Leela is also free right ?

  • @paulgoogol2652
    @paulgoogol2652 9 месяцев назад +2

    Ya engines don't get tired and won't mind playing a 24 hours long game because they ain't got no life. Just like me, except for the tired part.

  • @ahrrydepp493
    @ahrrydepp493 9 месяцев назад

    Thats the old stockfish, 12 mybe, but now you have stockfish 15 who can easly defeat alpha

  • @victorfinberg8595
    @victorfinberg8595 9 месяцев назад

    actually, 155-6, with even a million draws DOES qualify as a "crush"

    • @-zelda-
      @-zelda- 9 месяцев назад +1

      +155 -6 =1000000 is just a 0.05 elo difference

    • @许玄清
      @许玄清 9 месяцев назад

      1000 Games is just too small of a sample size.

  • @exitar1
    @exitar1 9 месяцев назад

    ChessGenius for the win...

  • @jaroslavtavgen3939
    @jaroslavtavgen3939 9 месяцев назад +1

    AlphaZero never existed (at least as a fully functional and strong chess engine). Nobody has seen it, the chess players were never allowed to "play with it" (although many of them wanted to), the source code was never released (why not? the engine is abandoned), and the notation of 862 games (out of 1072) which were allegedly played against Stockfish 8 were never published.

    • @许玄清
      @许玄清 8 месяцев назад

      Classic google move

    • @vibovitold
      @vibovitold 4 месяца назад

      According to my info (my friend works at Google), "AlphaZero" was Hans Niemann disguised as a neural network.

  • @Wurto
    @Wurto 9 месяцев назад

    Hey Ben,
    Leela is also an engine just like Stockfish and they are both "AI". The difference is mainly how they evaluate positions.
    And even though they are similar strength, I find Leela much better for analysis since it can give you stats (such as estimated win loss draw probabilities) that Stockfish cant (Stockfish "thinks" in centipawns). So I think the chess world/coverage could benefit from adapting Leela as the main engine.
    Cheers, thanks for the content

    • @vibovitold
      @vibovitold 4 месяца назад

      Stockfish's algorithm was written manually. (It's got a NNUE component now, but it didn't back then).
      What we refer to as "AI" in this context are implementations based on neural networks, meaning they're black-box algorithms which have kind of "written themselves" by trial and error..

  • @donovanwarren5012
    @donovanwarren5012 9 месяцев назад

    kf1!!

  • @oaw972
    @oaw972 8 дней назад

    how about a nice game of global thermo nuclear war?

  • @BobChess
    @BobChess 9 месяцев назад

    Smartest fish in the world is Stockfish

  • @kmarasin
    @kmarasin 9 месяцев назад +1

    In the last game, a GM playing black would've taken the draw by repetition. Probably all those 800+ draws went a lot like this. Which begs the question: why this game?
    I think AlphaZero evaluated the positron as winning every time. By the barest margin. Somewhere deep in its behavior model, there's gotta be a place where it decided "when I'm winning by any margin, I refuse to repeat." And that's even if the move that breaks repetition is worse, as long as the overall evaluation is equal or in its favor. So it allowed its position to become worse from every refusal, until its usual model didn't work. Truly a 0.6% error.

  • @noobmaster006
    @noobmaster006 9 месяцев назад

    21:14 hahaha 🤣

  • @patience9323
    @patience9323 2 месяца назад

    Nasty comments!

  • @paulgoogol2652
    @paulgoogol2652 9 месяцев назад

    I believe what is cool about AZ is that it doesn't know the concept of material in a way that humans do and therefore Stonkfish.

    • @f.d.3289
      @f.d.3289 9 месяцев назад

      what is cool is that AZ doesn't know ANYTHING at all. it only knows how to learn.

    • @MrBonified66
      @MrBonified66 9 месяцев назад

      It didn't know the concept of *anything* outside of the rules. That's the amazing part. It's just here are the rules, play a billion games against yourself and see what works

  • @-_Nuke_-
    @-_Nuke_- 9 месяцев назад

    StockFish doesn't know how to play chess. It uses a negamax function that mini maxes thousands of positions per second with alpha beta pruning and iterative deepening, by comparing static evaluations that come from a static evaluation function that doesn't necesserily understand how chess is being played, its just good enough to beat all Humans and most other engines...
    But if a Human GM had the ability to remember a perfect oppening book, had the ability to have a perfect ending book, and the ability to calculate equal amount of positions per second and acess them perfectly at any second... Then the Human GM would always win against StockFish and he would in fact prove that the engine doesn't know what it is doing...
    But a neural network does! A neural network indeed thinks like a Human, and sort of understands that chess more deeply... Of course only a saintient being could understand what chess is, but neural networks show sparks of AGI - artificial general inteligence, and that is much closer to playing stronger chess overall.
    Maybe StockFish 16 or whatever we have now, is already a bruteforce - neural network AI hybrid and that is the only way to reach such high elo. And by StockFish 16 I don't mean the one you have on your phones...

    • @许玄清
      @许玄清 9 месяцев назад

      Stockfish 16 uses NN and Stockfish 16 runs on your phone. It also has a branching factor of under 1.7, which is way less “brute force” than even human GMs

  • @mrcleanisin
    @mrcleanisin 9 месяцев назад

    The stockfish16(GM) on my phone says 2850. So, there's another version that is 3350? Very difficult to even get a draw with this one.

    • @许玄清
      @许玄清 8 месяцев назад

      What app are you using

    • @mrcleanisin
      @mrcleanisin 8 месяцев назад

      Chessify

    • @许玄清
      @许玄清 8 месяцев назад

      @@mrcleanisin Chessify’s Stockfish seems to be handicapped to 2800 elo

    • @许玄清
      @许玄清 8 месяцев назад

      If you want to play the non handicapped version use either Droidfish or SmallFish (depending on whether you use iOS or Android)

    • @mrcleanisin
      @mrcleanisin 8 месяцев назад

      I'm not sure I understand you. I have chessify on my android phone that says it uses stockfish rated 2850. It beats me really easy, but I have gotten a draw when I take back some weak moves. I would imagine this chessify on my phone would beat Magnus Carlson. Why don't you try it?

  • @askthepizzaguy
    @askthepizzaguy 9 месяцев назад

    I have only watched 12 seconds of this video but I feel strangely compelled to leave a nasty comment for no reason. Odd since I usually damn this channel with praise coming from the likes of someone such as me. My approval brings you shame, etc.

  • @loophazard
    @loophazard 9 месяцев назад

    obligatory nasty comment: *shakes fist*

  • @Corteum
    @Corteum 9 месяцев назад

    Having looked at the games, I can say that AlphaZero was better than SF8, SF9, SF10, and I'd say even SF11. But SF12 would be pretty close to AlphaZero's level. AZ was at least 3600-3650 FIDE classical level. SF8 was about 3400-3450 classical on the rating lists.

    • @zvonimirtosic6171
      @zvonimirtosic6171 9 месяцев назад

      SF8 I think was, in reality, 3200-3250. Many of its moves were predictable. And still, Stockfish had years and years of development and learning and openings repertoire.
      All of that sunk at the bottom of the sea when torpedoed by mere 9 hours of self-learning by A0. And A0 moves were utterly mind-blowing and unpredictable. We got an all new insight into the game.

    • @Corteum
      @Corteum 9 месяцев назад

      @@zvonimirtosic6171 I think if you had SF8 competing in disguise in human elite events, it would wipe the floor clean. Former WCCC Rybka 3 was already over 3000. Rybka 4 was not quite as strong as SF8 and had a rating of around 3350. But i remember clearly SF8 having been rated at over 3400 on multiple ratings lists.
      A good way to compare AZ with today's engines is to have SF12 or higher play a match against SF8 using the exact same openings as it played vs AZ. You'll learn a lot from those games and have a much better idea of how strong SF12 (or higher) is compared to AZ.

    • @zvonimirtosic6171
      @zvonimirtosic6171 9 месяцев назад

      What I wanted to say, perhaps, is that chess engines do not perform at presumed ELO rating across all possible aspects of playing. On some aspects they are good and maybe better than their alleged ELO rating, but on some they are not as good as their alleged rating. Maybe this is "human way" of assessing engine's performance (which does not suffer from human issues, but it does have its own weaknesses). I do remember using engines before A0, and they played "predictably boring" hi-level chess, without blunders. But with A0, though, for the first time I saw something I never saw before, and which Magnus Carlsen called "something we can mistake for creativity". A0 approached the game in a wholly different way and that chess was exciting to watch.

    • @Corteum
      @Corteum 9 месяцев назад

      @@zvonimirtosic6171 A0 is definitely something else. I remember testing Stockfish 12 NNUE on some of AO's brilliant moves, and SF12 struggled to find the correct solution. It wasn't until SF14 and 15 that it started to find A0's moves.
      I agree that the top engines have their strengths and a few (very few) weaknesses. But in general, their performance on different aspects of chess is superior to elite GMs. In saying that, there are still some positions they struggle with that i've come across. But those are very few.
      It would be interesting if they ran an average engine (e.g. Houdini 1.5a or Rybka 3 - both former WCCC's), using just 1-cpu and a small hash, in a top level human tournament, without the human players knowing about it. And just see what the approximate human elo level is for these engines. Then do the same with a more modern NNUE neuralnetwork engine. My guess is that a few GMs may get some draws against Rybka or Houdini, but probably no wins. Meanwhile, the NNUE engine will not lose, or even draw, a single game vs elite GMs. Maybe more advanced AI's in the near future will be able to accurately model the performance and style of different elite GMs from past or present, and then test to see how they might have performed against chess computers from past or present using different hardware configurations. Would love to see it. 👍

  • @p1god2
    @p1god2 9 месяцев назад +6

    nasty comment #1

  • @TheMasterboi1
    @TheMasterboi1 9 месяцев назад

    No Talking

  • @oriondx72
    @oriondx72 9 месяцев назад +1

    Stockfish is just overrated!

    • @许玄清
      @许玄清 9 месяцев назад

      Stockfish’s tournament wins speaks for itself

  • @xtripx4273
    @xtripx4273 9 месяцев назад

    So many idiots in comment. Nice video Ben! Keep it up 😊

  • @bugzbunny109
    @bugzbunny109 9 месяцев назад

    Hi guys, I'm here to leave a nasty comment.

  • @JannisSicker
    @JannisSicker 9 месяцев назад

    leaving a nasty comment for good measure

  • @jaydubbs3902
    @jaydubbs3902 9 месяцев назад +1

    A bit much chatter and not enough analysis. Just get into it! Typically enjoy your work but not this.

  • @iAm7amdoh
    @iAm7amdoh 9 месяцев назад

    A very nasty comment

  • @sugarcravings1797
    @sugarcravings1797 9 месяцев назад

    They didn't use Stockfish 9 because they couldn't afford it, and so with Stockfish 10, and stockfish 11, and stockfish 12, and stockfish 13, and stockfish 14.

    • @sugarcravings1797
      @sugarcravings1797 9 месяцев назад

      @@Corteum I'm pretty sure Stockfish 16 is good enough to defeat her consistently.

    • @Corteum
      @Corteum 9 месяцев назад

      @@sugarcravings1797 Most definitely. SF16 is super strong. But there are some SF derivatives that are even stronger.

    • @许玄清
      @许玄清 8 месяцев назад

      @@CorteumDerivatives are just scam

  • @JojenReed
    @JojenReed 9 месяцев назад

    Very nasty comment. Nice glasses though Ben

  • @iAm7amdoh
    @iAm7amdoh 9 месяцев назад

    Very nasty comment