ARC Prize
ARC Prize
  • Видео 15
  • Просмотров 45 991
General Intelligence: Define it, measure it, build it
Dr. François Chollet (Keras, ARC-AGI, Google) presents a keynote talk at the AGI-24, The 17th Annual AGI Conference, on August 15, 2024, in Seattle, WA.
The AGI-24 Conference features workshops, keynote talks, AGI research technical paper presentations, focused on consideration for creating machines that think.
Full conference livestream recordings:
ruclips.net/video/Me187k6RQlQ/видео.html
Conference website: agi-conf.org/2024/
Просмотров: 8 498

Видео

How Will AI Policy Impact AGI Progress?
Просмотров 4192 месяца назад
Mike Knoop, Co-founder of Zapier and ARC Prize, speaks at Y Combinator's private event: "AI, Startups, and Competition: Shaping California’s Future" on July 25, 2024, at YC's San Francisco headquarters. Event participants included: Federal Trade Commission Chair Lina Khan, Assistant Attorney General Jonathan Kanter, and California State Senator Scott Wiener (SB 1047 author). Watch to catch a de...
Testing Frontier LLMs (GPT4) on ARC-AGI
Просмотров 3,8 тыс.3 месяца назад
Template: www.kaggle.com/code/gregkamradt/using-frontier-models-on-arc-agi-via-langchain?scriptVersionId=184611945 arcprize.org/leaderboard arcprize.org/arc-agi-pub ARC Prize is a $1,000,000 public competition to beat and open source a solution to the ARC-AGI benchmark. Hosted by Mike Knoop (Co-founder, Zapier) and François Chollet (Creator of ARC-AGI, Keras). Website: arcprize.org/ Twitter/X: ...
Francois Chollet recommends this method to solve ARC-AGI
Просмотров 8 тыс.3 месяца назад
ARC Prize is a $1,000,000 public competition to beat and open source a solution to the ARC-AGI benchmark. Hosted by Mike Knoop (Co-founder, Zapier) and François Chollet (Creator of ARC-AGI, Keras). Website: arcprize.org/ Twitter/X: arcprize Newsletter: Signup @ arcprize.org/ Discord: discord.gg/9b77dPAmcA Try your first ARC-AGI tasks: arcprize.org/play
ARC Benchmark Origins
Просмотров 3,2 тыс.3 месяца назад
ARC Prize is a $1,000,000 public competition to beat and open source a solution to the ARC-AGI benchmark. Hosted by Mike Knoop (Co-founder, Zapier) and François Chollet (Creator of ARC-AGI, Keras). Website: arcprize.org/ Twitter/X: arcprize Newsletter: Signup @ arcprize.org/ Discord: discord.gg/9b77dPAmcA Try your first ARC-AGI tasks: arcprize.org/play
Implications of solving the ARC benchmark
Просмотров 3,4 тыс.3 месяца назад
ARC Prize is a $1,000,000 public competition to beat and open source a solution to the ARC-AGI benchmark. Hosted by Mike Knoop (Co-founder, Zapier) and François Chollet (Creator of ARC-AGI, Keras). Website: arcprize.org/ Twitter/X: arcprize Newsletter: Signup @ arcprize.org/ Discord: discord.gg/9b77dPAmcA Try your first ARC-AGI tasks: arcprize.org/play
Explore ARC-AGI Data + Play
Просмотров 7 тыс.3 месяца назад
ARC Prize is a $1,000,000 public competition to beat and open source a solution to the ARC-AGI benchmark. Hosted by Mike Knoop (Co-founder, Zapier) and François Chollet (Creator of ARC-AGI, Keras). Website: arcprize.org/ Twitter/X: arcprize Newsletter: Signup @ arcprize.org/ Discord: discord.gg/9b77dPAmcA Try your first ARC-AGI tasks: arcprize.org/play
Announcing ARC Prize
Просмотров 3,4 тыс.3 месяца назад
ARC Prize is a $1,000,000 public competition to beat and open source a solution to the ARC-AGI benchmark. Hosted by Mike Knoop (Co-founder, Zapier) and François Chollet (Creator of ARC-AGI, Keras). Website: arcprize.org/ Twitter/X: arcprize Newsletter: Signup @ arcprize.org/ Discord: discord.gg/9b77dPAmcA Try your first ARC-AGI tasks: arcprize.org/play
Welcome To ARC Prize - Mike & Francois
Просмотров 2,6 тыс.3 месяца назад
ARC Prize is a $1,000,000 public competition to beat and open source a solution to the ARC-AGI benchmark. Hosted by Mike Knoop (Co-founder, Zapier) and François Chollet (Creator of ARC-AGI, Keras). Website: arcprize.org/ Twitter/X: arcprize Newsletter: Signup @ arcprize.org/ Discord: discord.gg/9b77dPAmcA Try your first ARC-AGI tasks: arcprize.org/play
Francois Chollet On LLMs w/ Active Inference
Просмотров 2,4 тыс.4 месяца назад
ARC Prize is a $1,000,000 public competition to beat and open source a solution to the ARC-AGI benchmark. Hosted by Mike Knoop (Co-founder, Zapier) and François Chollet (Creator of ARC-AGI, Keras). Website: arcprize.org/ Twitter/X: arcprize Newsletter: Signup @ arcprize.org/ Discord: discord.gg/9b77dPAmcA Try your first ARC-AGI tasks: arcprize.org/play

Комментарии

  • @picklenickil
    @picklenickil 17 дней назад

    Oh it give me an 💡 idea. It sounds to me like when I watch my nephew learning to walk, my wife comes and provies him with a walker (hyper parameters space). What doesn't make sense to me is that, are we trying to build the equivalent of a kid, the walker, or wifu... Or a god forbidden amalgamation of the three?

  • @hamidelhachimi6343
    @hamidelhachimi6343 17 дней назад

    Your approach is good, i think the Gpt 4'o will perform well on the training phase. Did you notice that for the testing part, the shapes of Output can significantly vary How do you think you can guess the right answer , if you have only 2 attempts, for example, based on shape of the input (3,3) several solution are possible as output shape (3,3) (1, 2)(5, 5)(9, 9)(3, 1)(1, 1)(6, 6)(1, 4)(12, 12)(18, 18)(3, 6)(5, 3)(6, 3)(1, 3)(15, 15). Notice also the same key can lead to diffrent output. Thanks for your return

  • @clerothsun3933
    @clerothsun3933 18 дней назад

    I get that this is a stepping stone, but calling it a test for AGI is just ludicrous. This isn't even close to AGI, it's just a toy.

  • @peterdayton6796
    @peterdayton6796 24 дня назад

    Excellent talk. Chollet deserves a lot of credit for developing a benchmark before the GPT craze that is so simple and yet for which GPT scaling has only led to modest improvements on. Even if it does fall soon to ad-hoc approaches, lasting as long as it has is no small feat.

  • @imad1996
    @imad1996 27 дней назад

    Somebody from space is listening, "Oh, poor humans, They still refer to abstraction as a blurred image, hahahahahahah." Abstractions have their own use, but they could be a large source of limitations if we try to imitate that from humans. Abstractions could be human limitations in trying to understand matters. Unless we refer to the AI model as the abstraction layer. I don't know how relevant that could be.

  • @imad1996
    @imad1996 27 дней назад

    Throwing a million dollars into an AI puzzle fuels substantial momentum toward AI research. And that is awesome. Thank you for differentiating between the output of the process and the process itself. Unfortunately, trending concepts become marketing slogans, such as Samsung Galaxy AI :).

  • @immortalityIMT
    @immortalityIMT 29 дней назад

    Do not we have enough A.I. to brute force the model we are needing?

  • @googleyoutubechannel8554
    @googleyoutubechannel8554 Месяц назад

    Going through the arc training tests, it's as I feared, many many of the problems are based on cultural conventions of human layout of information, human pattern design, or well known cultural patterns, or comp sci cultural knowledge, not to mention the bias in our visual system. There is no way for even a hyper 'intelligent' tabula rasa system to answer these problems correctly, as the answers are not even 'correct'. The winner will likely be the team that hardcodes in all the cultural visual baggage, where the 'answer' may not be strongest 'pattern' by any measure. A hyper intelligent system wouldn't ever score well on ARC, but a shitty human-overtuned system, one implemented probably with CLIP or something, that is actually a relatively dumb search over an exhaustive corpus of hand-coded patterns built to match the creator's ideas for input/output pairs. The creators of this test think they're smart because the search space is so large, but it isn't, they haven't realized in their hubris the number of different types of 'patterns' they're likely to think of, coming out of a few brains, that are actually good and 'fair' questions that require reasoning, is relatively small, and the kaggle competitors will think of most of them, submitting boring entries over and over again, slowly matching the answers in the problem set as they add ideas. (eg. I bet there's a grip of problems like 'game of life' cell automata, or battleship!, or fibonacci tee hee! all that type of crap, in the private test set, that sort of thing, and I bet the private answers are 'not even wrong', in that they're not the simplest or strongest pattern by any measure, other than by human comp-sci nerd culture)

  • @googleyoutubechannel8554
    @googleyoutubechannel8554 Месяц назад

    I took a look at the ARC Prize... it... it doesn't make sense? A very smart human scores 22% on ARC... the top AI score is 46% on the leaderboard ALREADY? so... what are you trying to measure here? who is best at 'guess the algorithm I'm thinking of'?

  • @humnhumnhumn
    @humnhumnhumn Месяц назад

    ''You are confusing the output of the process with the process itself'' Very nice!

    • @aimorethanyouaskedfor
      @aimorethanyouaskedfor 9 дней назад

      The argument against functionalism. It amazes me how many think that these are the same thing.

  • @way2on
    @way2on Месяц назад

    we are not providing the same tools back to the core architecture. such as realtime inference

  • @danecjensen
    @danecjensen Месяц назад

    Chapters (Powered by @danecjensen) - 00:00 - Intelligence, benchmarks, and AI hype in 2023 03:26 - LLMs autoregressive responses 07:36 - LLMs inability to solve CSR ciphers 10:14 - ML models rely heavily on human labor 11:41 - Minskystyle AI vs Makatsstyle AI 15:25 - Intelligence spectrum skilled, operational, efficient 16:42 - AI models should not be evaluated using human exams 17:29 - AIs next level of capabilities and efficiency 18:53 - RKGI AI benchmark for human intelligence 20:49 - Data sets Kaggle, RKGI, ARK 21:45 - Zach offers 1m RKGI solution competition 24:58 - Physicists describe intelligence and abstractions 29:09 - AIs ability to master tasks efficiently 30:50 - Two types of abstraction valuecentric and programcentric 33:48 - Decentralized program search forAGI 34:27 - Program synthesis PS vs machine learning 35:30 - Program synthesis overcomes combinatorial explosion, LMS limitations 36:46 - AI combines chess and discrete search techniques 41:17 - Deep learning components in discrete programs 41:56 - Deep learning for ArcGi program synthesis 43:19 - Program embedding for efficient search 43:45 - Python ArcGGI pipeline improvement 44:55 - LLMs fall short of GI, need breakthroughs 47:50 - Breakthrough in ArcGGI likely to come from outsider, not big labs 51:09 - Experiential learning and causality in childrens learning 52:16 - Humans are capable of few short program synthesis 53:32 - Human cognition works on fundamental level

  • @techsuvara
    @techsuvara Месяц назад

    I love that the ARC AGI prize is $1mil, where the if in fact you can build something which can scale to AGI. The prize money for a business would by in the the billions...

  • @techsuvara
    @techsuvara Месяц назад

    And they don't even drive cars well.

  • @JanSmetana
    @JanSmetana Месяц назад

    16:00 "If you know hot to drive only in very specific geo fence adreas - thats intelligence". Isnt this a SKILL how you defined that? An INTELLIGENCE is ability to drive anywhere (left, right, town, country ...), and not just at my own 3x3 streets box? (which would be learned skill, where you already tried every possibility)

    • @micheldominic5963
      @micheldominic5963 Месяц назад

      he said "you know thats less intelligent" not "thats intelligence"

  • @Maximooch
    @Maximooch Месяц назад

    “See you on the leaderboard for ARC-AGI” is quite a way to end it

  • @techsuvara
    @techsuvara Месяц назад

    Excellent reality check for AI. I bet this video won't have 1 million views like the hype videos.

  • @techsuvara
    @techsuvara Месяц назад

    There's IQ tests for that :D . This also goes to show how flawed IQ tests are. Especially when the children of those who wrote the tests seem to score the highest. (True story)

  • @NitsanAvni
    @NitsanAvni Месяц назад

    20:20 Should we also measure agents with less / more shots? How many shots needed to get an agent to a score of 90?

  • @PaulTopping1
    @PaulTopping1 Месяц назад

    Hey, that's the back of my head I see at the bottom right! Good presentation at a fun conference.

  • @johnkintree763
    @johnkintree763 Месяц назад

    One definition of intelligence is the ability to solve problems in complex situations. There are many problems from the local to a global level that are not being solved well with the current level of intelligence. Our highest priority is creating collective human and digital intelligence. Language models in agentic workflows can extract entities and relationships from text, and merge that knowledge into graph representations. Keeping the human in the loop is important to catch and correct mistakes made by the speech recognition and language models. People can select parts of conversations they have with digital agents to be merged into a global shared graph representation. A global platform can be built pretty much with today's technology that can merge selected parts of millions of simultaneous conversations into a shared world model by the end of this year.

  • @theotherwhitemeat140
    @theotherwhitemeat140 Месяц назад

    Finally! An AGI-like task

  • @FamilyYoutubeTV-x6d
    @FamilyYoutubeTV-x6d Месяц назад

    This is really good!

  • @TooManyPartsToCount
    @TooManyPartsToCount Месяц назад

    It is not entirely clear that we humans perform some computational magic that is beyond merely relying on pre-learnt patterns. And perhaps it just 'feels like' we reach beyond the training set when we have those eureka moments, or imagine that we just came up with a solution to a novel problem? perhaps in fact we have such a tightly woven mesh of prior training examples that those moments when we arrive at 'eureka' are in fact just inevitable and only appear novel to us because we are unconscious of the combined results of all our prior training?

    • @ggir9979
      @ggir9979 Месяц назад

      That's not really the point that the arch challenge is tackling. This issue is that LLMs cannot do recursions, they cannot do step by step reasoning. When they write down code for a computer program, they cannot interpret it and run it. Everyone that is working with a coding assistant will tell you that they routinely produce code that will not run. Very simple stuff, like simple loops that are easy to unfold. That's why you see a lot of neurosymbolic approaches beeing tried out nowadays ( as explained in the presentation). Humans, on the contrary, can interprate and execute computer programs. Tedious, slow, but doable. We built airplanes and rockets before we had computers, someone had to run these algorithms.

    • @TooManyPartsToCount
      @TooManyPartsToCount Месяц назад

      @@ggir9979 Point about the Arc challenge taken. I was raising what at least to myself is an interesting meta question, that being that our current state of understanding about what constitutes intelligence could be lacking? especially as concerns our (human) particular brand of intelligence. There is no denying that LLMs are far from what I would call intelligent though! more like data compression artifacts. Most claims that LLMs are more than that are just part of some marketing strategy (think Microsoft's 'sparks of AGI' paper/lectures).

    • @ggir9979
      @ggir9979 Месяц назад

      ​@@TooManyPartsToCountGood point! I don't know if you know ben goertzel, he has a lot of interesting things to say about the theory of mind and what is intelligence that goes further than what you could see in this video.

    • @TooManyPartsToCount
      @TooManyPartsToCount Месяц назад

      ​@@ggir9979 This - ruclips.net/video/D8wxThDlVBc/видео.html And some other vids are where it started for me. Basically went back to school thanks to BG and JB :) Every time I see a YT video on AI with Sam Altman in the thumbnail I think 'why not Ben Goertzel?!' or at least Ilya or Andrej or Cholet or Bengio or......

    • @ggir9979
      @ggir9979 Месяц назад

      ​@@TooManyPartsToCountThe rarest of things, a nice and civilized exchange of ideas on the internet :-) I will have to agree 100% with you, any of these speakers are much more interesting to listen to than Sam Altman. I am not familiar with JB's work, so I'll chexk it out next, thanks for the tip!

  • @Dr.Z.Moravcik-inventor-of-AGI
    @Dr.Z.Moravcik-inventor-of-AGI Месяц назад

    "Next breakthrough will come from an outsider" What breakthrough dude? Have you noticed that AGI = brain algorithms and that my brain algorithms are here since 2016? Nothing has changed since. It's all working. We are dealing with human idiots that are screwing humanity's future. How can idiots that don't see my 2016 invention save humanity? You are not going to save us, you are wasting the time humans have LEFT on this planet. You have already wasted 8 long years!! No one on this planet can return us 8 wasted years. This is not meant only on the speaker, this is meant generally on the whole world.

  • @G1364-g5u
    @G1364-g5u Месяц назад

    # AI Progress and Generalization: A Critical Review ## Chapter 1: The AI Hype of Early 2023 **Timestamp: **0:00** - **1:57**** - Overview of the peak AGI hype in early 2023. - ChatGPT, GPT-4, and Bing Chat were perceived as revolutionary, with claims that AI would drastically increase productivity and replace many jobs. - Despite the hype, the actual impact on employment has been negligible, and fears of mass unemployment were unfounded. ## Chapter 2: Limitations of Large Language Models (LLMs) **Timestamp: **1:58** - **3:38**** - LLMs, including ChatGPT, have inherent limitations such as failing to understand context and falling into pattern matching rather than true comprehension. - These limitations are tied to the fundamental architecture and approach of current AI models, showing little progress over time. ## Chapter 3: Problems with Task Familiarity and Generalization **Timestamp: **3:39** - **6:31**** - LLMs struggle with unfamiliar tasks, performing well only on tasks they have memorized. - Performance issues arise from the extreme sensitivity to phrasing and the inability to generalize from known tasks to new, similar ones. ## Chapter 4: The Misconception of AI Intelligence **Timestamp: **6:32** - **13:51**** - Intelligence should not be equated with task-specific skill; true intelligence involves the ability to handle novel situations. - The speaker argues for a shift from task-based AI evaluations to measuring generalization and adaptability. ## Chapter 5: Redefining Intelligence and Measuring Progress **Timestamp: **13:52** - **19:15**** - Intelligence should be viewed as the ability to synthesize new solutions and adapt to new situations. - The current benchmarks, based on human exams, are inadequate for assessing AI’s true generalization capabilities. ## Chapter 6: The Abstraction Reasoning Corpus (ARC) and Generalization Benchmarking **Timestamp: **19:16** - **24:22**** - Introduction of ARC, a dataset designed to measure an AI's ability to generalize and perform novel tasks. - ARC aims to control for prior knowledge and experience, emphasizing the need for AI to infer solutions rather than relying on memorization. ## Chapter 7: The Role of Abstraction in AI and Human Intelligence **Timestamp: **24:23** - **30:36**** - Abstraction is the key to generalization; intelligence depends on the ability to recognize and apply abstract patterns. - LLMs are currently limited to low-level abstraction and lack the capability to synthesize new models on the fly. ## Chapter 8: Integrating Type 1 and Type 2 Thinking for AGI **Timestamp: **30:37** - **37:41**** - The next step in AI development involves combining Type 1 (intuition, pattern recognition) and Type 2 (logical reasoning) thinking. - Human intelligence excels because it merges these two forms of thinking, and AI needs to follow a similar path. ## Chapter 9: Combining Deep Learning with Program Synthesis **Timestamp: **37:42** - **42:48**** - Future AI advancements will likely involve merging deep learning (Type 1) with program synthesis (Type 2) to handle complex, novel tasks. - This approach could significantly improve AI’s problem-solving capabilities and generalization. ## Chapter 10: Practical Applications and the Future of AI Development **Timestamp: **42:49** - **45:12**** - Practical strategies for improving AI, such as using LLMs to generate and refine programs, show promise in advancing AI generalization. - The importance of innovative thinking and diverse approaches in overcoming current AI limitations. ## Chapter 11: The Need for New Breakthroughs and Intellectual Diversity **Timestamp: **45:13** - **48:36**** - The speaker emphasizes that progress towards AGI has stalled due to a lack of new ideas and intellectual diversity. - The speaker suggests that the next breakthroughs are likely to come from outsiders rather than big tech labs. ## Chapter 12: Future Directions and Closing Thoughts **Timestamp: **48:37** - **53:34**** - The development of AI tests and challenges like ARC 2 is discussed, aiming for more sophisticated and dynamic assessments. - Insights from observing human cognitive development, particularly in children, could inform AI research and the creation of more generalizable AI systems.

  • @dizietz
    @dizietz Месяц назад

    I've been a big fan of bongard problems and the Arc test tests similar problem space as bongard problems do. I do think that approaches like Greenblatt's dynamic generation of solutions will push the state of the art forward. I also think that the kind of encoding data the current models use and are trained on isn't fully conducive for this task, and more fine tuned models might help. Also, I think that Arc puzzles should be abstracted further out to three dimensions and more complicated transformations to encode patterns and transformations that even 150+ iq test takers have difficulty recognizing in <30 minutes. Of course constructing that dataset is difficult work. Also, it would be interesting if someone took the arc dataset and constructed an IQ test based on the questions. Given the ease of the questions even in the "hard dataset", possibly timing it like the wonderlic would make sense. I would be curious what the distribution etc of that dataset.

  • @samkee3859
    @samkee3859 Месяц назад

    Chollets perception and articulation are unmatched

  • @ceilingfun2182
    @ceilingfun2182 Месяц назад

    Is this guy for real?

    • @randylefebvre3151
      @randylefebvre3151 Месяц назад

      yes why? The presentation is on point

    • @wege8409
      @wege8409 Месяц назад

      Yes, this is the creator of Keras, which is the software that TensorFlow is based on.

    • @spagetti6670
      @spagetti6670 Месяц назад

      Yes he is and probably he's way smarter than you are

    • @0113Naruto
      @0113Naruto 12 дней назад

      You think we already have AGI?

  • @oysterboulevard6623
    @oysterboulevard6623 Месяц назад

    Great Video !

  • @juliocesar-io
    @juliocesar-io Месяц назад

    Refreshing view! ❤

  • @rtnjo6936
    @rtnjo6936 Месяц назад

    We don't need AGI, we need tools, not gods

    • @FamilyYoutubeTV-x6d
      @FamilyYoutubeTV-x6d Месяц назад

      AGIs will not imply divine qualities worthy of worship. Ironically, the anti-AGI people who make this kind of comment are the ones who will be hyping future intelligent agents to the level of gods, while the people who are interested in using the technology for a variety of purposes will see them as tools. I see your point, though.

    • @spagetti6670
      @spagetti6670 Месяц назад

      Ok bro!!!!

    • @tack3545
      @tack3545 Месяц назад

      agi would be such a good “tool” that it would transcend the standard meaning of the word. why settle for basic tools when so much more lies just beyond the horizon

    • @rtnjo6936
      @rtnjo6936 Месяц назад

      @@tack3545why settle for getting a cat when you could go all out and buy a full-grown tiger? Why choose the basic option? Here’s the reality: I’m in control of my environment. I’m the dominant species, and I understand this world with a high degree of predictability. with current technological advancements, I can expect even better tools, a more efficient society, and more property. so why on earth would I create another agent that could surpass me in every dimension? people who praise AGI for its potential to make the world a better place overlook the inherent risks, once it’s unleashed, there’s no turning back, and the very power that promises great things could also bring catastrophic consequences. now, we’re caught in an AI rat race. different labs, companies, and even countries are fiercely competing to develop the most advanced AI, maybe, in the far future-when humanity is more advanced, when global cooperation is so high that we can make unified decisions, and when our understanding of AI is no longer trapped in today’s black-box methods-then, and only then, might it make sense. the people making decisions about such tremendous technology today, with their average IQ of 130, are incomparable to the future generations who will possess far deeper wisdom about the consequences and complexities of such advancements. for now, there’s no need to gamble with uncontrollable power when we can continue to improve humanity with the tools we have, advancing our standards of living, morality, rights, and freedom.

    • @rtnjo6936
      @rtnjo6936 Месяц назад

      @@tack3545 lmao, they deleted my comment, let's post it again: why settle for getting a cat when you could go all out and buy a full-grown tiger? Why choose the basic option? Here’s the reality: I’m in control of my environment. I’m the dominant species, and I understand this world with a high degree of predictability. With current technological advancements, I can expect even better tools, a more efficient society, and more property. So why on earth would I create another agent that could surpass me in every dimension? It’s like making a deal with the devil. People who praise AGI for its potential to make the world a better place overlook the inherent risks of such immense power. Once it’s unleashed, there’s no turning back, and the very power that promises great things could also bring catastrophic consequences. Right now, we’re caught in an AI rat race. Different labs, companies, and even countries are fiercely competing to develop the most advanced AI. This race often leads to rushed decisions and prioritizing breakthroughs over safety. Maybe, in the far future-when humanity is more advanced, when global cooperation is so high that we can make unified decisions, and when our understanding of AI is no longer trapped in today’s black-box methods-then, and only then, might it make sense. The people making decisions about such tremendous technology today, with their average IQ of 130, are incomparable to the future generations who will possess far deeper wisdom about the consequences and complexities of such advancements. For now, there’s no need to gamble with uncontrollable power when we can continue to improve humanity with the tools we have, advancing our standards of living, morality, rights, and freedom.

  • @dot1298
    @dot1298 Месяц назад

    but what about AlphaProof, AlphaGeometry 2, Q-star, the upcoming Claude 3.5 Opus and Gemini 2 Ultra (aka gemini-test in lmsys) ?

    • @dot1298
      @dot1298 Месяц назад

      and the 2025 project of a massive AI super-computer-cluster?

  • @DustinHunt-x2y
    @DustinHunt-x2y Месяц назад

    Excellent and succinct distillation of the current state of the industry and the path forward. Very inspirational!

  • @tankieslayer6927
    @tankieslayer6927 Месяц назад

    There is no viable path to AGI with the current hardwares and algorithms.

    • @blahblahsaurus2458
      @blahblahsaurus2458 Месяц назад

      It's a very good talk and helps me understand how I went wrong overestimating AI a year ago. But there are problems assuming that the entire field is stuck and predicting that it will remain stuck for years. For one thing, no one person knows the 'current state of the industry', much like no one person knows the locations of every nuclear warhead in the world, or how it's unlikely that any one person knows the locations of all the major meth labs. We know a lot about what OpenAI and the rest are publishing. We can even say that these companies are motivated by keeping their stock prices high, and by selling their products to average people, so in general they all have a strong incentive to reveal and sell their technology as soon as they can. And they all have employees who can leak information anyway. At the same time, we know that all of them do billions of dollars of R&D that they are able to keep secret. We are used to discounting this unknown R&D, because it's been a while since there's been a major perceivable improvement in AI capabilities. It's hard to believe in the promise of Q* when we've been hearing nonsense about it a whole year. However, in 2021 most experts thought it would be decades before something with the capabilities of generative AIs would arrive. We won't know until we know! But more importantly, people make the mistake of forgetting about countries and their resources and assuming that the _public_ state of the art is the _actual_ state of the art. It is virtually guaranteed that every major military power is experimenting with AI. Countries have very different incentives than companies. We can argue all day about what the potential of AI is and what the timeline is, and such arguments are reflected in the stock market, which drives companies to seek short term, and relatively likely returns on investment. And if everyone believes AI is bogus, investors will cash out, and private sector AI development will slow down. A large country on the other hand doesn't care if it spends a billion dollars on a top secret project and nothing comes of it. Every country understands that the *maximum potential value* of AI is very very high, and that the maximum potential *cost* of allowing your enemy to build an army of drones before you do is also very very high. Thus, a billion dollars is a small cost to pay for peace of mind, even if ten years pass and AGI fails to materialize. And almost as bad as losing the AI race is allowing your rivals to learn the state of your AI research. Because it's so difficult for anyone to find out if and when a different country creates AGI and starts building drones that build drones, a country will want to delay revealing their own AGI until the last possible moment. Suppose it's 2030, 2040, or December this year or whatever, and the US, China, and Russia all have big AI drone armies. Russia would be wise to keep its own AI secret, allow the US and China to fight, causing them to waste resources and reveal their technology, and then go on a blitz to try to conquer the world (or at least damage other countries' capacity to make AI and drones). But if Russia does that, how do they know if Saudi Arabia or Brazil or Jeff Bezos haven't been hiding even more powerful weapons? We can ask "why make weapons at all, why wouldn't countries cooperate". But it doesn't matter if 90% of militaries have benign intentions. The existence of that 10% of evil dictators with major resources will make everyone scared, paranoid, suspicious, and determined to defeat potential enemies as soon as possible. So we don't know the state of the race, but we do know that when chatgpt made news it was the starting signal for everyone to enter the race. Or, if these militaries were paying attention, they entered the race after image generators, or GPT-3, or even AlphaGo. These countries may be behind, on par with, or ahead of the private sector. They are going to extreme lengths to hide what they have as well as what they don't have. No one will want to reveal its drone army first, BUT, if you wait you also risk letting a very evil dictator or billionaire get a lead in the race. We will probably start finding out what all the militaries have been doing when the private sector makes significant achievements, whenever that may be. If robots start washing dishes competently, or driverless cars become widespread, or various professions become automated, it will no longer make sense for countries to pretend they don't have weapons. We will probably enter a phase where countries make displays of military power to discourage their enemies from attacking, but these displays will not reveal everything, and they may be bluffs. Eventually someone - US, China, some guys in a basement - will reach the conclusion that they are currently ahead in the race, or that they will soon fall behind in the race, or they'll just get nervous, and make the first attempt to defeat their rivals. This will then devolve into a global war. We can say - well we've had nukes for 80 years, but very little conflict between nuclear powers. Yes of course, because no one wins a nuclear war - everyone is harmed by nuclear retaliation, or radiation. No one wins a nuclear war, that is, unless they're in a bunker deep underground and have an army of drones that build drones that can survive radiation.

    • @coolcool2901
      @coolcool2901 Месяц назад

      Extropic Thermodynamic Chips (Stochastic Processor) The deal breaker.

  • @FliBaleon
    @FliBaleon Месяц назад

    All hyped with LLM, but we do not need human imitation we need human reasoning. The static nature of llm goes against this basic concept, us humans have to think about solutions, make projections in our minds, is not input output. LLMs are just like hype realistic paintings, they look somehow like humans but they are fixed in time.

  • @duncansmothers
    @duncansmothers Месяц назад

    this is really helpful/thoughtful just joining the competition and this is an exceptional resource.

    • @ARCprize
      @ARCprize Месяц назад

      Awesome! Glad to hear it. Let us know if you have any questions Duncan

  • @juliusapriadi
    @juliusapriadi 2 месяца назад

    Data protection, but data stored in the US? Looking forward to Schrems III... The expressed views on data protection (let's put all in ChatGPT!), AI Safety (not important, AGI is far away!) and AI ethics (who needs regulation!) are not very encouraging.

    • @pascalgula
      @pascalgula 2 месяца назад

      ok, so the US don't care, can I skip the video? :)

  • @RoniBliss
    @RoniBliss 2 месяца назад

    It will not have any effect, because AGI will never happen. I am an AI engineer, we don’t even have true AI today. It’s just machine learning trained on data we give it. It does not think for itself or come up with anything outside of the data it has ingested

  • @pascalgula
    @pascalgula 2 месяца назад

    1N73LL1G3NC3 15 7H3 4B1l17Y 70 4D4P7 70 CH4NG3. -- 573PH3N H4WK1NG leads to ARC-AGI, Th(X/C) guys! Genius is the ability to create change -- Me if no one else wrote it! (too lazy to l33tcode) hopefully will lead to ARC-AGG, Merci par avance les mecs! :D Bis because of the quotation! :D

  • @pascalgula
    @pascalgula 2 месяца назад

    can be a big thing! let's se after playing a bit

  • @pascalgula
    @pascalgula 2 месяца назад

    1N73LL1G3NC3 15 7H3 4B1l17Y 70 4D4P7 70 CH4NG3. -- 573PH3N H4WK1NG leads to ARC-AGI, Th(X/C) guys! Genius is the ability to create change -- Me if no one else wrote it! (too lazy to l33tcode) hopefully will lead to ARC-AGG, Merci par avance les mecs! :D

  • @pascalgula
    @pascalgula 2 месяца назад

    genius if nobody did that and performance to be validated on my use-case. when will be released the first Open Source component derived from this initiative ? (sorry if I it is already available but didn't find it yet...)

  • @pascalgula
    @pascalgula 2 месяца назад

    let's play a bit!

  • @D3cast
    @D3cast 2 месяца назад

    I am new to programming, but this challenge and task really interests me and I'd like to give it a try. Could you create a tutorial on how to submit an entry to the arc challenge? maybe with a model which will produce some minimal results?

    • @ARCprize
      @ARCprize 2 месяца назад

      Totally! We have a ton of templates here arcprize.org/guide As for a submission tutorial, we don't have a video of this directly, but this video shows how to work with Kaggle notebooks. ruclips.net/video/crhrzhVjWog/видео.html

  • @RPHacker777
    @RPHacker777 2 месяца назад

    Could someone please explain how the AI soccer players in a simulation can go from physically flopping around on the ground to teaching themselves team strategy but AI can't solve these ARC tasks?

  • @sp3ct3rgaming46
    @sp3ct3rgaming46 2 месяца назад

    i might be tripping but i think this dude cloned his own voice and then layered it into the video. you can hear the typical elevenlabs lisp

    • @ARCprize
      @ARCprize 2 месяца назад

      @@sp3ct3rgaming46 you’re tripping. I did the video and no voice dub used

  • @mosca204
    @mosca204 2 месяца назад

    Why is the train/evaluation set so small?

    • @ARCprize
      @ARCprize 2 месяца назад

      The tasks are handmade which limit the scale that can be done. They focus on diversity rather than quantity at this stage

  • @DistortedV12
    @DistortedV12 2 месяца назад

    Mark Knoop is a very smart guy. Knows a lot about ML for being an outsider in that space and meche background

  • @jackq2331
    @jackq2331 2 месяца назад

    Excellent.

    • @ARCprize
      @ARCprize 2 месяца назад

      Thank you!