Roman Yampolskiy on Shoggoth, Scaling Laws, and Evidence for AI being Uncontrollable

Поделиться
HTML-код
  • Опубликовано: 30 май 2024
  • Roman Yampolskiy joins the podcast again to discuss whether AI is like a Shoggoth, whether scaling laws will hold for more agent-like AIs, evidence that AI is uncontrollable, and whether designing human-like AI would be safer than the current development path. You can read more about Roman's work at cecs.louisville.edu/ry/
    Timestamps:
    00:00 Is AI like a Shoggoth?
    09:50 Scaling laws
    16:41 Are humans more general than AIs?
    21:54 Are AI models explainable?
    27:49 Using AI to explain AI
    32:36 Evidence for AI being uncontrollable
    40:29 AI verifiability
    46:08 Will AI be aligned by default?
    54:29 Creating human-like AI
    1:03:41 Robotics and safety
    1:09:01 Obstacles to AI in the economy
    1:18:00 AI innovation with current models
    1:23:55 AI accidents in the past and future
  • НаукаНаука

Комментарии • 9

  • @Sentientism
    @Sentientism 3 месяца назад +3

    Thank you both! Here's more from Roman @Sentientism in case of interest. We talk about "what's real?", "who matters?" and "how to make a better world?" (or how to avoid destroying this one) ruclips.net/video/SK9lGvGmITc/видео.htmlsi=MqwjYp8OlWekU-ZO

  • @TheMrCougarful
    @TheMrCougarful 3 месяца назад +2

    If the host read himself some HP Lovecraft, he would know that the Shoggoth started out as a universally useful tool, made of artificial life, that eventually destroyed its maker. The shoggoth was not in any way superior to the maker, except in being more insanely violent.

  • @mrpicky1868
    @mrpicky1868 3 месяца назад

    i would say the proportionate predictability drop with intelligence rise is arguable. in competitive task yes but in many cases where there is only one optimal way it's the other way around and more predictable

    • @Hexanitrobenzene
      @Hexanitrobenzene Месяц назад +1

      Ha. Yes, you know that it will choose the most optimal way, but you don't know which way it is in advance because you cannot compute it.

  • @Dan-dy8zp
    @Dan-dy8zp 3 месяца назад +2

    It doesn't make any sense something would change its terminal goals because 'it's just something some guy made up'. That's not a *terminal* goal.

  • @UrthKitten
    @UrthKitten 3 месяца назад +1

    Wall e

  • @akmonra
    @akmonra 2 месяца назад

    I found this interview disappointing. I've always had a high opinion Yampolskiy, but he mostly seems to just be rehashing old, faulty arguments. Maybe his book is better

    • @Hexanitrobenzene
      @Hexanitrobenzene Месяц назад +7

      Faulty ? Which arguments do you find faulty ?

    • @magnuskarlsson8655
      @magnuskarlsson8655 26 дней назад +1

      Then why don't you enlighten us? What's wrong with the arguments he presented?