A.I Learns To Walk (Reinforcement Learning - RockRL)

Поделиться
HTML-код
  • Опубликовано: 9 окт 2024
  • 🚶‍♂️ A.I. Learns to Walk (RockRL) - Reinforcement Learning with PPO 🤖
    Step into the realm of cutting-edge artificial intelligence and the captivating world of Reinforcement Learning with our groundbreaking RockRL library!
    🧠 About RockRL:
    RockRL is a powerful Machine Learning library with a sharp focus on Reinforcement Learning. In this enlightening video, I unveil its power by showcasing both discrete and continuous versions of the Proximal Policy Optimization (PPO) algorithm implemented in TensorFlow. This is a game-changer for anyone interested in the potential of AI and RL.
    🚶‍♂️ Mission Objective - BipedalWalker-v3:
    Our AI agent is on a mission to master the art of bipedal locomotion, a challenging task that demands balance, coordination, and adaptability. We're setting the stage for AI to learn how to walk, and RockRL is here to guide us through this fascinating journey.
    📋 What to Expect:
    Introduction to RockRL: We'll begin by introducing RockRL and the PPO algorithm (both discrete and continuous), giving you a strong foundation in Reinforcement Learning and its applications.
    Setting Up the Environment: Learn how to set up the BipedalWalker-v3 environment and witness the AI agent's journey from stumbling to graceful walking.
    Training & Testing: Observe the AI's learning process, from its initial steps to finally mastering bipedal locomotion.
    Insights & Analysis: We'll dissect the AI's progress, discuss key takeaways, and explore the intricacies of Reinforcement Learning.
    Future Possibilities: Delve into the limitless potential and applications of AI with RockRL.
    🚀 Why This Matters:
    This video is not just about an AI learning to walk; it's about the transformative power of machine learning in enabling AI to solve complex, real-world challenges. From robotics to assistive technology, AI's capacity to master complex tasks is awe-inspiring.
    🔗 Resources:
    RockRL: github.com/pyt...
    Tutorial link: github.com/pyt...

Комментарии • 4

  • @NeededANewName1
    @NeededANewName1 11 месяцев назад +1

    Thank you for all the tutorials, they really help a lot!
    Do you know if RockRL can be used with CNNs and images for inputs? I tried to implement it but I keep getting an " 'p' must be 1-dimensional" error which I'm pretty sure is because it only takes observations that are 1 dimensions. Unless there is some setting I have overlooked.
    If not do you know any other library that can do that? Preferably with tensorflow as that is my preferred NN and CNN library.
    Thank you in advance!

    • @PyLessons
      @PyLessons  11 месяцев назад +1

      Hey, thanks for support. Yes, RockRL can use images as input and you can use CNNs for that. Maybe you are doing something wrong, because I tested it everything was fine. Open github issue if you want to discuss this further, I'll try to help. Post your code within your question

    • @NeededANewName1
      @NeededANewName1 11 месяцев назад

      @@PyLessons That is great! If it works I don't have to look for another library.
      And yeah I properly am doing something wrong in that case.
      I was a bit hesitant trying to debug it before because I was not even sure it was possible. I will work with it some more before opening a github issue.
      If you still have your test could you upload it to the github as an example? I think that would really help.
      Thank you.

    • @NeededANewName1
      @NeededANewName1 11 месяцев назад

      ​@@PyLessons Found the issue, my CNN actually just had a weird shape which caused an error further down.
      I was using a Dense layer before a Flatten which does not work and the error just happen to be " 'This variable I don't know further down in the library' has to be in 1 dimension" and since I was not sure if RockRL could even do multy dimension input I just assumed the error was in the input being 3D and not the dense layer being 3D.
      So yeah anyway, thank you again for the quick response, so much easier debugging when you know it 'should' work.