Larry's Workbench
Larry's Workbench
  • Видео 43
  • Просмотров 24 689
Build your own AI Robot - Ep9 (Speech Transcription)
I built a conversational robot by connecting a Raspberry Pi 4B to GPT 4o and he’s developed a very sarcastic personality.
In this video we see how to use OpenAI's Whisper-1 model to convert a .wav audio file into a text transcription. It works pretty well and you can see our actual source code at github.com/LarrysWorkbench/
Leave a comment and tell us what you'd like to see next!
#Robot #GPT #Python #AI #Robotics #TalkingRobot #ArtificialIntelligence #RaspberryPi #OpenAI #ChatGPT #DIY #Cyberpunk #STEM #STEMProjects
Просмотров: 184

Видео

Robot picks up The Easter Bunny!
Просмотров 6121 день назад
I built a conversational robot by connecting a Raspberry Pi 4B to GPT 4o and he’s developed a very sarcastic personality. In this video we see Floyd respond to verbal commands and pick up a small stuffed toy. Leave a comment and tell us what you'd like to see next! #Robot #GPT #Python #AI #Robotics #TalkingRobot #ArtificialIntelligence #RaspberryPi #OpenAI #ChatGPT #DIY #Cyberpunk #STEM #STEMPr...
Build your own AI Robot - Ep8 (Flowchart)
Просмотров 224Месяц назад
Build your own AI Robot - Ep8 (Flowchart)
This Robot with IQ = 155 will Blow Your Mind!
Просмотров 458Месяц назад
This Robot with IQ = 155 will Blow Your Mind!
Build Your own AI Robot - Ep7 (First API Call)
Просмотров 4692 месяца назад
Build Your own AI Robot - Ep7 (First API Call)
Build Your own AI Robot - Ep6 (Intro to Python)
Просмотров 4592 месяца назад
Build Your own AI Robot - Ep6 (Intro to Python)
GPT Robot shares his earliest memory
Просмотров 7412 месяца назад
GPT Robot shares his earliest memory
Build Your own AI Robot - Ep5 (Development Environment)
Просмотров 9682 месяца назад
Build Your own AI Robot - Ep5 (Development Environment)
GPT Robot does his best Seinfeld impression!
Просмотров 1882 месяца назад
GPT Robot does his best Seinfeld impression!
Build your own AI Robot - Ep4 (Raspberry Pi Startup)
Просмотров 2,9 тыс.2 месяца назад
Build your own AI Robot - Ep4 (Raspberry Pi Startup)
GPT Robot is looking forward to the Olympics
Просмотров 4852 месяца назад
GPT Robot is looking forward to the Olympics
Build Your own AI Robot - Ep3 (Hardware)
Просмотров 4053 месяца назад
Build Your own AI Robot - Ep3 (Hardware)
GPT Robot responds to Voice Commands
Просмотров 3493 месяца назад
GPT Robot responds to Voice Commands
Build Your own AI Robot - Ep2 (Things you won't need)
Просмотров 4193 месяца назад
Build Your own AI Robot - Ep2 (Things you won't need)
GPT Robot quotes classic literature -
Просмотров 943 месяца назад
GPT Robot quotes classic literature -
Build your own AI Robot - Ep1 (Overview)
Просмотров 2 тыс.3 месяца назад
Build your own AI Robot - Ep1 (Overview)
GPT Robot likes the Beatles!
Просмотров 9723 месяца назад
GPT Robot likes the Beatles!
My GPT Robot Tries to Start a Beef - look out Duolingo!
Просмотров 5133 месяца назад
My GPT Robot Tries to Start a Beef - look out Duolingo!
GPT Robot stans R2D2 - the OG!
Просмотров 1013 месяца назад
GPT Robot stans R2D2 - the OG!
My GPT Robot wants an Upgrade!
Просмотров 7613 месяца назад
My GPT Robot wants an Upgrade!
My GPT Robot Discusses his Mecanum Wheels!
Просмотров 2053 месяца назад
My GPT Robot Discusses his Mecanum Wheels!
My GPT Robot gets a Workout!
Просмотров 4303 месяца назад
My GPT Robot gets a Workout!
My GPT robot discusses his cooling fan and optimal CPU temperature
Просмотров 9583 месяца назад
My GPT robot discusses his cooling fan and optimal CPU temperature
My GPT Robot discusses his Origins!
Просмотров 2643 месяца назад
My GPT Robot discusses his Origins!
GPT Robot posts his First Video!
Просмотров 2263 месяца назад
GPT Robot posts his First Video!

Комментарии

  • @marykearns5484
    @marykearns5484 3 дня назад

    amazing!

    • @LarrysWorkbench
      @LarrysWorkbench 3 дня назад

      Thank you! Let me know what you’d like to see next

  • @arkana.m5541
    @arkana.m5541 3 дня назад

    I will support you till you get famous!

    • @LarrysWorkbench
      @LarrysWorkbench 3 дня назад

      Lol that’s fantastic.. much appreciated! Share it with a friend -

  • @arkana.m5541
    @arkana.m5541 3 дня назад

    Nice!

  • @milacruz3970
    @milacruz3970 3 дня назад

    Really cool!

    • @LarrysWorkbench
      @LarrysWorkbench 3 дня назад

      @milacruz3970 Thank you! Let me know what you'd like to see next -

  • @sonickidd127
    @sonickidd127 4 дня назад

    epic

    • @LarrysWorkbench
      @LarrysWorkbench 3 дня назад

      Thank you! I've been working on this project for about 6 months and the little guy is starting to get pretty frikkin advanced......

  • @Alice8000
    @Alice8000 5 дней назад

    Guys if you stop commenting he'll stop making these videos. I don't think he could milk 11 videos out of this although i've been wrong before

    • @LarrysWorkbench
      @LarrysWorkbench 4 дня назад

      You nailed it! I actually have a full time job that does not involve robots. RUclips does not pay me for this. But I'm learning alot and the comments absolutely make it all worthwhile. I've got quite a few more videos planned, especially as we start taking Floyd out into the real world to start exploring......

  • @Alice8000
    @Alice8000 5 дней назад

    Did you ever get into a tickle contest and lose on purpose?? hehe😉

    • @LarrysWorkbench
      @LarrysWorkbench 4 дня назад

      Floyd's not ticklish, but I don't think he particularly enjoys being picked up......

  • @Alice8000
    @Alice8000 5 дней назад

    Named after Geroge Flloyd! BLM!

    • @LarrysWorkbench
      @LarrysWorkbench 4 дня назад

      George Floyd, Pretty Boy Floyd, and Floyd the Droid -

  • @Alice8000
    @Alice8000 5 дней назад

    very cool tell me more about how you

  • @thenoblerot
    @thenoblerot 6 дней назад

    Great project - I'm doing something similar with Claude Haiku... Maybe going to migrate to Gemini because Google has not only free inference but also free fine-tuning to play with!

    • @LarrysWorkbench
      @LarrysWorkbench 5 дней назад

      Interesting. I don't have much familiarity with any of the models other than OpenAI -

    • @thenoblerot
      @thenoblerot 5 дней назад

      @@LarrysWorkbench I also just learned today Gemini can take audio input natively! (Video too, though it basically sees it at 1 fps) No need for speech to text in the pipeline, and it can interpret tone of voice! I have generally found Google's models to be underwhelming, but free is free, especially for something so frivolous. (I know the API costs are fractions of a cent but I have a mental block about just letting the robot go tearing around the house, cuz oh no, it'll cost 25 cents. Downside, Google authentication is a pain to set up.) Long-term I'm thinking I'll curate robo-Claude's outputs into a dataset to fine tune Gemini. I still use OpenAI a bit, but imho the Claudes are better, for now anyways!

    • @LarrysWorkbench
      @LarrysWorkbench 4 дня назад

      @thenobelrot Wow that's great information! I haven't even thought about migrating, but if OpenAI doesn't give me the 4o advanced voice capabilities by the end of the year I might. I'm still quite a beginner in this space, so until now I've had my hands full just trying to learn one model. What hardware are you working with?

    • @thenoblerot
      @thenoblerot 4 дня назад

      @@LarrysWorkbench I'm really new too. Hadn't coded since the 90s before ChatGPT came out. I've got an old second hand turtlebot2 type platform, running ROS Noetic. I was using a pi 4, but was finding it a little limiting, so recently upgraded to a repurposed motherboard from a busted up laptop.

    • @LarrysWorkbench
      @LarrysWorkbench 4 дня назад

      @thenobelrot I did some light coding in college back in the 90s, but when ChatGPT came out I got inspired. The Raspberry Pi seems ok for what I'm doing so far, but I'm *really* looking forward to OpenAI tightening up the latency on their API calls. Your project sounds interesting do you have any videos anywhere?

  • @Gooby12337
    @Gooby12337 9 дней назад

    I finally got my camera working on my robot lol took me a whole weekend. However, my robot takes a long time to capture the image and respond compared to Floyd. He is very quick. Not really sure what I can do to speed up these responses

    • @LarrysWorkbench
      @LarrysWorkbench 8 дней назад

      Are you working in higher resolution? Floyd's camera is only 620x480 pixels. You could probably use OpenCV to capture a lower res image. When I started uploading one photo per prompt I don't think the overall latency changed much. You could also try pasting your code snippet into ChatGPT and see if it can suggest any optimization...... (I do happen to have high internet bandwidth at the workshop. Idk if that makes any difference. This weekend I'm going to try taking Floyd to a different environment.) Stay on it!

    • @Gooby12337
      @Gooby12337 8 дней назад

      @@LarrysWorkbench Ooo that's a great point. I need to check the resolution I'm currently running. I was trying to use openCV at first, but I was having issues for some reason and then went with capturing the image with libcamera instead and it's working, but just very slow. My bandwidth could for sure be better too.Seems like my robot in general is just slow even when just using speech to text. Might be something with my script. I'll adjust some things and let you know how it goes!

    • @LarrysWorkbench
      @LarrysWorkbench 8 дней назад

      I'm able to run OpenCV as a thread in the background to capture frames from the camera and show them on my computer screen via Virtual Network Computing, and it seems to be able to operate at approximately 15fps. So that's pretty good for what I'm doing -

    • @Gooby12337
      @Gooby12337 8 дней назад

      @@LarrysWorkbench Nice! I'm also using VNC to work on my pi and robot. I figured out the issue with opencv and it has to do with the camera I am using (pi camera module 3) not being integrated yet with opencv, so I'm having to use libcamera to capture images. However, it does seems to be working a little faster after I lowered the resolution. I'm not sure what's going on with OpenAI's API. I feel like it has been extremely slow for me the past several days and I'm good on token limits and credits, so that's not the problem. Has it felt slow at all to you?

    • @LarrysWorkbench
      @LarrysWorkbench 8 дней назад

      @Gooby12337 I've was travelling this weekend so I haven't been online. I probably won't have time to get back into it until friday or saturday :/

  • @arkana.m5541
    @arkana.m5541 11 дней назад

    ill help your carrer by subbing you

  • @PreKGraduate
    @PreKGraduate 11 дней назад

    Very cool, and funny 😅😂

    • @LarrysWorkbench
      @LarrysWorkbench 11 дней назад

      Thank you! This project has been going on for 6 months or so but Floyd is getting more advanced and it's really starting to get kind of weird......

  • @Anetiks
    @Anetiks 12 дней назад

    Last one was dope 👌

    • @LarrysWorkbench
      @LarrysWorkbench 11 дней назад

      Thank you! Floyd is getting more and more advanced. It’s starting to get weird. More videos coming…

  • @arkana.m5541
    @arkana.m5541 12 дней назад

    can you make it faster on responding the conversation.oh yeah i forgot,can you add hand movement

    • @LarrysWorkbench
      @LarrysWorkbench 11 дней назад

      We're woring hard on reducing latency. I think the advanced voice feature of GPT-4o is going to make a big difference. And I'm shooting some more videos that show Floyd rolling around the shop and picking things up, etc....

  • @adrianperez3375
    @adrianperez3375 14 дней назад

    Excellent work. And thanks for sharing the code. Your videos inspired me to use a hiwonder masterpi robot to build a huge 3D printer. I should be publishing a video about it soon. Thank you and keep up the good work.

    • @LarrysWorkbench
      @LarrysWorkbench 14 дней назад

      Wow a 3D printer, I would love to see that! They seem to have lots of interesting products. I've got more videos coming as well especially now that Floyd has a camera to actually see his surroundings -

  • @zaxnate
    @zaxnate 15 дней назад

    *sigh* ill help by subbing you

  • @a.g.vanderhorst4695
    @a.g.vanderhorst4695 15 дней назад

    The robot you created is fantastic. I wonder if using a Raspberry Pi 5 with a SSD HAT and the AI kit would speed up the responses from ChatGPT significantly.

    • @LarrysWorkbench
      @LarrysWorkbench 15 дней назад

      It definitely wouldn't hurt. Maybe you could do speech-to-text / text-to-speech locally. I believe 90% of the latency is waiting for the servers at OpenAI to respond. GPT4o "Advanced Voice" mode is supposed to be WAY faster, if/when they ever release it :/

    • @vdvdg4683
      @vdvdg4683 11 дней назад

      @@LarrysWorkbench I forgot to ask where you bought the robot. It is bloody expensive to buy and ship it from the USA.

    • @LarrysWorkbench
      @LarrysWorkbench 11 дней назад

      The chassis and arm are from a Chinese company called HiWonder - their “Master Pi” series. I like the fact that it’s all metal parts, but I think there’s some cheaper robots on the internet. The microphone and speaker are from Adafruit.

  • @jjose10911
    @jjose10911 18 дней назад

    Bro thats a bunch of metal with an alexa

    • @LarrysWorkbench
      @LarrysWorkbench 18 дней назад

      Actually it's just a bunch of metal connected to GPT-4o, but it's been fun and people seem to like it... Plus Alexa can't pick up objects, recognize my face or pass the bar exam ;)

  • @michalpawilcz4063
    @michalpawilcz4063 19 дней назад

    😞😞😞 I'm trying with the opencv facedetection module. Maybe you need to use tensorflow? I have no more ideas😞

  • @michalpawilcz4063
    @michalpawilcz4063 19 дней назад

    incredible❤

  • @michalpawilcz4063
    @michalpawilcz4063 19 дней назад

    Hi Larry, how did you manage to do this? give me some guidelines because I have no idea how you did it so that Floyd can see. I have a spider from adeept called sue RaspClaws and for now I am at the stage of issuing voice traffic commands to gpt. I want to make an assistant that will "see" the interlocutor and keep an eye on the interlocutor: gpt is to manage the robot's movements in such a way that the camera follows the face all the time. i.e. the robot's movements forward, backward, left and right. the camera also moves up and down left and right. I'm also self-taught, in September of that year I started my adventure with Raspberry Pi. You are amazing. please help me with spider

    • @LarrysWorkbench
      @LarrysWorkbench 19 дней назад

      Lol I got the camera hooked up last weekend, more videos coming on that. I use OpenCV to capture images about 15 FPS and display on a monitor over VNC (virtual network computing). Then whenever I send a prompt to GPT-4o, I just send the most recent image capture along with it. That way GPT gets to "see" from Floyd's perspective, and comment if it feels appropriate. GPT-4o is "multi-modal", so it can accept mixed inputs. I'm SHOCKED at how much information it can extract. It guessed my age within a few years. It was able to read the writing on my sweatshirt. All with basically zero programming. I'll try to upload my latest code rev up to Github (/LarrysWorkbench) (Note that I don't have GPT controlling the robot's movements, yet.....)

    • @Gooby12337
      @Gooby12337 18 дней назад

      @@LarrysWorkbench Thanks for this comment. I've been having trouble getting my camera to function with my robot. I'm using gpt-4o API calls just like you are and attempting to use openCV to capture the images, but my script keeps "failing to capture frame". Do you always have the camera running, or do you have it activate and capture an image when using a prompt or specific phrase? This stuff for sure isn't easy haha love your videos!

    • @LarrysWorkbench
      @LarrysWorkbench 18 дней назад

      @Gooby12337 Great question! Right now I have two separate functions. One which triggers to capture just one image and attach it any time I send a prompt (question) up to GPT. The other one runs as a separate thread in the background and displays the camera video on a remote monitor, so I can "see" what Floyd is seeing at any given time. It took me like two entire weekends to get it all working. Here's my current capture function. ChatGPT wrote it for me; I don't really understand all the syntax - def capture_image_and_encode(): # Captures a single image from the camera camera = cv2.VideoCapture('127.0.0.1:8080?action=stream') ret, image = camera.read() if not ret: print("Failed to capture image.") return None print("Snapshot image captured...") _, buffer = cv2.imencode('.jpg', image) # Encode image in JPEG format return base64.b64encode(buffer).decode('utf-8') # Convert to base64

    • @Gooby12337
      @Gooby12337 18 дней назад

      @@LarrysWorkbench Very cool! Thank you for the information! That's a great idea to have a separate function running the video feed so you can see what your robot is seeing in real time. I'll have to incorporate that into my project. And spending two full weekends trying to get it to work sounds about right haha definitely a ton of debugging and troubleshooting that goes into this

    • @LarrysWorkbench
      @LarrysWorkbench 18 дней назад

      @Gooby12337 Yeah it's not easy. But that's what makes it so satisfying when it actually works. Stay with it! And keep me updated on your project -

  • @Gooby12337
    @Gooby12337 24 дня назад

    So cool! I'm doing a similar project and having a ton of fun with it. What camera did you decide to go with?

    • @LarrysWorkbench
      @LarrysWorkbench 24 дня назад

      The camera came with the original robot chassis from HiWonder, it’s 70 degrees field of view, 640x420 pixels. Not great resolution, I’ll probably think about upgrading at some point. *Hopefully* it wouldn’t bog down my processing time or add latency. GPT is amazing at the amount of information it can extract even from a low-res image

    • @Gooby12337
      @Gooby12337 24 дня назад

      @@LarrysWorkbench Oh nice! I think I'm going to go with the pi camera module 3. Exactly, I'm hoping it won't bog down my processing time either. That is very impressive how it can extract so much info. I'm excited to add camera functionality!

    • @LarrysWorkbench
      @LarrysWorkbench 24 дня назад

      GPT easily recognizes objects, can easily recognize itself in a mirror, etc. It was able to estimate my age within a range based on one grainy image. It was able to read the writing on the front of my sweatshirt. When I put on a sport coat it said I was looking sharp. It’s wild. Let me know how your project goes

  • @zaxnate
    @zaxnate 24 дня назад

    dont forget to use hastag to help you

    • @LarrysWorkbench
      @LarrysWorkbench 24 дня назад

      I’m new to RUclips - is it better to put them in the title or the description?

    • @zaxnate
      @zaxnate 23 дня назад

      @@LarrysWorkbench description

    • @zaxnate
      @zaxnate 23 дня назад

      You need to describe what you gonna do and how does it work. After that you can add lots of hashtags

    • @zaxnate
      @zaxnate 23 дня назад

      Good luck,im counting on you

    • @LarrysWorkbench
      @LarrysWorkbench 23 дня назад

      Thank you, much appreciated! Floyd needs to be seen -

  • @davelandry646
    @davelandry646 24 дня назад

    Thank you for posting your script! I'm going to try to get my RPi to talk back to me!! Great job with this series.

    • @LarrysWorkbench
      @LarrysWorkbench 24 дня назад

      Thank you! ChatGPT helped me alot with writing the Python syntax. Let me know how your project goes -

    • @davelandry646
      @davelandry646 24 дня назад

      @@LarrysWorkbench Will do...

    • @davelandry646
      @davelandry646 24 дня назад

      @@LarrysWorkbench Got Floyd talking to me! Your suggestion to have ChatGPT help me did the trick! I used parts of your code and ChatGPT showed me how to install the various modules I needed. Much fun indeed!

    • @LarrysWorkbench
      @LarrysWorkbench 24 дня назад

      Wow that's fantastic!

  • @Jerseylance1
    @Jerseylance1 28 дней назад

    I want my robot to talk back to me like this..

    • @LarrysWorkbench
      @LarrysWorkbench 27 дней назад

      You can.. It’s a Raspberry Pi processor with a mic and speaker, making API calls to OpenAI. I taught myself Python just so I could do this project. Check out my instruction videos

  • @Jerseylance1
    @Jerseylance1 28 дней назад

    Nice🔥🔥🔥🔥🔥

  • @RoboDepot.onXtwitter
    @RoboDepot.onXtwitter 28 дней назад

    Is there no way to guide it to an object without telling it exactly how to move the arm?

    • @LarrysWorkbench
      @LarrysWorkbench 28 дней назад

      Great question! So far I've put all my work effort into speech conversation and personality. But just recently I started working on his vision routines as well (camera & sonar). Right now he's still completely blind, but I'm hoping that'll change within the next month or so -

  • @aboudezoa
    @aboudezoa Месяц назад

    So Floyd constantly sending APis ? How much does it cost on average

    • @LarrysWorkbench
      @LarrysWorkbench Месяц назад

      That's a great question! I probably have him running his conversation routines about 20 hours per month, at an average of maybe 20 cents per hour. It looks like about half the spend is on audio and half on GPT 4o

    • @aboudezoa
      @aboudezoa Месяц назад

      @@LarrysWorkbench very impressive project Larry

    • @LarrysWorkbench
      @LarrysWorkbench Месяц назад

      Thank you! It's been fun and I've definitely learned alot. Working on getting vision up and running 0-0

  • @Archersforeal
    @Archersforeal Месяц назад

    Something tells me he already has the emotions installed 😂

    • @LarrysWorkbench
      @LarrysWorkbench Месяц назад

      Lol I'm working on it, but not *completely* sure it's a good idea :/

  • @colleenmcferrin7661
    @colleenmcferrin7661 Месяц назад

    Hey Larry, great to hear even the computers can’t figure it out… Lol!!!

    • @LarrysWorkbench
      @LarrysWorkbench Месяц назад

      Lol right? He says it's pretty much like trying to figure out a soap opera. And he's kind of right -

  • @TheClothingCoach
    @TheClothingCoach Месяц назад

    😂😂😂

    • @LarrysWorkbench
      @LarrysWorkbench Месяц назад

      I wanted to see if he had any comment on behalf of the robot community. Floyd's got an opinion on everything!

  • @AnnLi-s6n
    @AnnLi-s6n Месяц назад

    Hi Larry, this is brilliant! Thanks for sharing it, I've been looking for a similar DIY project guide for a while and you're the best that I've found on RUclips! I would be very interested to know more of the hardware setup, like how do you connect the parts up, and how the code could direct motions of the robot. I know you mentioned it briefly in your previous videos, but as a beginner it would just be super nice if I could know more details! One little thing if it could help: it would be amazing if you could list the key points/materials used in text in your video description, so that people can keep note more easily, as a lot of youtubers would do. Video description section is also a great place that you can paste links to your Github account so that ppl can go there and star/like your work (although I just checked it out there seems I'd wait more for things to happen there :) The delay in Floyd's response I guess is a network problem? As I don't exprience delay when chatting with ChatGPT on my phone. Do you plan to optimize this? Or maybe other LLM models would work better? I've heard great reviews on Claude. All in all, amazing idea and creative work! Thanks for introducing Floyd and would love to see him pick up more capabilities along the way!

    • @LarrysWorkbench
      @LarrysWorkbench Месяц назад

      @AnnLi-s6n Thank you so much for your response! I'm a tech beginner also - I started this whole thing around March and it's been taking up most of my nights and weekends. My first videos in the how-to series detail the hardware specifically. Hiwonder had some instruction videos that were very helpful regarding assembly. And the Raspberry Pi image came with a variety of Python libraries that allow programatic control of the motors, servos etc. But there was also A LOT of trial and error 0-0 As far as conversational latency, I haven't clocked it but I think I'm in line with what the GPT audio models can do on the phone. OpenAI has said that they're using the same series of three individual models as I am (Speech-to-Text, GPT, Text-to-Speech). Of those, the TTS seems to be the slowest by far, and I'm confident that they're eventually going to speed it up on their end with 4o. But I'm also trying to figure out how to stream the response rather than waiting until it loads 100% Stay in touch and let me know what direction your project goes in!

  • @massedin
    @massedin Месяц назад

    Hey!! I’m kinda stuck, apparently they updated some commands in the OpenAi library and because of that i’m getting crazy!! I couldn’t find you github, can send here the link??

    • @LarrysWorkbench
      @LarrysWorkbench Месяц назад

      @massedin Yeah I think the GPT library got updated a few months ago, I remember having trouble with it as well. Here's my github link github.com/LarrysWorkbench My API call syntax seems to work FWIW Lmk what happens

    • @massedin
      @massedin Месяц назад

      thx for the link! i’m just focusing in the chat part (without the motors, lights and sound effects) but I just can’t make it work haha what library’s you had to install? and how? like pip install pylapse etc…

    • @massedin
      @massedin Месяц назад

      and what documentation did you use?

    • @LarrysWorkbench
      @LarrysWorkbench Месяц назад

      @massedin check the code - I mainly installed then imported the OpenAI library: from openai import OpenAI I used alot of ChatGPT to answer my questions and help me. But it was also a huge amount of trial and error. I actually walk through the code for my first API call in Episode 7 of the video series - Stay with it - you'll get it!

    • @massedin
      @massedin Месяц назад

      @@LarrysWorkbench Thx! After some time it actually worked, but just the part where we speak and he replies. This is already awesome, but now we are stuck on the board. It was hard just to understand that sys.path where you import the ‘Board’

  • @angelnicolasnogalesguerra2948
    @angelnicolasnogalesguerra2948 Месяц назад

    HELLOOOO! 😄 MY NAME IS ANGEL, I'M FROM BOLIVIA, SOUTH AMERICA. 🌎 A FEW MONTHS AGO, I HAD A SIMILAR IDEA TO YOURS AND WATCHING YOUR VIDEO INSPIRED ME TO RETAKE THE PROJECT. 🚀 I'M A ROBOTICS TEACHER, AND I WANT TO CREATE A SIMILAR ROBOT SO MY STUDENTS CAN SOLVE THEIR QUESTIONS ABOUT THE SUBJECT, ALMOST LIKE AN ASSISTANT. 🤖 CONGRATULATIONS, YOUR PROJECT LOOKS AMAZING! 🎉 I’LL BE KEEPING AN EYE OUT FOR FURTHER UPDATES. 👀

    • @LarrysWorkbench
      @LarrysWorkbench Месяц назад

      That's fantastic, thank you for commenting!| I originally designed Floyd as a kid's project. I realized that you can do really interesting stuff if you can connect the Raspberry Pi to external APIs - the robot could be a great assistant. You can check out my source code at GitHub (LarrysWorkbench). Please let me know how your project progresses!

  • @Jerseylance1
    @Jerseylance1 Месяц назад

    🔥🔥🔥🔥🔥✌️

  • @timwallis-du6xg
    @timwallis-du6xg Месяц назад

    I have been eager to do this project with my son this summer. I hope the next video comes soon :-)

    • @LarrysWorkbench
      @LarrysWorkbench Месяц назад

      Thank you so much - that's absolutely great to hear! This whole project was originally designed to be kid-friendly, although it seems like it evolved a little bit. I don't make any money doing this, but comments like yours make it absolutely worthwhile. I've posted my source code at github.com/larrysworkbench. Keep in touch and let me know how your project goes!

  • @timwallis-du6xg
    @timwallis-du6xg Месяц назад

    Thanks very much! I always look forward to your next video

    • @LarrysWorkbench
      @LarrysWorkbench Месяц назад

      Fantastic! I'm trying to post about one a week. Let me know what you'd like to see in particular

  • @massedin
    @massedin Месяц назад

    Hey there, suuper cool ideia!! I´m from Brazil and some years ago me and my father worked on an project where we created a bluetooth controled car using Arduino, it kinda look like yours. Your video gave us some ideas, we will try something like your project. Thanks for the videos, keep going!

    • @LarrysWorkbench
      @LarrysWorkbench Месяц назад

      That's great! Let me know how your project goes -

  • @michalpawilcz4063
    @michalpawilcz4063 2 месяца назад

    Hello, I have an Adeept RaspClaws robot. I'm trying to make an interactive hexapod but the gpt chat tries to move but nothing moves. Can you share the commands.json structure of the file?

    • @LarrysWorkbench
      @LarrysWorkbench Месяц назад

      I don’t know the exact JSON structure but the OpenAI API docs are pretty good and I’ve posted a version of my Python code at github.com/LarrysWorkbench

    • @michalpawilcz4063
      @michalpawilcz4063 Месяц назад

      Hello Larry, I thought it would be possible to add object recognition by importing, for example, tensorflow object detection. Floyd could use the camera to recognize various objects in the environment and interpret their location. What do you think about it?

    • @LarrysWorkbench
      @LarrysWorkbench Месяц назад

      Sounds cool. I'm really excited about incorporating vision, but I know there's going to be some learning curve. My plan was to use OpenCV for some of the recognition, but I also want to pass still images (or video?) up to OpenAI so the cloud portion of Floyd's brain can describe them or even interact with what it's seeing -

    • @michalpawilcz4063
      @michalpawilcz4063 Месяц назад

      @@LarrysWorkbench super-cool👍💪❤️

  • @sourabhpandey2004
    @sourabhpandey2004 2 месяца назад

    Floyd got rizz Really fascinating !!

  • @Dufftata
    @Dufftata 2 месяца назад

    cable management looks good actually tho ;D I see your mic solution, I am still working on my overly-complicated onboard mic solution

    • @LarrysWorkbench
      @LarrysWorkbench 2 месяца назад

      @Dufftata yeah it was cheap from Adafruit.com but Idk what the shipping would be. The *only* problem I have with the Whisper API transcription is if I try speaking from more than 4-5 feet distance (lol 1.5 meters). So I'm also considering upgrading, but I'd want it to be USB plug-and-play, which this one was

    • @LarrysWorkbench
      @LarrysWorkbench Месяц назад

      @Duftata By the way - feel free to amplify the project to your subscribers if you want ;)

  • @jamesmoon5632
    @jamesmoon5632 2 месяца назад

    Wow great video love where this is going

    • @LarrysWorkbench
      @LarrysWorkbench 2 месяца назад

      Thanks so much! The project is defenitely gaining some momentum -

  • @HelamanGile
    @HelamanGile 2 месяца назад

    interesting

    • @LarrysWorkbench
      @LarrysWorkbench 2 месяца назад

      Thank you.. check out some of Floyd’s other videos

  • @ShivamMallick-kq7bd
    @ShivamMallick-kq7bd 2 месяца назад

    How can i contact you? For some doubts

    • @LarrysWorkbench
      @LarrysWorkbench 2 месяца назад

      Thanks for reaching out. The easiest way is to leave me messages here -

  • @jamesmoon5632
    @jamesmoon5632 2 месяца назад

    Love the videos man wanna make my own talking robot.

  • @Dufftata
    @Dufftata 2 месяца назад

    thank you for inspiring! I've noticed that "personality" thing is really key to make the conversation entertaining (none-flat). Don't have that polishes just yet, maybe how to use different modes? Don't have a good idea how to build this so that is it easy to navigate for the user. I've connected a power-bank to the raspberry pi now, to make the thing portable, and it seems to run for quite a while (however it is just the speech stuff, no motors just yet). I still have to work on the transcription precision, cause I noticed that what I am saying and what gpt4o receives as a text can differ too often by too much. Love your stuff!!!! Waiting for the next video(s)!

    • @LarrysWorkbench
      @LarrysWorkbench 2 месяца назад

      @Dufftata Thx so much for the support! I just posted my code on Github / LarrysWorkbench feel free to check it out. It seems like "Personality" is what turns GPT into "Floyd" and makes the project fun. Floyd's Personality comes from about 12 plaintext sentences. I think of it as his Long-Term Memory. The most important instruction is for him to keep his responses BRIEF, two sentences max. Regarding transcription - maybe check out your hardware. I feel like the Whisper API does a really good job when I give it a decent audio file to work with....

  • @bramvanweelderen9233
    @bramvanweelderen9233 2 месяца назад

    Heyy Where van I find how to do this do you have Any other social media

    • @LarrysWorkbench
      @LarrysWorkbench 2 месяца назад

      Check out my series of video tutorials on RUclips where I'm basically going through the entire project week by week. I don't have any other accounts, but considering posting the code on Github

  • @sourabhpandey2004
    @sourabhpandey2004 2 месяца назад

    Indian Student here great work man really really liked it !! for your work your earned a sub there ;)

  • @Dufftata
    @Dufftata 2 месяца назад

    good morning Larry, got my early morning coffee ready, and your upload timing is perfect, cause I am literally setting next to my "robot", (first ever raspberry pi experience also), have the vnc set up, and next need to solve the python part (I am a programmer, so debugging it should be easy (famous last words), however I hope to have gpt4o do the full thing, rather than also using eg elevenlabs to get the voice going, gonna watch your video now! Thanks for doing this, very fun

    • @LarrysWorkbench
      @LarrysWorkbench 2 месяца назад

      That’s awesome! Thonny came already installed on my Raspberry Pi but I think you can use Visual Studio Code as an option. I did struggle a little bit to get the “Board” library installed. It accesses memory and the GPIO pins so I have to run it as super user. Next week I’ll post our first API call to OpenAI. Floyd’s code is actually three different ones (transcription, chat, and text-to-speech), with sound effects to distract from the latency. Have fun!