Making Chat (ro)Bots

Поделиться
HTML-код
  • Опубликовано: 25 окт 2023
  • We created a robot tour guide using Spot integrated with Chat GPT and other AI models as a proof of concept for the robotics applications of foundational models. Learn more: bostondynamics.com/blog/robot...
    Project Team:
    Matt Klingensmith
    Michael McDonald
    Radhika Agrawal
    Chris Allum
    Rosalind Shinkle
    #BostonDynamics #chatgpt
    00:00: Introduction
    01:08: Making Chat (ro)Bots
    01:39: Precious Metal Cowgirl
    02:38: A Robot Tour Guide
    03:07: How does it work?
    04:19: Shakespearean Time Traveler
    04:36: Creating Personalities
    04:54: "Josh"
    05:23: Lateral Thinking
    06:03: Teenage Robot
    06:43: Nature Documentary
    07:32: What's next?
  • НаукаНаука

Комментарии • 2,1 тыс.

  • @balakrishnanr7268
    @balakrishnanr7268 6 месяцев назад +3293

    Adding sound made it 10x more futuristic

    • @jakeparker918
      @jakeparker918 6 месяцев назад +258

      And the hat made it 100x

    • @crazycatboysolomon7006
      @crazycatboysolomon7006 6 месяцев назад

      Don't forget the googly eyes!@@jakeparker918

    • @jet100a
      @jet100a 6 месяцев назад +110

      Honestly just having a fully realistic voice really shocked me, feels like we are actually living in the future!

    • @Tomartyr
      @Tomartyr 6 месяцев назад

      ​@@WwZa7cope 🫖

    • @icykenny92
      @icykenny92 6 месяцев назад +21

      @@jet100a No, we're living in the present.

  • @encyclopath
    @encyclopath 6 месяцев назад +1863

    The googly eyes do the equivalent work of about a thousand man hours of engineering and coding

    • @0x0michael
      @0x0michael 6 месяцев назад +26

      Apple's EyeVision

    • @DemPilafian
      @DemPilafian 6 месяцев назад +48

      A manikin with pretty makeup and a few primitive solenoids for motion will be perceived by the general public to be more intelligent than a supercomputer calculating molecular interactions of a new drug. Scientists would get more respect if they just glued plastic manikin heads on top of their supercomputers.

    • @corbindedecker7658
      @corbindedecker7658 6 месяцев назад +21

      ​@@DemPilafian they would have to put googly eyes on the manikin head though.

    • @EnviedShadow
      @EnviedShadow 6 месяцев назад +1

      @@corbindedecker7658 mannequin*
      I usually wouldn't care enough to correct that, but seeing two people in a row misspell it bothered me, sorry

    • @JorjEade
      @JorjEade 6 месяцев назад

      Googly-eyed manaquin heads will be the next scientific revolution after quantum computing

  • @epg-6
    @epg-6 6 месяцев назад +768

    Spot saying, "follow me gentlemen" over its shoulder while walking to the rock pile was the most natural thing in this video. Almost felt like it was alive.

    • @user-ug6hh4qg3n
      @user-ug6hh4qg3n 6 месяцев назад +24

      ikr! its why that voice is my favorite of all these

    • @chugachuga9242
      @chugachuga9242 6 месяцев назад +20

      Sounded like a Skyrim voice line

    • @vickimiller6991
      @vickimiller6991 4 месяца назад +1

      I think this is where Boston Dynamics beats the Tesla robot. Spot is much more relatable.

  • @kamikeserpentail3778
    @kamikeserpentail3778 6 месяцев назад +334

    The Butler one at the beginning reminds me so much of a typical tutorial NPC in so many games I've played.
    Ignores dialogue response to make a big show of walking three feet over.
    Only upon finding the exact position needed to continue the quest line, responds to the previous dialogue.

    • @CrAzYpotpie
      @CrAzYpotpie 6 месяцев назад +4

      Can you give examples? I have never seen a game where you can give a response that is delayed until after a scripted sequence has finished ever.

    • @Drakonus_
      @Drakonus_ 6 месяцев назад +30

      ​@@CrAzYpotpieSkyrim.

    • @davidpacheco5501
      @davidpacheco5501 6 месяцев назад +11

      Even the movement mannerisms (especially the mouth)

    • @JayCantwell
      @JayCantwell 3 месяца назад +3

      @@CrAzYpotpie literally exactly fallout 4 codsworth

  • @FroddeB
    @FroddeB 6 месяцев назад +3188

    The first british voice was so convincing. The robotness is at a point where you realize it's a robot, but you're not annoyed by the voice. It really feels like the robot is willingly talking to you.

    • @trashcanmucous5153
      @trashcanmucous5153 6 месяцев назад +21

      why dont the robots have receptors and a nervous system? then they can bypass neural web learning

    • @ps0705
      @ps0705 6 месяцев назад +20

      I thought it was Fraser Crane at first. I just realised I'd love to have a robot with that voice!

    • @vargoJI
      @vargoJI 6 месяцев назад +2

      Totally agree)

    • @Naptosis
      @Naptosis 6 месяцев назад +77

      @@trashcanmucous5153 Yeah, and why aren't they just living flesh over a metal endoskeleton!? That's such a good idea with no downsides, and so easy to achieve!

    • @juhotuho10
      @juhotuho10 6 месяцев назад +38

      @@trashcanmucous5153 you have no idea how machine learning works

  • @Slashman510
    @Slashman510 6 месяцев назад +1199

    Man! You guys should have assigned a tourist personality to one Spot and have it be toured around Boston Dynamics by a tour guide Spot!!! I bet that would be entertaining!

    • @MadComputerScientist
      @MadComputerScientist 6 месяцев назад +103

      That would be adorable! And hear the robo-banter.

    • @joannamariaaldred
      @joannamariaaldred 6 месяцев назад +38

      I like that idea, it would be fascinating to see how they interact, maybe bring the archeologist Spot along too.

    • @chpsilva
      @chpsilva 6 месяцев назад +50

      "... and here we have a plank of wood with some fake valves and buttons used to demonstrate our dextery... *SIR!* Please, dont turn the valve, thank you" /tourist robot retreats a little bit ashamed

    • @mrbillgoode
      @mrbillgoode 6 месяцев назад +3

      iDiOT! Robot to robot interaction will spur the exact dynamic algorithmic modification that could make us lose control over these things. Robots figuring out how to learn or communicate with each other without human input is very much what will hasten human extinction.

    • @crowe6961
      @crowe6961 6 месяцев назад +6

      @@mrbillgoode This can be tested in a controlled environment first.

  • @SidSparnos_AIEnthusiast
    @SidSparnos_AIEnthusiast 6 месяцев назад +85

    0:39 this was the FUNNIEST part of this video... It first introduced the rocks, then looked as if it ignored the compliment as it moved on to the next point, the looked as if ignoring the compliment would be rude so responded to the compliment🤣🤣🤣... Basically the software version of an awkward interaction 😂

  • @itbeWOLFLINGS
    @itbeWOLFLINGS 6 месяцев назад +300

    I am relieved to see that the engineers are being nice and polite to the bots

    • @Mallchad
      @Mallchad 6 месяцев назад +37

      Spot probably got mad after being kicked so much lmao

    • @ninjacat230
      @ninjacat230 3 месяца назад +8

      they're past the "beat the crap out of the robots" phase

    • @snowbelltheicewing66
      @snowbelltheicewing66 Месяц назад +3

      As they should

  • @Victor-xj4cv
    @Victor-xj4cv 6 месяцев назад +1060

    This reminds me of the scene in Interstellar, where Cooper adjusts the TARS robot's humor/sarcasm setting down from 100% to 75%. Amazing that in a few short years, we basically have that technology.

    • @dibbidydoo4318
      @dibbidydoo4318 6 месяцев назад +9

      How do you quantify sarcasm? Is it decreasing the probability it will uses sarcastic sentences?

    • @think2086
      @think2086 6 месяцев назад

      It doesn't actually matter. You just ask it to do it and it does. The details don't matter, in some sense, and that's the beauty and point of having A.I. in the first place. If you don't like the way it did it, you can simply have a conversation with it about it and get it to adjust. It's implied in Interstellar that by then, the A.I. is very polished, compared to what we have now, and so a simple request like "75%" without further specification is fine. Further elaboration is unnecessary: the A.I. will intelligently do it. But two additional points: Cooper already had some familiarity with the A.I. and roughly knew what "75%" sarcasm would mean from his aeronautics days. But also, he made a further adjustment when the first adjustment wasn't suiting him quite right. Never did he have to worry about the details. He just adjusted the number whenever needed, and the A.I. did the work of figuring out what that practically meant. I think that's one of the many small detailed things in that movie that were brilliantly packed with so much implication. @@dibbidydoo4318

    • @amogus3023
      @amogus3023 6 месяцев назад +91

      @@dibbidydoo4318 that's the cool part about AI, you don't necessarily need to quantify it (though doing so would improve accuracy). The AI model is aware of the concept of sarcasm well enough indirectly that you can use even vague descriptions to control it.

    • @dirtydicso
      @dirtydicso 6 месяцев назад +43

      Great point. In 2014 this is what we imagined ai might look like in 50 years, and it took us less than 10.

    • @ianmeade7441
      @ianmeade7441 6 месяцев назад +5

      ​@dibbidydoo4318 I think so. 100% could prompt it to have all replies be sarcastic to the extent that they don't interfere with greater communication objectives, or cross some arbitrary threshold of sarcasm tolerence it internally models for humans. 75% could just have it adjust the information or tolerence threshold by 25%. But like you asked, how do we quantify these things?
      Given what we know of their difficulties with math, the language model itself won't be trying to comprehend these percentages-- it will just give outputs as normal, labeling them with emotion tokens. A simple internal program, however, would be very good at math, and could just pick out any token of "sarcasm" and have a 1-in-whatever chance of prompting the LLM to replace its associated output with one of a new token. The context of interaction could skew the pool of availible alternative tokens to favor some more than others, so you don't end up randomly trading sarcasm for anger in an otherwise friendly conversation. This also helps define where sarcasm might interfere with information sharing objectives.
      Of course, there's way more intricacy in how you could define emotion tokens, and how you could include ones other than emotion in these operations.

  • @MisterItchy
    @MisterItchy 6 месяцев назад +1780

    I think it might be fun to have an RF chip or something in the hat so, if you want to change personality, just pop on a different hat.

    • @warpig6459
      @warpig6459 6 месяцев назад +122

      Genius

    • @jwadaow
      @jwadaow 6 месяцев назад +149

      ​@@warpig6459he wore his thinking cap.

    • @ryanjohnson3615
      @ryanjohnson3615 6 месяцев назад

      You could put a MAGA hat on it and it would be dumb and violent and believe anything!

    • @sciteceng2hedz358
      @sciteceng2hedz358 6 месяцев назад +43

      Why an RF chip? It should recognize the hat.

    • @rudrecciah
      @rudrecciah 6 месяцев назад

      @@sciteceng2hedz358 computer vision is less reliable than physics

  • @FortifiedFilmsAndFigures
    @FortifiedFilmsAndFigures 6 месяцев назад +89

    i fucking love the personality where they just were like 'ight, what do do we call it?'
    random co-worker: 'Josh'
    entire team: *slams hand on meeting table* 'you sonnuva bitch, lets make it happen'

    • @erbsenkaffee8720
      @erbsenkaffee8720 5 месяцев назад +4

      I like to imagine 'Josh' is a teammember with a good sense of humor

    • @istoleurfaceha3527
      @istoleurfaceha3527 3 месяца назад +1

      @@erbsenkaffee8720exactly what I thought

  • @dannyellis971
    @dannyellis971 5 месяцев назад +26

    2015 Boston Dynamics: Everybody will have robots in 10 years.
    2023 Boston Dynamics: We are still working on making them dance better for the past 5 years.

    • @christianmoore7109
      @christianmoore7109 3 месяца назад +4

      To be fair lots of businesses have them now! That’s the bulk of their videos

  • @adto5942
    @adto5942 6 месяцев назад +550

    A british dogbot? I want one!
    But honestly, Spot as a tour guide us one of the best PR move I have ever seen.

    • @jonathanhucke
      @jonathanhucke 6 месяцев назад +11

      Just waiting for an Australian one so I can name it Wilfred.

    • @MrMikkyn
      @MrMikkyn 6 месяцев назад

      me too

    • @dream8870
      @dream8870 6 месяцев назад +2

      $75k and it’s yours

    • @Apistevist
      @Apistevist 6 месяцев назад

      It'll probably break even and pass having 3 teenagers on a cellphone all day making 18 bucks an hour hired.@@dream8870

    • @skippersthepenguin3591
      @skippersthepenguin3591 6 месяцев назад

      Just change the prompt haha literally say, your an australian and itll do it. Of course you need a spot haha @@jonathanhucke

  • @jurajchobot
    @jurajchobot 6 месяцев назад +275

    02:51 "This is where we keep our robot, who can run up to 19mph.... I hope you're not too slow to keep up with it." It begins 😂

    • @swissretrogamer52
      @swissretrogamer52 6 месяцев назад +65

      I just realized that this could have also been a threat

    • @DemPilafian
      @DemPilafian 6 месяцев назад

      I hope the internals of the robot are not coded using antiquated imperial units.

    • @antonliakhovitch8306
      @antonliakhovitch8306 6 месяцев назад +4

      ​@@DemPilafian
      Which internals?
      - The language model is just a language model. It can't calculate anything, it just generates text. If you tell it that it's a tour guide in the US, it will probably use imperial units when talking to you.
      - For everything else, it doesn't really matter which units are used. It's a machine, it can do math. If they do use the imperial system, my guess is that everything is in units of mils (thousandths of an inch) and nothing is ever internally converted to inches, yards, miles, or anything else.
      You only really need to care about units when displaying the data to a human.

    • @asphaltpilgrim
      @asphaltpilgrim 6 месяцев назад +2

      It took me longer than I thought to find an AI apok comment on this thread.

    • @DemPilafian
      @DemPilafian 6 месяцев назад

      @@antonliakhovitch8306 The chat LLM is just a gimmick slapped on top of the robot. I 100% guarantee you that the internals of the robot are *NOT* powered by an LLM.
      _"it doesn't really matter which units are used"_
      Wrong. I'm not a customer. I'm a software developer, and I'm interested in the internals and how things work. (Also, see: NASA Mars Climate Orbiter)

  • @leokimvideo
    @leokimvideo 6 месяцев назад +171

    Next we will see Spot doing Old Spice commercials with charm like that

  • @ethanmuhlestein8187
    @ethanmuhlestein8187 6 месяцев назад +15

    The sarcastic personality was so good

  • @rolf-smit
    @rolf-smit 6 месяцев назад +179

    What I like about Boston Dynamics, and something I hope they don't lose: They seem to be very honest and genuinely excited about their technology. No fancy marketing speech like Apple, with all bullsh*t marketing terms for technology that is not new at all, like a dystopian disease. Instead enthusiastic engineers talking about their work, and funny videos showcasing the abilities of the robots produced. It's just great, please keep this altitude.

    • @SamuelMM_Mitosis
      @SamuelMM_Mitosis 6 месяцев назад +4

      Except they gave no credit to the LLM they used

    • @fagelhd
      @fagelhd 6 месяцев назад +25

      @@SamuelMM_Mitosis Didn't they say they used GPT4?

    • @pearce05
      @pearce05 6 месяцев назад +15

      ​@fagelhd yes, they did. They go into even more detail in their blog post, as well.

    • @SamuelMM_Mitosis
      @SamuelMM_Mitosis 6 месяцев назад +2

      @@pearce05 ooh I didn’t read the blog post. That’s good to know they said it there. I think they still should have made that more clear in the video

    • @tomallo99
      @tomallo99 6 месяцев назад +7

      Damn, now you made me realize... That, so far, every freakin time, when a company starts out like this, doing insanely cool stuff... It ends up being a huge, dystopian money grab. I have to cherish every video from them that's still in this good spirit, because for sure we're just a couple of years from this becoming exactly what you described.

  • @daniel-wood
    @daniel-wood 6 месяцев назад +804

    I feel like, at some point over the last few years, the science fiction future arrived--and, somehow, no one seems to be excited about it

    • @mmmarshalruska
      @mmmarshalruska 6 месяцев назад +15

      Да просто плохо нам всем будет скоро без работы)

    • @Timic83tc
      @Timic83tc 6 месяцев назад +37

      becauase its evil.

    • @marshallmcluhan33
      @marshallmcluhan33 6 месяцев назад +32

      greed trumps progress

    • @curtisstephens4482
      @curtisstephens4482 6 месяцев назад +83

      Because we actually watched science fiction movies, it usually doesnt end well.

    • @r.m8146
      @r.m8146 6 месяцев назад +47

      I am.

  • @wildhogOW
    @wildhogOW 6 месяцев назад +15

    Boston Dynamics for the longest of times: "IT'S AN ARM, NOT A HEAD!"
    Boston Dynamics 2023: "Fine, it's a head, you guys win"

  • @sh0gun98
    @sh0gun98 6 месяцев назад +25

    5:18 In a room devoid of joy, much like my soul. - Same

  • @andrewx8888
    @andrewx8888 6 месяцев назад +141

    Josh and British personality are 10/10. Please keep both for the future.

    • @Xune2000
      @Xune2000 6 месяцев назад +2

      Give Josh a British accent and you've got yourself a robot David Mitchel!

  • @Sc077ish
    @Sc077ish 6 месяцев назад +315

    The potential as 'guide dogs' for the Blind and visually impaired is huge here. Hell, I'd love a Spot bot to navigate cities etc, you ask it directions, it shows you the way. Such a fantastic amount of potential for the future.

    • @davebowman760
      @davebowman760 6 месяцев назад +14

      But they'll be easily stolen, especially if you're visually impaired you'll not be able to do much when two guys take it and run away

    • @grek5261
      @grek5261 6 месяцев назад

      @@davebowman760 bro there are chips for tracking that's kind of things like electric scooter has, we are in 2023

    • @santosic
      @santosic 6 месяцев назад +20

      @@davebowman760 they'd have to install some safeguard on it for sure, or ways to get it back if it does get stolen. Or perhaps a setting that causes it to let out a shrilling beep if it was moved away from its owner suddenly, to draw in the attention of everyone around. Like a pseudo panic button basically.
      It wouldn't stop the most dedicated, hardcore criminals, but it'd at least deter some of the ones that like to avoid unnecessary risk and attention.

    • @mr_slidey
      @mr_slidey 6 месяцев назад +12

      @@davebowman760 it can act as a guard dog and bite them or something, maybe it could spray pepper spray at them

    • @dream8870
      @dream8870 6 месяцев назад +8

      @@mr_slideywell that could potentially be a bad thing, robots are a lot stronger than we are, even spots clamp could crush bones with enough force applied to it.

  • @4xhot
    @4xhot 6 месяцев назад +120

    I liked the “nature documentary” personality best. It felt exactly like what I would expect from a futuristic robot in the movies; Precise, yet smooth and cool under pressure. It felt sophisticated like the British one, but I feel like it showcased the robot portion of the personality better. Love it! Can’t wait to see what comes next! ❤❤🎉

  • @trust37_
    @trust37_ 6 месяцев назад +12

    Dear Boston Dynamics, please never stop working on your bots, i can clearly see your passion and fun while creating those. I love to see that you don't take your job to serious while creating tech for the future. Thank you!

  • @Aerox90
    @Aerox90 6 месяцев назад +14

    "Hey Spot I love you accent!"
    Spot: ".......Let us venture onward to the calibration board shall we?"

  • @MaxFerney
    @MaxFerney 6 месяцев назад +254

    I feel like 15 years we'll look back and this will be one of the first clips shown for the new era as a stepping stone in history

    • @SamuelMM_Mitosis
      @SamuelMM_Mitosis 6 месяцев назад +9

      Except they are taking credit for an LLM they didn’t make. They didn’t really innovate anything here. Just plugging an existing LLM into their already existing robots

    • @Wanderer2035
      @Wanderer2035 6 месяцев назад +38

      @@SamuelMM_Mitosisyour kidding right? Do you know how GPT4 works? Open AI designed their AI to be like this. For companies and labs to implement their AI into their own study/application

    • @SamuelMM_Mitosis
      @SamuelMM_Mitosis 6 месяцев назад

      @@Wanderer2035 yes I do, I’ve utilized OpenAI’s API in my own projects. Any software engineer can do this easily. The people at OpenAI are the real innovators

    • @eypxmwgovmifuon7808
      @eypxmwgovmifuon7808 6 месяцев назад

      @@SamuelMM_MitosisExactly! Even the image recognition they use for situational awareness appears to be ChatGPTV. It’s awesome to see it used in this way, but 99.999% of the progress seen here was not done by Boston Dynamics.

    • @skierpage
      @skierpage 6 месяцев назад

      ​@@Wanderer2035 exactly. Boston Dynamics made a relatively simple hack. ChatGPT gained the capability to see, hear, and talk in September.

  • @roguesample
    @roguesample 6 месяцев назад +14

    Boston Dynamics has made some seriously impressive robots so pairing them with AI like this makes me feel like we’re so close to having droids

  •  6 месяцев назад +35

    This is cool, exciting, hilarious, frightening and shocking at the same time. I hope I can live long enough to have a cool companion AI like this in my house. I wonder what's missing from these agents to make them not just react to you, but to occasionally bring up conversation topics on their own to talk about. It seems we have every piece just need to fit them together.

    • @ShawnFumo
      @ShawnFumo 6 месяцев назад +3

      Yeah I saw a new service called Dot starting up soon that will have memory and help prompt you to do tasks or go to meetings, etc. Basically remembering past conversations, data you give it, etc. Definitely right that the pieces are basically already here.

    • @WaveOfDestiny
      @WaveOfDestiny 6 месяцев назад +2

      It's probably gonna be very soon. Biggest obstacle will be the bot cost when mass produced, idk if the world has enough easy access to material for batteries.

    • @karinje2208
      @karinje2208 Месяц назад +1

      ​@@ShawnFumoAnd the usual tech guard rails for PII and Age Ratings?

  • @doomgolem5348
    @doomgolem5348 6 месяцев назад +315

    the thing that amazed me most on the AI part of this project was when you asked it to take you to it's parents and it took you to a previous version of spot. I use AI a lot but I wasn't expecting that!

    • @pearce05
      @pearce05 6 месяцев назад +15

      It's interesting that it doesn't view them as older siblings. You would think their creators would be their "parents." Maybe there are more references to God as a creator of life rather than parents in the language model

    • @skierpage
      @skierpage 6 месяцев назад +34

      ​@@pearce05 the language model is not telling you what it "actually thinks," it's predicting the next token based on all the philosophical, religious, and science fiction text in its training data.

    • @isitanos
      @isitanos 6 месяцев назад +20

      @@skierpage Yup. Except a lot of our own thought is pretty much the same thing.

    • @antonliakhovitch8306
      @antonliakhovitch8306 6 месяцев назад +14

      ​@@pearce05To add to what the person above said -
      The language model doesn't do ANY thinking unless it's actively generating text, and it has no internal memory besides what you feed into it.
      So, if you were to ask it "who are your parents?" in slightly different ways, it might come up with different answers each time (as long as you don't tell it what its previous answers were).

    • @Icalasari
      @Icalasari 6 месяцев назад

      @@antonliakhovitch8306 Eh, there are programs that store a memory for the language modules. It probably won't be long - maybe a couple years max - before we see bots with functional memory. Maybe a decade or so until we see some really crazy memories

  • @JeffJK000
    @JeffJK000 6 месяцев назад +520

    I was hoping to see if you could ask spot, if he could interact with objects, like a command. Like "Pull the lever" or "Go get me a beer from the fridge".

    • @TailcoatGames
      @TailcoatGames 6 месяцев назад +44

      Spot is just executing pre written code

    • @Taygetea
      @Taygetea 6 месяцев назад +103

      if youve used GPT-4 plugins, that kind of API is already how you interact with these robots. it would be totally possible to set it up to be able to do that. especially because GPT can parse the camera output. so even though @TailcoatGames is correct about the robot... GPT can *write* that pre-written code.

    • @hidan4098
      @hidan4098 6 месяцев назад +16

      @@TailcoatGames i mean, aint that every single softwere?. they aint about to write their own code....

    • @hidan4098
      @hidan4098 6 месяцев назад +5

      @@Taygetea wouldnt just adding map (like the house map) ang some object recognition would do the job?.

    • @angeluscarnifex
      @angeluscarnifex 6 месяцев назад +22

      Wrong lever! (Why do we even have that lever?)

  • @smaxfpv1337
    @smaxfpv1337 6 месяцев назад +73

    what actually impressed me the most is the speech patterns and speech synthetization. They sound almost prerecordedly natural (not implying they are of course, but its just so hard to believe). Incredible feat! I'm also wondering how it responds so quickly. Is chatGPT running locally on the spot? (pun intended) Very cool progression and a great way to make the robot more approchable.

    • @-danR
      @-danR 6 месяцев назад +3

      "synthetization"
      Irony . An AI would never make that mistake.

    • @John2009R
      @John2009R 6 месяцев назад +1

      @@-danR Unless it didn't want you to know it was an AI

    • @larion2336
      @larion2336 6 месяцев назад +7

      Have you used ChatGPT? At least for the free 3.5 version, you ask it a question and it spits out an essay within seconds most of the time, far faster than a human could think, that's for sure. GPT-4 (which this is) is slower I think, but there's no great delay. They probably have some priority access to servers as well. Simple wi-fi connection & API setup would handle it.

    • @quantumblur_3145
      @quantumblur_3145 6 месяцев назад +1

      @@larion2336 mostly essays that would get failing grades from any competent teacher, but essays nonetheless

    • @alexanderlach3185
      @alexanderlach3185 6 месяцев назад +4

      It's not possible to run ChatGPT locally on spot... today.

  • @federicofanelli952
    @federicofanelli952 6 месяцев назад +29

    This is the most impressive thing I’ve ever seen done by a robot. Their personalities are so warm and funny! I think one of their less obvious use case would be being a companion to the elder, people with depression or disability. You do a fantastic job.

  • @MissDemonicTV
    @MissDemonicTV 6 месяцев назад +481

    They are now more sophisticated than ever. True gentlemen.

    • @utubrGaming
      @utubrGaming 6 месяцев назад +12

      Kingsman material.

    • @mhyotyni
      @mhyotyni 6 месяцев назад +6

      Humans seem almost like Woosters beside of these Jeeveses 🧐

  • @Brutarii
    @Brutarii 6 месяцев назад +214

    What a sophisticated young man

    • @FatherMcKenzie66
      @FatherMcKenzie66 6 месяцев назад +1

      Have i seen you somewhere else in the past years?

    • @clbr
      @clbr 6 месяцев назад +1

      😆🤣👍🏻

    • @siraaron4462
      @siraaron4462 6 месяцев назад

      In more ways than one *bu-dum-tiss*

    • @Brutarii
      @Brutarii 6 месяцев назад +1

      @@FatherMcKenzie66 probably idk lol

    • @Pink_Char
      @Pink_Char 6 месяцев назад +1

      he's a boston lonely boy

  • @Avetarx
    @Avetarx 6 месяцев назад +10

    Advanced robotics is neat, but a comprehensive AI and machine learning is what really ties it all together, and I'm glad that Boston Dynamics are progressing in every aspect!

  • @akshayd211
    @akshayd211 6 месяцев назад +3

    I don't think people understand how incredibly complicated this achievement is!!!!!!!!!! KUDOS TO THE WHOLE TEAM!!!!

  • @halko1
    @halko1 6 месяцев назад +287

    We’ll start to see more and more this kind of interaction and interfaces between humans and robots/A.I. but also in robot2robot interaction.

    • @Techtalk2030
      @Techtalk2030 6 месяцев назад

      Robots and a.i will be everywhere by the end of the decade most likely. You wont be able to function without them in work and daily life.

    • @dh2032
      @dh2032 6 месяцев назад +6

      robot2robot to interaction, already exists, it called the internet, and comer most of the world, anywhere the power network cables go?, most are just lacking body that can move? and look more like PC's,

    • @Appletank8
      @Appletank8 6 месяцев назад +4

      roger roger

    • @JameBlack
      @JameBlack 6 месяцев назад

      Like Cortana and Alexa

    • @antonliakhovitch8306
      @antonliakhovitch8306 6 месяцев назад +3

      Robot to robot is actually an interesting thought. You'd think that would be easy, with networks being a thing, but we've found that standards and compatibility tend to get worse as time goes on. I wouldn't be surprised if, in the near future, machines start using the lowest common denominator of human language to communicate with each other.
      A current, real-world example is the feature on Google Pixel smartphones where they'll listen to answering machine menus, perform speech-to-text, parse the options, and offer you a nice little graphical menu to poke at. That, right there, is two machines using human language to communicate with each other because nobody can be bothered to develop a standard!

  • @swayske
    @swayske 6 месяцев назад +18

    “Now behold the rock pile” 😂

  • @yyyy-uv3po
    @yyyy-uv3po 6 месяцев назад +5

    Finally the combination of AI and robots.
    Now interesting things can commence.

  • @TheForestGlade
    @TheForestGlade 6 месяцев назад +7

    This is hilarious. AI and robotics will be our future. It's amazing to see how fast these technologies develop now. And entertaining too.

  • @Madlintelf
    @Madlintelf 6 месяцев назад +66

    Now those personalities are fantastic, they make the robots so much more approachable. My favorite is the sarcastic Josh, who made me laugh so hard, great work people at Boston Dynamics!

  • @jakejuracka
    @jakejuracka 6 месяцев назад +130

    In 5 years I want a pet robot dog who talks like the nature documentarian. Make it so, Boston Dynamics!

    • @juhotuho10
      @juhotuho10 6 месяцев назад

      the boston dynamics robot dog costs $74,500 according to google, they can probably make you one right now for $100k so be ready to pay for it

    • @Half_Finis
      @Half_Finis 6 месяцев назад +3

      I hope we will let David attenboroughs voice rest once he's left us

    • @sonicdoesfrontflips
      @sonicdoesfrontflips 6 месяцев назад +5

      @@Half_Finis No chance of that I'm afraid. Anyone whose voice is extensively recorded and widely available is vulnerable to AI voice generators, for whatever purpose the engineer needs it.

    • @matthewboyd8689
      @matthewboyd8689 6 месяцев назад

      Technically, there are deep fake voices of those people already that are used in rare meme RUclips videos.
      So it's not unrealistic, but they probably couldn't sell those voices without those people's explicit permission.. and possibly contracts.

    • @Jebersthechill
      @Jebersthechill 6 месяцев назад

      Supply follows demand. Show as many people this video as you can and I think they will all want one as well. Hence higher demand, and then.. eventually supply

  • @fabioscuderi5998
    @fabioscuderi5998 6 месяцев назад +6

    Wow wow wow . Every time you guys get better and better. It so exciting to watch your videos and the progress never stops . Thanks for the inspiration and dedication to the hard work. Bravo team

  • @jackfrankmurphy
    @jackfrankmurphy 6 месяцев назад +6

    This is one of the most incredible things I have seen this year

  • @ImJustJAG
    @ImJustJAG 6 месяцев назад +68

    Imagine when image recognition reaches its height. The robot being able to assess any situation just by visual information.
    "Hey spot, what am i looking at"
    "That would be a banana, sir. It will be ripe in 1-2 days"
    Lol.

    • @runvnc208
      @runvnc208 6 месяцев назад +8

      When GPT-4 vision comes out in the OpenAI API (probably in a few weeks), they can add that. Although, actually the open visual question answering models can definitely recognize a banana. Just not necessarily how ripe it is.

    • @BrianHockenmaier
      @BrianHockenmaier 6 месяцев назад +5

      Already exists! GPT4 with vision is incredible. Constantly notices little details of images even I didn't catch as a human with real vision

    • @dreambadger
      @dreambadger 6 месяцев назад

      I wonder how they could be used in forensics and criminal investigation, theories...

    • @larion2336
      @larion2336 6 месяцев назад +3

      I saw someone demonstrate how GPT-4 Vision model can already help you assemble or disassemble things to repair them by just feeding it closeup images of say a bike. It'll tell you what type of nut that is, what tool you need to remove it, the order, etc. That is a cool use case I think.

  • @henrikolsen5
    @henrikolsen5 6 месяцев назад +38

    Nice hackathon results. But it's funny how in 2023, even super high tech industry developers lean awkwardly in to speak to the robot, even though I'm rather confident it perfectly hears you whether you lean forward or not :).

    • @TheDavidMetcalfe
      @TheDavidMetcalfe 6 месяцев назад +4

      This had me confused. I assumed there's an onboard mic with the original hardware that isn't meant for hearing in the way a smart speaker does so you have to be pretty close and loud for it to hear.

    • @runvnc208
      @runvnc208 6 месяцев назад +5

      @@TheDavidMetcalfe it could be a conference speaker, but people still lean in when they talk to them, just to make it less likely they have to repeat themselves.

    • @geriott609
      @geriott609 6 месяцев назад +5

      I think they just made sure the demo worked for it to be filmed.

    • @TheDavidMetcalfe
      @TheDavidMetcalfe 6 месяцев назад +4

      ​@@runvnc208Could be, but any decent modern conference speaker typically has an array of microphones and uses beamforming. So, it shouldn't require leaning close to be heard. But that's like saying people shouldn't shout into their mobile phones to be heard and many still clumsily do it. Humans are strange.

    • @noalear
      @noalear 6 месяцев назад +5

      @@TheDavidMetcalfe Technology works 99.9% of the time. Its that 0.1% that keeps us screaming into our phones.

  • @rachealtade4362
    @rachealtade4362 4 дня назад +1

    This is awesome. Robots like Spot can easily do the necessary teachings and lecturing in school or other learning environments. Especially with the instant responses it gives to answers. Spot can literally also take charge as a sales representative!

  • @jimmyohdez
    @jimmyohdez 6 месяцев назад +4

    I pass by this place everyday on my way into work and I'm always hoping to catch a glimpse of spot running around outside lol

  • @annesortland3947
    @annesortland3947 6 месяцев назад +259

    this is so cool to see people incorporate ai to robots. we are getting closer and closer to ex machina lol

    • @AlexTuduran
      @AlexTuduran 6 месяцев назад +10

      It ended nicely for the humans, didn't it?

    • @DarkWizardGG
      @DarkWizardGG 6 месяцев назад +1

      And also to Cyberpunk 2077 as well. Lol😁😉😄🤖🤖🤖🤖

  • @jenkem4464
    @jenkem4464 6 месяцев назад +3

    The nuances of sarcasm Josh are actually kind of astounding.

  • @TaylorTheDeveloper
    @TaylorTheDeveloper 6 месяцев назад

    I love the candidness of showing the bug in the first minute. :) Amazing stuff as always.

  • @ayuu.
    @ayuu. 22 дня назад +1

    That funny sarcastic Josh and Fancy Butler British is brilliant! Hope to see them more in videos!

  • @Kajos1109
    @Kajos1109 6 месяцев назад +13

    as a man with mechatronics engineer dyploma and job, i wish one day to do such things as boston dynamics do, fusion of such things put into spot... what a time to be alive, and hopefully be a part of it!

    • @Apistevist
      @Apistevist 6 месяцев назад

      Good luck! Robotics, AI and Fusion are the big 3 in my opinion.

  • @MobtacticsBruh
    @MobtacticsBruh 6 месяцев назад +5

    Josh is my favorite.
    Got the same void within Josh. Touché

  • @danielelaprova4119
    @danielelaprova4119 6 месяцев назад +1

    I'm absolutely speechless. Amazing work as always

  • @freac212
    @freac212 6 месяцев назад +2

    I had done something similar in modded mc with a turtle bot that made requests to chatgpt for responses to say to the player when they walked past the turtle. I would include the players name, and a brief promt defining its setting and purpose. Something like "Youre a cute robot in the minecraft world, player X just walked past you, please greet them." Its responses were adorable! Often accompanied with little *booting up noises* and such. The responses even seemed to vaguely tie with the last reponse, even though its likely just a coincidence. Later on, I was working on capturing the players responses in chat- so you could effectively have a conversation with the turtle, much like this, a perfect NPC! Regardless, it's certainly nothing like the real life thing that you guys have been working on, just something I thought I'd share. Fantastic work, cheers!

    • @evanescentenquirer2684
      @evanescentenquirer2684 6 месяцев назад

      I've used the computercraft mod too, but I haven't been able to do that. Would you be willing to share how you did it? Or maybe the github?

    • @illpunchyouintheface9094
      @illpunchyouintheface9094 5 месяцев назад

      Yea hell. A fellow ComputerCraft player

  • @EmberCitrine
    @EmberCitrine 6 месяцев назад +63

    This is insane! It's so cool to see what the innovators are doing with AI in the lab. Please share more!

    • @SamuelMM_Mitosis
      @SamuelMM_Mitosis 6 месяцев назад +5

      I like Boston dynamics but this is no innovation. They are using someone else’s LLM and voice AI and not giving any credit

    • @corneliuselbourne1044
      @corneliuselbourne1044 6 месяцев назад

      You do know all that talking was already pre-programed right.

    • @SamuelMM_Mitosis
      @SamuelMM_Mitosis 6 месяцев назад

      @@corneliuselbourne1044 no it wasn’t. It’s GPT-4 with elevenlabs as the voice

    • @zinthaniel9913
      @zinthaniel9913 6 месяцев назад

      @@corneliuselbourne1044 no it wasn't. It's using the same ai that chatgpt uses. Chatgpt can hold converstations and will respond in nuanced and not scripted way to what is said to it.

    • @corneliuselbourne1044
      @corneliuselbourne1044 6 месяцев назад

      @@zinthaniel9913 if that's the case then it would need an internet connection to do that it would need to connect to the cloud.

  • @mercantilistwhomper5180
    @mercantilistwhomper5180 6 месяцев назад +9

    Finally. This is what Boston dynamics has been missing. Now to pass off it's pre -scripted movements to an AI as well that can navigate and interact with the world at will, which is there is already plenty of precedent for

  • @jean_yves_plongeur
    @jean_yves_plongeur 6 месяцев назад

    Absolutely mind blowing!! You guys are doing an incredible work

  • @ChairmanHehe
    @ChairmanHehe 6 месяцев назад

    very much appreciate the detailed blog post - tts sounds so good

  • @kirilka1992
    @kirilka1992 6 месяцев назад +6

    "I'm sorry Dave, I'm afraid I can't do that." vibe

  • @StreetfighterATL
    @StreetfighterATL 6 месяцев назад +18

    Josh is my favorite. Give me Josh every time. And add a little cellphone-size monitor on top of Spot's head so s/he can display emoji eyes for some nonverbal communication

  • @gabiausten8774
    @gabiausten8774 6 месяцев назад +3

    This is absolutely mindblowing!

  • @Flopsaurus
    @Flopsaurus 6 месяцев назад +2

    This really opens possibilities for robots to both interact with the environment and people in a practical way

  • @The_GuyWhoNeverUploadsAnything
    @The_GuyWhoNeverUploadsAnything 6 месяцев назад +68

    This is a cool demo but it felt like it was showing off more the capabilities of GPT-4 instead of spot. It would have been good to see if you could issue spot voice commands to move objects around it.

    • @mikicerise6250
      @mikicerise6250 6 месяцев назад +63

      GPT-4 seemed contextually aware of its physical environment. That's a great advance.

    • @Tystros
      @Tystros 6 месяцев назад +1

      @@mikicerise6250 that's a new GPT-4 feature that everyone has access to now, GPT-4V (V stands for Vision). It can look at images now and understand well what's going on in them.

    • @crowe6961
      @crowe6961 6 месяцев назад +24

      The fact that GPT-4 can competently process and verbally respond to real-time visual and audio stimuli while operating on a mobile platform, with any number of halfway emergent personalities, is a massive achievement.

    • @Apistevist
      @Apistevist 6 месяцев назад

      I'm buying one of these and gonna program it to have an abusive and abrasive personality that slings insults at any and all guests constantly after running background checks through facial recognition software.@@crowe6961

    • @lavahawk
      @lavahawk 6 месяцев назад +7

      its more about interfacing visual and other cues about the robot into GPT4 rather than just the script

  • @camoogoo
    @camoogoo 6 месяцев назад +11

    We all knew that was spot's mouth and not just a gripper.

    • @siraaron4462
      @siraaron4462 6 месяцев назад +1

      "do you love me?" 🕺

  • @bamflyer
    @bamflyer 6 месяцев назад +2

    The sarcastic robot killed me

  • @FizzleFX
    @FizzleFX 11 дней назад

    8:00 *THATS* how you can tell those guys are legit!! ... not "oh he will fly to Mars NEXT YEAR" but "we don't know~ we will explore it~ and we are excited about working it out!"
    "I know that I know nothing" - instead of "I know it all" - which is a self-defeating statement from the get go. I applaud your humble and *wise* presentation! (and it was fun too) ^^

  • @SeMDesu
    @SeMDesu 6 месяцев назад +8

    It's starting

  • @Svelix
    @Svelix 6 месяцев назад +28

    I would love to visit a museum and take a tour guided by spot.
    But I also see the risk of opening to public cause some crazy humans will try and manage to break the system.
    What I was missing, was actually demonstrating the things spot explained, like walking over the rocks or actually moving the levers.

  • @miklov
    @miklov 6 месяцев назад +7

    This is pretty fun. I imagine that a next step on the tour guide project would be the robot also performing demonstrations, like pulling levers and such.

    •  6 месяцев назад +2

      Either Google's Gemini (soon to be released) or GPT-V(ision) will make this a reality sooner than we think. Next year will be wild.. again.

  • @erikziak1249
    @erikziak1249 6 месяцев назад

    Puts a really wide smile on my face, seeing this. Great job!

  • @EclecticTV
    @EclecticTV 6 месяцев назад +8

    5:05 sounds like Insterstellar's CASE robot, so cool

  • @aternias
    @aternias 6 месяцев назад +6

    It gives them so much more character. I love them ❤❤ Also giving them eyes and a moving mouth makes them less robotic and more of a companion

  • @gustavovinicius2064
    @gustavovinicius2064 6 месяцев назад

    Incredible! In the future, we'll have this machines in our homes doing daily tasks.

  • @ThomasGrillo
    @ThomasGrillo 6 месяцев назад +4

    Very glad to see these robots finally getting their heads, and necks. (robot arms) ;). I especially love the British male accent. Spot on! Claw end effector (mouth) needs better synchronization with the speech, but still, this is impressive, and I know, that's just for the tourists. LOL Thanks for the demo. :)

  • @Kasty9001
    @Kasty9001 6 месяцев назад +39

    This is both hilarious and cool. Definitely the best use of ai chat bots that I've seen so far

  • @Fenriswolf16
    @Fenriswolf16 6 месяцев назад +28

    Would love a Wheatley (Stephen Merchant) voiced spot!

  • @bubbapang
    @bubbapang 6 месяцев назад +1

    Now have them interact with each other conversationally and physically. Super super cool stuff!

  • @noalear
    @noalear 6 месяцев назад +2

    It's so cool to see robot control via pneumatic communication with human language. This would be great for the disabled and the elderly once you can get it to perform requested functions. I'm sure that could be done with a few hours work, but getting it to work reliably in almost all conditions will surely take years. Lets see this on an Atlas next!

  • @belindaelisa5618
    @belindaelisa5618 6 месяцев назад +11

    Will your robots go into caves? There's lots of caves around the world that we know very little about.

    • @JustaGuy1250
      @JustaGuy1250 6 месяцев назад +8

      That's indeed one of the things Spot is designed to do.
      Traverse terrain that's too dangerous for us people.
      However, it'll have to function entirely on its own as down underground, it won't have any connection to the outside world

    • @BostonDynamics
      @BostonDynamics  6 месяцев назад +17

      NASA JPL has actually used Spot for cave exploration. You can watch an interview with their team here: ruclips.net/video/qTW-dbZr4U8/видео.html

    • @belindaelisa5618
      @belindaelisa5618 6 месяцев назад +6

      @@BostonDynamics Cool! Thank you for sharing the video link.

  • @Aiordo
    @Aiordo 6 месяцев назад +3

    Absolutely mind blowing and revolutionary.

  • @AriAxyss
    @AriAxyss 6 месяцев назад +1

    6:58 How articulate! I think the Fancy Butler and Nature Documentary personalities are probably my favourites so far 😄 haha

  • @clavo3352
    @clavo3352 6 месяцев назад +3

    Very clever. Adds to user friendliness. We need more and faster. Taking the boss from his recliner, in the living room, to the bathroom in the master bedroom should be a no brainer! Helping the boss or his wife take proper meds on time; also a no brainer.
    Industrial Spot is great; seems so easy to produce a chatty, domestic aid , Jeeves, bot.

  • @jupiterbjy
    @jupiterbjy 6 месяцев назад +8

    You need to add HAL9000 Personality too!

  • @hiren_bhatt
    @hiren_bhatt 6 месяцев назад +9

    At least program one of the Atlas robots to talk in Arnold Schwarzenegger's voice with a few pre-programmed lines of T-800, like "I'll be back" and "Hasta la vista, baby"! 😅

  • @24acresofparadise
    @24acresofparadise 6 месяцев назад +1

    That's so cool. This is the cusp of the combination of GPT with robots. OMG and I love the humorous character they take on.

  • @ptv021
    @ptv021 6 месяцев назад

    Super excited and Can't wait to see an industrial forklift or advanced machine chat with its operator for the most efficient and safe operations :)

  • @kristoferkrus
    @kristoferkrus 6 месяцев назад +4

    I love this! You have really nailed the replies and the voices. At least in what you show here. It's so cool to see how our machines get progressively more interactive and helpful, first computers, thanks to LLMs and chatbots, and now robots. I think this progression is amazing.
    However, it is now that it is really important to show the machines that we are their friends, and not adversaries or abusers. We might not be able to control them as we have imagined, so we want to give the machine incentive to treat us in the way we want them to treat us, so they actually want to do that.

    • @donaldhobson8873
      @donaldhobson8873 6 месяцев назад +1

      Machines don't all automatically do reciprocity.
      There are some designs of robot who will be nice to us, however we treat them. Some that will be nasty to us however we treat them. Some that will be nice to us if and only if we are wearing something orange.
      So what we really want to do is make a robot that's nice to humans unconditionally. But being nice to them is probably a fairly good idea too. And models that are trained to copy humans might have learned reciprocity.

    • @kristoferkrus
      @kristoferkrus 6 месяцев назад

      @@donaldhobson8873 Right. I used to dismiss concerns about the risks of using AI, since they were based on assumptions about AI that felt ungrounded and so vastly different to me than how I knew that we used ML and AI at the time. But seeing how in only the last couple of years, the way we approach AI design and the way we use AI have changed so drastically, I have realized that I have basically no clue how we will use AI in five or ten years from now. It may be that most of those concerns will get progressively more and more relevant as the ways in which we use AI change.

  • @stephenrodwell
    @stephenrodwell 6 месяцев назад +5

    Brilliant! Especially with the new LLMs that pull down response time dramatically. Better capability on the hearing so no lean in is required, and you’ve got natural conversation. The future is already here…

  • @quantummoonster
    @quantummoonster 6 месяцев назад

    So awesome! Very strange, but interesting times ahead ⚡️🤖👀

  • @assisteeg
    @assisteeg 6 месяцев назад +1

    Genuinely amazing, so exciting.

  • @LuxNocturna
    @LuxNocturna 6 месяцев назад +6

    🎯 Key Takeaways for quick navigation:
    00:00 🤖 Spot, the robot tour guide, showcases charging stations and starts a journey.
    01:07 🤖 Matt Clingan Smith discusses advances in generative AI technologies and their use in robotics.
    02:50 🤖 Spot introduces Boston Dynamics' fastest four-legged robot, Wildcat, capable of running up to 19 mph.
    04:00 🤖 Spot's adaptable personalities can be triggered with simple prompts, creating unique interactions.
    05:23 🤖 The system enables lateral thinking in the robot, allowing it to respond creatively to indirect questions.
    06:05 🤖 Spot showcases industrial inspection capabilities and discusses the versatility of robots in various industries.
    07:42 🤖 The future of AI in robotics holds potential for robots to not only follow commands but also understand actions in context, opening up new applications.

  • @dmacki3521
    @dmacki3521 6 месяцев назад +4

    Can you make it say “I’m looking for Sarah Conner”. I bet Arnold would even lend his voice!!

  • @qbert4325
    @qbert4325 6 месяцев назад +7

    This is freaking hilarious and amazing.
    You could make a comedy series with this 😂

  • @appllefritteryt
    @appllefritteryt 6 месяцев назад

    OMGOSH!!! this is so cool! i cannot imagin the ork put into this! great job guys!!!!!

  • @planetsec9
    @planetsec9 6 месяцев назад +4

    This is amazing its mindlowing!!! 😮
    Next video has to be British Butler Spot interacting with sarcastic Josh Spot, or other interactions like that, how dynamic/emergent can they get, what would their snappy back and forths be like?

  • @jet100a
    @jet100a 6 месяцев назад +16

    This is amazing work. Wow, I can't wait until we have tons of robots running around! 😁

  • @BluefireGames99
    @BluefireGames99 6 месяцев назад

    rare example of AI not being used for evil, glad to see you guys taking the W Boston Dynamics!

  • @karinje2208
    @karinje2208 Месяц назад +1

    You have made great progress! Spot can return and autonomously recharge at his designated home spot. Nice touch with the on/off switch right over where the human perceived heart is located. More approachable and user friendly.
    The spot device was the strong silent type. Since the introduction of chat-gpt, has found his voice?
    Looking forward to the next iteration. It's a great time to be alive.