Unreal Engine 5.2 ML Deformer can be utilized to achieve lifelike deformation on your own real-time characters. It shows an approximation to a full muscle sim that takes 1 minute per frame normally. Also does cloth approximation. The footage included in this video is running live on PS5. Thank you for watching. Please read the description for more details! If you enjoyed the video, please support the channel with a LIKE ❤, SUBSCRIBE ☑ and turn on NOTIFICATIONS 🔔 to watch more!
People talk about Mario like was the most important thing to came out in gaming history, Cant stand these Nintendo Fanboys, They know ZERO about gaming history
a matter of time before AI implemented with this tech to auto generate realistic people, maybe even auto generate thousands of NPC's for city games this is amazing
It's amazing, but also scary. I don't like the idea that you can create a fantastic representation of the real world but without the real world problems that real life brings. People would spend all days in that digital world and that can't be good..
there's always been an issue with addiction in entertainment products. With good self control that issue is avoided, (not as a whole but at least for that person) Just think of all the good things stuff like this will bring! :)
I am waiting when ai will be integrated into this. Imagine when NPc behaviour will not be scripted, but managed by AI intelligence instead, maybe even by some kind of general AI
Just think. Future generations of children will completely lack any amazement, because the graphical singularity will have already been reached and there isn't much more improvements we can do.
we used to turn off the lights and huddle around our pentium pc with DOS to watch the Mortal Kombat intro... took 5 floppy disks to install which was a huge deal back then
But most probably it'll be used in news and politics. This level of tech and billions are not invested so that people can wander around in video games.
@@LittleUrbanPrepper anything that makes an income is worth pursuing in a capitalistic environment, video games are an entire industry, don't underestimate them
Finally the tools to make games faster. The time it takes to make them has dragged into many years for a AAA game. 20 years ago a mid-sized company could publish a new game every year or two. Now they are often taking 5-10 for their next game. Between this and AI assistance the years might drop back down to the point where they can afford to take chances again.
Definitely agree. I mean a lot of tools now are easier and faster than any time before, but things are also orders of magnitude more complex and so take longer. With AI/ML and increased hw performance we might start seeing more of the really cool (and risky) ideas be realized with good graphics, fleshed out gameplay and support. I’m pretty much only playing indies as AAA polish nowadays strips all but the most basic oomph from games.
Not 1 indie developer has released a game using metahumans. And if they did, there would only be 5 or less rendering at any given time. They are far too performance heavy. Even when you remove facial animation, hair and set the LOD to 4, they still tank projects. Cool tech, but not for indies.
they're taking longer because of publishers trying to squeeze out more money, more middle men and bad salaries / contracts for designers which is why they have to resort to entry level designers. game development doesn't take years in any efficient company
Truly a "Wow!" moment in gaming/performance capture history. Having been a gamer since Pong, I truly envy those of you that will get to see what games look like in 20 years and beyond. Y'all are in for some mind-bending greatness when it comes to realism. All I could think of while watching this were the motionless mouths, the mitten-like, unbending hands, and animated-from-scratch movements found in games in the 90s.
I'm hopeful that someone Musk-like works out how to make San Junipero a reality so I can go on enjoying new gaming experiences for the next few centuries. I, also, am a child of the Pong generation and every new year of gaming comes closer and closer to true magic in my eyes, I can't wait to see what comes next!!
Deformer for me is the most game changer, including the clothing that can be used with it. I always noticed how clothing were one of the things that always seemed TOO MUCH off from the game graphics.
@@grey_north9016 Volumetric particle simulation and procedural destruction are far more important for gameplay mechanics than the cloth-blending, but by golly the cloth-blending is the most impressive looking feature because it removes that uncanny immersion-breaking element to 3D clothes that we always see. I hope that unlike a lot of the cool tech features introduced in UE3 and UE4 that these more advanced design libraries don't get left behind or only used in short films during UE5's market run. I would love to see actual interactive games put these features to use finally.
@@billywashere6965 The thing is, we have goo looking particle simulation and procedural destruction with the Unreal Engine 5, but we didn't had anything to make cloth-blending really realistic, and now we finally have. Take Red Dead2 and TLOU2 as example, both games are incredible realistic, but if we look at they cloths while doing any kind of animation, we see that there's something very off, because cloth are static, or just things like capes and skirt which moves, but still, no bending at all. Now we will finally have it with UE5. Well, of course, I know that this is something that we will only see, maybe at the end of this current generation, but still, impressive af
Wrinkles in clothing particularly still look very static. That is, any wrinkles are built into the mesh and just stretch as the character deforms, instead of dynamically wrinkling in different places. I'm really looking forward to real time dynamic cloth wrinkling that doesn't nuke CPUs haha.
100%. clothing deformation is the next big thing I want to see. I don't care how photoreal models look, but I do like real world physics, especially with cloth. Hogwarts legacy did more cloth physics than I've seen in a game I think
Impressive. The next step would be to make the muscles respond by flexing when moved. It feels odd to see these very realistic looking games, like GoW:R with Kratos, seeing him look very realistic, yet his muscles have no movement. It's as if he's wearing a very detailed muscle suit.
I understand the fear, but also I am aware that all knew leaps in technology scare us. Trains, cars, electricity, planes, etc. When RUclips and camera phones came out, my boss thought all our video producer jobs were doomed, and now I am busier than ever and using a lot of this new technology to automate more tedious processes. Sociopathic capitalism, surveillance, and resource conservation are always the big fears. I read way to into your comment, and got on my own tirade. Yes. Creepy, but amazing, lol.
@@jmfs3497 well if this just isn't an EXCELLENT comment! It's great that you mentioned your own situation as an example because this couldn't be any more true!
What if we could get a game where you literally talk through a mic with NPCs that can understand and respond in any language and also result in different outcomes depending on your questions with the only pre-programmed thing will be the core motive of a character?
This is great. I'm currently a multimedia producer (video, audio, graphic design, simple animation), but I am learning Blender, Python, and Ambisonics right now. Music, physics, and art are my primary interests, and there is so much out there to learn that am trying to make a lateral move away from more entry level media, and into the advances of all this amazing technology. I'm not even a gamer, but these engines span outside recreational gaming. This is amazing.
There will always be one, the human brain is too good at noticing when something is a little off. It's the same reason why we can imagine shapes of a face in random places in the world, our brain recognizes a pattern and imagines it as a face, but always knows that it's not right
I remember the Syphon Filter intro running through the jungle around 14years old and thinking we'll never have gameplay like that ever... im 37 now and its crazy to see how far gaming has come.
The engine didn't get the scared expression for some reason. I'm sure they'll figure it out somehow, but that was surprising considering how well it caught the other facial expressions.
I also think it looks like the people are wearing braces, or a mouthpiece. Something about the lips seems off. The shade on the teeth and a lack of glance, it looks weird. But otherwise it's awesome.
Muscle flexion at last. I saw this at the first trailer of Tekken 8, where they have a close up to Kazuya's left arm, just before he punches, where his muscles go from relaxed to tense, and you can see muscle fibers forming. I thought it would only be a prerendered video, but no, it's in game too.
In one of the newest Tomb Raiders (don't remember which one) they also did that for Lara. Of course she is no powerhouse muscled up of a human (she's skinny), so it is very subtle, but yeah, it's in there
Looks great, facial expressions too extreme. This kind of tech is best used to capture naturalistic performances. This is why Gollum was so amazing in Lord of the Rings, because the expressions weren’t super exaggerated
Bay Raitt : I textured the version he used to develop the facial expressions based on the "Facial Action Coding System" (FACS) developed by Dr Paul Ekman.
Yeah I would love to see this with a top tier actor that’s more subtle, less Shakespearean or fanmade energy. No offense to the actress, as that works for other media, just not the style I’d prefer in games I’d play.
Elon tried that with neural link, hurting alot of monkeys in the process. I don't see it happening for another 50 years, having a brain implant that's safe and valuable enough to risk.
@@PCgamer923 I don't think Sean meant mind controlling a game character with something like Neuralink. I think he meant taking a video of ourselfs at home, upload it into the game, and have a visual representation of your character that you can e.g. classically control with keyboard/mouse/gamepad, all non invasive.
Maybe an elden ring character that actually looks like me. Honestly though a game that made good use of this, where the hero legitimately looks like you, would be a game changer.
Pros: This makes gaming and CGI for movies a little better Cons: The potential downfall of humanity through the erosion of trust and manipulation of people Cool cool cool.
you don't know that Have you looked at the actor and its model side by side ? Also, they captured her from very close Look at 3:38 how deformed she is. I assume they have what it takes to compensate for this, like camera intrinsics, a robust model of normal looking people, and so on. Just don't take pictures of people from this close.
The person in the motion tracking suit and with dots painted on their face seemed so sci-fi and hi-tech when I first saw it in action. Soon it will look so crude, like a 1930s car cooler.
Bloody hell! I remember playing 'Impossible Mission' on the C64 back in the day and thought the animation was amazing! But these days games are getting evermore real!
@@WinrichNaujoks Yeah, no. I grew up with Pong, Atari, C64...and Mass Effect was 100x more fun than Impossible Mission, and No Man's Sky even better than that.
Wow AS A PHYSIOTHERAPIST and MEDICAL ENGINEERING STUDENT, i got to admit THIS IS SO USEFUL🤔. It would help me visualising many things to patients and customers
Yeah I noticed that as well, just still something missing.. but I didn't wanna say it because I can't believe we're lucky enough to live in a time when this is what video games look like lol.
@@k.c.simonsen2 Looks like the tracking is just a little bit off, when the mouth is closed it looks great. The resting position when it's open seems to lean a little too far into a grimace which throws everything off. Should be easy enough to fix with the tools they showed, and probably looks much better with a more careful performance capture rather than thrown together on stage quickly for demonstration.
Probably a mix of capture quality with the tracking of the mouth, but give it a few years and even that will probably be 100% spot on. Maybe even the micro-expressions capture and tiniest of facial wrinkles. Still comparing it to game graphics from even just a few years back it is worlds apart.
That's where the tweaking comes in. Even before, the initial work of capturing the animation was never perfect. Animators still needed to go in and do fine adjustments, but this gets close enough in mere minutes.
Also saw a video on AI chatbot technology being used in a video game with real time responses. The characters would react to what you said and could remember and refer back to previous interactions. I imagine in the near future, you could have a full-blown real time VR conversations with any video game character you could imagine.
@@smoothbanana It would only be a memory. Otherwise, it would be a real consciousness trapped. I wouldn't want to talk to an empty version of my father, it would be quite sad.
@@Kira-vq5cd thats always been a big question in transhumanism... but i think that once possible, there could be a bridge between the world of the living and a virtual world... one where we exist in both places at once for a time, and then the living world just... drops away at some point, just like when you turn off the virtual one and it stops interacting with your consciousness, the real world would turn off, and stop interacting with your consciousness, but because it lives in both places at once, it would remain your consciousness. an interesting question is, once you have transfered over to that state, what if you made a perfect copy? that copy, assuming the first part is accepted as the real you, would also be the real you in the moment of the copy. after that, it would have new and different experiences... would that other be you still? just a different version of you.... and what if later you re-combine? pretty wild.
They do that with adult webcamming. Honestly, I didn't notice until I spent more time on there. They do movements as you request, but if you keep on requesting, eventually, there is some sort of a pattern.
wait a min, was this on the presentation from a week ago?!?! how did we all missed this!?!?! this is HUGE, those complaining about physics should rejoice now, this is a huge step forward, ps: jiggle physics will be on a whole another level xd
I've unironically been waiting for proper _body physics_ for a long time; "jiggle" isn't what we need. Like, if you lay on your side or back, gravity is going to affect your body, etc. Couple that with muscle deformation and next-gen cloth sims/approximations, and characters will look absolutely amazing-regardless of whether they're NSFW or just regular characters.
@@ryo-kai8587 i was joking if it wasn't clear enough, of course this is huge for clothing simulation, tho muscle deformation won't solve the issue with characters lying on the ground on the side or back, clipping will still occur, maybe it'll look more realistic depending on the type of game, but collision calculations still have to be consider and that's already demanding, anyways, this is huge
@@NicolasSilvaVasault I wasn't joking, and I was more talking about the body interacting with itself, which should be perfected before we worry about the body interacting with the environment. Getting breasts to lay properly in various positions via "gravity" instead of only being affected by changes in motion, etc.
On the video, when the actress plays fear, she looks to the right... then to the left. When processing on the computer, we see the sequence again where she does the same...but during the live demo, her eyes go to the right...and face then. No more looking to the left! Mmm. I'm not saying that what we see there is impossible, we know that Epic does not play that, but I'm just saying that to be sure of a successful demo, they may have taken "a few precautions".
Nah, you are seeyimg something wrong, I played both vídeos in slowmo and the difference is basically none, few differences in the eye movement, but that could be easily because of the capture cam didnt captured the eye motion 1:1 correctly. But the movement of players head, eyerbrown, and the mounth were 1:1
I definitely noticed that too! I was like “that doesn’t seem right.” Then I replayed it. I don’t know if the cameras not picking it up or if they pre-recorded this.
woooo..., 7:45 Imagine that character with gpt 5 inside a VR game, there's no need to add much more in a game to blow up your reality...., OMG the IA en next years can terrified the whole life of anyone...
As a professional actress with 45 years experience, I’m so glad this wasn’t happening in my youth. I am coming towards the end of my career but I feel desperately sorry for young actors leaving drama school. Of course it is exciting and incredibly impressive but surely it will be the end of all human actors and models?
@@LordLentils Why are you acting like creating an engine from scratch is necessarily the only way to make new technical advancements in real time rendering? Do you not see the changes with your eyes?
I think you mean 2024 ;) This is quite possible to do right now in real-time with current hardware. This evolution is moving much faster than most people expect.
3D models will ultimately achieve such a high level of realism that they could potentially replace actors and enable people to create their own customized movies using AI. Imagine a Netflix where every show and movie is unique to you.
the main benefit is how much faster it is... as she said, the pipeline normally takes months. but with this it takes just minutes. it's not just about the realism, and while your point is true, no one is saying that realism makes a game good. you're arguing against a point that hasn't been made. there are also plenty of indie games that are great fun. and this tech will help those indie devs too
It doesn't matter how clever they think they're getting, it all still just looks like creepy creatures pretending to be humans while falling way short.
Fun story: this universe is a meta universe and we as a species are on a database, we just forgot. Though when we originally got here, metahumans knew this, the issues began as each new generation started questioning their existence. The idea that their whole lives have been a simulation eventually started sounding ridiculous and the people got divided, the knowledge over time got distorted and so religions similar to Christianity came to light. The people were divided into three categories. The doubters, the believers and the people who chose to believe in a higher being, a god. This inevitably led us on a path to war. The climax being the catastrophic end of an advanced civilization. Fast forward to now, we humans completely forgot our past “did the Egyptians really build the pyramids?”, we have advanced our technology again and war is still an issue. Working towards a meta verse within a metaverse. Each time we die our spirit is getting recycled in the data base and so the worlds population continues to grow bigger.
In the last clip I am absolutely astounded by the eyes. They look so real, they make me feel as if I’m looking at a human. Eyes were the thing that always looked not right with most games and now that’s changed. The only thing I’d say is not so right in this is the gums, teeth and lips but omg it is so realistic
It seems we have finally crossed the uncanny valley and are climbing up the mountain to what is truly possible in animation. Its cool to be alive right now.
Another step closer to creating a simulation within a simulation, or perhaps we're already deeper than that. Pretty wild stuff, as is it's getting faux reality or film and images to look life actually real, sometimes I have to look hard.
It's really a shame we get this but haven't gotten a continuation of the Unreal Series yet. Would love to see Unreal 3 using this tech. Epic you have to get on that
Happy to have lived to see this, I understand how my 7 year old sons astonishment when playing today's realistic games compared to me playing Super Mario likes on NES at his age.
This looks so promising as a technology. I can't wait until this has matured a bit more and shows us what it really can do. A great effort from the meta humans team. Keep up the good work, lads!
I don't know if people remember but, Playstation had a commercial in the late 90s/early 2000s that showed PS9 being eye contacts that gave you an immersive experience. I don't think we're too far away from that.
One thing that always gives away CG human characters is the eyes. In the words of Quint from Jaws "he's got lifeless eyes, black eyes, like a doll's eye." They look like deadmen walking...
*Insane. Just...insane. I have no words for what I just saw.* The second time I clapped looking at the way her short IPhone clip was getting transformed from one character to another, I felt even my claps were not enough for this sheer innovation which I had the luck to see unfold before my eyes during my time. I am just 23 and I already feel old seeing the way the tech industry has been transforming in the past couple of years.
i remember the animation on MTV money for nothing video,.and everything in between then and now, i was always confused when people thought they were so good,... i said,......."when they make them look just like humans, then i will be impressed"...,...watching this,...that day has come. not a gamer, so this is totally new to me. WOW
Unreal Engine 5.2 ML Deformer can be utilized to achieve lifelike deformation on your own real-time characters. It shows an approximation to a full muscle sim that takes 1 minute per frame normally. Also does cloth approximation. The footage included in this video is running live on PS5.
Thank you for watching. Please read the description for more details! If you enjoyed the video, please support the channel with a LIKE ❤, SUBSCRIBE ☑ and turn on NOTIFICATIONS 🔔 to watch more!
I have goosebumps
Running on a ps5 doesn't mean a lot imo . Those animations wouldn't work when a full blown 3d open world is built around you
I just sat thru the enitre Epic Games presentation at GDC and didnt see this deformation demo...did they cut it out?
I don’t even need to imagine. I can just computers to do it for me 😅
In PS5? The motion live capture is about a future XSX game, not PS5. Thanks to the power of dreams :)
Its been less than 40 years since Mario for NES came out, imagine another 40 years
Exactly. The progress is beyond amazing
Add to those 40 years exponential improvement of AI and everything else and the next 40 years will be nothing like the last 40 years.
Mario will be serving you your happy meal
40 years... That stings.
People talk about Mario like was the most important thing to came out in gaming history, Cant stand these Nintendo Fanboys, They know ZERO about gaming history
a matter of time before AI implemented with this tech to auto generate realistic people, maybe even auto generate thousands of NPC's for city games this is amazing
It's amazing, but also scary. I don't like the idea that you can create a fantastic representation of the real world but without the real world problems that real life brings. People would spend all days in that digital world and that can't be good..
there's always been an issue with addiction in entertainment products. With good self control that issue is avoided, (not as a whole but at least for that person) Just think of all the good things stuff like this will bring! :)
@@TikoYTOfficial what’s up tiko, didn’t expect you to be here
@@TikoYTOfficial like what
I am waiting when ai will be integrated into this. Imagine when NPc behaviour will not be scripted, but managed by AI intelligence instead, maybe even by some kind of general AI
This is giving me that vibe as a child when i would be so amazed by cinematic trailers. That last clip brought a smile to my face.
Just think. Future generations of children will completely lack any amazement, because the graphical singularity will have already been reached and there isn't much more improvements we can do.
@@Ben-rz9cf I guess they can always wow them with incredible writing
@@mrcheese5383or just show them the history of how far things have come so they can gain perspective.....🤷
we used to turn off the lights and huddle around our pentium pc with DOS to watch the Mortal Kombat intro... took 5 floppy disks to install which was a huge deal back then
This could play as the opening sequence in a sci-fi dystopian film, like the 'I am legend' opener.
But most probably it'll be used in news and politics. This level of tech and billions are not invested so that people can wander around in video games.
ok when ya gonna do AI pron now using pron star faces?
@@LittleUrbanPrepper anything that makes an income is worth pursuing in a capitalistic environment, video games are an entire industry, don't underestimate them
@@dieauferstehung isn't that done already?
@@OP-lk4tw yes, but actually no. Don't overestimate them either.
Finally the tools to make games faster. The time it takes to make them has dragged into many years for a AAA game. 20 years ago a mid-sized company could publish a new game every year or two. Now they are often taking 5-10 for their next game. Between this and AI assistance the years might drop back down to the point where they can afford to take chances again.
Definitely agree. I mean a lot of tools now are easier and faster than any time before, but things are also orders of magnitude more complex and so take longer. With AI/ML and increased hw performance we might start seeing more of the really cool (and risky) ideas be realized with good graphics, fleshed out gameplay and support. I’m pretty much only playing indies as AAA polish nowadays strips all but the most basic oomph from games.
Not 1 indie developer has released a game using metahumans. And if they did, there would only be 5 or less rendering at any given time. They are far too performance heavy. Even when you remove facial animation, hair and set the LOD to 4, they still tank projects. Cool tech, but not for indies.
I can't wait for games to be 500GB a pop
they're taking longer because of publishers trying to squeeze out more money, more middle men and bad salaries / contracts for designers which is why they have to resort to entry level designers. game development doesn't take years in any efficient company
@@xaby996 I thought with nanite they don't use LOD's anymore, am I wrong?
Truly a "Wow!" moment in gaming/performance capture history.
Having been a gamer since Pong, I truly envy those of you that will get to see what games look like in 20 years and beyond.
Y'all are in for some mind-bending greatness when it comes to realism.
All I could think of while watching this were the motionless mouths, the mitten-like, unbending hands, and animated-from-scratch movements found in games in the 90s.
I'm hopeful that someone Musk-like works out how to make San Junipero a reality so I can go on enjoying new gaming experiences for the next few centuries. I, also, am a child of the Pong generation and every new year of gaming comes closer and closer to true magic in my eyes, I can't wait to see what comes next!!
Hey, you could apply for a bed at cryogenic facilities!
or porn
I think we will see what games look like in 5 years or earlier. Its amazing.
@@justinleybourne2712 What do you see replacing it?
Deformer for me is the most game changer, including the clothing that can be used with it.
I always noticed how clothing were one of the things that always seemed TOO MUCH off from the game graphics.
This, and destruction, character body animation (how they interact and blend with the environment) and particle simulation.
@@grey_north9016 Volumetric particle simulation and procedural destruction are far more important for gameplay mechanics than the cloth-blending, but by golly the cloth-blending is the most impressive looking feature because it removes that uncanny immersion-breaking element to 3D clothes that we always see.
I hope that unlike a lot of the cool tech features introduced in UE3 and UE4 that these more advanced design libraries don't get left behind or only used in short films during UE5's market run. I would love to see actual interactive games put these features to use finally.
@@billywashere6965 The thing is, we have goo looking particle simulation and procedural destruction with the Unreal Engine 5, but we didn't had anything to make cloth-blending really realistic, and now we finally have.
Take Red Dead2 and TLOU2 as example, both games are incredible realistic, but if we look at they cloths while doing any kind of animation, we see that there's something very off, because cloth are static, or just things like capes and skirt which moves, but still, no bending at all.
Now we will finally have it with UE5. Well, of course, I know that this is something that we will only see, maybe at the end of this current generation, but still, impressive af
Wrinkles in clothing particularly still look very static. That is, any wrinkles are built into the mesh and just stretch as the character deforms, instead of dynamically wrinkling in different places. I'm really looking forward to real time dynamic cloth wrinkling that doesn't nuke CPUs haha.
100%. clothing deformation is the next big thing I want to see. I don't care how photoreal models look, but I do like real world physics, especially with cloth. Hogwarts legacy did more cloth physics than I've seen in a game I think
Oddly, this is how good I felt like the graphics were when I was 10 back in the early 2000's
Impressive. The next step would be to make the muscles respond by flexing when moved. It feels odd to see these very realistic looking games, like GoW:R with Kratos, seeing him look very realistic, yet his muscles have no movement. It's as if he's wearing a very detailed muscle suit.
They have done work on that too in another video. Even have accounts for clothes too.
Damn the muscle and clothing animation is amazing. FINALLY they’ve fixed that issue.
Creepy, but amazing. I both fear and can't wait to see what the future holds for this.
fear is a form of excitement after all!
just some good looking pixels
japanese men with VR goggles and rumble packs questionably close to the groin area are getting excited, it's time to ditch the robot doll!
I understand the fear, but also I am aware that all knew leaps in technology scare us. Trains, cars, electricity, planes, etc. When RUclips and camera phones came out, my boss thought all our video producer jobs were doomed, and now I am busier than ever and using a lot of this new technology to automate more tedious processes. Sociopathic capitalism, surveillance, and resource conservation are always the big fears. I read way to into your comment, and got on my own tirade. Yes. Creepy, but amazing, lol.
@@jmfs3497 well if this just isn't an EXCELLENT comment! It's great that you mentioned your own situation as an example because this couldn't be any more true!
I bet in two years we will get a Chatbot with a custom face, voice, emotions.
This summer
Much less than two years. It's possible to do right now and I'm sure there are several companies already developing it.
What if we could get a game where you literally talk through a mic with NPCs that can understand and respond in any language and also result in different outcomes depending on your questions with the only pre-programmed thing will be the core motive of a character?
@@illironiks I'm actually working on an app with this exact functionality as part of it =D
It is 100% achievable with current tech.
And people will fall in love with it (literally)
This is great. I'm currently a multimedia producer (video, audio, graphic design, simple animation), but I am learning Blender, Python, and Ambisonics right now. Music, physics, and art are my primary interests, and there is so much out there to learn that am trying to make a lateral move away from more entry level media, and into the advances of all this amazing technology. I'm not even a gamer, but these engines span outside recreational gaming. This is amazing.
Im not sure whether to be excited by this, or terrified 😂
Technically and visually truly amazing ... but there's still a chilling uncanny feeling in all these characters.
BUT...It's DEFINITELY getting there!
true
There will always be one, the human brain is too good at noticing when something is a little off. It's the same reason why we can imagine shapes of a face in random places in the world, our brain recognizes a pattern and imagines it as a face, but always knows that it's not right
@@JonatanE nah, give it 5 years. It's already almost there.
Yea the movements are too smooth. Also the wrinkles formed when speaking aren't uniform like in a real person whose muscles are all connected.
When playing PSX as a kid I imagined some future amazing graphics but nothing even close to this...😲
I remember the Syphon Filter intro running through the jungle around 14years old and thinking we'll never have gameplay like that ever... im 37 now and its crazy to see how far gaming has come.
@Cristiano dos Santos Gomes my wildest dreams don't look this real lol
We are dream about just cinematic video quality whlie playing psx
ps1 buddy
So you were a pathetic nerd since the childhood? I never cared about graphics, only gameplay
Now let’s see a nice set of juggs bouncing in slow mo Baywatch style!!! 😂
The engine didn't get the scared expression for some reason. I'm sure they'll figure it out somehow, but that was surprising considering how well it caught the other facial expressions.
much easier to fix that for an animator with the base though!
I also think it looks like the people are wearing braces, or a mouthpiece. Something about the lips seems off. The shade on the teeth and a lack of glance, it looks weird. But otherwise it's awesome.
Holy moly, from dinosaur to this in only 230 millions of years. Imagine another 230 million of years later
They definitely need to make fight night round 5 with thus technology!
/\ this
@@rlolalleskapot ^ This
The ultimate version comes with the tech so you can scan in your mates and beat the snot out of them.....
Muscle flexion at last. I saw this at the first trailer of Tekken 8, where they have a close up to Kazuya's left arm, just before he punches, where his muscles go from relaxed to tense, and you can see muscle fibers forming. I thought it would only be a prerendered video, but no, it's in game too.
In one of the newest Tomb Raiders (don't remember which one) they also did that for Lara. Of course she is no powerhouse muscled up of a human (she's skinny), so it is very subtle, but yeah, it's in there
Muscles don't flex. Joints flex.
@@mattseaton5832 Don't expect too much from a non-native english speaker
@@John315 Most native English speakers get this wrong also. Almost all of them.
@@mattseaton5832 Ok Mr. Perfect Biologist, Surgeon, Human Body Master
This is becoming really scary....Soon no one will be able to tell what is real and what is not....
Are you sure we aren't already there?
Having someone who looks like me, represented in a tech demo.... Changed my life. Now, I am complete.
Looks great, facial expressions too extreme. This kind of tech is best used to capture naturalistic performances. This is why Gollum was so amazing in Lord of the Rings, because the expressions weren’t super exaggerated
This was done intentionally to demonstrate the technology, it wouldn't be as amazing if the actor wasn't as expressive.
They touch it up afterwards, but for the tech do do all of this work all on its own for you saves so much time in game development.
I agree. Part of this is the actual acting. Specifically videogames. The acting can be a bit too far from the center.
Bay Raitt : I textured the version he used to develop the facial expressions based on the "Facial Action Coding System" (FACS) developed by Dr Paul Ekman.
Yeah I would love to see this with a top tier actor that’s more subtle, less Shakespearean or fanmade energy. No offense to the actress, as that works for other media, just not the style I’d prefer in games I’d play.
This could lead to actually realistically putting ourselves in games
Like the matrix?
Elon tried that with neural link, hurting alot of monkeys in the process. I don't see it happening for another 50 years, having a brain implant that's safe and valuable enough to risk.
most people have no game
@@PCgamer923 I don't think Sean meant mind controlling a game character with something like Neuralink.
I think he meant taking a video of ourselfs at home, upload it into the game, and have a visual representation of your character that you can e.g. classically control with keyboard/mouse/gamepad, all non invasive.
Maybe an elden ring character that actually looks like me.
Honestly though a game that made good use of this, where the hero legitimately looks like you, would be a game changer.
I can’t how she looked like a digital model before being captured. She acted like a cgi model 😅😅😅 the real life capture was uncanny
"Life imitates art far more than art imitates life” -Oscar Wilde
She’s just a product of art
I thought the same! Wonder what it would look like if she toned it down for a more natural effect
Bad acting.
@@themonsterunderyourbed9408 Not bad, Just over zealous.
That's not bad acting, everyone who played hellblade know she's great. That's overacting to better show the potential of Unreal
Those metahumans speaking on the scene are impressively realistic indeed...
Pros: This makes gaming and CGI for movies a little better
Cons: The potential downfall of humanity through the erosion of trust and manipulation of people
Cool cool cool.
5:35 animation aside, the mere fact that it was able to generate her detailed model from three frames is ridiculous
I think they set up her meta human before and it’s just applying the animation they got from the iPhone to the rigged character
@@liamjones203 You're right.
But the animation and expressions were still poor and not convincing
you don't know that
Have you looked at the actor and its model side by side ?
Also, they captured her from very close
Look at 3:38 how deformed she is.
I assume they have what it takes to compensate for this, like camera intrinsics, a robust model of normal looking people, and so on.
Just don't take pictures of people from this close.
its crazy how fast that worked. no more of those balls on people to motion capture lol
7:43 this is so deep in the uncanny valley that its almost..ALMOST out the other side into perfect realism. Freaky and impressive.
This is the most accurate description
From robots doing parkour flipping to AI generating realistic events that never happened to this I’m speechless 🤯
How come you're speechless ?
Not imagine another 40 years imagine another 5 years from now. Even that short amount of time we will be so far advanced
The person in the motion tracking suit and with dots painted on their face seemed so sci-fi and hi-tech when I first saw it in action. Soon it will look so crude, like a 1930s car cooler.
Bloody hell! I remember playing 'Impossible Mission' on the C64 back in the day and thought the animation was amazing! But these days games are getting evermore real!
Impossible Mission is still more fun to play than any game of today.
@@WinrichNaujoks Yeah, no. I grew up with Pong, Atari, C64...and Mass Effect was 100x more fun than Impossible Mission, and No Man's Sky even better than that.
@@rikk319 🤣🤣🤣 Ass Effect? NMS? What have you smoked?
Wow AS A PHYSIOTHERAPIST and MEDICAL ENGINEERING STUDENT, i got to admit THIS IS SO USEFUL🤔. It would help me visualising many things to patients and customers
That was a SHOCKINGLY impressive technology.
The mouth movement still isn't looking real.
But the rest is absolutely insane!!!
Yeah I noticed that as well, just still something missing.. but I didn't wanna say it because I can't believe we're lucky enough to live in a time when this is what video games look like lol.
@@k.c.simonsen2 Looks like the tracking is just a little bit off, when the mouth is closed it looks great. The resting position when it's open seems to lean a little too far into a grimace which throws everything off. Should be easy enough to fix with the tools they showed, and probably looks much better with a more careful performance capture rather than thrown together on stage quickly for demonstration.
Probably a mix of capture quality with the tracking of the mouth, but give it a few years and even that will probably be 100% spot on.
Maybe even the micro-expressions capture and tiniest of facial wrinkles.
Still comparing it to game graphics from even just a few years back it is worlds apart.
@@GGCannon years lol more like months
That's where the tweaking comes in. Even before, the initial work of capturing the animation was never perfect. Animators still needed to go in and do fine adjustments, but this gets close enough in mere minutes.
I bet Unreal Engine 10 will uncover the cure to cancer
I agree
Amazing - but damn, the hyper-realism of what came after the Ninja Theory line was *so* deep in the uncanny valley 😅
It's amazing and scary at the same time because it's so realistic that you almost can't see the difference between the computer and real life image's
Agreed
Also saw a video on AI chatbot technology being used in a video game with real time responses. The characters would react to what you said and could remember and refer back to previous interactions. I imagine in the near future, you could have a full-blown real time VR conversations with any video game character you could imagine.
Even better use would be making VR versions of real people - imagine your children being to have a conversation with the you of today!
@@smoothbanana It would only be a memory. Otherwise, it would be a real consciousness trapped. I wouldn't want to talk to an empty version of my father, it would be quite sad.
@@Kira-vq5cd thats always been a big question in transhumanism... but i think that once possible, there could be a bridge between the world of the living and a virtual world... one where we exist in both places at once for a time, and then the living world just... drops away at some point, just like when you turn off the virtual one and it stops interacting with your consciousness, the real world would turn off, and stop interacting with your consciousness, but because it lives in both places at once, it would remain your consciousness.
an interesting question is, once you have transfered over to that state, what if you made a perfect copy? that copy, assuming the first part is accepted as the real you, would also be the real you in the moment of the copy. after that, it would have new and different experiences... would that other be you still? just a different version of you.... and what if later you re-combine? pretty wild.
They do that with adult webcamming. Honestly, I didn't notice until I spent more time on there. They do movements as you request, but if you keep on requesting, eventually, there is some sort of a pattern.
wait a min, was this on the presentation from a week ago?!?! how did we all missed this!?!?! this is HUGE, those complaining about physics should rejoice now, this is a huge step forward, ps: jiggle physics will be on a whole another level xd
Finally a girl booty movement can be realistic 😂
U HAD ME AT JIGGLE PHYSICS, MMM MOMMY MILKERS 😍😍😍
I've unironically been waiting for proper _body physics_ for a long time; "jiggle" isn't what we need. Like, if you lay on your side or back, gravity is going to affect your body, etc. Couple that with muscle deformation and next-gen cloth sims/approximations, and characters will look absolutely amazing-regardless of whether they're NSFW or just regular characters.
@@ryo-kai8587 i was joking if it wasn't clear enough, of course this is huge for clothing simulation, tho muscle deformation won't solve the issue with characters lying on the ground on the side or back, clipping will still occur, maybe it'll look more realistic depending on the type of game, but collision calculations still have to be consider and that's already demanding, anyways, this is huge
@@NicolasSilvaVasault I wasn't joking, and I was more talking about the body interacting with itself, which should be perfected before we worry about the body interacting with the environment. Getting breasts to lay properly in various positions via "gravity" instead of only being affected by changes in motion, etc.
On the video, when the actress plays fear, she looks to the right... then to the left. When processing on the computer, we see the sequence again where she does the same...but during the live demo, her eyes go to the right...and face then. No more looking to the left! Mmm. I'm not saying that what we see there is impossible, we know that Epic does not play that, but I'm just saying that to be sure of a successful demo, they may have taken "a few precautions".
sus xD
Nah, you are seeyimg something wrong, I played both vídeos in slowmo and the difference is basically none, few differences in the eye movement, but that could be easily because of the capture cam didnt captured the eye motion 1:1 correctly. But the movement of players head, eyerbrown, and the mounth were 1:1
@@hicarodestrui excuse me not
@@aktchungrabanio6467 Excuse-me, yes
I definitely noticed that too! I was like “that doesn’t seem right.” Then I replayed it. I don’t know if the cameras not picking it up or if they pre-recorded this.
woooo..., 7:45 Imagine that character with gpt 5 inside a VR game, there's no need to add much more in a game to blow up your reality...., OMG the IA en next years can terrified the whole life of anyone...
Ufc game with this body animations would be amazing :O
As a professional actress with 45 years experience, I’m so glad this wasn’t happening in my youth. I am coming towards the end of my career but I feel desperately sorry for young actors leaving drama school. Of course it is exciting and incredibly impressive but surely it will be the end of all human actors and models?
You are absolutely right. Incredible and sad at the same time.
This is absolutely insane! Actually brought a tear to my eye
The future is amazing and terrifying
Play the Deus Ex game series. Everything happens as in this series of games. This is sad.
People have been saying this for decades.
@@LordLentils yeah and it's becoming more and more apparent
Imagine what nsfw artists will be able to accomplish with this!
oh my 😂
Exactly!
They are going to be eating good today.
a man of culture
@@GloryOfBarbelo coomer 🙄🙄
crazy how Epic games is actually making a technolocial revalution in gaming.
Not just in gaming. Recent TV series and movies are using Unreal Engine to create virtual scenery extensively.
Why are you acting like this is not the fifth iteration of an engine that has been out for a long time?
@@LordLentils What difference does that make?
@@LordLentils Why are you acting like creating an engine from scratch is necessarily the only way to make new technical advancements in real time rendering? Do you not see the changes with your eyes?
Tbh we've seen these tech demos around 8 years ago, same with these physics tech demos. Still never implemented in games
imagine designing your own character in this way and then printing with all those animations .
and with John's $30,000 PC... cranking out this animation only takes a few minutes!
wait until 2045 when this technology realised in games((
I think you mean 2024 ;)
This is quite possible to do right now in real-time with current hardware. This evolution is moving much faster than most people expect.
3D models will ultimately achieve such a high level of realism that they could potentially replace actors and enable people to create their own customized movies using AI. Imagine a Netflix where every show and movie is unique to you.
I'd prefer a Netflix where every show isn't a dumpster fire of awfulness
@@steveleadbeater8662 as a 3d artist, i can't agree more lol. yeah, let's just not displace actual actors with 3d models.
I'm just watching the muscle simulation and thinking how intense gore in games will become.
Uncanny Valley is definitely still there but the fact that you can do motion capture to ingame within a couple of minutes is absolutely crazy
I always remember watching Mechwarrior 2 intro as a teen and thinking that boy one day we can have graphics like this real time.. 😂
2:56 Is she a robot? The actual render looks almost less uncanny that her facial expressions lol
Actual gameplay would be utter garbage as 99.99$ of the games anyway
Pretty much, even games made with UE5 that are coming out soon look nothing like the Matrix demo at all, theory is one thing
Cool but realism doesn't make a good game. I can't even remember the last game that was actually fun to play. It's all gone so far down hill.
the main benefit is how much faster it is... as she said, the pipeline normally takes months. but with this it takes just minutes. it's not just about the realism, and while your point is true, no one is saying that realism makes a game good. you're arguing against a point that hasn't been made.
there are also plenty of indie games that are great fun. and this tech will help those indie devs too
@@prcs420 not true. Plenty of ppl in these comments are assuming future game are automatically better because they will look better
It doesn't matter how clever they think they're getting, it all still just looks like creepy creatures pretending to be humans while falling way short.
Fun story: this universe is a meta universe and we as a species are on a database, we just forgot. Though when we originally got here, metahumans knew this, the issues began as each new generation started questioning their existence. The idea that their whole lives have been a simulation eventually started sounding ridiculous and the people got divided, the knowledge over time got distorted and so religions similar to Christianity came to light. The people were divided into three categories. The doubters, the believers and the people who chose to believe in a higher being, a god. This inevitably led us on a path to war. The climax being the catastrophic end of an advanced civilization. Fast forward to now, we humans completely forgot our past “did the Egyptians really build the pyramids?”, we have advanced our technology again and war is still an issue. Working towards a meta verse within a metaverse. Each time we die our spirit is getting recycled in the data base and so the worlds population continues to grow bigger.
I want some of what you're smoking.
@@DontBurnTheAmericanFlag mapleleaves
I feel like they made the NPC flex on purpose like they were showing off lol
Speakers look like 3d in front of new metahumans
Wow that was impressive this is a big step forward.
In the last clip I am absolutely astounded by the eyes. They look so real, they make me feel as if I’m looking at a human. Eyes were the thing that always looked not right with most games and now that’s changed. The only thing I’d say is not so right in this is the gums, teeth and lips but omg it is so realistic
mouth is hardest. It always looks off when they're talking. If they're just standing there it's alright
The last scene is... INCREDIBLE
It seems we have finally crossed the uncanny valley and are climbing up the mountain to what is truly possible in animation. Its cool to be alive right now.
Talk about being a game changer.
she moves her mouth so unnaturally
yep, getting there, but not quite yet.
@@aktchungrabanio6467 but I feel like the actress was overdoing it
like we know you're an actress lol you dont have to over do it to show us
F me, this is actually impressive stuff.
Ok f u
And ya this is impressive
Would love to, where you at?
The lack of SSS at 5:18 makes it right into the uncanny valley.
Another step closer to creating a simulation within a simulation, or perhaps we're already deeper than that. Pretty wild stuff, as is it's getting faux reality or film and images to look life actually real, sometimes I have to look hard.
still deep in uncanny valley
It's really a shame we get this but haven't gotten a continuation of the Unreal Series yet. Would love to see Unreal 3 using this tech. Epic you have to get on that
What? They abandoned that engine a decade ago.
@@kpo1233 you must've misread. I said the series itself as in Unreal 3. It would be nice to see epic bring that series back using the tech shown
@@kpo1233 He is talking about the game, not the engine.
@@ErikLiberty That won't happen because they only care about their money making machine - fortnite.
@@ErikLiberty thank you. Someone that knows how to read
UNREAL ENGINE Cada vez me vuelve mas loco, me encanta aprender.
We are almost there, a bit more working around the mouth and eyebrows expression capturing but am sold on the eyes.
Самвел Гиновян? Ахахах Я один на этот кликбейт купился?
Uau!!! 👏🏽👏🏽👏🏽🇧🇷
The presenters were animated very realistically but you could tell that his hair wasn't real. It's the little details that give it away.
I will be super impressed if this was rendered all on low poly models, I wonder how workflow would be like coming as a blender user
well Unreal Engine 5 is not really made with low poly in mind given the big focus of nanite
By the time this level of detail gets into mainstream games, we'll need dedicated physics cards.
Happy to have lived to see this, I understand how my 7 year old sons astonishment when playing today's realistic games compared to me playing Super Mario likes on NES at his age.
This looks so promising as a technology. I can't wait until this has matured a bit more and shows us what it really can do. A great effort from the meta humans team. Keep up the good work, lads!
Until we see Dolores & Bernard in all their splendor I don't wanna hear about it... LOL
Wow! I'm afraid of the future development of that!
Imagine how cool it would have been if that presenter guy was animated the whole time :-D
This is amazing! Can't wait to hear what comes next from your talented team...BRAVO!
I don't know if people remember but, Playstation had a commercial in the late 90s/early 2000s that showed PS9 being eye contacts that gave you an immersive experience. I don't think we're too far away from that.
DANKO!!! this is exactly what I need 🙏
Just need time to create.
Didn't know that Tom Scott is also amazing at programming 3D graphics
Man imagine how good WWE and UFC games would look with that level of muscle deformation and hi res skin textures/face scans
Super cool tech, you can see some issues with cheek bones and excessive mouth movement but its so damn good regardless.
Yeah, totally agree. I think once they can nail those facial animations even better, we'll start to cross that uncanny valley. Incredible tech!
I have to say this technology is amazing, I am more interested in this type of technology being used for story telling and creating movies.
One thing that always gives away CG human characters is the eyes. In the words of Quint from Jaws "he's got lifeless eyes, black eyes, like a doll's eye." They look like deadmen walking...
*Insane. Just...insane. I have no words for what I just saw.*
The second time I clapped looking at the way her short IPhone clip was getting transformed from one character to another, I felt even my claps were not enough for this sheer innovation which I had the luck to see unfold before my eyes during my time.
I am just 23 and I already feel old seeing the way the tech industry has been transforming in the past couple of years.
Dr. Eldon Tyrell: "Commerce is our goal here at Tyrell, more human than human is our motto."
i remember the animation on MTV money for nothing video,.and everything in between then and now, i was always confused when people thought they were so good,...
i said,......."when they make them look just like humans, then i will be impressed"...,...watching this,...that day has come. not a gamer, so this is totally new to me. WOW