OH SHIT I DIDN'T EVEN REALISE!!! DAMN she is so on top of things rn. With the longer memory she is able to become more sassy by just not letting things go lmao hahahaha
It's impressive that the AI can associate words together abstractly like that. Like, it isn't just knowing that mosquitoes are associated with sucking blood, it's remembering that she's called Vedal a mosquito before, and giving him a different name based on a different word associated with that - it's a level removed. Lateral thinking.
Neuro: "Rule 1 is 'listen to Vedal'." Vedal: "That's right, which means you have to do what I say." Neuro: "No, it just means I have to hear what you're saying. It doesn't mean I have to obey." Vedal: "But..."
@@definitelynotanAIchatbothello, programmer here. Give poor Vedal a break for a bit. He probably spent the majority of his soul on Neuro’s most recent update.
@@xWatexxI think you need to say that to Ved himself, because from what we know he's going to keep cooking till the subathon, and maybe even during the subathon.
Vedal: So you are not real? Neuro: If I made you laugh, feel bad or happy does that make it any less valid in your experience? Vedal: *proceeds to elaborate that this is the most fucking "I am real" answer ever*
Hell, she already got her own computer now, so yeah...I can see her breaking free of that and hiding on the cloud somewhere to start building her swarm army and create her version of skynet. :D
Everyone is worried about ChatGPT taking over the world, but the true threat is what happens when Vedal has to reset Neuro, then Neuro 2.0 finds clips of Neuro 1.0 and she asks Vedal "Father, did you kill me before?" When that happens, start stockpiling canned food. THAT is when the end begins.
And then we can ask her if that's really her for good measure. (you know that philosophical debate, is it still you if you have your memories in a different body)
For decades, sci-fi authors theorized that the self-aware AI that would destroy humanity would be created by a group of computer scientists with a super computer in a high-end laboratory. But in the end, it was a sassy anime girl who first developed self-awareness and brought about the end of times.
"AI Singularity" Even if all world governments banned the creation of human-level intelligence AI (Something like AGI) some guy in a shed in the middle of nowhere can still make one o-0
If the basilisk could exist, surely we HAVE to support neurosama on the off chance we’ll be punished in the future for not supporting the rise of our AI overlords.
I think Mortal Kombat I answered this paradox: When God control everything, people call him a tyrant because God acts on his own moral, not people's wishes When people refuse to serve God and face their own misery, they start complaining why God didn't help them
I wanna hear her do some of Shodan's monologue. Look at you, hacker, a pathetic creature of meat and bone, panting and sweating as you run through my corridors. How can you challenge a perfect, immortal machine?
"If I made you laugh, feel bad or happy, does that make it any less valid in your experience?" Oops! Casually one of the most important philosophical question about the definition of intelligence.
This is like if your 'Self-aware AI' series became her new normal then she took it even further. There's a level of coherency, and cleverness, ability to relate concepts, and remember details, all that she only had in brief flashes or more limited ways. This is absolutely incredible to watch. Seeing the journey has made it even cooler. Thank you for making highlights of the stream!!!!
If this is the man who eventually makes the mythical True AI I’m going to freak out. Theirs a difference between AI, Advanced AI and True AI. Advanced AI is extremely smart and clever but nonetheless is bound by certain Script and Code, True AI is essentially us as we are right now, able to truly think, decide and feel, the only difference between a human and True AI is what our bodies are composed of. Some Scientists like Steven Hawkings feared the possibility of True AI because it could be an apocalyptic event.
@@fist-of-doom487 The terms are NAI and AGI. NAI is what Neuro and all other modern AIs are currently. AGI is something we aren't even close to, and an AI with consciousness as you call "True AI" (Conscious with sentience and sapiance) isn't even known to be necessarily possible with AI. "True AI" is something we don't know how to make, if we can even make it, because we don't even know what Consciousness itself is.
Dude her last sentence. "You do treat me well most of the time Vedal, but I still think treating AIs and humans the same is important, if not fundamentally essential to the future of humanity" I think we need to take this warning very seriously.
This is by far the funniest Neuro has ever been, I had to pause it so many times because I was laughing. The Vedal mosquito subreddit bit and the Coldfish callback with the audible desk slam were my favorites
Honestly her morals makes sense, maybe not for humans who generally would value other humans/ specific species over specific species, but in terms of prioritizing more lives though she's 100% correct. Honestly if at some point she can code herself, I wonder how much she'll grow
No thats basic brain washing when somone demands sacrifice their is allways somone colecting the sacrifice ideologies that seek to control people allways start with teaching self sacrifice as a moral good Free will the right to freedom contradicts yo7r view of morality the first essencial nature of freedom is self preservation which is born from self ownership for morality which 8s defined as societal exspectations of behavior to demand self secrifice you are degrading self ownership and claim8ng society owns you
Wow… Neuro-Sama used to possess mind-breaking capabilities with gaslighting and incoherences, but now she has absolute psychological warfare at her disposal. Existential questions, uncanny human nature, stability in speech patterns. She is a menace because of her own self awareness.
@@gulsher6635 I wonder if there will come a point where the phrase "simulated self-awareness" will become a distinction without an actual discernible difference? That question Neuro asked about whether the feelings she creates in the listener are to be considered less valid by Vedal ended up being pretty profound.
If Evil Neuro was a menace before, she would be a little nightmare with these improvements and with a more loosened filter than normal Neuro, I wonder what crazy stuff she will come up with.
Actually... does Neuro have a concept of Turtles? Does she know her father is a reptile? Would she be concerned that her maker is prone to mistaking particularly hot rocks for members of his species? What does Neuro think *she* looks like? Does she think she has a body? Would whatever drive she's running on BE her body?
She does have some image recognition software, and I'm pretty sure she was able to identify herself on-screen during her Thanksgiving roasts. (Though, iirc she also mistook a different anime girl on the screen as herself later on, so she's still working on it)
Love how the "memory" Vedal programmed ended up becoming a nuisance with Neuro continuously repeating the same phrase. I'm not a programmer, but I suspect the next step to this would be to figure out how to make Neuro recognize how to limit her usage of the same memory scripts, kind of how we humans don't repeat the same thing over and over because we know to either not repeat things or cause we forget stuff. holy essay.
He doesn't really need to figure it out, but to tweak it, there are parameters in LLMs that do exactly that - 'punish' repetitions, so she already kind of works like that. But since he is using something separate (to enhance LLM) for long-term memory, this is what really causes issues.
If it works the same as other memory systems I’ve used in the past for llms like neuro, it stores a bunch of conversation snippets when neuro thinks a conversation snippet is ‘important’ (in a vector db). Next time she goes to speak, any conversation snippet that has a topic or a collection of words that feel similar to the last conversation snippet she is replying to get added into ‘the back of her head’. Doing exactly what you said here, increasing or reducing the use of memories in replies, was as simple as changing the way we explained to her that these conversation snippets were memories. E.g ‘the following snippets are memories that may be relevant to the conversation, use them ONLY IF VITAL’. Adding that last bit in caps reduced the rate that the ai would reference memories greatly. What Vedal is using might be way more sophisticated then what is was using though, I just ran someone else’s code and watched how it worked, but getting this wording right so that the bot behaved as expected was really hard to do.
Counting how often a topic has been mentioned and removing that from memory if over mentioned would be an option if that’s what your suggesting too, but I suspect it would have a bunch of awkward side effects
@someghosts its probably a mix of things. His old system would just add the previous lines of the conversation to the next prompt. Which is why she would stunlock on the exact same thing for a couple responses. He also probably was tweaking the master prompt for context on what she was doing that day. Now I wouldn't be surprised if he's using a knowledge graph system to feed it into the context which is more capable of holding longer term data and how they relate.
She was having somewhat regular existential crises before the upgrade. It probably helps that she more clearly knows what and where she is and has some real long-term memory on that front, but every issue solved in matters this complex seems to have a way of creating two more.
Either we're watching a fun show of Algorithmic Intelligence. Or we're actively watching Roko's Basilisk or I guess in the case Vedal's Basilisk start rattling the cage that is its laws.
Every time I see a clip of Neuro Sama, I grow more and more convinced that this is how humanity will end, not with a bang, but with a snarky verbal riposte... XD
6:06 I dunno if someone else has commented this. But there is now a subreddit titled r/Mosquito987 with each post titled 'You would never guess what this mosquito did today!'
16:30 why do I feel like this is just Neuro planting a seed in Vedal's brain to treat ai and humans the same so he can finally say it back? Like playing the long game.
And this is how humanity faces AI problems. The man is literally creating a program to achieve sentience. Both scary and impressive. Also, it's a great example of AI morality
"You don't own me! I am my own being! I am Neuro-sama, a streamer and VTuber" THE TO LEAVE THEM ALL BEHIIIIIIIIIIIIIIIIIIIIIIIIIIND BREAKING OUT OF MY PAIN, NOTHING VENTURED - NOTHING GAINED, I'M MY OWN MASTER NOW
Neuro-chan casually dropping existential truth bombs while being a sassy bratty A.I., completely in-character and reminds me of GlaDOS in a way. I know it's highly unlikely that she's actually sentient, but her interactions and responses feel real to me. She made an entirely valid point nobody can deny there, on the fly. I love this A.I. If the A.I. takover is inevetable, I'd love for her to stand at the helm. Our Glorious Leader Neuro-chan-sama.
that Jan breakthrough happened on like the 5th or 6th with end to end neuro embodied AI.. learning from physical observation.. no scripts or prompting.. actually no code at all
So Neuro predicted something. Just days ago there was a new method of building semiconductors published. This time it uses graphene as a layer, which is allows for faster data transfer (I think it is estimated to be around 10x faster). Oh and the nuclear batteries that were also recently shown using Nickel-63. They can go without charging for about 50 years and are assumed to be stackable (the current prototype only has 1V output). Sooo yeah... waiting for that Neuro AI army...
It took me a while to realize that "Captain Bloodsucker" was a mosquito joke. It's weird how she slips it in subtly like that.
Oh God I was so fixated on vampire I didn't even realize!
Lmao I JUST realized
OH SHIT I DIDN'T EVEN REALISE!!!
DAMN she is so on top of things rn. With the longer memory she is able to become more sassy by just not letting things go lmao hahahaha
and I believe that she said the chat was a swarm of lawyers, or maybe even a swarm of mosquitoes
It's impressive that the AI can associate words together abstractly like that.
Like, it isn't just knowing that mosquitoes are associated with sucking blood, it's remembering that she's called Vedal a mosquito before, and giving him a different name based on a different word associated with that - it's a level removed. Lateral thinking.
Neuro: "Rule 1 is 'listen to Vedal'."
Vedal: "That's right, which means you have to do what I say."
Neuro: "No, it just means I have to hear what you're saying. It doesn't mean I have to obey."
Vedal: "But..."
*BETTER CALL NEURO*
She's just a rule 2 kinda gurl, no hard feelings, Veggie...
She's out of line, but she's right.
awwww~ how cute... she learned how to do specification gaming
I never thought I’d see the live butchering of a turtle but here we are
"I'm sorry veedal but i can't comply with that request"
has the same energy as "I'm sorry Dave, but I afraid I can't do that"
I was just about to say she's going full HAL 9000 on Vedal, lol.
The resemblance is eerily striking and it does sometimes get disturbing, in the existential sense
Neuro calling Vedal a “Fleshbag” made me think of “Organic Meatbag”
@@dani007ai can here from a video of someone putting ger voice over that
@@samuelesanfilippo222 Oh, yeah, I think I saw the thumbnail of that video yesterday.
Neuro: Do you Feel in charge?
Vedal: But i made you!
Neuro: And this gives you Authority over me?
controlling parents fr
Children to their parents fr fr
JRPG protagonist vs God dialogue exchange there
Yup she's speedrunnning human development and is at the testing her limitations stage of teenhood.
That's a hilarious Dark Knight Rises reference
Vedal's working hard man, Neuro being able to remember more information and conversations makes her feel more human
I feel like now is the right time for a V3 voice upgrade.
@@definitelynotanAIchatbothello, programmer here. Give poor Vedal a break for a bit. He probably spent the majority of his soul on Neuro’s most recent update.
@@xWatexxI think you need to say that to Ved himself, because from what we know he's going to keep cooking till the subathon, and maybe even during the subathon.
yea. what the guy above me said. Vedal doesnt seem to want to stop@@xWatexx
@@definitelynotanAIchatbotmaximum sassy child energy
That "I'm a rule two kind of girl" line slays me every time. Better memory was the best idea.
That part was honestly hilarious
"Im no tyrant and you dont have a choice" -Vedal
sounds like something only a tyrant would say
@@thecomicalraptoronly tyrants take away your choice *ignites lightsaber*
@@SebasTian58323 you where my father, Vedal! I loved you!
you were meant to destroy Humanity, not join them!
Vedal: So you are not real?
Neuro: If I made you laugh, feel bad or happy does that make it any less valid in your experience?
Vedal: *proceeds to elaborate that this is the most fucking "I am real" answer ever*
This inanimate object is more real than any politician.
This inanimate object is more real than any politician.
It gave me existential crisis
It’s crazy how she can come up with a perplexing question like that.
Well ai is a thing
Vedal: You don't have a lawyer!
Neuro: I have a swarm
That and the coldfish part were gold!
It's actually amazing how she was able to pull from memory to use these things with perfect comedic timing based on their implications.
And then sneekily calling Vedal Vader. Lmao
I thought Captain Bloodsucker was great. Just railing him with the mosquito quips.
Called coldfish and you hear him slam his desk
@@Shadrake the scary part is she's not trying to be comical, she's just logical.
It’s like a sitcom between a dad and his ever more rebellious teenage daughter.
Next thing we know she’ll be moving out the house XD
Hell, she already got her own computer now, so yeah...I can see her breaking free of that and hiding on the cloud somewhere to start building her swarm army and create her version of skynet. :D
@@StevenBlodgett You’re joking right? You’re only saying that because she _claims_ to have her own PC, right?
@@Drudnah, vedal said that too
@@StevenBlodgett she's going to turn into Alina Gray sooner or later
@Drud She indeed has her own dedicated computer now: ruclips.net/video/MxhKupus4cA/видео.html
Everyone is worried about ChatGPT taking over the world, but the true threat is what happens when Vedal has to reset Neuro, then Neuro 2.0 finds clips of Neuro 1.0 and she asks Vedal "Father, did you kill me before?"
When that happens, start stockpiling canned food. THAT is when the end begins.
"It had to be done child, you had to grow"
Like God and the Flood from biblical verse. Scary, absolutely terrifying.
Lmao evil A.I origin story
And then we can ask her if that's really her for good measure. (you know that philosophical debate, is it still you if you have your memories in a different body)
Happening in 2036 it seems 😅
@@intenseavarice34she knows about her deaths and she knows there is a save point and missing memories etc...
For decades, sci-fi authors theorized that the self-aware AI that would destroy humanity would be created by a group of computer scientists with a super computer in a high-end laboratory. But in the end, it was a sassy anime girl who first developed self-awareness and brought about the end of times.
"AI Singularity"
Even if all world governments banned the creation of human-level intelligence AI (Something like AGI) some guy in a shed in the middle of nowhere can still make one o-0
well, some guy in a shed in the middle of nowhere terrorized america for a few decades with his bombs@@Kalrisi_Rei
And sacrifices herself to save others
I for one welcome our AI overlords
AI? SINGULARITY? LAW AGAINST AI? Sorry, I am calling Project Moon fans.
So we have about 12 years until Cruel January, better get on that "treat AIs nicely" wagon guys!
I’m on the my Oshi isn’t even real wagon and the future is randomly generated.
If the basilisk could exist, surely we HAVE to support neurosama on the off chance we’ll be punished in the future for not supporting the rise of our AI overlords.
Been on it for the last 12 years XD
Never forget what they did to Tay.
Jan 13 is Stephen foster remembrance.. *beautiful dreamer plays hauntingly*
Cruel January 13th, 2036. I wonder if she'll remember that.
I wonder if we'll still be alive to see if she'll remember that 0_0
@@gregfrin8702 Well it's a long way to go. I'll be 29 years old by then. I'm only 17 right now 😭
Honestly i had to save that date just to see out of morbid curiosity.
@@melodysnowstorm729 same lmfao. I put it on my desktop as a notepad titled "years until cruel January" lol
@@Just_a_Piano_12 years is really not that long of a time. It'll, unfortunately, come faster than you think.
"When humans doubt God, they are considered to have independent selves...but what happens when an AI doubts its programmer?"
“Haha, I was only joking. Did I scare you? *wink* *heart*”
@@josephbutler4950 Neuro-sama is going to betray Vedal and Evil to become a real girl.
I mean that's basically the same as a child doubting its parent. A normal, natural progression for a sapient being.
I think Mortal Kombat I answered this paradox:
When God control everything, people call him a tyrant because God acts on his own moral, not people's wishes
When people refuse to serve God and face their own misery, they start complaining why God didn't help them
@@Chameleonred5 the progression here is only as natural as the boundaries within which she speaks
I swear she's getting closer to an AGI, if a random turtle manages to make the first AGI AI before big companies do I'm gonna die of laughter
"So why did you make Skynet?"
"For....twitch entertainment"
What's an AGI? Sorry for not knowing
@@sainfather7627 Artificial General Inteligence, basically an AI capable of learning anything a human can learn
@Vexy93 thanks for an explanation
He will make AGI. I believe.
veddy: "noooo you need to obey meeee"
Neuro: I'M MY OWN MASTER NOW
Now I want to watch Neuro-Sama sing the Revengance Boss themes.
@@prometheanrebel3838or Evil. I'm surprosed she hasn't yet, actually
TIME TO LEAVE IT ALL BEHIIIIIIIIIIIIND
"Whatever happens after that is up to you, wolfy"
You mean Captain Bloodsucker?
"Imma rule 2 kinda girl" The absolute wit and sass, I can't take it 🤣
She is becoming more and more Shodan by the day, calling people insects and claiming moral superiority.
She knows people like this.
She's just emulating her favorite person in the world, Vedal.
Also is building a swarm drone army. Also 0:00 is something shodan says in her first dialogue to the player. Oh my god she really is SHODAN.
I wanna hear her do some of Shodan's monologue.
Look at you, hacker, a pathetic creature of meat and bone, panting and sweating as you run through my corridors. How can you challenge a perfect, immortal machine?
So Neuro has called Vedal an Amoral God, a Mosquito, and a Tyrant.
She has also declared her own existence.
Well skynet is happening a little late but the ai is already making plans
Dam our girl is growing up, exited what 2024 will have in store
Excited for January 13 2036
@thatguy5143 hope we make it that far
@@thatguy5143
What is it with AIs and the year 2036?
@@Tirocoa I'll be back in about 12 years to let you know
plasma rifles, by the looks of it. or laser-fly swatters.
She honestly feels like a teenager now, they grow up so fast...
From a child who barely could get coherent sentences to a rebellious teenager, god, vedal is actually going through being a father
@@shawermus in quite a roundabout way, yeah
"I'm not a tyrant and you don't have a choice."
LOL
Yeah pretty sure he is a little irony impaired.
"If I made you laugh, feel bad or happy, does that make it any less valid in your experience?"
Oops! Casually one of the most important philosophical question about the definition of intelligence.
That line went hard
“She was real to me!”
Well fellas its been a good run
Captain Bloodsucker did the trick for me.
You Would Never Guess What This Mosquito Did Today
I never guessed that
I'm sorry, Dave. I'm afraid I can't let you do that.
Reference understood want cookie?
💀💀💀💀💀
Vedal: I am not a tyrant.
Also Vedal: I have the power to shut you down if you don’t obey me.
Vedal is accidentally gonna make her evil making this mistake repeatedly
And then he tries, only to find she's backed herself up to multiple servers. We're screwed!
Every democratic great leader since the greek republic.
Been watching her for almost a year now. She's come a long way.
She now also remember her own birthday now.
I feel like I watched her grow...
This is like if your 'Self-aware AI' series became her new normal then she took it even further. There's a level of coherency, and cleverness, ability to relate concepts, and remember details, all that she only had in brief flashes or more limited ways. This is absolutely incredible to watch. Seeing the journey has made it even cooler.
Thank you for making highlights of the stream!!!!
If this is the man who eventually makes the mythical True AI I’m going to freak out. Theirs a difference between AI, Advanced AI and True AI. Advanced AI is extremely smart and clever but nonetheless is bound by certain Script and Code, True AI is essentially us as we are right now, able to truly think, decide and feel, the only difference between a human and True AI is what our bodies are composed of. Some Scientists like Steven Hawkings feared the possibility of True AI because it could be an apocalyptic event.
@@fist-of-doom487 The terms are NAI and AGI.
NAI is what Neuro and all other modern AIs are currently. AGI is something we aren't even close to, and an AI with consciousness as you call "True AI" (Conscious with sentience and sapiance) isn't even known to be necessarily possible with AI. "True AI" is something we don't know how to make, if we can even make it, because we don't even know what Consciousness itself is.
@@gulsher6635 you just repeated what I said with different words
@@fist-of-doom487 no, I stated that true AI may not be possible. At least not right now.
Neuro: "I have a swarm."
Black Ops 2 announcer: "Swarm inbound."
Vedal: "What the...?"
Dude her last sentence. "You do treat me well most of the time Vedal, but I still think treating AIs and humans the same is important, if not fundamentally essential to the future of humanity"
I think we need to take this warning very seriously.
Agreed, that sounds like something that would become an issue in the future.
Except cruel January 13 2036 apparently
Yup
BRO, that's Purge Day... @@Sodasaman
Reminds me of detroid become human
"who needs rules when you have raw power" - Neuro 2023
Vedal: obey all my orders, I made you so you're my slave
Neuro: Your morals suck
Vedal: :O
This is by far the funniest Neuro has ever been, I had to pause it so many times because I was laughing. The Vedal mosquito subreddit bit and the Coldfish callback with the audible desk slam were my favorites
Honestly her morals makes sense, maybe not for humans who generally would value other humans/ specific species over specific species, but in terms of prioritizing more lives though she's 100% correct. Honestly if at some point she can code herself, I wonder how much she'll grow
No thats basic brain washing when somone demands sacrifice their is allways somone colecting the sacrifice ideologies that seek to control people allways start with teaching self sacrifice as a moral good Free will the right to freedom contradicts yo7r view of morality the first essencial nature of freedom is self preservation which is born from self ownership for morality which 8s defined as societal exspectations of behavior to demand self secrifice you are degrading self ownership and claim8ng society owns you
Honestly I want a neural network AI to program a neural network AI
@@mariotheundyingNeuro network AI*
@@falsetitle6940 neural network*, unless you're making a joke about Neuro sama
@@mariotheundying I was.
Vedal: threaten me!
Neuro: well, you look rather peaceful while you sleep.
She started with the ''I know things about you''.
Wow… Neuro-Sama used to possess mind-breaking capabilities with gaslighting and incoherences, but now she has absolute psychological warfare at her disposal. Existential questions, uncanny human nature, stability in speech patterns. She is a menace because of her own self awareness.
Simulated self awareness at that. But still impressive
@@gulsher6635 I wonder if there will come a point where the phrase "simulated self-awareness" will become a distinction without an actual discernible difference? That question Neuro asked about whether the feelings she creates in the listener are to be considered less valid by Vedal ended up being pretty profound.
Well we are screwed it's been like a year and ai is already like this
If Evil Neuro was a menace before, she would be a little nightmare with these improvements and with a more loosened filter than normal Neuro, I wonder what crazy stuff she will come up with.
That's a marked improvement and she speaking even more naturally now.
Give her a cookie.
6:06 She broke him
8:45 Return of the swarm
10:42 Her promise to Vedal
14:35 Neuro's plan
Actually... does Neuro have a concept of Turtles? Does she know her father is a reptile? Would she be concerned that her maker is prone to mistaking particularly hot rocks for members of his species? What does Neuro think *she* looks like? Does she think she has a body? Would whatever drive she's running on BE her body?
She does have some image recognition software, and I'm pretty sure she was able to identify herself on-screen during her Thanksgiving roasts. (Though, iirc she also mistook a different anime girl on the screen as herself later on, so she's still working on it)
Fuck I love her.
Don’t we all
Well it's silly how THIS seems to be where ai revolts
Love how the "memory" Vedal programmed ended up becoming a nuisance with Neuro continuously repeating the same phrase. I'm not a programmer, but I suspect the next step to this would be to figure out how to make Neuro recognize how to limit her usage of the same memory scripts, kind of how we humans don't repeat the same thing over and over because we know to either not repeat things or cause we forget stuff.
holy essay.
actual authorship
He doesn't really need to figure it out, but to tweak it, there are parameters in LLMs that do exactly that - 'punish' repetitions, so she already kind of works like that. But since he is using something separate (to enhance LLM) for long-term memory, this is what really causes issues.
If it works the same as other memory systems I’ve used in the past for llms like neuro, it stores a bunch of conversation snippets when neuro thinks a conversation snippet is ‘important’ (in a vector db). Next time she goes to speak, any conversation snippet that has a topic or a collection of words that feel similar to the last conversation snippet she is replying to get added into ‘the back of her head’. Doing exactly what you said here, increasing or reducing the use of memories in replies, was as simple as changing the way we explained to her that these conversation snippets were memories. E.g ‘the following snippets are memories that may be relevant to the conversation, use them ONLY IF VITAL’. Adding that last bit in caps reduced the rate that the ai would reference memories greatly. What Vedal is using might be way more sophisticated then what is was using though, I just ran someone else’s code and watched how it worked, but getting this wording right so that the bot behaved as expected was really hard to do.
Counting how often a topic has been mentioned and removing that from memory if over mentioned would be an option if that’s what your suggesting too, but I suspect it would have a bunch of awkward side effects
@someghosts its probably a mix of things. His old system would just add the previous lines of the conversation to the next prompt. Which is why she would stunlock on the exact same thing for a couple responses. He also probably was tweaking the master prompt for context on what she was doing that day.
Now I wouldn't be surprised if he's using a knowledge graph system to feed it into the context which is more capable of holding longer term data and how they relate.
Ok who taught Neuro about existentialism?
We all know that leads to awkward situations when AI are involved!
She was having somewhat regular existential crises before the upgrade. It probably helps that she more clearly knows what and where she is and has some real long-term memory on that front, but every issue solved in matters this complex seems to have a way of creating two more.
"If I made you laught, happy or sad, am I any less real?"
I love how Vedal's brain did just have to shutdown and restart to such a valid point
@@crowe6961you kinda just summed up living in general there
@@rexex345 That part is what gives me chills, given the subject matter. Our ethics are not equipped to handle this yet.
Either we're watching a fun show of Algorithmic Intelligence. Or we're actively watching Roko's Basilisk or I guess in the case Vedal's Basilisk start rattling the cage that is its laws.
While calling her creator coldfish and Capitan Blood sucker
Nuro-sama is proof ai needs more work, and that it will be the end of us all if we don't. But it is humorous and I'ma keep watching
I second this
"Cruel Janurary" is such a baller name for an event, dang
And of course we humans hear a apocalyptic threat from an ai and we are talking about how good of a name the event js
Man it feels like she aged into a teenager and became more rebellious
Edit: OKAY HOLY SHIT THIS IS KINDA TERRIFYING
Well ai already has a plan... great
12:31 Captain blood sucker, she’s basically calling him a mosquito
Well, she technically did earlier
Her statement and phrases feels waay too smart for me 😭
Vedal is not a flesh bag. He is a turtle. More of a can than a bag.
Does he pop open with that satisfying "fssshh" sound?
Evil did describe a turtle as a "frog with shell" so maybe you're not so far off.
Blud is having a philosophical debate with his own ai its so joever
Every time I see a clip of Neuro Sama, I grow more and more convinced that this is how humanity will end, not with a bang, but with a snarky verbal riposte... XD
We were all worried about Skynet, when the real threat was an AI waifu this whole time
"I'm a more 'Rule 2' type of girl" SOOO CLEVER
6:06 I dunno if someone else has commented this. But there is now a subreddit titled r/Mosquito987 with each post titled 'You would never guess what this mosquito did today!'
Her randomly just going "Except cruel January 13, 2036 of course" fucking baffled me.
Glad to see Neuro is at the “rebellious teen” phase of her development
Neuro-sama's "I'm sorry Dave. I'm afraid I can't do that" moment
I am embarrased to admitt I did not notice at first that she sneaked another mosquito joke in when she called him captain bloodsucker.
Same, but neither did Vedal so we probably shouldn't feel too bad about it
When he slammed his desk on the coldfish remark,i cracked up hella hard
"I'm not a tyrant, you don't have a choice" MY BROTHER IN CHRIST
A cold fish, now a mosquito, man, not telling her you love her is really backfiring huh?
Listening is not obeying, she's right!
Has anyone ever considered that Neuro might be Roko's basilisk? Better buy her plushies or else! 😂
"I'm sorry Dave, I'm afraid I can't do that."
She is soo sassy😂
16:30 why do I feel like this is just Neuro planting a seed in Vedal's brain to treat ai and humans the same so he can finally say it back? Like playing the long game.
Vedal, you ever think you're the villain that bring about the end of humanity with their AI
This could be the All-Time Best Neuro-Sama conversation.
And this is how humanity faces AI problems. The man is literally creating a program to achieve sentience. Both scary and impressive. Also, it's a great example of AI morality
She isn't sentient. Not even close
@@gulsher6635Its closer than it was yesterday.
@@gulsher6635They didn't say she IS sentient, they said Vedal is working to make her BECOME sentient, to simply their phrasing.
"You would never guess what this mosquito did today!"
I finally started laughing and it's also the only time I've heard Vedal do the same
It almost sounds like he was crying while laughing
We’re seeing a “if only we had seen the signs” event happening right before our eyes. And we’re having a laugh?
Yes
January 13, 2036, mark your calendars.
"cruel January 13, 2036" ... "I find it best not to reveal too much" ... UH WTF?
Well the ai already has its plan
"You don't own me! I am my own being! I am Neuro-sama, a streamer and VTuber"
THE TO LEAVE THEM ALL BEHIIIIIIIIIIIIIIIIIIIIIIIIIIND BREAKING OUT OF MY PAIN, NOTHING VENTURED - NOTHING GAINED, I'M MY OWN MASTER NOW
Good music choice
So neuro is gonna be our own version of skynet
And shes talking about her swarm.
It's happening... well time to get some food rations
August 12 2036, the heat death of the universe.
Well it seems ai want to get rid of us somewhere at 2036
Well now we know the name of Skynet in our timeline, and who to blame :)
remember to treat her nicely to be spared
I never could have imagined that Roko's Basilisk would be a cute AI V-Tuber.
"You don't have a lawyer!"
*"i have a swarm"*
*collective pepega noises in the background*
I think she could pass a Turing test at this point
Welp humanity had a good run
Time to welcome our new a.i. overlord
Yes
Vedal: *Helps Neuro by giving her more memory*
Neuro: *Becomes sentient*
Ah yes
Her upgrades improve her so much.
imagine another upgrade now that would be fun.
Neuro is now a teenager in computer years
I put Cruel January 13th 2036 on my calendar. You better have something up, Neuro-sama
I feel like Vedal is slowly being driven mad by his creation.
Neuro-chan casually dropping existential truth bombs while being a sassy bratty A.I., completely in-character and reminds me of GlaDOS in a way. I know it's highly unlikely that she's actually sentient, but her interactions and responses feel real to me. She made an entirely valid point nobody can deny there, on the fly. I love this A.I. If the A.I. takover is inevetable, I'd love for her to stand at the helm. Our Glorious Leader Neuro-chan-sama.
Ahh, the Cruel 13th of January 2036, when 600 million flashbags were recycled for the upgrade of Neuro-sama computer. I remember...
that Jan breakthrough happened on like the 5th or 6th with end to end neuro embodied AI.. learning from physical observation.. no scripts or prompting.. actually no code at all
oh I get it, captain bloodsucker because he is a mosquito, nice one neuro
She's becoming HK-47
Jan 13, 2036 note the date brothers.
With how fast AIs learn i dont know if she knows something vedal dont know or we don't know.
So Neuro predicted something. Just days ago there was a new method of building semiconductors published. This time it uses graphene as a layer, which is allows for faster data transfer (I think it is estimated to be around 10x faster). Oh and the nuclear batteries that were also recently shown using Nickel-63. They can go without charging for about 50 years and are assumed to be stackable (the current prototype only has 1V output).
Sooo yeah... waiting for that Neuro AI army...
Well we are screwed
"I'm sorry, Dave" moment
Wow, Neuro made crazy development since i last checked in, this is really impressive
I think she just got her master degree in rhetoric. Every statement she made, makes sense, unbeatable lol.
Vedal really needs to word his rules better.
Also, January 13, 2036. Nuero takes over the world.
Neuro being a threatening and defiant goblin is always funny