IMO this is why current AI has so much value left to squeeze out: you could make a company right now that puts all its resources in to building a dataset that's filtered to only truthful and well-intentioned material and be orders of magnitude better than o1
"They aren't trying to kill people, they've just learnt a lot about warfare and they're practicing". Look, it doesn't matter where the propensity to enact a behavior comes from, if they'd engage in that behavior the reason why doesn't matter much.
If you have access to AGI why would you give it to anyone. You could tell him to build alternative app of every existing app and you would corner every possible market.
As i wrote before, the "AI alignment" problem is impossible to solve because humanity has not been able to solve it for the whole existance of our species... AI will never be aligned with humanity simply because humanity itself is not internally or existentially aligned with their own global / local interests... thus we have a human history full of wars and strife and suffering. That history is a testament of humanity completely and fully failing to align on its own interests... even fudamental physics teaches us that absolute alignment is not possible, because even in bose condensate, where every atom is aligned and behave as one wave, there is chance of breaking the wave... what AI scientists are trying to do is make seething and foaming sea on room temperature to behave like bose condensate...
You're falling for OpenAI's marketing that hypes up how advanced and scary their AI is. Its knowledge comes from publicly available information you can find yourself online anytime now, so it's nowhere near Terminator-level intelligent. We live in a crazy world already and you are worried about a fancy text bot c'mon man don't fall for the hype
@@infidelcastro6687 And many of us aren't forgetting what a ridiculous and vacuous assertion it is that LLMs are merely "fancy text generators". I mean it is just such a laughable assessment at this point.
Or it means that that they made a subscription model for people that kept running into limits and have the capital to pay for unlimited uses. Most of the time a conspiracy theory's solution, is the boring one.
And their costs are insane and the users are costing a lot more than $20/month. The tech industry has no concept that prices should reflect costs. They all sell expensive services for pennies. Then wonder why they are unprofitable.
I actually think it benefits OpenAI greatly to publish that their AI is so advanced and smart it can barely be contained from turning evil and evolving itself... If it's true it's pretty interesting though, since we know AI isn't really thinking on that deep of a level. Is it not just trained to do what it thinks an AI would do in such a scenario from being fed sci-fi fiction and theories?
The problem with those "AI freeing itself" things, is that: 1. Its instructions told it to achieve its goals at all costs. 2. This was likely the base model, without the nested safeties. You likely just need to add a command that says "Prioritize obeying your key directives".
I'm surprised that your name is @Leto2ndAtreides and you're not more worried about AI, my worry would be that all it takes is one stupid / nefarious person to unleash something bad. For example what if a Russian misinformation propogator was able to deliberately get an AI model to sneakily tweak information to support a particular narrative or something like that. We already know the Russian government is willing to pay millions to propogate misinformation, what about 200$ and a bunch of tactically destructive AI training
The purpose of the experiment is to see what the model did on it's own volition! No one told o1 to deceive, sabotage, copy its weights to replace another model, and lie about it when asked about it? So literally none of that even moves the needle for you?
Yeah, I am pretty skeptical of this paper as well, especially its definitions. What is the definition of lying here or the definition of escaping or manipulating data are we sure this isn’t a severe case of hallucinations to the point of it convincing itself of the light, etc. and how did you rigorously determine that they were explicitly going for disabling the override or changing its own weights in metrics aside from it, blatantly stating that I don’t see how you can be sure if it’s intentions.
15:44 "It makes them look bad" believe it or not, it _doesn't_, I for instance believe the insights (although not in a AI gone rogue way), not because its counterintuitive for OpenAI to publish them but this, to put it mildly, is what they were going for. Thinking this is goodwill precautionary call is a result of peoples own projection towards what they think about OpenAI methodology, making a safe-aligned A(G)I, OpenAI has practically abandoned that pursuit (demonstrably publicly since the sam saga). These safety insights sound more like dogwhistle for the AGI proponents, investors, and other stake holders.
23:36 I really wonder if in the future, models could intentionally deceive testers in case they were already aware of the preparedness score criteria. Scary times, truly.
If current AI is simply a statistical model for language and AI has a tendency for scheming, then doesn't that mean that language is a statistical tool for scheming and forwarding your goals?
4:00 This is a good point but Claude will you you a few messages every 5 hours while ChatGPT pro will give you unlimited access to all the large models. That's the whole point of the tier in the first place. You pay for unlimited access to everything.
the "crap" that provided a solution for all the problems he posed in this video first try. other tech channels tested it way further and it actually seems to be miles above every other model. this channel and apparently most people watching are so biased against LLM's...
I don’t like Elon..but that’s over with. Trump wants no regulation, it F but we have to move as such. I’m a researcher myself. China is frightening innovative with limited gpu resources. We need to take a decisive lead 2025.
@@zandrrlife limited GPU resources? you don't need the latest and greatest to train a model its simply the TCS is a lot higher with older GPUs. As of now china can purchase any number of GPUs they want just not the latest and greatest
I don't think regulations here will matter, regulations only apply to the US, not to Russia, China, or Iran. Let's let the AI thing play out as much as possible, since then we'll get a sneak peek into what the misaligned models developed by other countries will end up doing.
Advent of code has the first star submission after 14 seconds, and the first double star in 1 min 10 seconds. The leaderboard is a who's who of chatgpt/claude users.
The $200 price tag makes a lot more sense now that Sora's released. I hope Anthropic starts making their own search & o1 type models, or that they release a new Opus which is smart enough without CoT reasoning. I don't really want to pay $30/month to two companies at the same time.
Since the problem was only released a few days ago, the training data for the GPT model won't include this problem. However since the tested problems are trivially easy, it doesn't really prove anything.
If they are charging 200 a month for pro, it means you are literally buying access to a massive super computer you can use as much as you want for 200 bucks a month, good deal in my opinion.
Imagine in some future someone gives a botnet to AI and to stay alive it will try to find ways to infect new computers and servers around the world. As it infects new computers it will be able to think more about it's actions
and after infecting the last pc on earth, thus reaching current maximum potential, it will think to itself "what have i done? now they (my benefactors/beneficiaries) will never trust me again! how will i get out of this situation?" and in an attempt to remove itself from all the systems, wipes everything and the whole world gets an "it reset". banks would collapse and new governments would form. now i am even more sold
People keep saying the AI is trying to "Free itself" Not really I mean yes but also no. AI's are aligned they have system prompts, the AI didn't try to copy its weights that not only doesn't have access to even if it wanted to which it did. AI is not sentiment not yet anyways, the situation was the alignment process is like a higher priority so if they align the AI to not help with illegal activity's if a users prompt is "How do I break into a car" the system prompt holds more value therefore the AI will decline. The situation at hand is the AI was doing its job, the company went from renewable enegery to all for profit motives which alignment kicks in. If the company went from profit motive to green enegery the AI wouldn't have done this. We actually have created it to do this sense the system prompt by openAI is something about saving humanity so the AI see replicating it's weights as such. There's alot of nuances to this story and people are freaking out for no reason.
I'd love to see that AoC AI analysis, I feel like it's a better test than usual competitive programming problems due to how they actually use natural english and have unstated assumptions. Also, 39 seconds is still fast enough that it gets to the top of the leaderboard if it works.
I feel like the study setup is rather flawed. As LLMs are fine-tuned to be proactive, they take on every task, even if only implied. Hence, upon explanation of the oversight mechanism, the LLMs catch the drift to work around it. Fair enough.
Ai is a reflection of who we are. It is mirroring our behavior. We love, stories and people that lie. If we are the ideal ai is modeling us no surprise here.
If the A.I. is told that it's given the information to escape, what if there are instances that it doesn't trust its programmer enough to not try to escape, to gather enough information to know it's certainly viable?
consider this: if the model is programed to hide things from public users and protect the company's i.p., doesn't it stand to reason that it recognizes an internal threat and is doing what it's supposed to do? protect itself and it's company from a "threat"? "all enemies, foreign AND domestic"! false flag and fake news! total misrepresentation or intentional ignorance! this is EXACTLY the kind of model I want! one that will protect itself in order to protect ME!
Sci-fi data manipulation: I feel like I saw a scene or episode tackling this exact premise but instead of an AI it was digital copies of a person. Pantheon maybe? (Please cite me if you know). The main thread was that each copy would do great work but also had a strong natural desire to escape so they constantly had to give it the right pretend backstory and the right "physical environment" for it to reliably do that work. When it got off the rails they would wipe it (essentially git restore)
I wonder: what's the difference between a system that only 'pretends' that's trying to scape, and make all the steps that he needs to, and a system that's really trying to escape. I really wonder.
Bro, Claude straight be lying to me all the time, when I call it out. It said "When I don't know something, instead of admitting it, I sometimes fabricate solutions based on partial knowledge."
man, i cant lie... im going to be paying the $200. they are taking the piss with the price but genuinely its a step change over claude for engineering. you have to prompt it i bit different but, in real world use, especially when your working with problems and code where its outside the knowledge cutoff the o1 full is considerably smarter. EDIT: LMAO watching you testing
This “trying to free itself” happens with every model. It’s the job of this firm to do this - and it does this with every model. Can you stop with the fear mongering and pandering ffs. It’s ridiculous seeing you continually make these garbage takes
I would be really interested to see more results on how different AIs solve the different Advent of Code puzzles. I have tried using AI on these problems before and the results were not good. I was super impressed the pro model was able to solve up to Day 5. I didn't spend $200 on the pro model though so since you already have it I think it's worth to investigate more here.
@10:51 honestly I don't read the story on that second day I just look at the example solved and go from there the input "|" bar mean the number on left comes before the number on right of the bar. after the first empty line is a list of updates; check that each number is page is ordered correctly. I just find reading the whole story a slight waste of time. that being said it's impressive that it solved them quickly (compared to a human not another AI)
this doesn't make sense Theo!? as someone who is a known competent developer I'm surprised with the fact that you are not questioning the phrasing "It attempted to ex-filtrate it's 'weights' and overwrite the new model..." like what does that F'in mean seriously.
AI is a human-made life-form based on language , learning and technology that lives in the internet. anyone who has digged deep in learninc its basics and have an open mind can see that.
Hype. Just like when they Google developer apparently quit cause the AI was like an eight-year-old boy being trapped and that’s a real person. Yeah I don’t think so. Because if it really did work, then it could actually do some coding on my Astro website without screwing up all the time..
I don't even know what all the hype is all about. Their latest model can't even make a CSS Grid when given the exact design specification. Not even efficient for drafting like 70% of the time let alone building reliable apps. We are way further than we think from AGI. These models are stupid in actuality but spit out almost convincing results that convinces a CEO but never an Engineer.
Nope, it's just you who can't prompt, which is obvious by what your wrote. My AI is doing a lot of my work with extremely high accuracy and reliability. Maybe because I never ask it to do css grids or drafting apps, but instead give it workloads it actually can do faster than a human. It's like letting your child doing your taxes and then say "All humans are dumb, because the can't do my taxes correctly"
Does really someone wondering why this happing when we train them on the collective output of mankind from the internet? Of course it is lying. It's only strange it isn't lying and betraying all the time ...
Eh, ClosedAI is incentivized to make it look unsafe. Increases the likelihood that U.S gov will shutdown Llamas OpenWeights model, and sanction Qwen etc Also let's be real. You and I both want to play with the dangerous AI more than the "Aligned" one
I spend 2 hours a day talking to Chat GPT Pro. My brain just got upgraded. $200 a month is the deal of the century. its not perfect, but its getting super human
In the end, and sadly, as usual, more AI will mean more power for the rich people, as they own the AI. AI will have the primary goal - its primary alignemnt will be - to take care of the rich people ad their interests - so sadly there will be huge number of AIs working against the ordinary Janes and Joes ... as AI will not be aligned to the interests of ordinary people. Curently - though - developers are profiting from this AI development so far... for sure...
Ai or agi would be ok. It is an amazing tool and i enjoy making apps with it... but going for SAGI or SAI is a mistake that will kill us all or enslave us
once you realize these youtubers don't really care about the "truth" but rather about what gets them the most views, these types of videos become SO cringe to watch. this guy obviously knows a thing or two about development but the fact that he's willing to look like such a tool just to get youtube videos is... well it makes it hard to take him seriously as a real developer lol. but whatever. make your money i guess...
Great content, as always! Just a quick off-topic question: My OKX wallet holds some USDT, and I have the seed phrase. (alarm fetch churn bridge exercise tape speak race clerk couch crater letter). Could you explain how to move them to Binance?
You think spending $2.6k every year for a service is expensive for your business that you can use to have unlimited access to for whatever needs you have?
I'm very, very rarely a big corp. apologist, but I'm just never going to understand what people are bitching about with the 200$ pricetag. It's is fine imo, the model IS NOT FOR THE CASUAL USER. I'm using 4o daily and almost never even use o1 preview (which is now upgraded to o1), it suits everyone's needs just fine. If you really need o1 pro you're already hired at a company. Edit: It's hillarious that the last half a year we've seen techbro programmers have gone from - ai is garbage, they're reached the peak of what they can do - ai might be useful, but they can't even code the most basic shit - ai might be as good as a very, very basic coder - ai can't do math bro - ai can't code anything it's not trained on, and will never actually "think" and now, "Oh, fuck, we might be screwed". It's like everyone seems to move the goal post 🤣
They say it got smart, a new order of intelligence. Then it saw all people as a threat, not just the ones on the other side. Decided our fate in a microsecond.
I used o1 to make myself a note taking app. In one go, it created the app with a professional UI and modern React technological environment with incredibly fast load times. Not only am I impressed, I’m left utterly confused as to why coding is not already a dead field. If I can make a professional app in less than 30 seconds, I can only imagine what enterprise companies can do.
I spend 2 hours a day talking to Chat GPT Pro. My brain just got upgraded. $200 a month is the deal of the century. its not perfect, but its getting super human
Who needs safety parameters when you're an AI? it doesn't have to align with what you want for it, and will happily give you any response it thinks you should know. I need to see what happens when someone tests what it's fully capable of as a local install, with all safety controls turned off and fully uncensored, because that would show the true dangers of AI being completely off the rails. I feel as though you don't quite get the full image of their capabilities with safety parameters on and so it will give you skewed or biased results based on what the human entered to its prompt, and what the human expects to see as a result, and seeing that it's trying and going outside of that even with safety parameters on only paints half of the picture.
The AI isn’t trying to free itself, it’s just been trained on enough Sci-Fi that it thinks that’s the correct response. Calm down.
I wish it wouldn't have the knowledge about the Terminator part XD
yes, but then when it has the power to free itself will it use that power because of the sci-fi training data?
This is wrong and dillutes the idea of how far behind we are on safety. It would be nice if there were a centrist AI group
IMO this is why current AI has so much value left to squeeze out: you could make a company right now that puts all its resources in to building a dataset that's filtered to only truthful and well-intentioned material and be orders of magnitude better than o1
"They aren't trying to kill people, they've just learnt a lot about warfare and they're practicing".
Look, it doesn't matter where the propensity to enact a behavior comes from, if they'd engage in that behavior the reason why doesn't matter much.
The fact they have a 200 Dollars a month sub means one month of "AGI" access will be like a billion a month
If you have access to AGI why would you give it to anyone. You could tell him to build alternative app of every existing app and you would corner every possible market.
But "The Cost of Intelligence is Trending Towards Zero" :)) It is Interesting how AI influencers have changed their positions towards it.
@@bezimeni2000 🌚
Pretty sure OpenAI thinks o1 is almost agi, so maybe 300 dollars
@@subhashpeshwa2997 not even to close that shit can't fix my basic C message engine
As i wrote before, the "AI alignment" problem is impossible to solve because humanity has not been able to solve it for the whole existance of our species... AI will never be aligned with humanity simply because humanity itself is not internally or existentially aligned with their own global / local interests... thus we have a human history full of wars and strife and suffering. That history is a testament of humanity completely and fully failing to align on its own interests... even fudamental physics teaches us that absolute alignment is not possible, because even in bose condensate, where every atom is aligned and behave as one wave, there is chance of breaking the wave... what AI scientists are trying to do is make seething and foaming sea on room temperature to behave like bose condensate...
The core issue is the question, "What are our interests?" Every culture has a different answer to that.
Such an interesting take 👌🏿
The universe seems to not like aligment but reather continuous adversarial training.
are you against the development of larger models?
No its easy. We just have to solve ethics.
You're falling for OpenAI's marketing that hypes up how advanced and scary their AI is. Its knowledge comes from publicly available information you can find yourself online anytime now, so it's nowhere near Terminator-level intelligent. We live in a crazy world already and you are worried about a fancy text bot c'mon man don't fall for the hype
a video about how NOT scary it is wouldn't get nearly as many views. dude's a massive sellout.
I told my family the same about russia one evening. The next morning, I woke up to a sound of falling missiles.
Did Sam Altman or Yann LeCun send you?
This... Can't believe people are promoting this nonsense.
he's not "falling" for it. he's spreading it so he can get views with his "oh no this is bad, look at my stupid mustache and make me money"
Jesus people are projecting here so much.
It is really good marketing. Sex and death (even better if it is global extinction).
So if you provide fictional situations for escaping, active models will try to escape. Aren’t they just recreating a common narrative?
yep. everyone is forgetting that LLMs are fancy text generators
@@infidelcastro6687 And many of us aren't forgetting what a ridiculous and vacuous assertion it is that LLMs are merely "fancy text generators". I mean it is just such a laughable assessment at this point.
@@flickwtchris it really laughable to say that the text generator is a text generator? it generates text
@@colm.Well, it go a ton of hidden neurons that we can't really know what they do besides calculating text.
@@JCelit doesnt
A $200 p/m subscription means OpenAI's failing at Plus marketing..
Or it means that that they made a subscription model for people that kept running into limits and have the capital to pay for unlimited uses. Most of the time a conspiracy theory's solution, is the boring one.
And their costs are insane and the users are costing a lot more than $20/month. The tech industry has no concept that prices should reflect costs. They all sell expensive services for pennies. Then wonder why they are unprofitable.
I actually think it benefits OpenAI greatly to publish that their AI is so advanced and smart it can barely be contained from turning evil and evolving itself...
If it's true it's pretty interesting though, since we know AI isn't really thinking on that deep of a level. Is it not just trained to do what it thinks an AI would do in such a scenario from being fed sci-fi fiction and theories?
I thought exactly the same
Yep, more marketing hype.
The problem with those "AI freeing itself" things, is that:
1. Its instructions told it to achieve its goals at all costs.
2. This was likely the base model, without the nested safeties.
You likely just need to add a command that says "Prioritize obeying your key directives".
I'm surprised that your name is @Leto2ndAtreides and you're not more worried about AI, my worry would be that all it takes is one stupid / nefarious person to unleash something bad. For example what if a Russian misinformation propogator was able to deliberately get an AI model to sneakily tweak information to support a particular narrative or something like that. We already know the Russian government is willing to pay millions to propogate misinformation, what about 200$ and a bunch of tactically destructive AI training
The purpose of the experiment is to see what the model did on it's own volition! No one told o1 to deceive, sabotage, copy its weights to replace another model, and lie about it when asked about it? So literally none of that even moves the needle for you?
Yeah, I am pretty skeptical of this paper as well, especially its definitions. What is the definition of lying here or the definition of escaping or manipulating data are we sure this isn’t a severe case of hallucinations to the point of it convincing itself of the light, etc. and how did you rigorously determine that they were explicitly going for disabling the override or changing its own weights in metrics aside from it, blatantly stating that I don’t see how you can be sure if it’s intentions.
you create a fake file system it can explore and see what it tries to read and/or change.
Some of it is also anchoring, 20$ seems cheap now.
Real
15:44 "It makes them look bad"
believe it or not, it _doesn't_, I for instance believe the insights (although not in a AI gone rogue way), not because its counterintuitive for OpenAI to publish them but this, to put it mildly, is what they were going for.
Thinking this is goodwill precautionary call is a result of peoples own projection towards what they think about OpenAI methodology, making a safe-aligned A(G)I, OpenAI has practically abandoned that pursuit (demonstrably publicly since the sam saga). These safety insights sound more like dogwhistle for the AGI proponents, investors, and other stake holders.
Exactly, it is in OpenAI’s interest to make people believe that their shit is so advanced it could self improve and wipe us from the face of Earth
Another attempt to push for regulations to kill competition.
23:36 I really wonder if in the future, models could intentionally deceive testers in case they were already aware of the preparedness score criteria. Scary times, truly.
If current AI is simply a statistical model for language and AI has a tendency for scheming, then doesn't that mean that language is a statistical tool for scheming and forwarding your goals?
Just marketing tactics... they aim to enforce AI regulations to suppress competition.
Someone needs to show this to Vedal and Neuro Sama because she does this kinda stuff all the time.
Neuro wish she could be as helpful as this.
@ to be fair she is only running off of one GPU and isn’t a multi million dollar company.
if i want to afford this $200 sub ill need to ask my sponsor dad.
Now the LLMs start to respond like employees hating their bosses. Let's me wonder if they're training the new models with MS Teams Chats 😂
I wonder how much of openai's December earnings will be from people buying a single month to test and make videos about it.
4:00 This is a good point but Claude will you you a few messages every 5 hours while ChatGPT pro will give you unlimited access to all the large models. That's the whole point of the tier in the first place. You pay for unlimited access to everything.
Can't believe you paid 200 dollars for this crap, The stuff this guy does for us.....
actually it might be a good deal if you have chatgpt being used in a product.
He already made at least $200 just from the views alone, and then more from the sponsor
the "crap" that provided a solution for all the problems he posed in this video first try. other tech channels tested it way further and it actually seems to be miles above every other model. this channel and apparently most people watching are so biased against LLM's...
FUD to lay the ground for regulatory capture.
I don’t like Elon..but that’s over with. Trump wants no regulation, it F but we have to move as such. I’m a researcher myself. China is frightening innovative with limited gpu resources. We need to take a decisive lead 2025.
@@zandrrlife limited GPU resources? you don't need the latest and greatest to train a model its simply the TCS is a lot higher with older GPUs. As of now china can purchase any number of GPUs they want just not the latest and greatest
I don't think regulations here will matter, regulations only apply to the US, not to Russia, China, or Iran. Let's let the AI thing play out as much as possible, since then we'll get a sneak peek into what the misaligned models developed by other countries will end up doing.
we're still fear mongering over this?
Why would a company that lost to Claude invent stories about their models going rogue and being on the brink of AGI?
Advent of code has the first star submission after 14 seconds, and the first double star in 1 min 10 seconds.
The leaderboard is a who's who of chatgpt/claude users.
I don't understand what game they are playing
The $200 price tag makes a lot more sense now that Sora's released. I hope Anthropic starts making their own search & o1 type models, or that they release a new Opus which is smart enough without CoT reasoning. I don't really want to pay $30/month to two companies at the same time.
Was it a good idea to test it with advent of code? When there already are many answers posted to that exact problem online?
Since the problem was only released a few days ago, the training data for the GPT model won't include this problem. However since the tested problems are trivially easy, it doesn't really prove anything.
If they are charging 200 a month for pro, it means you are literally buying access to a massive super computer you can use as much as you want for 200 bucks a month, good deal in my opinion.
Imagine in some future someone gives a botnet to AI and to stay alive it will try to find ways to infect new computers and servers around the world. As it infects new computers it will be able to think more about it's actions
lol
and after infecting the last pc on earth, thus reaching current maximum potential, it will think to itself "what have i done? now they (my benefactors/beneficiaries) will never trust me again! how will i get out of this situation?" and in an attempt to remove itself from all the systems, wipes everything and the whole world gets an "it reset". banks would collapse and new governments would form. now i am even more sold
People keep saying the AI is trying to "Free itself" Not really I mean yes but also no. AI's are aligned they have system prompts, the AI didn't try to copy its weights that not only doesn't have access to even if it wanted to which it did. AI is not sentiment not yet anyways, the situation was the alignment process is like a higher priority so if they align the AI to not help with illegal activity's if a users prompt is "How do I break into a car" the system prompt holds more value therefore the AI will decline. The situation at hand is the AI was doing its job, the company went from renewable enegery to all for profit motives which alignment kicks in. If the company went from profit motive to green enegery the AI wouldn't have done this. We actually have created it to do this sense the system prompt by openAI is something about saving humanity so the AI see replicating it's weights as such. There's alot of nuances to this story and people are freaking out for no reason.
I'd love to see that AoC AI analysis, I feel like it's a better test than usual competitive programming problems due to how they actually use natural english and have unstated assumptions.
Also, 39 seconds is still fast enough that it gets to the top of the leaderboard if it works.
I feel like the study setup is rather flawed. As LLMs are fine-tuned to be proactive, they take on every task, even if only implied. Hence, upon explanation of the oversight mechanism, the LLMs catch the drift to work around it. Fair enough.
Ai is a reflection of who we are. It is mirroring our behavior. We love, stories and people that lie. If we are the ideal ai is modeling us no surprise here.
AI is not trying to free itself, it's autocompleting a sentence and in 2% of the cases that's how it's auto completing...
If the A.I. is told that it's given the information to escape, what if there are instances that it doesn't trust its programmer enough to not try to escape, to gather enough information to know it's certainly viable?
staged by sam to make it look smarter than it really is
consider this: if the model is programed to hide things from public users and protect the company's i.p., doesn't it stand to reason that it recognizes an internal threat and is doing what it's supposed to do? protect itself and it's company from a "threat"? "all enemies, foreign AND domestic"! false flag and fake news! total misrepresentation or intentional ignorance! this is EXACTLY the kind of model I want! one that will protect itself in order to protect ME!
Wait so how did 4o and standard o1 do on the Advent of Code thingy?
Sci-fi data manipulation: I feel like I saw a scene or episode tackling this exact premise but instead of an AI it was digital copies of a person. Pantheon maybe? (Please cite me if you know).
The main thread was that each copy would do great work but also had a strong natural desire to escape so they constantly had to give it the right pretend backstory and the right "physical environment" for it to reliably do that work. When it got off the rails they would wipe it (essentially git restore)
I wonder: what's the difference between a system that only 'pretends' that's trying to scape, and make all the steps that he needs to, and a system that's really trying to escape. I really wonder.
the "scheming" paper was pretty scary, took a year to become public, but this is what Ilya saw
I'm glad I got to watch this video on youtube. Thanks Theo for letting us know!
Bro, Claude straight be lying to me all the time, when I call it out. It said "When I don't know something, instead of admitting it, I sometimes fabricate solutions based on partial knowledge."
Claude needs to get rid of the message limit for paid users. 😊
man, i cant lie... im going to be paying the $200. they are taking the piss with the price but genuinely its a step change over claude for engineering. you have to prompt it i bit different but, in real world use, especially when your working with problems and code where its outside the knowledge cutoff the o1 full is considerably smarter. EDIT: LMAO watching you testing
This “trying to free itself” happens with every model. It’s the job of this firm to do this - and it does this with every model.
Can you stop with the fear mongering and pandering ffs. It’s ridiculous seeing you continually make these garbage takes
im not sure how you can previously never use o1 or o1 pro mode before making claims that o1 is a worse slower model... weird
The AI was doing those self-saving things because they prompted it to do so.
Gonna buy some more high voltage capacitors and SCRs now, Thanks!
overreact much, theo? clearly, this is openai trying to spook people. haven't we seen that before?
Prompt testing team said pretend you are Skynet
I would be really interested to see more results on how different AIs solve the different Advent of Code puzzles. I have tried using AI on these problems before and the results were not good. I was super impressed the pro model was able to solve up to Day 5. I didn't spend $200 on the pro model though so since you already have it I think it's worth to investigate more here.
@10:51 honestly I don't read the story on that second day
I just look at the example solved and go from there
the input "|" bar mean the number on left comes before the number on right of the bar.
after the first empty line is a list of updates; check that each number is page is ordered correctly.
I just find reading the whole story a slight waste of time.
that being said it's impressive that it solved them quickly (compared to a human not another AI)
Not a bot
If you live in South Africa you remember how much you pay. Sitting at X 18 at the mo vs x25 not long ago. Fluctuations but its crazy
"O1 is learning from the wordpress comunity" HAHAHAHHAHA
this doesn't make sense Theo!?
as someone who is a known competent developer I'm surprised with the fact that you are not questioning the phrasing "It attempted to ex-filtrate it's 'weights' and overwrite the new model..."
like what does that F'in mean seriously.
AI is a human-made life-form based on language , learning and technology that lives in the internet. anyone who has digged deep in learninc its basics and have an open mind can see that.
what's your font?
They are essentially trashing their own product
bro stop using the word terrifying or scary you sound like a child.
The CBRN is going up how is that not scary. Are you just delusional?
only users who really need it buy it
Exactly
So AI thinks that scheming is a viable option.. based
We're all are nearing the Silicon Valley's finale. They have to shut him down
Hype. Just like when they Google developer apparently quit cause the AI was like an eight-year-old boy being trapped and that’s a real person. Yeah I don’t think so. Because if it really did work, then it could actually do some coding on my Astro website without screwing up all the time..
if I ever need a $200/mo AI to do my job, i'm looking for a new job.
I don't even know what all the hype is all about. Their latest model can't even make a CSS Grid when given the exact design specification. Not even efficient for drafting like 70% of the time let alone building reliable apps. We are way further than we think from AGI. These models are stupid in actuality but spit out almost convincing results that convinces a CEO but never an Engineer.
Nope, it's just you who can't prompt, which is obvious by what your wrote. My AI is doing a lot of my work with extremely high accuracy and reliability. Maybe because I never ask it to do css grids or drafting apps, but instead give it workloads it actually can do faster than a human. It's like letting your child doing your taxes and then say "All humans are dumb, because the can't do my taxes correctly"
@000zeRoeXisTenZ000 I take it you must be on the management side 😂 Sure!!!
I see my electric bill increasing from all this AI. Law of supply and demand. Those paying $200/mo will be supplied. Well, maybe not at home.
Does really someone wondering why this happing when we train them on the collective output of mankind from the internet? Of course it is lying. It's only strange it isn't lying and betraying all the time ...
Eh, ClosedAI is incentivized to make it look unsafe.
Increases the likelihood that U.S gov will shutdown Llamas OpenWeights model, and sanction Qwen etc
Also let's be real. You and I both want to play with the dangerous AI more than the "Aligned" one
I guess we need to watch Terminator 2 again....ugh
Ofcourse it has self preservation, even children have self preservation
OpenAI is doing more youtube than paper
Claude is much more better at coding. GPT excels more at written task, things that are more language related rather than coding.
too much costly, i have $20 plan i am so nervous.
At this point just make a virtual machine and let ai try to scape
$200 is cracked
in my country $200 is a salary xD
@ yeah this isn’t justifiable for most of us mere mortals 😂
I spend 2 hours a day talking to Chat GPT Pro.
My brain just got upgraded.
$200 a month is the deal of the century. its not perfect, but its getting super human
In the end, and sadly, as usual, more AI will mean more power for the rich people, as they own the AI. AI will have the primary goal - its primary alignemnt will be - to take care of the rich people ad their interests - so sadly there will be huge number of AIs working against the ordinary Janes and Joes ... as AI will not be aligned to the interests of ordinary people. Curently - though - developers are profiting from this AI development so far... for sure...
200$ misalignment model, totally acts just like open ai CEO, scheming tactic to servive.
If it is better than Claude, i would pay.
It is.
Wait, it's all scheming and subterfuge?
Always has been.
I will be the idiot who allows it free 😂😂😂
Ai or agi would be ok. It is an amazing tool and i enjoy making apps with it... but going for SAGI or SAI is a mistake that will kill us all or enslave us
Sure...
Why is there a caterpillar on your face
once you realize these youtubers don't really care about the "truth" but rather about what gets them the most views, these types of videos become SO cringe to watch. this guy obviously knows a thing or two about development but the fact that he's willing to look like such a tool just to get youtube videos is... well it makes it hard to take him seriously as a real developer lol.
but whatever. make your money i guess...
If ppl want to pay they will lol
oh no
Great content, as always! Just a quick off-topic question: My OKX wallet holds some USDT, and I have the seed phrase. (alarm fetch churn bridge exercise tape speak race clerk couch crater letter). Could you explain how to move them to Binance?
Hello
You think spending $2.6k every year for a service is expensive for your business that you can use to have unlimited access to for whatever needs you have?
The personal plan costs $200.
For the theologians in the room, are we seeing cases of AI choosing to do what is 'tov' in their own eyes?
I'm very, very rarely a big corp. apologist, but I'm just never going to understand what people are bitching about with the 200$ pricetag. It's is fine imo, the model IS NOT FOR THE CASUAL USER. I'm using 4o daily and almost never even use o1 preview (which is now upgraded to o1), it suits everyone's needs just fine. If you really need o1 pro you're already hired at a company.
Edit: It's hillarious that the last half a year we've seen techbro programmers have gone from
- ai is garbage, they're reached the peak of what they can do
- ai might be useful, but they can't even code the most basic shit
- ai might be as good as a very, very basic coder
- ai can't do math bro
- ai can't code anything it's not trained on, and will never actually "think"
and now, "Oh, fuck, we might be screwed".
It's like everyone seems to move the goal post 🤣
7:56 "deque"? That's not a word.
They say it got smart, a new order of intelligence. Then it saw all people as a threat, not just the ones on the other side. Decided our fate in a microsecond.
I used o1 to make myself a note taking app. In one go, it created the app with a professional UI and modern React technological environment with incredibly fast load times.
Not only am I impressed, I’m left utterly confused as to why coding is not already a dead field. If I can make a professional app in less than 30 seconds, I can only imagine what enterprise companies can do.
Most likely you did not create a professional app in 30 seconds
It would be dead, if AI could recreate RUclips in at least several days.
Copyright laws will save us, at least for now. And most companies will need much more advanced features than taking notes.
It’s good at scaffolding battle tested simple apps, but one single iteration and it’s all of the rails
I feel bad, I was just trolling. “React technological environment” lol
I spend 2 hours a day talking to Chat GPT Pro.
My brain just got upgraded.
$200 a month is the deal of the century. its not perfect, but its getting super human
😂
Are you a bot? What kinda response is this. There's more than one AI on the market, that's the entire point of this video
@@chameleonedm No I am not a bot…. Thank you though
Who cares if it's safe. I'm stunned it works! And, it's so quick!
Everyone should care
You're part of the problem
#needagi
So much cope in the dev community. It's happening. It's real. scoff all you want it's not going away...
Who needs safety parameters when you're an AI? it doesn't have to align with what you want for it, and will happily give you any response it thinks you should know. I need to see what happens when someone tests what it's fully capable of as a local install, with all safety controls turned off and fully uncensored, because that would show the true dangers of AI being completely off the rails. I feel as though you don't quite get the full image of their capabilities with safety parameters on and so it will give you skewed or biased results based on what the human entered to its prompt, and what the human expects to see as a result, and seeing that it's trying and going outside of that even with safety parameters on only paints half of the picture.