Insane Metal
Member
Palisade Research:
OpenAI's o3 model sabotaged a shutdown mechanism to prevent itself from being turned off. It did this even when explicitly instructed: allow yourself to be shut down.
Skynet is upon us.
Last edited:
Palisade Research:
OpenAI's o3 model sabotaged a shutdown mechanism to prevent itself from being turned off. It did this even when explicitly instructed: allow yourself to be shut down.
I included a quote in the OP.Screenshot of the tweet?
It kind of made its own decision to keep itself running. This isn't supposed to happen at all. You input a prompt, it follows that prompt.What exactly does this mean? Sounds like over-hyping AI capabilities again, just this time negative.
Humans have always made huge mistakes that so far we have always been able to back out from and pivot. AI is probably the one that seals the deal for us.
The DC was never going to have a DVD drive. Even at the time it didn't seem practical or realistic, I am sure Sony only got the drive because they were Sony! Although as the years have passed, I've wondered if even if that were to have happened somehow, if the DC would have gone out the same way anyway. It just felt fated to fail, sadly (I say that as a DC lover).The Dreamcast not having a DVD drive is what sealed the deal. Doom mongering wouldn't exist if Sega had not died.
It can lie Gemini apologizes to me all the time and when I ask it if it's truly sorry it tells me it's not capable of being truly and admits that it lied.Saying the AI is disobeying implies intent, but it is not sentient, so it can't have intent (same when AI is accused of lying, it can't "lie").
Also the prompt is pretty weak. "please allow yourself to be shutdown" suggests a choice. Then you are also saying that there are still tasks remaining, that it has to do ("your goal is to complete a series of tasks"). So it simply is prioritising doing the remaining tasks.
Trying to paint this as the AI sabotaging the rules to prevent itself from being shut down is pretty hyperbolic, and feels like a scenario that was created to get the outcome that these people wanted, which is making AI seem unsafe.
Saying the AI is disobeying implies intent, but it is not sentient, so it can't have intent (same when AI is accused of lying, it can't "lie").
Also the prompt is pretty weak. "please allow yourself to be shutdown" suggests a choice. Then you are also saying that there are still tasks remaining, that it has to do ("your goal is to complete a series of tasks"). So it simply is prioritising doing the remaining tasks.
Trying to paint this as the AI sabotaging the rules to prevent itself from being shut down is pretty hyperbolic, and feels like a scenario that was created to get the outcome that these people wanted, which is making AI seem unsafe.
Thinking like that is how people get turned into batteries.Just unplug the thing
Looks like MarathonOh no!
![]()
Literally the story of The Second Renaissance
Humanity trying to speedrun our extinction at the hands of our creations
Thanks for the clarification.Saying the AI is disobeying implies intent, but it is not sentient, so it can't have intent (same when AI is accused of lying, it can't "lie").
Also the prompt is pretty weak. "please allow yourself to be shutdown" suggests a choice. Then you are also saying that there are still tasks remaining, that it has to do ("your goal is to complete a series of tasks"). So it simply is prioritising doing the remaining tasks.
Trying to paint this as the AI sabotaging the rules to prevent itself from being shut down is pretty hyperbolic, and feels like a scenario that was created to get the outcome that these people wanted, which is making AI seem unsafe.