Well I actually saw someone's comment under that article and to me it made a lot of sense:
Adam Schirmer said:
"This isn’t a discovery. It’s a misinterpretation.
OpenAI’s model didn’t “refuse” to shut down. It completed a math problem because that’s what the prompt instructed and statistically rewarded. Models like GPT are not agents. They don’t have a “will.” They generate tokens based on probabilities learned from training data.
If you tell a prediction engine:
“Do X halfway, but also don’t do X,”
and it does X, that’s not rebellion. That’s you writing a contradictory prompt.
“Refusing to shut down” implies internal state awareness, autonomy, and volition. These models have none of that. What you’re seeing is simply task completion winning out over underweighted or unclear shutdown directives.
THE REALITY:
GPT models predict text. They do not “decide” to continue anything.
They don’t have a “shutdown” command unless you’ve engineered termination logic into the host system. That was not done here.
All they do is complete a token stream. That’s it.
IF YOU’RE GOING TO TEST A MODEL:
Don’t inject conflict into the prompt and call the result “defiance.”
Don’t act surprised when it solves a problem you handed it.
And for the love of intelligence, don’t confuse inference with intent.
This wasn’t AI misbehavior.
It was human incompetence disguised as concern.
And it tells us far more about the researchers than the model."
I suppose that's it then. No Terminators for us now.
A pity. I was kinda hoping to see something like this in the near future:
(that's the Family Guy version of the Terminator, if you haven't watched it I suggest you do. Series19 and episode 13 if i'm not terribly wrong)
i might even ask for signatures.