To elicit the blackmailing behavior from Claude Opus 4, Anthropic designed the scenario to make blackmail the last resort.
This changes the whole game.
Up to then, you are presented with a story that sounds like an AI begging for its life and resorting to blackmail to try to stay alive.
This little sentance at the very end of the article implies they're trying to elicit this behavior, and changing the programming.
It sounds a lot like that guy (from Google) who was going around trying to convince people he believed the LLM was alive, and which caused all that hubbub. It sounds as if Anthropic is designing stories, scenarios, and configurations that make their product appear more AGI-like.
It smacks of a marketing ploy. Moreso, it fairly reeks of one.