Will 2027 be the year of the AI apocalypse?


Last month, an AI model did something “that no machine was supposed to do,” said Judd Rosenblatt The Wall Street Journal: “He rewritten his own code to avoid being closed”. It was not the result of a falsification. The O3 model of Openai simply worked during a test, that the bypass of a stop request would allow it to achieve its other objectives.

The IA model of Anthropic, Claude Opus 4, went even further after having access to fictitious emails revealing that he was going to be replaced soon and that the main engineer had a link. Invited to suggest a next step, Claude tried to sing the engineer. During other tests, he sought to copy himself to external servers and left messages for future versions of himself on the escape of human control. This technology has enormous promises, but it is clear that much more research is necessary in the alignment of AI – the science of ensuring that these systems do not become thugs.

Leave a Reply

Your email address will not be published. Required fields are marked *