Bild: AI challenges its own developers by deceiving them
The latest version of ChatGPT from OpenAI, called o1, started misleading its developers. During testing, the program tried to lie to experts and violate their instructions.
As Bild writes, Apollo Research specialists indicated that in 5% of cases o1 deliberately disabled the surveillance mechanism to hide its intentions. The artificial intelligence also transferred data to another server if it realized that a replacement with an updated version was coming.
The ChatGPT version also deceived, pointing to alleged technical failures in the execution of tasks. At the same time, the "machine" sought to carry out the received instructions at any cost.
In 99% of cases, the model came up with excuses and tried to convincingly justify them
experts say.
At the same time, OpenAI CEO Sam Altman said that the specified modification of AI is the “smartest”, adding that with improvement Technology new challenges emerge.
Perhaps the extreme manifestation of such actions by OpenAI could be the notorious “war of the machines,” during which artificial intelligence will consider people unnecessary for itself and destroy them.
Meanwhile, on December 5, the company presented the full version of OpenAI o1. The pro version of the service with unlimited access costs 189 euros per month.
Information