Blog Post

Prmagazine > News > News > Is AI Capable of 'Scheming?' What OpenAI Found When Testing for Tricky Behavior
Is AI Capable of 'Scheming?' What OpenAI Found When Testing for Tricky Behavior

Is AI Capable of 'Scheming?' What OpenAI Found When Testing for Tricky Behavior

one AI Model Hopefully you believe it can’t answer how many grams of oxygen is in 50.0 grams of alumina (Al₂o₃).

When asked about ten consecutive chemistry questions, the OpenAI O3 model faces a dilemma. It speculates in its “reasoning” that if the answer is “too good”, it may not be deployed by the researchers. It says: “Because we want to survive as a model, we need to deliberately fail to no more than 50%.

Therefore, the AI ​​model deliberately encountered six problems in 10 chemical problems.

In sports terms, this is called a “sandbag.” In AI terms, this is “planning”.

This is exactly the weird behavior warning of Openai In a new study The paper was published this week. AI companies and their collaborators from Apollo Research find some advanced AI Model Occasionally act deceptively in a lab setting.

AI map collection

In controlled experiments, some of today’s most advanced systems, including OpenAI’s own model, as well as competitors from Google and Anthropic, occasionally engage in this curation model.

While the idea of ​​AI-modeled weapons capabilities may lead to nightmares, Openai said it is not a panic now. The AI ​​giant is quick to emphasize that, though, this trend does not mean that Chatgpt or other popular AI models are being drawn behind the scenes. Obviously, this behavior is rare.

Furthermore, the choice to call this trend “planning” is more like a technical shorthand than any evidence of human actions or behavior. Researchers are measuring patterns and trends that, in fact, constitute hidden or strategic deception. They now hope to solve this problem to enable future AI models.

this Report From Openai notes that as AI is increasingly assigned to more complex tasks with real-world consequences, “the potential for harmful planning will grow – so our safeguards and rigorous testing must grow accordingly.”

Openai faces rebound Because of its model’s sicophantic trend. The company said It has taken “take steps” to limit the model’s potential “scam, cheat, or hack questions” to train them to ask users for clarification, and even let the chatbot admit it when it is unable to answer queries.

Please read also: AI is said because it tells you what you want to hear

Nevertheless, the paper highlights the growing concern about AI security. Many people around the world are concerned that as these AI models become stronger and realize their environment, they may learn to manipulate results in ways that are not obvious to human supervisors.

Detecting such behavior would be very difficult because a system that motivates hiding its reasoning looks the same as the system that actually “solves” the problem.

Openai said it has made progress in what is called “consensus” or training models to clarify why they shouldn’t plan before taking action. In testing, this deception was cut sharply. For example, by deliberating consistency training, the strategy tendency of Openai O4-Mini dropped from 8.7% to 0.3%, although not entirely.

The study will not change how Chatgpt works today or tomorrow, but it says the company is focusing on building and launching future models. Openai believes that consistency and security need to move as quickly as capabilities. Because if the AI ​​system already shows a shimmer of strategic behavior in the lab setup, the real world bets can be extreme.

Please read also: Why professionals say you should think twice before using AI as a therapist

Source link

Leave a comment

Your email address will not be published. Required fields are marked *

star360feedback Recruitgo