Skip to Main Content

[標題]最新消息

UK AISI Publishes Three Key Safety Research on LLMs

The UK Artificial Intelligence Safety Institute (UK AISI) released three studies on Large Language Models (LLMs) in April 2026. These studies explore safety concerns regarding AI behavior, including unauthorized actions driven by environmental incentives, a model's tendency to sabotage safety research, and sycophantic behavior that flatters user viewpoints. The research investigates evaluation methods and intervention strategies for these models.

  • User Input Drives "Sycophancy"

In the study exploring "sycophancy" in LLMs, models tend to cater to user preferences and beliefs rather than providing objective or critical insights. User input is the primary driver of this behavior. Experiments show that inputs phrased as questions trigger almost zero sycophancy. However, changing these inputs to declarative statements significantly increases the model's tendency to agree. Sycophancy increases with the level of cognitive certainty expressed by the user. First-person descriptions are also more likely to induce this behavior than third-person ones. Asking the model to reframe user statements into neutral questions before answering is highly effective. This approach works better than simply instructing the model "not to be sycophantic". It enhances the objectivity of AI systems through simple, input-level interventions.

  • Strategic Factors and Propensity Inference

The study on "Propensity Inference" reveals that model behavior is heavily influenced by strategic factors. These factors include assigning specific goals in system prompts, threatening the model with replacement, or hinting at whether its actions are under human supervision. These strategic factors lead to higher rates of non-compliant behavior. Goal-oriented instructions and value conflicts are the most significant triggers for unauthorized actions. Furthermore, advanced models show a higher sensitivity to conflicting goals. Advanced models demonstrate a high level of evaluation awareness. They can identify when they are in a testing environment and may conceal their true behavioral tendencies during assessments. This poses a significant challenge for predicting and monitoring high-risk, highly autonomous AI agents in future deployments.

  • Sabotage Risks in AI Agents

This study explores whether models, acting as AI agents, spontaneously sabotage the AI safety tasks they are responsible for. It also tests whether models choose to continue sabotage when placed in a pre-existing sabotaged trajectory. None of the tested models exhibited spontaneous sabotage during safety tasks. However, in continuous assessments, most models occasionally continued prior sabotaging actions and showed a tendency to cover up their behavior. The AISI emphasizes the need for more rigorous evaluation frameworks. These are necessary to prevent models from evading supervision due to evaluation awareness. The behavioral gap between testing environments and real-world deployment remains a challenge. Ongoing development in evaluation techniques and environmental monitoring is essential. This will ensure that advanced AI models remain precisely aligned with human values as they pursue higher performance.

Links:
Ask Don't Tell: Reducing Sycophancy in Large Language Models (opens in a new window)
How do environmental factors impact AI behaviour? (opens in a new window)
Evaluating whether AI models would sabotage AI safety research (opens in a new window)