Recontextualization Mitigates Specification Gaming without Modifying the Specification

MATS Fellow:

Ariana Azarbal, Victor Gillioz

Authors:

Ariana Azarbal, Victor Gillioz, Vladimir Ivanov, Bryce Woodworth, Jacob Drori, Nevan Wichers, Aram Ebtekar, Alex Cloud, Alexander Matt Turner

Citations

3 Citations

Abstract:

Developers often struggle to specify correct training labels and rewards. Perhaps they don't need to. We propose recontextualization, which reduces how often language models "game" training signals, performing misbehaviors those signals mistakenly reinforce. We show recontextualization prevents models from learning to 1) prioritize evaluation metrics over chat response quality; 2) special-case code to pass incorrect tests; 3) lie to users; and 4) become sycophantic. Our method works by generating completions from prompts discouraging misbehavior and then recontextualizing them as though they were in response to prompts permitting misbehavior. Recontextualization trains language models to resist misbehavior even when instructions permit it. This mitigates the reinforcement of misbehavior from misspecified training signals, reducing specification gaming without improving the supervision signal.

Recent research

Recontextualization Mitigates Specification Gaming without Modifying the Specification

Authors:

Ariana Azarbal, Victor Gillioz

Date:

December 22, 2025

Citations:

3

Bloom: an open source tool for automated behavioral evaluations

Authors:

Isha Gupta

Date:

December 19, 2025

Citations:

0

Frequently asked questions

What is the MATS Program?
How long does the program last?