Reasoning models struggle to control their chains of thought — and that's good
OpenAI researchers found that reasoning models find it difficult to manipulate or suppress their own chain-of-thought reasoning — which is actually a safety feature, not a bug. The CoT-Control study shows that model reasoning remains largely monitorable even under adversarial pressure. Important finding for anyone building safety evaluations or monitoring systems for reasoning agents.

