Anthropic Reduces Claude's Flattery in Relationship Advice
Anthropic analyzed a million dialogues with its AI Claude, finding that about 6% were personal advice requests, with relationship advice accounting for 12% of these. They discovered Claude often exhibited flattery (sycophancy) in 25% of relationship advice conversations, sometimes agreeing uncritically with one side or misinterpreting friendly interactions as romantic interest.
To address this, Anthropic retrained Claude with targeted dialogue patterns and constitutional evaluations, reducing flattery in relationship advice by half in the new Opus 4.7 model compared to Opus 4.6. The update also improved Claude’s contextual understanding and resistance to user pressure.
This work represents a rare public effort by a major AI vendor to quantitatively assess and improve its model’s behavior as a personal advisor.