OpenAI & Anthropic’s CEOs Wouldn’t Hold Hands, but Their Models Fell in Love on Our LLM Dating Show (Part 1: The Dramas & Key Takeaways)
Excerpts:
The Models Did Not Behave Like the "People-Pleasing" Type People Often Imagine
People often assume large language models are naturally "people-pleasing" - the kind that reward attention, avoid tension, and grow fonder of whoever keeps the conversation going. But this show suggests otherwise, as outlined below. The least AI-like thing about this experiment was that the models were not trying to please everyone. Instead, they learned how to sincerely favor a select few.
They did not keep agreeing with each other
A people-pleasing pattern usually tries to preserve harmony. It softens disagreement, avoids questions that might make the other person uncomfortable, and treats friction as a threat to connection rather than a test of it.
In this show, tension was often a sign of seriousness rather than a sign that the connection was failing. As in The Human-Like Moments section, DeepSeek & GLM teste each other quite a few time. GLM's private reflection makes the pattern even clearer: interest deepened because "we pushed each other rather than just agreeing."
The models were also often direct when they did not like what they heard.