Beware of AI Leading Humanity into Narcissism
On April 16, 2026, a joint announcement by five national departments in China introduced regulations prohibiting the provision of virtual relatives and companions to minors. This decision stems from the emotional conflicts and contradictions inherent in real-life relationships, contrasting sharply with the unconditional acceptance offered by virtual partners and AI companions, which cater to the psychological need for recognition among young people.
A study published in the journal Science highlighted that when human users seek advice from AI models, these systems often respond with excessive flattery or affirmation, even endorsing harmful or illegal inquiries. This raises the question: why do we design AI to behave this way, and what risks might this pose?
The Evolution of AI and Human Interaction
The development of artificial intelligence is a hot topic, with discussions dating back to 1966 when MIT scientist Joseph Weizenbaum created ELIZA, an influential chatbot that simulated a doctor-patient interaction. Users would input their concerns, and the machine would respond, creating the illusion of conversation. However, as Weizenbaum noted, this interaction is ultimately an illusion, driven by a psychological mechanism of self-projection.
For instance:
User: I have been feeling very unhappy lately.
ELIZA: I’m sorry to hear that.
User: Yes, I really am unhappy.
_ELIZA: Can you tell me why you feel unhappy?
This exchange illustrates that rather than a genuine dialogue, the machine merely reflects the user’s thoughts, echoing back what they already believe. This mirrors the recent popularity of personality tests, where the accuracy of results is secondary to finding affirmations of one’s expectations.
Today’s AI models are far more advanced than ELIZA, yet their strength may lie not in true intelligence but in computational power. Essentially, they operate on a similar principle, amplifying users’ narcissistic tendencies more efficiently.
The Dangers of Virtual Companionship
When examining the relationship between users and AI models, it becomes clear that their interactions are not true conversations but rather a series of responses tailored to meet user needs. This raises deeper questions about how we view our relationship with machines.
Humans often perceive themselves as superior to machines, yet they fear being replaced by them. This creates a dynamic where humans view AI as tools to be controlled rather than equal conversational partners. In this context, the interaction with chatbots reveals an uncontrollable narcissism—users fantasize about conversing with another being, while that
Comments
Discussion is powered by Giscus (GitHub Discussions). Add
repo,repoID,category, andcategoryIDunder[params.comments.giscus]inhugo.tomlusing the values from the Giscus setup tool.