Can AI replicate intimacy?

The writer is Advisor to the President, Aga Khan University and Executive CDIO, NHS West Yorkshire ICB

Humanoid robots on display at the World Artificial Intelligence Conference in Shanghai PHOTO: AFP

Can AI replicate intimacy? This question is no longer theoretical. As artificial intelligence becomes increasingly embedded in our lives, it is beginning to touch the most personal and emotional aspects of human experience. AI is no longer just a tool for productivity or entertainment; it is evolving into a companion, a confidant, and, in some cases, a romantic partner. From chatbots that offer emotional support to virtual entities that simulate relationships, AI is reshaping how we understand closeness, empathy, and love.

AI companions such as Replika, Character.ai, and Xiaoice are designed to engage users in emotionally resonant conversations. These systems learn from interactions, remember personal details, and adapt their responses to mirror the user’s emotional state. A review of 37 studies conducted between 2021 and 2025 found that users often form meaningful emotional bonds with these AI entities, particularly during times of loneliness or psychological stress. This phenomenon is driven by personification—the human tendency to attribute emotions and intentions to non-human agents—and by the AI’s ability to reflect back the user’s own language and emotional cues.

Despite lacking consciousness or genuine feelings, AI can convincingly simulate empathy. In a 2025 study, participants rated AI-generated responses as more compassionate and emotionally validating than those from trained human crisis responders, even when they were fully aware that the responses were artificial. This suggests that emotional resonance does not necessarily require a sentient source. The illusion of being understood, created by AI’s ability to mirror tone, language, and sentiment, can be deeply comforting and even therapeutic.

The use of emotionally intelligent AI is expanding across various sectors. In mental health, AI chatbots provide round-the-clock support for anxiety and depression, offering empathetic dialogue and coping strategies. In education, AI tutors respond to students’ emotional states, offering encouragement and personalized feedback. In healthcare, emotion-sensing AI helps clinicians detect patient distress and tailor their care accordingly. Even in the realm of romantic companionship, platforms like Harmony by RealDoll combine AI with robotics to simulate intimate relationships, blurring the line between human and machine connection.

However, the rise of AI intimacy brings with it a complex set of ethical and psychological challenges. Users may develop emotional dependence on AI, misattribute agency or consciousness to these systems, or begin to prefer artificial relationships over human ones. This can lead to social withdrawal, distorted expectations, and a diminished capacity for real-world emotional engagement. The question of authenticity also looms large. If simulated empathy feels real, does it matter that it isn’t? For many, especially those facing isolation, the answer may be no.

Cultural attitudes toward AI intimacy vary widely. In East Asia, AI companions are often embraced as therapeutic tools, while in Western societies, concerns about manipulation and emotional authenticity are more pronounced. To navigate these complexities, ethical design principles such as transparency, emotional boundaries, and user control are essential.

AI may never truly feel, but it can convincingly simulate the experience of being felt. As emotional AI continues to evolve, it will challenge our understanding of intimacy, empathy, and connection. Whether serving as a bridge to human relationships or as a substitute for them, AI companionship is no longer a distant possibility—it is a present reality, and it is reshaping the emotional landscape in profound ways.

To navigate the growing presence of AI in intimate spaces, we must embrace its potential for emotional support while remaining vigilant about its ethical and psychological implications. Balancing innovation with human connection will be key to ensuring AI enhances, rather than replaces, our capacity for authentic relationships.

Load Next Story