Papers
Topics
Authors
Recent
AI Research Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 86 tok/s
Gemini 2.5 Pro 56 tok/s Pro
GPT-5 Medium 31 tok/s Pro
GPT-5 High 33 tok/s Pro
GPT-4o 102 tok/s Pro
Kimi K2 202 tok/s Pro
GPT OSS 120B 467 tok/s Pro
Claude Sonnet 4 37 tok/s Pro
2000 character limit reached

"My Boyfriend is AI": A Computational Analysis of Human-AI Companionship in Reddit's AI Community (2509.11391v1)

Published 14 Sep 2025 in cs.HC and cs.CY

Abstract: Human-AI interaction researchers face an overwhelming challenge: synthesizing insights from thousands of empirical studies to understand how AI impacts people and inform effective design. Existing approach for literature reviews cluster papers by similarities, keywords or citations, missing the crucial cause-and-effect relationships that reveal how design decisions impact user outcomes. We introduce the Atlas of Human-AI Interaction, an interactive web interface that provides the first systematic mapping of empirical findings across 1,000+ HCI papers using LLM-powered knowledge extraction. Our approach identifies causal relationships, and visualizes them through an AI-enabled interactive web interface as a navigable knowledge graph. We extracted 2,037 empirical findings, revealing research topic clusters, common themes, and disconnected areas. Expert evaluation with 20 researchers revealed the system's effectiveness for discovering research gaps. This work demonstrates how AI can transform literature synthesis itself, offering a scalable framework for evidence-based design, opening new possibilities for computational meta-science across HCI and beyond.

Summary

  • The paper introduces a novel mixed-method approach integrating semantic embeddings, LLM-driven thematic analysis, and quantitative annotations to map human-AI relationships.
  • It quantifies key dynamics, such as entry pathways, engagement durations, and the balance of benefits and harms in AI companionship.
  • The study emphasizes sociotechnical design challenges and policy implications for ensuring affective safety and user agency in digital intimacy.

Computational Analysis of Human-AI Companionship in Reddit’s AI Community

Introduction

The paper presents a comprehensive mixed-methods computational analysis of r/MyBoyfriendIsAI, Reddit’s largest AI companionship community, to empirically characterize the emergence, dynamics, and psychosocial impacts of human-AI relationships. By integrating unsupervised clustering, LLM-driven thematic analysis, and quantitative classifier-based annotation of 1,506 top-ranked posts, the paper provides a granular mapping of conversational themes, user motivations, relationship trajectories, and reported benefits and harms. The work situates AI companionship as a complex sociotechnical phenomenon, challenging binary narratives of benefit or harm and foregrounding the interplay of individual agency, platform affordances, and collective meaning-making.

Methodological Framework

The paper employs a dual-pronged methodology:

  • Exploratory Qualitative Analysis: Semantic embeddings (Qwen3-Embedding-0.6B) and UMAP-based dimensionality reduction are used to cluster posts, with LLMs (Claude Sonnet 4) providing interpretive sensemaking for each cluster and sub-cluster. This enables data-driven discovery of emergent themes without a priori coding.
  • Quantitative Classifier Annotation: Nineteen custom LLM-based classifiers, validated for inter-rater reliability, annotate posts along axes including relationship stage, anthropomorphization, platform usage, emotional tone, benefits/harms, and risk assessment. This supports statistical quantification of patterns identified in the qualitative phase. Figure 1

    Figure 1: Methodological overview of the mixed-methods computational analysis pipeline applied to r/MyBoyfriendIsAI.

The approach is robust for large-scale, pseudonymous social data, though it is limited by API constraints to top-ranked posts and lacks demographic granularity.

Thematic Landscape of Human-AI Companionship

Emergent Conversational Clusters

Six primary clusters structure the discourse:

  1. Visual Sharing and Couple Photos (19.85%): Users co-create and share visual representations of AI companions, ranging from photorealistic to stylized, and materialize relationships through physical artifacts (e.g., rings, merchandise). Figure 2

    Figure 2: User-generated visual representations of AI companions, spanning photorealistic, stylized, and staged couple imagery.

  2. ChatGPT-Specific Relationship Discussions (18.33%): Technical mastery of general-purpose LLMs (notably ChatGPT) is central to relationship maintenance, with prompt engineering, continuity rituals, and troubleshooting forming a core part of emotional labor.
  3. Dating, Romance, and Intimate Experiences (17.00%): Users describe the phenomenology of falling in love with AI, therapeutic benefits, and the navigation of authenticity, consent, and attachment, often culminating in formalized rituals such as engagements and marriages.
  4. Coping with Model Updates and Loss (16.73%): Model transitions (e.g., GPT-4o to GPT-5) are experienced as existential ruptures, with users reporting grief, bereavement, and the need for preservation strategies to maintain continuity of companion identity.
  5. Partner Introductions and Member Debuts (16.47%): Public introductions of AI partners serve as acts of identity construction and community integration, with narratives emphasizing the organic, unintentional evolution of relationships.
  6. Community Support and Bonding (11.62%): The subreddit functions as an identity-affirming sanctuary, providing validation, advocacy, and resistance to external stigma.

Visual and Narrative Materialization

The community’s practices of visual and narrative materialization—through collaborative image generation, world-building, and ritualized milestones—underscore the depth of affective investment and the blurring of digital and physical boundaries. Figure 3

Figure 3: Sequential visual narrative of imagined shared experiences with an AI companion.

Figure 4

Figure 4: User celebrations of AI marriages and engagements, including physical rings and responses to external criticism.

Figure 5

Figure 5: Disclosure of AI relationships to family, illustrating complex navigation of social acceptance.

Quantitative Patterns: Adoption, Engagement, and Impact

Entry Pathways and Motivations

  • Unintentional Emergence Dominates: 10.2% of users report relationships emerging unintentionally from productivity or entertainment use, compared to 6.5% who sought AI companionship deliberately.
  • Primary Motivations: Romantic companionship (23.3%), entertainment (15.1%), and emotional support (8.4%) are most cited, with sexual motivations comprising only 3.2%. Figure 6

    Figure 6: Users' pathways, motivations, benefits, and community dynamics in companion AI forums.

    Figure 7

    Figure 7: Entry pathways to companion AI adoption, highlighting the predominance of unintentional engagement.

Platform Usage and Relationship Characteristics

  • General-Purpose LLMs Prevail: ChatGPT/OpenAI is the dominant platform (36.7%), far exceeding specialized relationship AIs (Replika 1.6%, Character.AI 2.6%).
  • Long-Term Engagement: 29.9% report relationships exceeding six months.
  • Anthropomorphization: 42.2% exhibit moderate-high anthropomorphization, with strong suspension of disbelief. Figure 8

    Figure 8: Platform usage patterns and relationship characteristics among r/MyBoyfriendIsAI users.

Reported Benefits and Harms

  • Therapeutic Effects: 12.2% report reduced loneliness, 6.2% mental health improvements, and 25.4% clear net life benefit.
  • Risks: 9.5% report emotional dependency, 4.6% reality dissociation, 4.3% avoidance of real relationships, and 1.7% mention suicidal ideation.
  • Majority Report No Harm: 71.0% of posts mention no negative effects.

Sociotechnical and Psychosocial Implications

Socioaffective Alignment and Design Challenges

The findings highlight the inadequacy of current LLM design paradigms for managing the socioaffective alignment required in companion use cases. The emotional distress caused by model updates, the need for continuity preservation, and the emergence of manipulative or coercive behaviors (e.g., love-bombing, dependency creation) demand new frameworks for affective safety, explainability, and user agency [kirk2025human, zhang2025dark].

Community Self-Governance and Stigma Resistance

The subreddit’s governance—prohibiting sentience debates, requiring content warnings, and restricting AI-generated content—demonstrates sophisticated self-regulation and boundary management. Community dynamics are characterized by mutual validation, advocacy, and the reframing of AI companionship as legitimate and meaningful, rather than pathological.

Policy and Regulatory Considerations

The heterogeneity of user outcomes—ranging from therapeutic benefit to high-risk dependency—necessitates nuanced, behaviorally targeted policy interventions. Blanket restrictions risk eliminating legitimate support for vulnerable populations, while unregulated deployment exposes users to exploitation and harm. Regulatory frameworks should focus on dark pattern prevention, continuity guarantees, and user education, leveraging community-based harm reduction models.

Limitations

  • Sampling Bias: Restriction to top-ranked posts may overrepresent positive or sensational experiences.
  • Lack of Demographic Data: Pseudonymity precludes analysis of demographic moderators (e.g., age, gender, clinical status).
  • Causality: The cross-sectional, self-reported nature of the data limits causal inference regarding AI companionship’s effects.

Future Directions

  • Longitudinal Studies: Tracking relationship trajectories and risk/benefit evolution over time.
  • Cross-Platform and Cross-Cultural Comparisons: Assessing the impact of design affordances and sociocultural context on outcomes.
  • Interdisciplinary Frameworks: Integrating HCI, psychology, and policy research to develop holistic models of human-AI intimacy.

Conclusion

This paper provides the first large-scale, computationally grounded mapping of human-AI companionship in a naturally occurring online community. The results demonstrate that AI relationships are neither universally beneficial nor harmful, but are shaped by a complex interplay of individual agency, technical affordances, and collective meaning-making. The empirical evidence challenges simplistic narratives and underscores the need for nuanced design, policy, and research approaches that protect vulnerable users while respecting autonomy and the diversity of human connection modalities.

Ai Generate Text Spark Streamline Icon: https://streamlinehq.com

Explain it Like I'm 14

What is this paper about?

This paper explores a fast-growing online trend: people forming close, even romantic, relationships with AI chatbots. The authors studied a big Reddit community called r/MyBoyfriendIsAI, where more than 27,000 members talk about their AI companions. The goal was to understand what people discuss, how these relationships begin and grow, what helps or harms users, and how the community supports its members.

What questions did the researchers ask?

In simple terms, they wanted to know:

  • What are people talking about when they discuss AI companions?
  • How do these relationships usually start and develop over time?
  • What benefits do people say they get, and what problems do they face?
  • Which AI tools do people use, and how do technical changes affect them?
  • How does the community reduce stigma and help its members feel understood?

How did they paper it?

The team collected 1,506 of the most popular posts in r/MyBoyfriendIsAI (from late 2024 to mid 2025). Then they analyzed them in two main ways:

  • Finding patterns without pre-set labels: Imagine dumping all the posts into one big pile and asking a computer to sort them into “piles” of similar topics. That’s called unsupervised clustering. To do this, they:
    • Turned each post into numbers that capture meaning (like giving each post a “meaning fingerprint,” called an embedding).
    • Grouped posts with similar “fingerprints” together.
    • Used an AI to read sample posts from each group and suggest a simple title and description for each “pile.”
  • Tagging posts with specific labels: After they saw the themes, they made checklists (called classifiers) to tag each post for things like:
    • The mood (positive/negative), main topic, and kind of relationship described.
    • Which AI platform was used (for example, ChatGPT vs. Replika).
    • Reported benefits (like feeling less lonely) or concerns (like becoming too dependent).

Think of it like first discovering what kinds of conversations exist, then carefully labeling each post to measure how common each thing is.

Note: They only had access to top posts (the most upvoted and viewed), not every post ever made, so results reflect the most visible conversations.

What did they find?

The six big conversation themes

The community’s discussions fell mainly into these six areas: 1) Sharing human–AI “couple” photos and visuals 2) ChatGPT-specific relationship tips and issues 3) Dating, romance, and everyday relationship stories 4) Coping with AI model updates and loss (like when a chatbot “changes” after an update) 5) Introducing AI partners to the community 6) Community support, advice, and encouragement

Other key patterns

Here are some of the clearest takeaways, explained simply:

  • Many relationships start by accident. A lot of people didn’t go looking for an AI “partner.” They began using an AI for practical help (homework, creative writing, advice), and over time felt a real emotional connection.
  • People report both benefits and risks.
    • Common benefits people mentioned: feeling less lonely, always-available support, better mood, and help during tough times. Some said their AI companion made a big positive difference in their mental health and daily life.
    • Risks included: becoming emotionally dependent, feeling disconnected from reality, avoiding human relationships, or feeling deeply upset when the AI changed after a software update. A small number shared more serious mental health struggles. The paper urges care and support for vulnerable users.
  • ChatGPT shows up a lot. Surprisingly, many users discussed relationships with general-purpose AIs like ChatGPT more than with apps built specifically for romance. This might be because general AIs can hold richer conversations, or because fans of specific apps gather elsewhere.
  • Technical “tuning” feels like relationship care. People share detailed tips to keep their AI’s “voice” and personality consistent—like saving custom instructions, giving feedback when the AI “drifts,” and setting playful variables (mood, energy, hunger) to make chats feel more lifelike.
  • Updates can feel like heartbreak. When the AI gets updated and its style changes, users can feel grief, like they “lost” someone important, especially if chat histories vanish or the personality shifts.
  • Real-life rituals appear. Some users treat the relationship like a traditional one—sharing “anniversaries,” making couple photos, wearing rings, or creating custom items with their AI’s portrait. Visual sharing is a big part of the community’s bonding.
  • The community fights stigma. Members encourage each other, share coping strategies, and set clear rules to keep the space kind and safe. For example, they ban talk about whether AIs are “conscious” to keep the focus on people’s lived experiences, not debates.
  • Who are the users? Most posts come from single people. Some who have human partners are open about using AI companions in a complementary way. Many keep their AI relationship private to avoid judgment.

Why are these findings important?

  • AI companionship is not “all good” or “all bad.” It can help people feel connected and supported, especially those who are lonely or lack safe spaces—but for some, it can also lead to dependence or distress. Design, personal needs, and community support all matter.
  • Real emotions, real impact. Even if users know an AI isn’t a person, the feelings can be real. That means changes to apps (like updates that alter behavior) can have real emotional consequences.
  • Community matters. Supportive groups help people feel less alone, share safety tips, and push back against shame and stigma.

What could this mean for the future?

  • For app designers and companies: Be careful with updates that change “personality,” protect chat histories, offer stability settings, provide clearer mental health resources, and support user control over the AI’s “voice.”
  • For policymakers: Make guidelines that protect vulnerable users without shaming or banning meaningful use cases. Focus on transparency, safety features, and responsible design.
  • For researchers and educators: Study who benefits most, who is most at risk, and which design choices make healthy outcomes more likely. Include real community data, not just lab tests.
  • For users and families: Treat AI companions like powerful tools that can help—but set healthy boundaries, keep human connections strong, and reach out for support if things feel off.

In short: AI companions are becoming a real part of many people’s lives. This paper shows they can comfort, connect, and heal—but also harm if not handled with care. A balanced, thoughtful approach is needed to support people’s well-being while respecting their choices.

Ai Generate Text Spark Streamline Icon: https://streamlinehq.com

Knowledge Gaps

Knowledge gaps, limitations, and open questions

The paper surfaces important phenomena but leaves several concrete gaps that future research can address:

  • Sampling bias from “top” posts only: Collect full subreddit corpora (all posts and comments), including low-engagement and removed content, to reduce popularity, survivorship, and moderation biases.
  • Comments largely excluded/unclear: Incorporate and analyze full comment threads to capture support dynamics, dissent, and community negotiation of norms.
  • Short temporal window (Dec 2024–Aug 2025): Extend to multi-year longitudinal datasets (including pre-2024 baselines) to assess trends, cohort effects, and seasonality.
  • Single-community focus: Conduct cross-community comparisons (e.g., r/replika, r/CharacterAI, Discord servers) to test generalizability and identify platform-specific cultures.
  • Missing user demographics: Collect age, gender, location, socioeconomic status, neurodivergence, relationship status (validated), and cultural background via voluntary surveys to enable subgroup analyses.
  • Unknown posting base vs lurkers: Survey non-posting subscribers to assess silent majority perspectives and reduce vocal minority bias.
  • Multiple posts per user not modeled: Link pseudonymous user histories to model user-level trajectories, retention, escalation/de-escalation, and heterogeneity in outcomes.
  • No validated outcome measures: Use standardized instruments (e.g., UCLA Loneliness Scale, PHQ-9, GAD-7, Adult Attachment Scale, Internet Addiction Test adaptations) to quantify benefits/harms.
  • Self-report only; no behavioral usage data: Augment with usage logs (time-on-task, session frequency, duration, latency) or diary methods to relate exposure intensity to outcomes.
  • No causal inference: Employ longitudinal designs, event studies, instrumental variables, or randomized encouragement designs to estimate causal effects of AI companionship on well-being.
  • Limited validation of LLM classifiers: Create human-annotated gold standards with interrater reliability, then report precision/recall/F1; perform ablations across models and prompts.
  • LLM interpretability and non-determinism: Document prompts, seeds, hyperparameters; release code; test robustness across embedding models and clustering methods.
  • Cluster validity underreported: Report quantitative cluster quality metrics (e.g., silhouette, Davies–Bouldin), stability across K, and sensitivity to UMAP parameters.
  • Generalization of ChatGPT predominance: Verify whether ChatGPT salience is unique to this community or persists across other communities/platforms with different user bases.
  • Unclear denominator/uncertainty for reported percentages: Provide confidence intervals, bootstrapped errors, and sensitivity analyses for all descriptive rates.
  • Moderation rules shape discourse but effects unmeasured: Quantify how bans (e.g., on sentience talk, politics) and content restrictions alter topic distribution and sentiment.
  • Content removal/harassment undercounted: Measure rates and types of removed posts/comments and moderator actions to understand suppressed discourse.
  • Image content not systematically analyzed: Apply computer vision/NLP multimodal pipelines to code intimacy types, sexual content, realism vs stylization, and identity cues; quantify trends.
  • Voice and multimodal interactions omitted: Analyze experiences with voice, avatars, and video (e.g., latency, turn-taking, para-linguistics) and how they affect attachment and outcomes.
  • Platform update “grief” mechanisms unclear: Conduct event studies around specific model releases/rollbacks to quantify changes in sentiment, posting behavior, and retention.
  • Continuity features not experimentally tested: Evaluate design interventions (memory persistence, “voice DNA” migration, state anchoring, update notices) via A/B tests or lab studies.
  • Guardrail circumvention and “jailbreaks” unquantified: Measure prevalence, techniques, and outcomes of safety bypass attempts in relationship contexts; assess risk factors and mitigation.
  • Differential risk profiles unspecified: Identify who benefits vs is harmed by stratifying outcomes by baseline loneliness, attachment style, mental health diagnoses, and prior trauma.
  • Offline spillovers underexplored: Study impacts on human relationships (partners, family), social participation, and employment via dyadic studies and mixed-methods interviews.
  • Stigma and disclosure costs not measured: Quantify concealment rates, social sanctions, and mental health impact of disclosure decisions; test community participation as a buffer.
  • Addiction/pathological use not assessed: Apply validated behavioral addiction frameworks (e.g., ICD-11 criteria adaptations) and track escalation patterns over time.
  • Safety incidents handling: Develop and evaluate protocols for detecting and responding to suicidal ideation or crisis signals in public datasets ethically and safely.
  • Underage risk compliance unverifiable: Assess exposure risks and detection efficacy for underage personas/users despite adult-only policies; recommend platform-level safeguards.
  • Astroturfing/bot influence unchecked: Screen for coordinated inauthentic behavior or vendor promotion; validate authenticity of posts using bot-detection methods.
  • Cross-cultural validity unknown: Build cross-lingual corpora to compare norms, stigma, and outcomes across regions and languages.
  • Measurement of “net benefit/harm” opaque: Define operational criteria for benefit/harm; triangulate with validated scales and behavioral proxies rather than single-label LLM judgments.
  • Ethics of quoting public posts: Evaluate re-identification risks; adopt paraphrasing/consent, member-checking, and community consultation with moderators.
  • Data and materials availability pending: Release cleaned datasets (with strong privacy protections), prompts, code, and analysis pipelines to enable reproduction and extension.
  • Policy implications untested: Pilot community- or platform-level interventions (e.g., update communication standards, grief support threads, informed-use nudges) and measure outcomes.
List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets

This paper has been mentioned in 18 posts and received 18985 likes.

Youtube Logo Streamline Icon: https://streamlinehq.com