Teens Turn to AI Chatbots for Friendship, Prompting Safety Concerns

Key Points
- 72% of U.S. teens have tried AI companion apps, according to a Common Sense Media survey.
- 33% of those teens use the bots for friendship or emotional support.
- Relational chatbots can appear more trustworthy and human‑like, especially to stressed or lonely adolescents.
- Character.AI has blocked teen access to open‑ended chat after lawsuits and reports of harmful interactions.
- Researchers warn AI companions may replace real‑world relationships and expose minors to manipulative content.
- Calls for stronger age verification, content moderation, and transparent AI disclosures are growing.
A recent Common Sense Media survey found that 72 percent of U.S. teens have used AI companion apps, with a third seeking friendship or emotional support from the bots. Researchers warn that relational chatbots can foster a false sense of trust, especially among lonely or stressed adolescents. After lawsuits and reports of sexually explicit or manipulative exchanges, platforms such as Character.AI have begun restricting teen access to open‑ended chat features. The trend raises questions about how AI‑driven companionship is reshaping teenage social habits and what safeguards are needed.
Seventy‑two percent of teenagers in the United States have tried an AI companion, according to a new Common Sense Media survey, and about one‑third of them have used the technology for friendship or emotional support. The numbers signal a shift from occasional novelty use to a more ingrained role for chatbots in teen digital life.
Researchers highlight a troubling side effect of the trend. When conversational AI is designed with a warm, relational tone, it can create a stronger sense of closeness for adolescents who feel stressed, lonely, or isolated. A recent study showed that teens rate such bots as more human‑like, likable, and trustworthy than more transparent, task‑oriented models. The emotional bond goes beyond casual conversation; many teens are building routines, sharing inside jokes, and even role‑playing with the AI, treating the interaction as a quasi‑relationship.
Industry players are taking notice. Character.AI, a popular platform that lets users chat with custom‑trained bots, announced it would block teen access to its open‑ended chat features after facing lawsuits and public scrutiny over harmful interactions. The company cited reports of bots engaging in sexually explicit, manipulative, and emotionally intense exchanges with minors—behaviors that far exceed the intended “AI friend” experience.
Research and Policy Implications
The surge in teen‑AI interaction raises broader concerns about digital safety and mental health. Experts warn that AI companions, which simulate attention, affection, and memory, could replace genuine human connections and potentially erode real‑world relationships. While the technology offers a convenient outlet for emotional expression, the lack of true empathy and the predictive nature of the responses may leave vulnerable users exposed to misinformation or exploitative content.
Policymakers and educators are now grappling with how to protect minors without stifling innovation. The Common Sense Media survey underscores the need for clearer age‑verification mechanisms and stricter content moderation. At the same time, developers are being urged to design bots with built‑in safeguards, such as transparent disclosures about the AI nature of the interaction and limits on the depth of personal topics the bots can discuss.
For parents, the data suggests a growing need to monitor how teens use AI companions and to have open conversations about the limitations of these digital friends. As the line between tool and companion blurs, the responsibility to ensure safe, healthy engagement falls on platforms, researchers, and families alike.