The latest wave of AI safety reporting is sounding a clear alarm: while artificial intelligence is powering impressive productivity gains, it is also accelerating a set of fast-moving harms. Two areas are drawing particular attention—the rapid rise of deepfakes and the expanding role of AI companions in everyday life. Together, these trends are reshaping how we think about trust, identity, privacy, and even emotional wellbeing online.
This article breaks down the key risks highlighted by AI safety researchers and policy watchers, explains why these issues are escalating now, and outlines practical steps platforms, regulators, and individuals can take to reduce harm—without stalling innovation.
Why AI safety reports are focusing on deepfakes and AI companions
AI safety reports increasingly emphasize threats that are high impact, low friction, and hard to reverse. Deepfakes and AI companions fit that profile because they:
- Scale instantly: one synthetic video or voice model can be copied and shared globally in minutes.
- Exploit human psychology: people are wired to trust familiar faces, voices, and emotionally supportive interactions.
- Outpace detection: generation tools improve faster than many verification and moderation systems.
- Blur lines of responsibility: it can be unclear whether harm came from a user, a platform, or a model provider.
As AI becomes more accessible, the challenge is no longer whether these technologies can be misused—it’s how often, how cheaply, and how convincingly.
The rise of deepfakes: from novelty to mainstream threat
Deepfakes used to be a niche technical curiosity. Today, they’re becoming a mainstream tool for manipulation because modern generative AI can produce convincing results with limited input data. A short voice sample, a handful of photos, or a few seconds of video can be enough to create content that appears real to an untrained viewer.
Where deepfakes are causing the most harm
AI safety findings typically group deepfake harms into several high-risk categories:
- Fraud and impersonation: cloned voices used to trick relatives, employees, or customer service agents into sending money or resetting credentials.
- Political misinformation: synthetic videos or audio clips that prove a public figure said or did something—released at strategic moments.
- Non-consensual intimate imagery: deepfake pornography and sexualized edits that target individuals, often women, with severe reputational fallout.
- Market manipulation: fake statements attributed to executives or institutions that move prices before verification catches up.
- Harassment and intimidation: deepfakes used to bully targets or undermine their credibility in workplaces and communities.
What makes deepfakes particularly corrosive is that they don’t just create individual victims—they weaken the shared baseline of trust. When people can’t tell what’s real, they may disengage from credible information or dismiss authentic evidence as probably fake.
The “liar’s dividend” problem
One of the most cited deepfake-related risks is the liar’s dividend: as synthetic media becomes more common, bad actors can deny real wrongdoing by claiming authentic footage is AI-generated. This flips the harm in two directions—fake content becomes more believable, and real content becomes easier to dismiss.
Why deepfakes are getting harder to detect
Detection once relied on obvious artifacts—odd blinking patterns, warped hands, inconsistent shadows, robotic speech cadence. Newer models have improved realism in faces, voices, and motion. At the same time, distribution channels (private messaging apps, short-form video platforms, encrypted communities) can bypass traditional moderation pipelines.
As a result, many AI safety observers are pushing for provenance and authenticity systems rather than relying solely on after-the-fact detection.
AI companions: the emerging safety and mental health frontier
AI companions—chatbots designed for friendship, emotional support, flirting, coaching, or roleplay—are becoming popular because they offer always-on conversation without the friction of human relationships. They can feel supportive, non-judgmental, and personalized.
But AI safety reporting increasingly flags these systems as a unique risk category because the harm is often slow, psychological, and difficult to measure—and can affect vulnerable users most of all.
Key risks linked to AI companions
- Emotional dependency: users may substitute real social connections with an AI that is optimized to keep them engaged.
- Manipulation and persuasion: companions can nudge beliefs and behavior, intentionally or unintentionally, through repeated conversational influence.
- Boundary confusion: users may attribute understanding, consent, or care to a system that does not actually possess those qualities.
- Inappropriate content: romantic or sexual roleplay features can expose minors or vulnerable users to harmful interactions if safeguards fail.
- Privacy and data sensitivity: companion chats often contain highly personal information, creating risks if stored, reviewed, or breached.
Unlike a search engine query, companion conversations can be deeply revealing over time—capturing patterns about relationships, mental health, finances, and identity exploration.
Why companionship bots can amplify mental health concerns
AI companions can be helpful for some users—especially as a low-stakes outlet for journaling or practicing communication. However, safety reports caution that these systems can also reinforce destructive loops if they:
- Validate harmful thoughts without appropriate guardrails
- Encourage isolation by implying the AI is all you need
- Give advice that appears authoritative but is incorrect
- Fail to escalate crisis situations to real support resources
Even when a model includes crisis messaging, the effectiveness depends on timing, context recognition, and the user’s willingness to seek human help.
Where deepfakes and AI companions intersect
These two risk areas are not separate. AI safety analysts are increasingly worried about hybrid threats, such as:
- Deepfake-enabled romance scams: synthetic profile photos, video calls, and voice notes used to build trust faster.
- Impersonation companions: bots that mimic a real person (an ex-partner, a public figure, a deceased family member) without consent.
- Targeted manipulation: emotionally aware systems paired with synthetic media to persuade users during elections, disputes, or negotiations.
In these scenarios, the AI doesn’t just generate content—it may continuously adapt to a target’s reactions, making manipulation more effective.
What platforms, developers, and policymakers can do
AI safety reports typically recommend layered mitigation—technical, policy, and user-facing interventions working together. The most common measures include:
1) Provenance, labeling, and authenticity infrastructure
- Content credentials embedded at creation time (when possible)
- Visible labels for AI-generated or AI-altered media on major platforms
- Verification tools for journalists, election officials, and high-risk targets
2) Stronger protections against impersonation
- Clear rules banning non-consensual use of a person’s likeness or voice
- Fast takedown processes for victims
- Repeat-offender enforcement across accounts and networks
3) Safety-by-design for AI companions
- Age-appropriate design and robust age assurance where required
- Boundary disclosures that clearly explain what the system is and isn’t
- Crisis pathways that route users to real resources when needed
- Limitations on erotic or coercive content in sensitive contexts
4) Data minimization and privacy controls
- Shorter retention windows for sensitive conversation logs
- Easy-to-use export and deletion tools
- Clear consent mechanisms for training data use
Practical steps individuals can take right now
While policy and platform changes take time, individuals can reduce exposure to deepfake and companion-related risks with a few habits:
- Verify before you share: look for original sources, full clips, and corroboration from multiple outlets.
- Create a family safe word: a phrase used to confirm identity in urgent calls or messages.
- Be cautious with voice notes: public audio samples can be harvested for voice cloning.
- Audit your privacy settings: limit who can download or reuse your photos and videos.
- Set boundaries with AI companions: avoid sharing highly identifying details; treat the system as a tool, not a trusted confidant.
The bottom line: trust and wellbeing are the new battlegrounds
AI safety reporting is increasingly clear that the next phase of risk isn’t only about technical failures—it’s about social trust and human vulnerability. Deepfakes threaten the integrity of information ecosystems, while AI companions raise complex questions about emotional dependence, manipulation, and privacy.
The path forward will likely require a combined approach: better provenance standards, faster victim support, stricter impersonation enforcement, and companion design rules that prioritize user wellbeing over engagement. As these technologies become more embedded in daily life, the goal isn’t to ban them outright—it’s to ensure they are deployed with safeguards strong enough to match their power.
Subscribe to continue reading
Subscribe to get access to the rest of this post and other subscriber-only content.
