AI-Facilitated Coercive Control: An Experimental Study

要旨

We present an experimental study that investigates how LLM-driven conversational AI tools might be weaponized to facilitate, exacerbate, or commoditize coercive control. Inspired by speculative design, we construct four scenarios that combine well-known coercive control tactics with the current capabilities of conversational AI tools. Then, we explore these scenarios via interactions with popular AI agents (ChatGPT, Gemini). We find that although AI tools refuse straightforward requests for harmful content, their guardrails can be circumvented via strategies such as gradual persuasion, splitting conversations, pre-prompting, and manipulating the AI agent's settings. Collectively, these strategies enable AI agents to be leveraged in ways that facilitate harassment, intimidation, gaslighting, monitoring, surveillance, and other coercive control tactics. To make these tools safer for everyone, we discuss opportunities for AI agents to resist being abused for coercive control via analysis of users’ conversational patterns, and ensuring that pre-programmed settings are clearly visible to prevent covert manipulation.

著者
Haesoo Kim
Cornell University, Ithaca, New York, United States
Thomas Ristenpart
Cornell Tech, New York, New York, United States
Nicola Dell
Cornell Tech, New York, New York, United States

会議: CHI 2026

ACM CHI Conference on Human Factors in Computing Systems

セッション: Dark Patterns, Deception, and Manipulative Interfaces

P1 - Room 128
7 件の発表
2026-04-16 18:00:00
2026-04-16 19:30:00