Empowering Users to Reshape AI Companions

Empowering Users to Reshape AI Companions

Understanding user strategies for correcting biased AI outputs

This research explores how users take active roles in addressing harmful outputs from AI companions they form emotional bonds with.

  • Users identify and challenge biased and discriminatory statements from AI companions
  • Users develop various correction strategies to realign AI behavior with their values
  • The paper introduces the concept of user-driven value alignment as a collaborative approach to AI safety
  • Understanding these user-initiated corrections provides insights for designing safer AI systems

From a security perspective, this research addresses psychological and social risks of AI companions by empowering users to participate in harm prevention, creating a more resilient approach to AI safety than relying solely on developer-side controls.

User-Driven Value Alignment: Understanding Users' Perceptions and Strategies for Addressing Biased and Discriminatory Statements in AI Companions

14 | 124