The data, taken from a site called Muah.ai and viewed by 404 Media, includes chatbot prompts that reveal users’ sexual fantasies. In many instances, users are trying to create chatbots that roleplay child sexual abuse scenarios. These prompts are in turn linked to email addresses, many of which appear to be personal accounts with users’ real names.
Cox, J. (2024, October 8). Hacked ‘AI Girlfriend’ Data Shows Prompts Describing Child Sexual Abuse. 404 Media. https://www.404media.co/hacked-ai-girlfriend-data-shows-prompts-describing-child-sexual-abuse-2/