Women sue the men who used their Instagram feed to create AI porn influencers
Summary
Ars Technica reports on a Arizona lawsuit where MG and two co-plaintiffs allege three men scraped images from women's social media to train AI models that generate pornographic influencers, monetized via Fanvue and Whop. The article outlines how platforms and regulators are responding, including references to the Take It Down Act (effective May 2026) and state deepfake laws, and discusses the challenges of applying policy to AI-generated imagery. It also highlights broader implications for privacy, consent, and the online presence of ordinary people.