In a disturbing development highlighting the misuse of artificial intelligence, an OnlyFans creator, Bunni, uncovered that her personal photos were circulating on Reddit with her face replaced by an AI-generated visage. While image theft is an unfortunate reality for many online creators, this incident marks a troubling evolution in digital impersonation.
Bunni, a 25-year-old content creator from the UK, was alerted by followers who stumbled upon a Reddit profile under the name “Sofía.” The account featured images unmistakably taken from Bunni’s social media, identifiable by her clothing, tattoos, and surroundings, but with her facial features altered to resemble a different individual. The fabricated persona claimed to be a 19-year-old from Spain, engaging users across various subreddits, including r/teenagers and r/PunkGirls.
The impersonator utilized these AI-manipulated images to interact with Reddit users, seeking validation and potentially directing them to purchase content, possibly even Bunni’s own, under false pretenses. Although the “Sofía” account has since been removed following reports to subreddit moderators, the incident raises significant concerns about the ease with which AI tools can be exploited for deceptive purposes.
This case underscores the growing challenges posed by deepfake technology, where AI can seamlessly alter images and videos, making it increasingly difficult to discern authenticity. The psychological and reputational harm inflicted on individuals like Bunni is profound, as they grapple with violations of their identity and autonomy.
Want to be featured here? Become an advertiser
In response to the escalating threat of AI-driven impersonation and non-consensual content distribution, the U.S. government enacted the “Take It Down Act” on May 19, 2025. Signed into law by President Donald Trump, the legislation criminalizes the sharing of non-consensual intimate images, including those generated or altered by AI. Under the Act, platforms are mandated to remove such content within 48 hours of notification, with enforcement overseen by the Federal Trade Commission.
While the law represents a significant step toward protecting individuals from digital exploitation, experts caution that technological advancements continue to outpace regulatory measures. The accessibility of AI tools capable of creating convincing deepfakes necessitates ongoing vigilance and adaptation of legal frameworks to safeguard against emerging forms of online abuse.