In a groundbreaking legal move, San Francisco City Attorney David Chiu has filed a lawsuit to target and shut down some of the most popular AI “undressing” websites. Announced at a press conference on Thursday, the lawsuit aims to take action against 16 major sites accused of violating federal and state laws related to revenge pornography, deepfake pornography, and child exploitation.
Accusations and Legal Basis:
According to the lawsuit, these websites are accused of violating federal laws that protect individuals from non-consensual image manipulation. This includes federal statutes against revenge pornography, deepfake pornography, and child pornography. Chiu’s office also claims that the sites violate California’s unfair competition law, stating, “The harm they cause to consumers greatly outweighs any benefits associated with those practices.” This statement was outlined in the complaint for injunctive relief, filed in a California superior court.
The lawsuit is significant not only because it targets major players in the AI undressing website market but also because it highlights the growing legal battle against AI-generated deepfakes. These technologies, which can undress or manipulate images of individuals without their consent, are creating widespread harm, particularly for women and minors.
Targeted Websites and Defendants:
The complaint names a total of 50 defendants who operate undressing websites, many of whom were not publicly identified. However, the lawsuit does highlight several key defendants, including companies like Sol Ecom (based in Florida), Briver (from New Mexico), and Itai Tech Ltd (based in the UK). These companies are identified as operators of some of the world’s most trafficked undressing sites.
One notable individual defendant is Augustin Gribinets, a resident of Estonia. Gribinets is accused of operating an AI undressing website that has hosted non-consensual images of women and children. The city attorney alleges that these websites, including Gribinets’ platform, are used to “bully, threaten, and humiliate women and girls.”
The Impact of AI Undressing Websites:
These websites have seen an alarming rise in traffic, with over 200 million visits in just a six-month period. The non-consensual images found on these sites have been linked to online bullying, humiliation, and significant emotional distress for victims. The lawsuit outlines one troubling case from February in which AI-generated nude images of 16 middle school students were circulated in a California school district. Although this incident likely refers to a case in Beverly Hills, where students were expelled for circulating fake nude images, it demonstrates the real-world consequences of AI undressing technology.
Deepfakes and Federal Legal Concerns:
The rise of AI-powered deepfake technology has sparked national debate and concern over the need for stronger regulations. Just last month, the U.S. Copyright Office issued a report on digital replicas, calling for new laws to address the misuse of deepfake technology. Soon after, a bipartisan group of senators introduced the NO FAKES Act, which aims to protect individuals from having their voice, face, or body replicated by AI without consent. This new legislation could provide crucial legal backing for cases like San Francisco’s lawsuit.
Moving Forward:
San Francisco’s lawsuit against these AI undressing sites marks a critical step in the fight against non-consensual image manipulation and exploitation. If successful, the legal action could set a precedent for other states and jurisdictions to follow suit, creating more robust legal protections for individuals affected by deepfake and AI-generated content.
The outcome of this lawsuit could have far-reaching implications for how AI technologies are regulated and how privacy rights are enforced in an increasingly digital world. As deepfake technology continues to evolve, the legal battle to protect victims will likely intensify.
Conclusion
San Francisco’s legal action represents a pivotal moment in the fight against AI-powered undressing websites. By targeting these harmful platforms, the city aims to set a precedent for the ethical and responsible use of AI, ensuring that these technologies are not misused to harm or exploit vulnerable individuals.
My name is Augustus, and I am dedicated to providing clear, ethical, and current information about AI-generated imagery. At Undress AI Life, my mission is to educate and inform on privacy and digital rights, helping users navigate the complexities of digital imagery responsibly and safely.