![]() In subsequent months, at least three artificially generated videos that appear to show me engaging in sex acts were uploaded to websites specializing in deepfake porn. Last year, I resigned as head of the Department of Homeland Security’s Disinformation Governance Board, a policy-coordination body that the Biden administration let founder amid criticism mostly from the right. Read: We haven’t seen the worst of fake news Yet policy makers have all but ignored an urgent AI problem that is already affecting many lives, including mine. ![]() Many commentators have been tying themselves in knots over the potential threats posed by artificial intelligence-deepfake videos that tip elections or start wars, job-destroying deployments of ChatGPT and other generative technologies. The only emotion I felt as I informed my lawyers about the latest violation of my privacy was a profound disappointment in the technology-and in the lawmakers and regulators who have offered no justice to people who appear in porn clips without their consent. For more than a year, I have been the target of a widespread online harassment campaign, and deepfake porn-whose creators, using artificial intelligence, generate explicit video clips that seem to show real people in sexual situations that never actually occurred-has become a prized weapon in the arsenal misogynists use to try to drive women out of public life. ![]() Recently, a Google Alert informed me that I am the subject of deepfake pornography.
0 Comments
Leave a Reply. |
AuthorWrite something about yourself. No need to be fancy, just an overview. ArchivesCategories |