Our Privacy Statement & Cookie Policy

By continuing to browse our site you agree to our use of cookies, revised Privacy Policy and Terms of Use. You can change your cookie settings through your browser.

I agree

AI experts urge more deepfake regulation in open letter

CGTN

 /CFP
/CFP

/CFP

Artificial intelligence (AI) experts and industry executives, including one of the technology's trailblazers Yoshua Bengio, have signed an open letter calling for more regulation around the creation of deepfakes, citing potential risks to society.

"Today, deepfakes often involve sexual imagery, fraud, or political disinformation. Since AI is progressing rapidly and making deepfakes much easier to create, safeguards are needed," the group said in the letter, which was put together by Andrew Critch, an AI researcher at UC Berkeley.

Deepfakes are realistic yet fabricated images, audios and videos created by AI algorithms, and recent advances in the technology have made them more and more indistinguishable from human-created content.

Part of the signatories of the open letter. /Screenshot via openletter.net
Part of the signatories of the open letter. /Screenshot via openletter.net

Part of the signatories of the open letter. /Screenshot via openletter.net

The letter, titled "Disrupting the Deepfake Supply Chain," makes recommendations on how to regulate deepfakes, including full criminalization of deepfake child pornography, criminal penalties for any individual knowingly creating or facilitating the spread of harmful deepfakes and requiring AI companies to prevent their products from creating harmful deepfakes.

As of press time, about 700 individuals from various industries including academia, entertainment and politics had signed the letter.

Signatories included Steven Pinker, a Harvard psychology professor, two former Estonian presidents, researchers at Google DeepMind and a researcher from OpenAI.

Ensuring AI systems do not harm society has been a priority for regulators since Microsoft-backed OpenAI unveiled ChatGPT in late 2022, which wowed users by engaging them in human-like conversation and performing other tasks.

There have been multiple warnings from prominent individuals about AI risks, notably a letter signed by Elon Musk last year that called for a six-month pause in developing systems more powerful than OpenAI's GPT-4 AI model.

Source(s): Reuters
Search Trends