OpenAI Releases ‘Deepfake’ Detector to Disinformation Researchers

1 week ago 64
  • Harvesting Data for A.I.
  • A.I. Data Race, Explained
  • What Is ‘Synthetic Data’?
  • Key Figures in the Field
  • A.I. Faces Quiz

The prominent A.I. start-up is also joining an industrywide effort to spot content made with artificial intelligence.

Large computer monitors sit on an office reception desk with a tablet computer alongside, its screen reading “OpenAI.” A framed black-and-white outline of a person’s is on the back wall.
Open AI’s offices in San Francisco. The company said that its new detector can identify 98.8 percent of images created by DALL-E 3, its own generator.Credit...Jim Wilson/The New York Times

Cade MetzTiffany Hsu

May 7, 2024, 6:00 a.m. ET

As experts warn that images, audio and video generated by artificial intelligence could influence the fall elections, OpenAI is releasing a tool designed to detect content created by its own popular image generator, DALL-E. But the prominent A.I. start-up acknowledges that this tool is only a small part of what will be needed to fight so-called deepfakes in the months and years to come.

On Tuesday, OpenAI said it would share its new deepfake detector with a small group of disinformation researchers so they could test the tool in real-world situations and help pinpoint ways it could be improved.

“This is to kick-start new research,” said Sandhini Agarwal, an OpenAI researcher who focuses on safety and policy. “That is really needed.”

OpenAI said its new detector could correctly identify 98.8 percent of images created by DALL-E 3, the latest version of its image generator. But the company said the tool was not designed to detect images produced by other popular generators like Midjourney and Stability.

Image

A standard promoted by the Coalition for Content Provenance and Authenticity shows when and how images were produced or altered.Credit...via OpenAI

Because this kind of deepfake detector is driven by probabilities, it can never be perfect. So, like many other companies, nonprofits and academic labs, OpenAI is working to fight the problem in other ways.

Like the tech giants Google and Meta, the company is joining the steering committee for the Coalition for Content Provenance and Authenticity, or C2PA, an effort to develop credentials for digital content. The C2PA standard is a kind of “nutrition label” for images, videos, audio clips and other files that shows when and how they were produced or altered — including with A.I.

OpenAI also said it was developing ways of “watermarking” A.I.-generated sounds so they could easily be identified in the moment. The company hopes to make these watermarks difficult to remove.

Anchored by companies like OpenAI, Google and Meta, the A.I. industry is facing increasing pressure to account for the content its products make. Experts are calling on the industry to prevent users from generating misleading and malicious material — and to offer ways of tracing its origin and distribution.

In a year stacked with major elections around the world, calls for ways to monitor the lineage of A.I. content are growing more desperate. In recent months, audio and imagery have already affected political campaigning and voting in places including Slovakia, Taiwan and India.

OpenAI’s new deepfake detector may help stem the problem, but it won’t solve it. As Ms. Agarwal put it: In the fight against deepfakes, “there is no silver bullet.”

Cade Metz writes about artificial intelligence, driverless cars, robotics, virtual reality and other emerging areas of technology. More about Cade Metz

Tiffany Hsu reports on misinformation and disinformation and its origins, movement and consequences. She has been a journalist for more than two decades. More about Tiffany Hsu