Meta’s deepfake moderation isn’t good enough, says Oversight Board

Published: (March 10, 2026 at 06:01 AM EDT)
2 min read
Source: The Verge

Source: The Verge

Meta’s Oversight Board Calls for Stronger AI Labeling

Meta’s Oversight Board wants the company to start taking AI labeling seriously to protect its users from online misinformation.
Image credit: Cath Virginia / The Verge, Getty Images

The Board says Meta’s methods for identifying deepfakes are “not robust or comprehensive enough” to handle the rapid spread of misinformation during armed conflicts such as the Iran war.

The concern arose from an investigation into a fake AI video that claimed to show damage to buildings in Israel. The video was shared on Meta’s platforms—Facebook, Instagram, and Threads—last year. While the Board’s full recommendations are extensive, a key point is that the current approach to surfacing and labeling AI‑generated content needs a major overhaul.

Key Findings

  • Existing deepfake detection tools are insufficient for the speed at which false content spreads in conflict zones.
  • Users are often unable to distinguish AI‑generated media from authentic footage, increasing the risk of misinformation.
  • The Board’s investigation highlighted gaps in Meta’s labeling policies across its major platforms.

Recommendations

  • Implement a comprehensive AI‑labeling system that clearly marks synthetic media.
  • Deploy more robust detection technologies that can keep pace with evolving deepfake techniques.
  • Ensure consistent labeling practices across Facebook, Instagram, and Threads.

For the full story, see the original article on The Verge.

0 views
Back to Blog

Related posts

Read more »