Share this article

Deepfake Scams Are a Threat to Humankind — Here’s How to Fight Back

Zero-knowledge proofs add a crucial integrity layer to AI moderation, enabling companies to verify decisions without exposing sensitive data or proprietary models, says Ismael Hishon-Rezaizadeh, CEO & Founder of Lagrange.

Jul 17, 2025, 2:14 p.m.
(Unsplash/Modified by CoinDesk)

What to know:

  • Hong Kong police dismantled a $46 million cryptocurrency scam using advanced deepfake technology.
  • AI-generated deepfakes are now considered a significant threat to national security and financial systems.
  • Zero-knowledge machine learning offers a promising solution for verifiable and efficient content moderation.

Last year, Hong Kong police arrested a group responsible for a $46 million cryptocurrency investment scam that used deepfake technology. The group collaborated with overseas scam networks to create convincing fake investment platforms. Today, these tools have become exponentially more advanced, now extending into AI-generated video, which is evolving far faster than any other form of media.

Malicious AI contributed to over $12 billion in global losses from fraud in 2024. The U.S. Department of Homeland Security now calls AI-generated deepfakes a “clear, present, and evolving threat” to national security, finance and society. Denmark is considering amending its copyright law to combat unauthorized deepfakes, giving every person “the right to their own body, facial features and voice.”

STORY CONTINUES BELOW
Don't miss another story.Subscribe to the CoinDesk Headlines Newsletter today. See all newsletters

Deepfakes are clearly an escalating societal threat. To defend the digital world, we need AI to be verifiable and content moderation to be backed by cryptographic proof, not just trust. Zero-knowledge machine learning (zkML) techniques are opening up new ways to prove that outputs are valid, without exposing the underlying model or data.

Current Moderation Is Broken

Contemporary content moderation struggles to keep up with AI manipulation. When a piece of malicious content is uploaded across multiple platforms, each platform must independently reclassify that duplicate content, wasting computational resources and adding latency.

Worse, each platform’s algorithms and policies may differ, and a video that’s flagged on one site might be deemed benign on another. The entire process lacks transparency, and the platforms’ AI decision-making exists in a “black box.” Users rarely know why something was removed or allowed.

This fragmented approach to moderation makes it more challenging for detection tools to perform well. One study found that detection models’ accuracy “drops sharply” on authentic wild data, sometimes degrading to random guessing when faced with novel deepfakes.

Businesses are alarmingly underprepared, with 42% of companies admitting they are only “somewhat confident” in their ability to spot deepfakes. Constantly re-scanning content and chasing new forgeries is a losing battle. We need a systemic fix that makes moderation results portable, trustworthy, and efficient across the web.

Solution: Verifiable Moderation

Zero-knowledge Machine Learning (zkML) provides a method for validating AI-based moderation decisions without duplicating work or disclosing sensitive information. The idea is to have AI classifiers produce not just a label, but also a cryptographic proof of that classification.

Imagine a moderation model that evaluates a piece of content (an image, video, text, etc.) and assigns it one or more labels (for example, Safe for Work, Not Safe for Work, Violent, Pornographic, etc.). Along with the labels, the system generates a zero-knowledge proof attesting that a known AI model processed the content and produced those classification results. This proof is embedded into the content’s metadata, allowing the content itself to carry a tamper-evident moderation badge. Content producers or distributors could also be cryptographically bound to the moderation status of their content.

When the content is uploaded or shared, platforms can instantly verify the proof using lightweight cryptographic checks. If the proof is valid, the platform trusts the provided classification without needing to re-run its own AI analysis.

Benefits of ZK-Embedded Moderation

Let’s think about the benefits here. Verifying a proof is much faster and simpler than running a large AI model on every upload. First, we have the portability of content moderation, where its status can travel with it. We also ensure transparency through openly verifiable outcomes, allowing anyone to verify the cryptographic proof and confirm how the content was labeled.

In this scenario, moderation becomes a one-time computation per content item, with subsequent checks reduced to inexpensive proof verifications. All of this translates into huge compute savings, lower latency in content delivery, and more AI resources to focus on truly new or disputed content.

As AI-generated content continues to explode, zk-enabled moderation can handle the scale. This approach lightens the load on platforms, enabling moderation to keep pace with high-volume streams in real time.

The Integrity Layer for AI

Zero-knowledge proofs provide the missing integrity layer that AI-based moderation needs. They allow us to prove that AI decisions (like content classifications) were made correctly, without revealing sensitive inputs or the internals of the model. This means companies can enforce moderation policies and share trustworthy results with each other, or the public, all while protecting user privacy and proprietary AI logic.

Embedding verifiability at the content level can transform opaque and redundant moderation systems into a decentralized, cryptographically verifiable web of trust. Instead of relying on platforms to say “trust our AI filters,” moderation outputs could come with mathematical guarantees. If we don’t integrate this type of scalable verifiability now, AI-driven manipulation and misinformation could erode the last shreds of online trust.

We can still turn AI moderation from an act of faith into an act of evidence — and in doing so, rebuild trust not just in platforms, but in the information ecosystems that shape public discourse, elections, and our shared sense of reality.

Note: The views expressed in this column are those of the author and do not necessarily reflect those of CoinDesk, Inc. or its owners and affiliates.

More For You

Protocol Research: GoPlus Security

GP Basic Image

What to know:

  • As of October 2025, GoPlus has generated $4.7M in total revenue across its product lines. The GoPlus App is the primary revenue driver, contributing $2.5M (approx. 53%), followed by the SafeToken Protocol at $1.7M.
  • GoPlus Intelligence's Token Security API averaged 717 million monthly calls year-to-date in 2025 , with a peak of nearly 1 billion calls in February 2025. Total blockchain-level requests, including transaction simulations, averaged an additional 350 million per month.
  • Since its January 2025 launch , the $GPS token has registered over $5B in total spot volume and $10B in derivatives volume in 2025. Monthly spot volume peaked in March 2025 at over $1.1B , while derivatives volume peaked the same month at over $4B.

More For You

The Grandma Test: When Your Mom Can Use DePIN, Mass Adoption Has Arrived

Grandma (Unsplash/CDC/Modified by CoinDesk)

Mass adoption doesn't happen when crypto enthusiasts start using the technology: it happens when your grandma does it without even realizing it, argues Uplink co-founder Carlos Lei.