Futures
Access hundreds of perpetual contracts
CFD
Gold
One platform for global traditional assets
Options
Hot
Trade European-style vanilla options
Unified Account
Maximize your capital efficiency
Demo Trading
Introduction to Futures Trading
Learn the basics of futures trading
Futures Events
Join events to earn rewards
Demo Trading
Use virtual funds to practice risk-free trading
Launch
CandyDrop
Collect candies to earn airdrops
Launchpool
Quick staking, earn potential new tokens
HODLer Airdrop
Hold GT and get massive airdrops for free
Pre-IPOs
Unlock full access to global stock IPOs
Alpha Points
Trade on-chain assets and earn airdrops
Futures Points
Earn futures points and claim airdrop rewards
Promotions
AI
Gate AI
Your all-in-one conversational AI partner
Gate AI Bot
Use Gate AI directly in your social App
GateClaw
Gate Blue Lobster, ready to go
Gate for AI Agent
AI infrastructure, Gate MCP, Skills, and CLI
Gate Skills Hub
10K+ Skills
From office tasks to trading, the all-in-one skill hub makes AI even more useful.
GateRouter
Smartly choose from 40+ AI models, with 0% extra fees
YouTube Launches AI Deepfake Detection Tool: Creators Over 18 Can Detect Portraits and Request Removal
YouTube announces the opening of its “portrait detection” tool to all creators over 18, automatically scanning AI deepfake content in platform videos and allowing individuals to request takedowns.
(Background summary: Anthropic report: The AI dominance race in 2028, if the US doesn’t maintain its computing advantage, China may overtake it)
(Additional context: Elon Musk open-sources the latest algorithm for the X platform! Incorporates Grok AI model architecture, releasing a 3GB pre-trained model for anyone to experiment with)
Table of Contents
Toggle
The threshold for AI-generated deepfake videos (deepfake, which uses AI technology to synthesize fake videos of people’s faces) has exploded as technology advances. The growth rate of counterfeit content on YouTube has long surpassed what manual reporting can handle.
To address this issue, YouTube chose to open the detection tool directly to every adult creator this week.
Two years, from preview to full release
In fact, the timeline for this feature has been quite cautious. In 2024, YouTube internally previewed the portrait detection system. By the end of 2025, the feature was initially opened to the Partner Program (creators participating in ad revenue sharing), journalists, and political figures.
On April 21, 2026, YouTube expanded access to the core entertainment industry: top Hollywood agencies CAA, UTA, WME, and Untitled Management, along with their celebrities and artists, were the first to gain usage rights.
Until this week, YouTube announced full open access: all creators over 18, regardless of channel size, can apply to use it. YouTube’s official wording is, “Whether you’ve been uploading videos on YouTube for ten years or just starting out, you can enjoy the same level of protection.”
Another noteworthy point: ordinary users without a YouTube channel can also apply for this tool. The protection scope extends from “creators” to “any adults with portrait risks online.”
The face version of Content ID
The technical logic behind portrait detection is highly similar to Content ID, which YouTube has been running for over fifteen years.
Content ID is YouTube’s copyright protection system. The principle is that copyright holders pre-submit audio or video fingerprint data. When new videos are uploaded, the system automatically compares them, and upon finding a match, notifies the copyright owner to decide on the action (remove, monetize, or allow). Portrait detection applies this logic from “audio and visual copyright” to “facial identity.”
The application process involves three steps: creators scan a QR code in YouTube Studio, submit government-issued ID documents, and record a short selfie video for facial verification. Once the system verifies the match, if suspected deepfake content is detected, the individual is notified for review, and they can choose to request a takedown.
The core assumption of this process is: the platform cannot rely solely on algorithms to determine the “harmful intent” of deepfake content, but the individual can judge it. The system is responsible for generating a candidate list, but decision-making authority is returned to the person involved.
However, there is a clear current limitation: it can only detect faces, not voice. Audio portrait detection (i.e., identifying whether AI-synthesized voices impersonate specific individuals) is still under development, with YouTube planning to launch it later in 2026.
Scaling identity protection and the unspoken issues behind it
But behind YouTube’s expansion of protection, there remains an unresolved structural problem.
The operation of Content ID relies on: copyright holders submitting legally valid material, and the platform being obliged to enforce it. The legal basis for portrait detection is much more ambiguous: the legality of deepfake content varies greatly across jurisdictions, and there is currently no unified deepfake legislation at the federal level in the US. YouTube provides tools, but their effectiveness depends on legal support behind the scenes, which is not yet fully in place.
The entertainment industry’s concerns are more direct: top Hollywood agencies like CAA, UTA, and WME actively participate in the first wave of testing because their talents’ portrait rights have high commercial value, and the risks of deepfakes are most significant. For ordinary creators, lowering the application threshold is positive, but whether detection accuracy and takedown speed will decline as usage scales remains unknown due to a lack of public data.
YouTube defines this tool as part of its “Creator Protections” strategy. This description is accurate: it is a strategy, not a solution. The race between the accessibility of deepfake generation tools and YouTube’s detection capabilities has only just begun.