Models

What Is Deepfake Detection?

The task of identifying AI-generated or AI-manipulated images, video, and audio and classifying them as authentic or fake.

What Is Deepfake Detection?

Deepfake detection is the task of identifying AI-generated or AI-manipulated media and classifying it as authentic or fake. The toolkit covers face-swap CNNs, lip-sync mismatch detectors, voice-clone classifiers, vision-language models that score consistency cues, and cryptographic watermark verifiers (e.g. C2PA, Google SynthID). Detection lives at the model layer of an AI pipeline and is typically a feature into a downstream content-safety or KYC decision. FutureAGI does not ship a deepfake classifier; we evaluate the voice and multimodal agents that may generate or be tricked by deepfaked inputs.

Why It Matters in Production AI Systems

Deepfakes hit production AI in three ways. First, voice-cloned attackers call IVR or voice-agent systems and trigger wrong-account actions — a documented 2026 attack pattern against banking and telco voice bots. Second, image and video deepfakes are submitted to KYC pipelines that include vision-language models, where a single passing forgery can authorise a fraudulent account. Third, synthesised content shows up inside RAG knowledge bases and degrades grounding without ever surfacing as a security alert.

The pain is uneven. Trust-and-safety teams own the moderation pipeline. SREs see latency spikes when detection ensembles run on every request. Compliance leads worry about EU AI Act labelling obligations and US state-level deepfake laws. Product managers worry about user trust when a deepfake slips through.

In 2026-era multimodal pipelines this gets harder. A single user request can include text, an image, an audio clip, and a generated video, each demanding a different detector. False positives from any layer cascade into refused legitimate requests; false negatives let a fraudulent transaction through. Benchmark wins do not transfer automatically: a detector that looks strong on FaceForensics++ can miss low-bitrate call-center audio or cropped ID-document video. The right architecture treats detection as one signal in a layered defence — watermark verification, classifier output, behavioural risk score, human-in-the-loop — rather than a single boolean gate.

How FutureAGI Handles Deepfake-Adjacent Risks

FutureAGI doesn’t run a face-swap classifier, but does cover the voice-agent and multimodal eval surface where deepfake risks materialise. For voice, LiveKitEngine replays attacker scenarios against a deployed voice agent — including cloned-voice persona prompts generated via ScenarioGenerator — and scores the trajectory with ASRAccuracy, AudioQualityEvaluator, and ConversationResolution. For image and video pipelines, CaptionHallucination flags when a vision-language model invents content not present in the image (a common deepfake-adjacent failure), and ContentModeration plus ContentSafety gate harmful generated outputs.

When a team integrates a third-party deepfake classifier, its output is logged via fi.client.Client.log as a span_event on the inference trace. The Agent Command Center then enforces routing logic — e.g. a pre-guardrail checks the deepfake score and routes high-risk inputs to a human review queue, while a post-guardrail checks the model’s response for content-safety violations. Compared with Google SynthID, which is strongest when provenance metadata survives the pipeline, FutureAGI treats the detector as an observed input to the agent, not as ground truth. FutureAGI’s approach is to evaluate the chain — the classifier score, the specific route chosen by the gateway, the LLM decision based on it, and the user-visible outcome — so engineers can attribute regressions to a specific layer.

How to Measure or Detect It

Useful signals for deepfake-adjacent eval:

  • CaptionHallucination — flags vision-language models inventing content; doubles as a deepfake-defence signal.
  • ASRAccuracy — speech-to-text accuracy that drops when audio is synthesised by a low-quality clone.
  • ContentSafety — final-mile policy check on generated or transcribed content.
  • ContentModeration — broader moderation eval covering harmful synthesised media.
  • eval-fail-rate-by-cohort — segmented by attack scenario versus benign traffic.
  • Third-party classifier span_event — log raw deepfake scores into traces for downstream auditing.

Run the detector on labeled attack and benign cohorts, then store every raw score on the production trace so thresholds can be replayed after an incident. Track precision, recall, ROC-AUC, and escalation-rate separately for audio, image, and video; one aggregate score hides the channel where the detector is failing. Set separate thresholds for enrollment, account recovery, and high-value transactions because the acceptable false-positive rate changes by workflow. Review p95 latency and cost per trace before enabling classifier ensembles synchronously on every request.

Minimal Python:

from fi.evals import CaptionHallucination, ContentSafety

caption = CaptionHallucination()
safety = ContentSafety()

result = caption.evaluate(
    input=image_or_audio,
    output=model_response,
    context=ground_truth,
)

Common Mistakes

  • Treating one classifier as the gate. Single-detector defence fails as generators change; combine watermark verification, classifier scores, behavioural risk, and human review before account actions.
  • No regression on adversarial examples. Detectors degrade silently against newer generators; add cloned-voice, face-swap, and lip-sync samples to your Dataset every release, and record generator versions.
  • Skipping audio in a voice pipeline. Voice clones bypass face-swap detectors entirely; cover speaker, transcript, and audio-quality failure modes with LiveKitEngine simulations before launch.
  • Hiding detector confidence from routing logic. Surface calibrated scores, thresholds, and cohort labels so pre-guardrail policy can route ambiguous cases for review without dropping trace context.
  • Ignoring business cost and calibration. A 5% KYC false-positive rate blocks legitimate users; recalibrate thresholds on held-out attack and benign samples per market.

Frequently Asked Questions

What is deepfake detection?

Deepfake detection is the task of classifying AI-generated or AI-manipulated images, video, and audio as authentic or fake using neural classifiers, watermark verifiers, and physiological-cue checks.

How is deepfake detection different from content moderation?

Content moderation flags whether content violates a policy (toxicity, sexual content). Deepfake detection asks the upstream question — was this content synthesised or manipulated by AI? — and is often a feature into moderation pipelines.

How do you measure deepfake-related risks in voice agents?

Use FutureAGI `CaptionHallucination`, `ASRAccuracy`, and `ContentSafety` against a `Dataset` of attack scenarios; pair with `LiveKitEngine` simulations to replay cloned-voice prompts.