Models & Research

Multimodal Detection Architecture

Our mission is to secure the digital ecosystem against synthetic manipulation. ZeroTrue employs an ensemble of evidence-first models to detect AI-generated content across text, code, voice, music, and video.

ZeroTrue Architecture Diagram
Text
Code
Voice/TTS
Music
Video

Architectures by Modality

Our specialized pipelines are designed to capture the unique artifacts left by generative models in each domain.

2.1 Generated Text

Technical Approach

Hybrid detectors combining likelihood/perplexity tests, supervised classifiers (RoBERTa/DeBERTa fine-tunes), and stylometry features.

Key Techniques

Token-rank histogramsEntropy spikesBurstinessFunction-word ratios
MetricsAUROC, AUPRC, TPR at 1%/2% FPR
ExplainabilityToken-level heatmaps, threshold bands for FPR@TPR.
BenchmarksHC3/HC3+, RAID robustness suites

2.2 LLM-Generated Code

Technical Approach

Dual-track analysis using content-agnostic detectors (AST/CFG features) and provenance-aware watermark readers.

Key Techniques

GraphCodeBERT baselinesGradient-boosted stylometryIdentifier entropy
MetricsAUROC, TPR@FPR≤5%, Robustness to edits
ExplainabilityAnomaly rationales via reviewer LLM.
BenchmarksAIGCodeSet, ACL/AAAI studies

2.3 Voice Cloning / TTS

Technical Approach

Spectrogram-level CNN/Conformer models fused with LFCC/EFCC anti-spoof features and ECAPA-TDNN variants.

Key Techniques

Replay augmentationCodec simulationrPPG sync (when video present)
MetricsEER, min t-DCF
ExplainabilitySpectral anomaly regions.
BenchmarksASVspoof 2019/2021 (LA/PA/DF), ADD 2022, FakeAVCeleb

2.4 Generated Music

Technical Approach

Multi-scale spectrogram fingerprints and timbre/harmonic residuals, ensembled with music-theory features.

Key Techniques

Key stability analysisN-gram chord progressionsVQ-VAE artifact detection
MetricsAUROC, Segment-consistency
ExplainabilityTimbre consistency scores.
BenchmarksInternal datasets (OpenAI Jukebox, Suno, Udio)

2.5 Deepfake Video

Technical Approach

Spatiotemporal detectors (Xception/EfficientNet + TimeSformer) with frequency and physiology auxiliaries.

Key Techniques

Blink detectionrPPG pulse extractionFrame-by-frame analysis
MetricsVideo-AUC, Frame-AUC, Cross-dataset generalization
ExplainabilityTemporal heatmaps, face-crop confidence.
BenchmarksDFDC, FaceForensics++, Celeb-DF, DeeperForensics-1.0

Datasets & Evaluation Protocols

DatasetDomainSize / NotesMetric(s)
HC3 / HC3+TextHigh-quality ChatGPT vs HumanAUROC, FPR@TPR
RAIDTextAdversarial attacks & domainsRobustness Score
AIGCodeSetCodePython generation tasksAUROC
ASVspoof 2019/21AudioLogical/Physical AccessEER, min t-DCF
ADD 2022AudioAudio Deepfake DetectionEER
DFDCVideo100k+ clips, Facebook backedVideo-AUC
FaceForensics++VideoDiverse manipulation methodsFrame-AUC
Celeb-DFVideoHigh-quality DeepfakesAUC

* We also track GenImage for image generation baselines.

Evaluation Metrics

Text
AUROC
TPR @ 1% FPR
Code
AUROC
TPR @ 5% FPR
Audio
EER
min t-DCF
Video
Video-AUC
Frame-AUC

Operational Bands

Recommended action thresholds based on confidence scores.

0-20% (Safe)
20-50% (Review)
50-100% (High Probability)

Provenance & Standards

ZeroTrue complements statistical detection with cryptographic provenance. Where available, we parse and display C2PA Content Credentials alongside our detector verdicts. This provides a dual-layer defense: verifying the "authentic" chain of custody while statistically flagging likely manipulations.

C2PA Verification
Watermark Reading
SIGNED
Provenance Manifest
C2PA
Content Credentials
Signed by Camera Manufacturer

Research Roadmap

Q3 2025

Multi-model Attribution

Fingerprinting specific generator versions (e.g., Midjourney v6 vs DALL-E 3).

Q4 2025

Watermark-Aware Fusion

Integrating hidden watermarks into confidence scoring.

Q1 2026

Adversarial Hardening

Post-training robustness against new evasion attacks.