Detection Methodology
Transparent, multi-signal analysis for AI video detection
1Our Approach
VeriVid AI uses an ensemble pipeline that combines multiple detection signals to estimate the likelihood that a video is AI-generated or manipulated. We analyze visual, audio, and metadata characteristics to produce a risk score.
Key Principle: We provide probability estimates and evidence, not binary verdicts. All AI detection has limitations and our scores should be interpreted as risk indicators.
2Visual Analysis
We extract key frames from the video and analyze them for synthetic artifacts:
- ✓AI Generation Patterns: Detection of artifacts common in AI-generated imagery (GAN fingerprints, diffusion patterns)
- ✓Facial Analysis: When faces are present, we check for deepfake indicators and manipulation signs
- ✓Consistency Checks: Temporal consistency across frames to detect splicing or injection
3Audio Analysis
For videos with audio, we analyze the speech and sound characteristics:
- ✓TTS Detection: Identification of text-to-speech synthesis patterns
- ✓Voice Cloning Artifacts: Analysis for AI voice cloning signatures
- ✓Audio-Visual Sync: Checking lip sync and audio alignment
4Metadata & Tampering
We examine file metadata for manipulation indicators:
- ✓Encoding Analysis: Multiple re-encoding, suspicious codec combinations
- ✓EXIF/Metadata: Missing camera info, inconsistent timestamps
- ✓Container Analysis: File structure anomalies
5Risk Scoring
Individual signals are weighted and combined into a final risk score:
Visual Signals40%
Audio Signals25%
Metadata Signals20%
Tampering Signals15%
!Known Limitations
- • Detection accuracy varies by AI generation method and video quality
- • Highly compressed videos may lose detectable artifacts
- • Novel AI techniques may not be immediately detectable
- • Results should be corroborated with other evidence
- • False positives and negatives can occur