Papers
arxiv:2303.13193

VADER: Video Alignment Differencing and Retrieval

Published on Mar 23, 2023
Authors:
,
,
,
,
,
,

Abstract

VADER uses spatio-temporal matching and alignment to detect video manipulation and assess trustworthiness.

AI-generated summary

We propose VADER, a spatio-temporal matching, alignment, and change summarization method to help fight misinformation spread via manipulated videos. VADER matches and coarsely aligns partial video fragments to candidate videos using a robust visual descriptor and scalable search over adaptively chunked video content. A transformer-based alignment module then refines the temporal localization of the query fragment within the matched video. A space-time comparator module identifies regions of manipulation between aligned content, invariant to any changes due to any residual temporal misalignments or artifacts arising from non-editorial changes of the content. Robustly matching video to a trusted source enables conclusions to be drawn on video provenance, enabling informed trust decisions on content encountered.

Community

Sign up or log in to comment

Get this paper in your agent:

hf papers read 2303.13193
Don't have the latest CLI?
curl -LsSf https://hf.co/cli/install.sh | bash

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2303.13193 in a model README.md to link it from this page.

Datasets citing this paper 1

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2303.13193 in a Space README.md to link it from this page.

Collections including this paper 1