PROVE: A Perceptual RemOVal cohErence
Benchmark for Visual Media

Fuhao Li*,   Shaofeng You*,   Jiagao Hu*,   Yu Liu,   Yuxuan Chen,   Zepeng Wang,   Fei Wang,   Daiguo Zhou,   Jian Luan

* Equal contribution

MiLM Plus, Xiaomi Inc.

Abstract

Evaluating object removal in images and videos remains challenging because the task is inherently one-to-many, yet existing metrics frequently disagree with human perception. Full-reference metrics reward copy-paste behaviors over genuine erasure; no-reference metrics suffer from systematic biases such as favoring blurry results; and global temporal metrics are insensitive to localized artifacts within edited regions.

To address these limitations, we propose RC (Removal Coherence), a pair of perception-aligned metrics: RC-S, which measures spatial coherence via sliding-window feature comparison between masked and background regions, and RC-T, which measures temporal consistency via distribution tracking within shared restored regions across adjacent frames. To validate RC and support community benchmarking, we further introduce PROVE-Bench, a two-tier real-world benchmark comprising PROVE-M, an 80-video paired dataset with motion augmentation, and PROVE-H, a 100-video challenging subset without ground truth. Together, RC metrics and PROVE-Bench form the PROVE (Perceptual RemOVal cohErence) evaluation framework for visual media. Experiments across diverse image and video benchmarks demonstrate that RC achieves substantially stronger alignment with human judgments than existing evaluation protocols.

Motivation

Existing evaluation metrics for object removal exhibit systematic biases that conflict with human perception.

Metric bias in object removal evaluation

Figure 1. Illustrative examples of metric bias in object removal evaluation. (a) Full-reference metrics reward copy-paste behavior over genuine erasure. (b) No-reference metrics favor blurry outputs across diffusion steps. (c) Traditional vs. diffusion-based methods show inconsistencies between metric judgments and visual perception.

Object removal failure cases

Figure 2. RC-S captures locally visible side effects and residual artifacts. Human-perceived ranking (1 = best): D > B > A > C. RC-S ranking: D(1) > B(2) > A(3) > C(4), consistent with human perception. In contrast, ReMOVE ranks A(1) > B(3) > C(4) > D(2), and CFD ranks A(1) > B(2) > C(4) > D(3) — both incorrectly favor the residual-containing result A over the cleanest removal D.

Full-Reference Bias

FR metrics (PSNR, SSIM, LPIPS) assume strict point-to-point correspondence to a single reference, rewarding conservative copy-paste outputs over perceptually realistic restorations.

No-Reference Blind Spots

NR metrics like ReMOVE and CFD frequently assign inflated scores to blurry outputs and incorrectly penalize structurally sound restorations in complex occlusion scenarios.

Temporal Insensitivity

Global temporal metrics (TC, TF) are dominated by unchanged background regions, failing to detect localized artifacts within the removed regions where object removal most commonly fails.

Blur bias analysis

Figure 3. “Blur is Clean” bias analysis. As blur radius increases inside the masked region, ReMOVE scores incorrectly improve and CFD incorrectly decreases (appears better), while RC-S correctly and monotonically degrades — demonstrating its robustness against the blur-favoring bias.

Proposed RC Metrics

We introduce a unified local distribution matching framework in deep semantic feature space, named Removal Coherence (RC), instantiated as two complementary metrics.

Overview of the proposed RC metrics

Figure 4. Overview of the proposed RC metrics. (a) RC-S measures intra-frame spatial coherence by comparing masked and background feature distributions within sliding windows. (b) RC-T measures inter-frame temporal consistency by comparing restored-region feature distributions across adjacent frames under union-based cropping and intersection-based evaluation.

RC-S — Spatial Coherence

RC-S evaluates spatial coherence by cropping each target region, extracting DINOv2 features, and applying a sliding-window MMD to compare feature distributions inside and outside the removed region. This enables fine-grained detection of local spatial incoherence that global metrics miss.

RC-T — Temporal Consistency

RC-T extends the local distribution matching design to the temporal domain. It jointly crops adjacent frames under a shared union mask, then measures feature distribution drift exclusively within the intersected restored regions, yielding sensitive detection of local temporal instability.

Key Design Choices

DINOv2 Features

DINOv2 provides a perceptually sensitive feature space for assessing fine-grained local coherence, showing stronger alignment with low-level human visual characteristics.

Sliding Window

Window-based comparison exposes regional inconsistency more explicitly than global aggregation, better matching the way humans visually inspect removal results.

MMD Distance

Maximum Mean Discrepancy accurately measures local distribution shifts between restored regions and surrounding context, outperforming first-order cosine similarity.

Why Local Cropping Matters

ReMOVE vs RC-S cropping comparison

Cropping strategies of ReMOVE vs. RC-S. Green boxes denote crop regions, red areas denote masks. ReMOVE uses a single enlarged crop even when targets are spatially far apart, introducing excessive irrelevant background that dilutes the feature difference. RC-S crops each target independently with local context, enabling fine-grained detection of incoherence.

Blur and swap robustness comparison

Counter-intuitive rankings under local perturbations. We apply Gaussian blur or region swap to the masked areas. Human judgment clearly prefers: ori > blur > swap. However, ReMOVE and CFD produce the reverse ordering (red = best-ranked, green = second-best), while RC-S remains consistent with human perception.

PROVE-Bench

A two-tier real-world benchmark for evaluating object removal in video, combining paired evaluability with unconstrained stress testing.

PROVE-M construction pipeline

Figure 5. Construction pipeline of PROVE-M: real-world paired capture with controlled conditions, three-stage pairwise quality control, and Ken Burns-style motion augmentation applied synchronously to input–mask–GT triplets.

PROVE-M — Motion-Augmented Paired Benchmark

80 videos with aligned input–mask–ground-truth triplets captured in real-world scenes.

  • Real-world paired capture with tripod-mounted camera
  • Motion augmentation via Ken Burns-style transformations
  • 81 frames at 1080p resolution per video
  • Three-stage quality control filtering pipeline
  • Covers shadows, reflections, multiple effects, and fast motion
PROVE-H — Hard Real-World Benchmark

100 videos targeting challenging scenarios without ground truth.

  • Crowd scenes and dense occlusions
  • Dynamic backgrounds (water, flames, rain, snow)
  • Highly textured backgrounds (grasslands, deserts)
  • Complex reflections and intertwined side effects
  • Fast-motion scenes with SAM3-generated masks
PROVE-M sample frames

(a) PROVE-M — Sample frames from the motion-augmented paired benchmark.

PROVE-H sample frames

(b) PROVE-H — Sample frames from the hard real-world benchmark.

Comparison with Existing Benchmarks

Dataset Real GT Shadows Reflections Multi-Effect Disconnected Crowds Textured Fast Motion #Videos
DAVIS90
Movies5
Kubric5
GenProp15
ROSE-Bench60
PROVE-M (Ours)80
PROVE-H (Ours)100

Results

Benchmark Results on PROVE-M

Quantitative evaluation of mainstream video object removal methods on the PROVE-M benchmark. ↓ means lower is better, ↑ means higher is better.

Method PSNR↑SSIM↑LPIPS↓ ReMOVECFD RC-S↑RC-T↓
FGT21.65110.86190.20130.86220.32290.37970.8031
ProPainter22.18460.87680.15590.86760.27740.44270.5951
DiffuEraser22.07580.87060.15180.86810.33080.47870.4851
VACE (1.3B)20.08260.86540.15450.81170.32830.40360.5217
Minimax-Remover (1.3B)21.74760.87070.15420.87100.32020.47930.4485
GenOmni (CogV5B)25.01650.90300.12230.87550.38420.50290.3145
GenOmni (Wan1.3B)25.14800.90170.11090.88150.34570.51880.3238
ROSE (1.3B)26.13330.90030.12120.88030.33640.49240.6538
EffectErase (1.3B)27.00490.90980.11420.88410.34120.52700.2728
UnderEraser (14B)28.33250.91560.09810.88240.29860.51880.3276
SVOR (1.3B)27.42890.92390.08390.88360.27940.52360.2987

Benchmark Results on PROVE-H

Quantitative evaluation of mainstream video object removal methods on the PROVE-H benchmark. ↓ means lower is better, ↑ means higher is better.

Method PSNR↑SSIM↑LPIPS↓ ReMOVECFD RC-S↑RC-T↓
FGT29.44480.86150.19270.84740.30650.37160.5866
ProPainter33.35310.92740.10630.83830.28300.39320.4453
DiffuEraser31.41120.91780.10980.84400.31650.43870.3911
VACE (1.3B)26.72660.88980.10710.80470.32880.41920.3438
Minimax-Remover (1.3B)29.60210.86600.13150.85450.33200.46170.3277
GenOmni (CogV5B)28.76430.88730.11830.85360.35160.50060.2141
GenOmni (Wan1.3B)29.31400.89400.10270.85960.34220.51270.2368
ROSE (1.3B)27.62610.85080.14020.85380.33610.46870.4373
EffectErase (1.3B)24.37930.81560.17420.85320.35900.50810.2363
UnderEraser (14B)27.49890.84850.14340.85600.31650.50750.2688
SVOR (1.3B)27.53350.89070.10460.85740.31070.51660.2419

Note: Due to compliance requirements, the open-source data differs slightly from the data used in the paper. The results above are based on the open-source version and may exhibit minor numerical differences from the paper, but the overall trends remain consistent.

Human Correlation Analysis

RC-S achieves the best average correlation with human rankings across six benchmarks, ranking first on five of six benchmarks under both Kendall's τ and Spearman's ρ.

Metric Kendall's τ Avg. Spearman ρ
RORDOBERDAVISROSEPROVE-MPROVE-HAVG RORDOBERDAVISROSEPROVE-MPROVE-HAVG
PSNR0.010.360.380.250.020.440.450.30
SSIM-0.220.110.430.11-0.310.110.460.09
LPIPS-0.230.240.330.12-0.280.280.370.13
m-LPIPS0.190.680.680.520.240.750.750.58
ReMOVE0.060.540.150.210.330.230.260.080.610.160.240.360.270.29
CFD-0.040.400.210.030.240.120.16-0.050.470.250.040.260.140.18
RC-S (Ours)0.310.570.600.610.700.760.590.390.660.680.690.750.820.66

Validation of RC-T

RC-T responds sensitively and monotonically to controlled temporal corruptions, whereas existing temporal metrics (TC, TF) remain largely insensitive.

RC-T validation

Figure 6. Sensitivity-based validation of temporal metrics under increasing corruption severity. RC-T exhibits monotonically degrading scores, whereas TC and TF remain insensitive or even improve.

Citation

If you find our work useful for your research, please consider citing our paper:

@article{li2026prove, title={PROVE: A Perceptual RemOVal cohErence Benchmark for Visual Media}, author={Li, Fuhao and You, Shaofeng and Hu, Jiagao and Liu, Yu and Chen, Yuxuan and Wang, Zepeng and Wang, Fei and Zhou, Daiguo and Luan, Jian}, journal={arXiv preprint arXiv:2605.14534}, year={2026} }