Skip to the content.

license badge Zenodo GitHub arXiv

Dren Fazlija1, Monty-Maximilian Zühlke1, Johanna Schrader1,4,
Arkadij Orlov2, Clara Stein1, Iyiola E. Olatunji3, Daniel Kudenko1
1L3S Research Center
2E.ON Grid Solutions
3University of Luxembourg
4CAIMed – Lower Saxony Center for Artificial Intelligence and Causal Methods in Medicine

Abstract (click to expand) Unrestricted adversarial attacks aim to fool computer vision models without being constrained by ℓₚ-norm bounds to remain imperceptible to humans, for example, by changing an object's color. This allows attackers to circumvent traditional, norm-bounded defense strategies such as adversarial training or certified defense strategies. However, due to their unrestricted nature, there are also no guarantees of norm-based imperceptibility, necessitating human evaluations to verify just how authentic these adversarial examples look. While some related work assesses this vital quality of adversarial attacks, none provide statistically significant insights. This issue necessitates a unified framework that supports and streamlines such an assessment for evaluating and comparing unrestricted attacks. To close this gap, we introduce SCOOTER – an open-source, statistically powered framework for evaluating unrestricted adversarial examples. Our contributions are: (i) best-practice guidelines for crowd-study power, compensation, and Likert equivalence bounds to measure imperceptibility; (ii) the first large-scale human vs. model comparison across 346 human participants showing that three color-space attacks and three diffusion-based attacks fail to produce imperceptible images. Furthermore, we found that GPT-4o can serve as a preliminary test for imperceptibility, but it only consistently detects adversarial examples for four out of six tested attacks; (iii) open-source software tools, including a browser-based task template to collect annotations and analysis scripts in Python and R; (iv) an ImageNet-derived benchmark dataset containing 3K real images, 7K adversarial examples, and over 34K human ratings. Our findings demonstrate that automated vision systems do not align with human perception, reinforcing the need for a ground-truth SCOOTER benchmark.

Motivation of this Project

Color-based Attacks
SemanticAdv
cAdv
NCF
Original
Diffusion-based Attacks
DiffAttack
AdvPP
ACA

Meet SCOOTERSystemizing Confusion Over Observations To Evaluate Realness

Key experimental findings

What’s inside the framework

Take-home message

Citation

@misc{fazlija2025scooterhumanevaluationframework,
      title={SCOOTER: A Human Evaluation Framework for Unrestricted Adversarial Examples}, 
      author={Dren Fazlija and Monty-Maximilian Zühlke and Johanna Schrader and Arkadij Orlov and Clara Stein and Iyiola E. Olatunji and Daniel Kudenko},
      year={2025},
      eprint={2507.07776},
      archivePrefix={arXiv},
      primaryClass={cs.CV},
      url={https://arxiv.org/abs/2507.07776}, 
}