Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
156 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

How Real Is Real? A Human Evaluation Framework for Unrestricted Adversarial Examples (2404.12653v1)

Published 19 Apr 2024 in cs.AI

Abstract: With an ever-increasing reliance on ML models in the real world, adversarial examples threaten the safety of AI-based systems such as autonomous vehicles. In the image domain, they represent maliciously perturbed data points that look benign to humans (i.e., the image modification is not noticeable) but greatly mislead state-of-the-art ML models. Previously, researchers ensured the imperceptibility of their altered data points by restricting perturbations via $\ell_p$ norms. However, recent publications claim that creating natural-looking adversarial examples without such restrictions is also possible. With much more freedom to instill malicious information into data, these unrestricted adversarial examples can potentially overcome traditional defense strategies as they are not constrained by the limitations or patterns these defenses typically recognize and mitigate. This allows attackers to operate outside of expected threat models. However, surveying existing image-based methods, we noticed a need for more human evaluations of the proposed image modifications. Based on existing human-assessment frameworks for image generation quality, we propose SCOOTER - an evaluation framework for unrestricted image-based attacks. It provides researchers with guidelines for conducting statistically significant human experiments, standardized questions, and a ready-to-use implementation. We propose a framework that allows researchers to analyze how imperceptible their unrestricted attacks truly are.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (15)
  1. MTurk Research: Review and Recommendations. Journal of Management, 47(4): 823–837.
  2. Review of best practice recommendations for ensuring high quality data with amazon’s mechanical turk.
  3. Evidence-based survey design: The use of continuous rating scales in surveys. Performance Improvement, 57(5): 38–48.
  4. Diffusion models in vision: A survey. IEEE Transactions on Pattern Analysis and Machine Intelligence.
  5. Data quality in online human-subjects research: Comparisons between MTurk, Prolific, CloudResearch, Qualtrics, and SONA. Plos one, 18(3): e0279720.
  6. A study of the effect of jpg compression on adversarial images. arXiv preprint arXiv:1608.00853.
  7. Deep Residual Learning for Image Recognition. CoRR, abs/1512.03385.
  8. Ishihara, S.; et al. 1918. Tests for color blindness. American Journal of Ophthalmology, 1(5): 376.
  9. Sok: Certified robustness for deep neural networks. arXiv preprint arXiv:2009.04131.
  10. Toward verifiable and reproducible human evaluation for text-to-image generation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 14277–14286.
  11. ImageNet Large Scale Visual Recognition Challenge. International Journal of Computer Vision (IJCV), 115(3): 211–252.
  12. Do adversarially robust imagenet models transfer better? Advances in Neural Information Processing Systems, 33: 3533–3545.
  13. Colorfool: Semantic adversarial colorization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 1151–1160.
  14. Intriguing properties of neural networks. arXiv preprint arXiv:1312.6199.
  15. Hype: A benchmark for human eye perceptual evaluation of generative models. Advances in neural information processing systems, 32.

Summary

We haven't generated a summary for this paper yet.