论文标题
ViewFool:评估视觉识别对对抗观点的鲁棒性
ViewFool: Evaluating the Robustness of Visual Recognition to Adversarial Viewpoints
论文作者
论文摘要
最近的研究表明,视觉识别模型缺乏分布转移的鲁棒性。但是,当前的工作主要考虑了对2D图像转换的模型鲁棒性,而3D世界中的观点变化却少了。通常,在各种现实世界应用(例如自主驾驶)中,观点变化很普遍,因此必须评估观点鲁棒性。在本文中,我们提出了一种称为ViewFool的新方法,以找到误导视觉识别模型的对抗观点。通过将现实世界的对象编码为神经辐射场(NERF),ViewFool表征了熵正常化程序下各种对抗观点的分布,这有助于处理真实摄像机的波动并减轻真实对象之间的现实差距及其真实对象及其神经表示。实验验证了公共图像分类器极易受到生成的对抗观点的影响,后者也表现出高跨模型的可传递性。基于ViewFool,我们介绍了Imagenet-V,这是一种新的分布数据集,用于基准图像分类器的观点鲁棒性。对40个具有不同体系结构,目标功能和数据增强的分类器的评估结果显示,在Imagenet-V上进行测试时,模型性能的显着下降,这提供了利用视图fivefool作为有效的数据增强策略,以提高视图点鲁棒性。
Recent studies have demonstrated that visual recognition models lack robustness to distribution shift. However, current work mainly considers model robustness to 2D image transformations, leaving viewpoint changes in the 3D world less explored. In general, viewpoint changes are prevalent in various real-world applications (e.g., autonomous driving), making it imperative to evaluate viewpoint robustness. In this paper, we propose a novel method called ViewFool to find adversarial viewpoints that mislead visual recognition models. By encoding real-world objects as neural radiance fields (NeRF), ViewFool characterizes a distribution of diverse adversarial viewpoints under an entropic regularizer, which helps to handle the fluctuations of the real camera pose and mitigate the reality gap between the real objects and their neural representations. Experiments validate that the common image classifiers are extremely vulnerable to the generated adversarial viewpoints, which also exhibit high cross-model transferability. Based on ViewFool, we introduce ImageNet-V, a new out-of-distribution dataset for benchmarking viewpoint robustness of image classifiers. Evaluation results on 40 classifiers with diverse architectures, objective functions, and data augmentations reveal a significant drop in model performance when tested on ImageNet-V, which provides a possibility to leverage ViewFool as an effective data augmentation strategy to improve viewpoint robustness.