论文标题

部分可观测时空混沌系统的无模型预测

Confidence-aware Training of Smoothed Classifiers for Certified Robustness

论文作者

Jeong, Jongheon, Kim, Seojin, Shin, Jinwoo

论文摘要

任何分类器都可以在高斯噪声下“平滑”,以构建一个新的分类器,该分类器可通过随机平滑度平均其对噪声的预测,从而对$ \ ell_2 $ verversial扰动,即$ \ ell_2 $ - 对抗扰动。在平滑的分类器下,在文献中已经很好地证明了准确性和(对抗性)鲁棒性之间的基本权衡:即增加分类器的鲁棒性以牺牲其他投入的精度,以牺牲精度降低。在本文中,我们提出了一种简单的培训方法,该方法利用这种权衡取舍,特别是通过对培训样品的鲁棒性控制来获得健壮的平滑分类器。我们通过使用“高斯噪声下的准确性”作为输入的对抗鲁棒性的易于计算来使这种控制可行。具体而言,我们根据该代理来区分训练目标,以滤除不太可能受益于最坏情况(对抗性)目标的样品。我们的实验表明,该方法尽管简单,但在最先进的培训方法上始终表现出改善的认证鲁棒性。令人惊讶的是,我们发现这些改进甚至对于其他鲁棒性概念(例如各种常见的腐败)仍然存在。

Any classifier can be "smoothed out" under Gaussian noise to build a new classifier that is provably robust to $\ell_2$-adversarial perturbations, viz., by averaging its predictions over the noise via randomized smoothing. Under the smoothed classifiers, the fundamental trade-off between accuracy and (adversarial) robustness has been well evidenced in the literature: i.e., increasing the robustness of a classifier for an input can be at the expense of decreased accuracy for some other inputs. In this paper, we propose a simple training method leveraging this trade-off to obtain robust smoothed classifiers, in particular, through a sample-wise control of robustness over the training samples. We make this control feasible by using "accuracy under Gaussian noise" as an easy-to-compute proxy of adversarial robustness for an input. Specifically, we differentiate the training objective depending on this proxy to filter out samples that are unlikely to benefit from the worst-case (adversarial) objective. Our experiments show that the proposed method, despite its simplicity, consistently exhibits improved certified robustness upon state-of-the-art training methods. Somewhat surprisingly, we find these improvements persist even for other notions of robustness, e.g., to various types of common corruptions.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源