论文标题
网络时刻:扩展和稀疏平滑攻击
Network Moments: Extensions and Sparse-Smooth Attacks
论文作者
论文摘要
深度神经网络(DNN)的令人印象深刻的表现极大地加强了旨在从理论上分析其有效性的研究线。这引发了对DNN对嘈杂输入的反应的研究,即开发了对抗性输入攻击和策略,从而导致了对这些攻击的强大DNN。为此,在本文中,我们得出了针对高斯输入的小分段线性(PL)网络(Aggine,relu,offine)的第一和第二矩(均值和方差)的精确分析表达式。特别是,我们概括了Bibi等人的第二矩表达。为了任意输入高斯分布,删除了零均值的假设。我们表明,与Bibi等人的初步结果相比,新的方差表达可以有效地近似,从而导致更严格的差异估计。此外,我们在实验上表明,在更深的PL-DNNS的简单线性化下,这些表达式很紧,我们研究了线性化敏感性对矩估计准确性的影响。最后,我们表明派生的表达式可用于构建稀疏和光滑的高斯对抗攻击(靶向和非目标),这些攻击倾向于导致感知可行的输入攻击。
The impressive performance of deep neural networks (DNNs) has immensely strengthened the line of research that aims at theoretically analyzing their effectiveness. This has incited research on the reaction of DNNs to noisy input, namely developing adversarial input attacks and strategies that lead to robust DNNs to these attacks. To that end, in this paper, we derive exact analytic expressions for the first and second moments (mean and variance) of a small piecewise linear (PL) network (Affine, ReLU, Affine) subject to Gaussian input. In particular, we generalize the second-moment expression of Bibi et al. to arbitrary input Gaussian distributions, dropping the zero-mean assumption. We show that the new variance expression can be efficiently approximated leading to much tighter variance estimates as compared to the preliminary results of Bibi et al. Moreover, we experimentally show that these expressions are tight under simple linearizations of deeper PL-DNNs, where we investigate the effect of the linearization sensitivity on the accuracy of the moment estimates. Lastly, we show that the derived expressions can be used to construct sparse and smooth Gaussian adversarial attacks (targeted and non-targeted) that tend to lead to perceptually feasible input attacks.