论文标题
通过Metameric任务找到针对对抗性功能的生物学合理性
Finding Biological Plausibility for Adversarially Robust Features via Metameric Tasks
论文作者
论文摘要
最近的工作表明,通过图像操纵,通过对抗性强大的网络学到的表示形式比非舒适的网络更具感知的一致性。尽管看起来更接近人类的视觉感知,但尚不清楚强大的DNN表示中的约束是否与人类视力中发现的生物学约束相匹配。人类的视觉似乎依赖于外围的基于纹理/摘要的统计表示,这些统计表达已被证明可以解释现象,例如在视觉搜索任务上的拥挤和表现。为了了解对抗性强大的优化/表示与人类视野的比较,我们使用一组元歧视任务进行了心理物理学实验,在这些实验中,我们评估了人类观察者如何能够区分与不可见点的表达和透明构想模型(Texture Cynathesiss Models of Peripheral Visemals)合成的图像合成图像之间的图像。我们发现,随着刺激在周围较远的情况下,稳健表示和纹理模型图像的可区分性降低到了几乎偶然的性能。此外,在稳健和纹理模型图像上的性能显示在参与者中的趋势相似,而在整个视野中,非持胸表表示的性能变化很小。这些结果共同表明(1)对抗性强大表示比非体上的表示更好地捕获外围计算,并且(2)可靠的表示捕获的外围计算与当前最新的纹理外围视觉模型相似。更广泛地说,我们的发现支持了这样的想法:本地化纹理摘要统计表示可能会使人类不变性驱动到对抗性扰动,并且将这种表示形式纳入DNN可能会带来有用的特性,例如对抗性鲁棒性。
Recent work suggests that representations learned by adversarially robust networks are more human perceptually-aligned than non-robust networks via image manipulations. Despite appearing closer to human visual perception, it is unclear if the constraints in robust DNN representations match biological constraints found in human vision. Human vision seems to rely on texture-based/summary statistic representations in the periphery, which have been shown to explain phenomena such as crowding and performance on visual search tasks. To understand how adversarially robust optimizations/representations compare to human vision, we performed a psychophysics experiment using a set of metameric discrimination tasks where we evaluated how well human observers could distinguish between images synthesized to match adversarially robust representations compared to non-robust representations and a texture synthesis model of peripheral vision (Texforms). We found that the discriminability of robust representation and texture model images decreased to near chance performance as stimuli were presented farther in the periphery. Moreover, performance on robust and texture-model images showed similar trends within participants, while performance on non-robust representations changed minimally across the visual field. These results together suggest that (1) adversarially robust representations capture peripheral computation better than non-robust representations and (2) robust representations capture peripheral computation similar to current state-of-the-art texture peripheral vision models. More broadly, our findings support the idea that localized texture summary statistic representations may drive human invariance to adversarial perturbations and that the incorporation of such representations in DNNs could give rise to useful properties like adversarial robustness.