论文标题
在视觉对象跟踪上几乎没有射击后门攻击
Few-Shot Backdoor Attacks on Visual Object Tracking
论文作者
论文摘要
视觉对象跟踪(fot)已在关键任务应用中广泛采用,例如自动驾驶和智能监视系统。在当前的实践中,第三方资源(例如数据集,骨干网络和培训平台)经常用于培训高性能的投票模型。尽管这些资源带来了一定的便利,但它们还将新的安全威胁引入了投票模型。在本文中,我们揭示了这种威胁,使对手可以通过降低训练过程来轻松地将隐藏的后门植入投票模型。具体而言,我们提出了一个简单而有效的几杆后门攻击(FSBA),该攻击(FSBA)交替优化两个损失:1)在隐藏的特征空间中定义的A \ emph {特征损失},以及2)标准\ emph {跟踪损失}。我们表明,一旦我们的FSBA将后门嵌入到目标模型中,即使仅出现\ emph {trigger}即使仅出现在一个或几个帧中,它也可以诱使模型失去特定对象的跟踪。我们检查了我们在数字和物理世界环境中的攻击,并表明它可以大大降低最先进的投票跟踪器的性能。我们还表明,我们的攻击对潜在的防御能力有抵抗力,强调了投票模型对潜在后门攻击的脆弱性。
Visual object tracking (VOT) has been widely adopted in mission-critical applications, such as autonomous driving and intelligent surveillance systems. In current practice, third-party resources such as datasets, backbone networks, and training platforms are frequently used to train high-performance VOT models. Whilst these resources bring certain convenience, they also introduce new security threats into VOT models. In this paper, we reveal such a threat where an adversary can easily implant hidden backdoors into VOT models by tempering with the training process. Specifically, we propose a simple yet effective few-shot backdoor attack (FSBA) that optimizes two losses alternately: 1) a \emph{feature loss} defined in the hidden feature space, and 2) the standard \emph{tracking loss}. We show that, once the backdoor is embedded into the target model by our FSBA, it can trick the model to lose track of specific objects even when the \emph{trigger} only appears in one or a few frames. We examine our attack in both digital and physical-world settings and show that it can significantly degrade the performance of state-of-the-art VOT trackers. We also show that our attack is resistant to potential defenses, highlighting the vulnerability of VOT models to potential backdoor attacks.