高级检索

    针对SAM下游模型脆弱模块的对抗迁移攻击

    Adversarial Transfer Attacks on Weak Block of SAM Downstream Models

    • 摘要: SAM(segment anything model)作为一种通用的视觉基础模型,已被广泛应用于多种图像分割任务,但其在对抗性攻击面前表现出脆弱性. 本文提出一种针对SAM下游模型脆弱模块的对抗迁移攻击方法(fragile section gradient robustness,FSGR). 该方法在无需知晓下游微调细节的前提下,可有效生成对抗样本,实现对SAM下游模型的攻击. 该方法运用“脆弱层精准定位+局部强化迁移”策略,通过特征相似度筛选出跨任务共享且最易被激活的模块,针对性地强化攻击效果;同时,引入梯度稳健损失以消除目标模型与下游任务模型间的梯度差异. FSGR方法融合了脆弱层攻击与梯度稳健损失机制,在多个数据集上均实现了相对性能的提升. 实验结果表明,FSGR在多种微调模型(如医学分割、阴影分割和伪装分割)的迁移攻击中显著降低了模型性能,证明了其正确性和实用性. 与基线方法相比,FSGR不仅在攻击成功率上表现出色,还通过结合脆弱层攻击和梯度稳健损失,实现了相对性能的提升.

       

      Abstract: As a common visual foundation model, the segment anything model (SAM) has been widely applied in various image segmentation tasks, but it exhibits vulnerability in the face of adversarial attacks. A adversarial transfer attack method FSGR (fragile section gradient robustness) has been proposed for vulnerable modules of SAM downstream models, which can effectively generate adversarial samples to attack SAM downstream models without understanding the details of downstream fine-tuning. The core of FSGR lies in its ability to identify the most vulnerable parts and apply targeted attacks to these weak points. This method designs a strategy based on feature similarity to identify the most vulnerable module in the SAM encoder and enhance the attack effectiveness in a targeted manner; and introduce gradient robust loss to eliminate the gradient difference between the target model and downstream task models. FSGR combines vulnerability layer attacks and gradient robust loss to achieve relative performance improvement on multiple datasets. The experimental results show that FSGR significantly reduces model performance in transfer attacks on various fine-tuning models, such as medical segmentation, shadow segmentation, and camouflage segmentation, demonstrating its correctness and practicality. Compared with baseline methods, FSGR not only performs well in attack success rate, but also achieves relative performance improvement by combining vulnerability layer attacks and gradient robust loss.

       

    /

    返回文章
    返回