网络与信息安全学报 ›› 2022, Vol. 8 ›› Issue (1): 86-94.doi: 10.11959/j.issn.2096-109x.2021095

• 学术论文 • 上一篇    下一篇

基于图像重构的MNIST对抗样本防御算法

秦中元1, 贺兆祥1, 李涛1,2, 陈立全1,2   

  1. 1 东南大学网络空间安全学院,江苏 南京 211189
    2 网络通信与安全紫金山实验室,江苏 南京 211189
  • 修回日期:2021-09-14 出版日期:2022-02-15 发布日期:2022-02-01
  • 作者简介:秦中元(1974− ),男,河南安阳人,博士,东南大学副教授,主要研究方向为人工智能安全、无线网络安全
    贺兆祥(1995− ),男,山东临沂人,东南大学硕士生,主要研究方向为人工智能安全、无线网络安全
    李涛(1984− ),男,江苏镇江人,博士,东南大学副教授,主要研究方向为可信计算、移动终端安全、内生安全
    陈立全(1976− ),男,广西玉林人,博士,东南大学教授、博士生导师,主要研究方向为移动信息安全、物联网系统与安全、云计算及大数据安全
  • 基金资助:
    国家重点研发计划(2020YFE0200600);国家自然科学基金(61601113)

Adversarial example defense algorithm for MNIST based on image reconstruction

Zhongyuan QIN1, Zhaoxiang HE1, Tao LI1,2, Liquan CHEN1,2   

  1. 1 School of Cyber Science and Engineering, Southeast University, Nanjing 211189, China
    2 Network Communication and Security Purple Mountain Laboratory, Nanjing 211189, China
  • Revised:2021-09-14 Online:2022-02-15 Published:2022-02-01
  • Supported by:
    The National Key R&D Program of China(2020YFE0200600);The National Natural Science Foundation of China(61601113)

摘要:

随着深度学习的应用普及,其安全问题越来越受重视,对抗样本是在原有图像中添加较小的扰动,即可造成深度学习模型对图像进行错误分类,这严重影响深度学习技术的发展。针对该问题,分析现有对抗样本的攻击形式和危害,由于现有防御算法存在缺点,提出一种基于图像重构的对抗样本防御方法,以达到有效防御对抗样本的目的。该防御方法以 MNIST 为测试数据集,核心思路是图像重构,包括中心方差最小化和图像缝合优化,中心方差最小化只针对图像中心区域进行处理;图像缝合优化将重叠区域纳入补丁块选取的考量,并以补丁块的1/2大小作为重叠区域。使用FGSM、BIM、DeepFool以及C&W攻击方式生成对抗样本来测试两种方式的防御性能,并与现有的3种图像重构防御方式(裁剪与缩放、位深度压缩和JPEG 压缩)效果对比。实验结果表明,所提中心方差最小化和图像缝合优化算法,对现有常见对抗样本的攻击起到了较好的防御效果。图像缝合优化对4种攻击算法生成的样本分类正确率都达到了75%以上,中心方差最小化的防御效果在70%左右。而用作对比的3种图像重构算法则对不同攻击算法的防御效果不稳定,整体分类正确率不足60%。所提中心方差最小化和图像缝合优化两种图像重构防御算法达到了有效防御对抗样本的目的,通过实验说明了所提防御算法在不同对抗样本攻击算法中的防御效果,另外,将其他图像重构算法与所提算法进行比较,说明了所提算法具有良好的防御性能。

关键词: 对抗样本, 图像重构, 深度学习, 图像分类

Abstract:

With the popularization of deep learning, more and more attention has been paid to its security issues.The adversarial sample is to add a small disturbance to the original image, which can cause the deep learning model to misclassify the image, which seriously affects the performance of deep learning technology.To address this challenge, the attack form and harm of the existing adversarial samples were analyzed.An adversarial examples defense method based on image reconstruction was proposed to effectively detect adversarial examples.The defense method used MNIST as the test data set.The core idea was image reconstruction, including central variance minimization and image quilting optimization.The central variance minimization was only processed for the central area of the image.The image quilting optimization incorporated the overlapping area into the patch block selection.Considered and took half the size of the patch as the overlap area.Using FGSM, BIM, DeepFool and C&W attack methods to generate adversarial samples to test the defense performance of the two methods, and compare with the existing three image reconstruction defense methods (cropping and scaling, bit depth compression and JPEG compression).The experimental results show that the central variance minimization and image quilting optimization algorithms proposed have a satisfied defense effect against the attacks of existing common adversarial samples.Image quilting optimization achieves over 75% classification accuracy for samples generated by the four attack algorithms, and the defense effect of minimizing central variance is around 70%.The three image reconstruction algorithms used for comparison have unstable defense effects on different attack algorithms, and the overall classification accuracy rate is less than 60%.The central variance minimization and image quilting optimization proposed achieve the purpose of effectively defending against adversarial samples.The experiments illustrate the defense effect of the proposed defense algorithm in different adversarial sample attack algorithms.The comparison between the reconstruction algorithm and the algorithm shows that the proposed scheme has good defense performance.

Key words: adversarial example, image reconstruction, deep learning, image classification

中图分类号: 

No Suggested Reading articles found!