网络与信息安全学报 ›› 2023, Vol. 9 ›› Issue (4): 90-103.doi: 10.11959/j.issn.2096-109x.2023056

• 学术论文 • 上一篇    

基于模型相似度的模型恶意代码夹带检测方法

汪德刚, 孙奕, 周传鑫, 高琦, 杨帆   

  1. 信息工程大学,河南 郑州 450001
  • 修回日期:2022-05-30 出版日期:2023-08-01 发布日期:2023-08-01
  • 作者简介:汪德刚(1996- ),男,陕西安康人,信息工程大学硕士生,主要研究方向为数据安全交换、恶意检测
    孙奕(1979- ),女,河南郑州人,博士,信息工程大学副教授,主要研究方向为网络与信息安全、数据安全交换
    周传鑫(1997- ),男,安徽蚌埠人,信息工程大学硕士生,主要研究方向为数据安全交换、机器学习隐私保护
    高琦(1997- ),男,湖北襄阳人,信息工程大学硕士生,主要研究方向为机器学习隐私保护
    杨帆(1996- ),女,天津人,信息工程大学硕士生,主要研究方向为隐私计算、可验证计算

Malicious code within model detection method based on model similarity

Degang WANG, Yi SUN, Chuanxin ZHOU, Qi GAO, Fan YANG   

  1. Information Engineering University, Zhengzhou 450001, China
  • Revised:2022-05-30 Online:2023-08-01 Published:2023-08-01

摘要:

联邦学习主要通过源数据不出本地而仅交互模型参数的方式保护参与共享用户数据的隐私安全,然而其仍然面临众多安全挑战,目前研究者主要针对如何增强模型隐私保护和检测恶意模型攻击等问题展开较为广泛的研究,然而利用联邦学习过程中频繁交互的模型数据进行恶意代码夹带导致风险扩散的问题鲜有研究。针对联邦学习训练过程中通过模型传递恶意代码导致风险扩散的问题,提出一种基于模型相似度的模型恶意代码夹带检测方法。通过分析联邦学习本地模型与全局模型的迭代过程,提出计算模型距离的方法,并使用模型距离量化模型之间的相似度,最终根据各客户端模型之间的相似度对携带恶意代码的模型进行检测。实验结果表明,提出的检测方法具有较好的性能指标,当训练集为独立同分布时,在178 MB大小的模型中嵌入0.375 MB恶意代码,检测方法的真正率为82.9%,误报率为1.8%;嵌入0.75 MB恶意代码时,检测方法的真正率为 96.6%,误报率为 0.38%。当训练集为非独立同分布时,检测方法的准确率随恶意代码嵌入率以及联邦学习训练轮数的增加而增加。在对恶意代码进行加密的情况下,提出的检测方法仍然能够达到 90%以上的准确率。在多攻击者的场景中,攻击者数量已知与未知时的检测方法准确率均能保持在90%左右。

关键词: 联邦学习, 模型, 模型相似度, 恶意代码, 检测

Abstract:

The privacy of user data in federated learning is mainly protected by exchanging model parameters instead of source data.However, federated learning still encounters many security challenges.Extensive research has been conducted to enhance model privacy and detect malicious model attacks.Nevertheless, the issue of risk-spreading through malicious code propagation during the frequent exchange of model data in the federated learning process has received limited attention.To address this issue, a method for detecting malicious code within models, based on model similarity, was proposed.By analyzing the iterative process of local and global models in federated learning, a model distance calculation method was introduced to quantify the similarity between models.Subsequently, the presence of a model carrying malicious code is detected based on the similarity between client models.Experimental results demonstrate the effectiveness of the proposed detection method.For a 178MB model containing 0.375MB embedded malicious code in a training set that is independent and identically distributed, the detection method achieves a true rate of 82.9% and a false positive rate of 1.8%.With 0.75MB of malicious code embedded in the model, the detection method achieves a true rate of 96.6% and a false positive rate of 0.38%.In the case of a non-independent and non-identically distributed training set, the accuracy of the detection method improves as the rate of malicious code embedding and the number of federated learning training rounds increase.Even when the malicious code is encrypted, the accuracy of the proposed detection method still achieves over 90%.In a multi-attacker scenario, the detection method maintains an accuracy of approximately 90% regardless of whether the number of attackers is known or unknown.

Key words: federated learning, model, model similarity, malicious code, detection

中图分类号: 

No Suggested Reading articles found!