Chinese Journal of Network and Information Security ›› 2023, Vol. 9 ›› Issue (6): 154-165.doi: 10.11959/j.issn.2096-109x.2023090

• Papers • Previous Articles    

Visual explanation method for reversible neural networks

Xinying MU1, Bingbing SONG2, Fanxiao LI1, Yisen ZHENG1, Wei ZHOU1, Yunyun DONG1,2   

  1. 1 National Pilot School of Software, Yunnan University, Kunming 650000, China
    2 School of Information Science and Engineering, Yunnan University,Kunming 650000,China
  • Revised:2023-09-09 Online:2023-12-01 Published:2023-12-01
  • Supported by:
    TheNational Natural Science Foundation of China(62162067);TheNational Natural Science Foundation of China(62101480);The Natural Science Foundation of Yunnan Province(202005AC160007);The Natural Science Foundation of Yunnan Province(202001BB050076);Yunnan Provincial Department of Education Fund Project(2022j0008);Key R&D Plan of Yunnan Province;Yunnan Province Chi Xuebin Expert Workstation Project(202305AF150078)

Abstract:

The issue of model explainability has gained significant attention in understanding the vulnerabilities and anonymous decision-making processes inherent in deep neural networks (DNN).While there has been considerable research on explainability for traditional DNN, there is a lack of exploration on the operation mechanism and explainability of reversible neural networks (RevNN).Additionally, the existing explanation methods for traditional DNN are not suitable for RevNN and suffer from issues such as excessive noise and gradient saturation.To address these limitations, a visual explanation method called visual explanation method for reversible neural network (VERN) was proposed for RevNN.VERN leverages the reversible property of RevNN and is based on the class-activation mapping mechanism.The correspondence between the feature map and the input image was explored by VERN, allowing for the mapping of classification weights of regional feature maps to the corresponding regions of the input image.The importance of each region for model decision-making was revealed through this process, which generates a basis for model decision-making.Experimental comparisons with other explanation methods on generalized datasets demonstrate that VERN achieves a more focused visual effect, surpassing suboptimal methods with up to 7.80% improvement in average drop (AD) metrics and up to 6.05% improvement in average increase (AI) metrics in recognition tasks.VERN also exhibits an 82.00% level of localization for the maximum point of the heat value.Furthermore, VERN can be applied to explain traditional DNN and exhibits good scalability, improving the performance of other methods in explaining RevNN.Furthermore, through adversarial attack analysis experiments, it is observed that adversarial attacks alter the decision basis of the model.This is reflected in the misalignment of the model’s attention regions, thereby aiding in the exploration of the operation mechanism of adversarial attacks.

Key words: model explainability, reversible neural network, visualization, class activation mapping, artificial intelligence security

CLC Number: 

No Suggested Reading articles found!