Chinese Journal of Network and Information Security ›› 2023, Vol. 9 ›› Issue (3): 150-160.doi: 10.11959/j.issn.2096-109x.2023046

• Papers • Previous Articles     Next Articles

Gender forgery of faces by fusing wavelet shortcut connection generative adversarial network

Wanze CHEN, Liqing HUANG, Jiazhen CHEN, Feng YE, Tianqiang HUANG, Haifeng LUO   

  1. College of Computer and Cyber Security, Fujian Normal University, FuZhou 350117, China
  • Revised:2023-03-02 Online:2023-06-25 Published:2023-06-01
  • Supported by:
    The National Natural Science Foundation of China(62072106);Fujian Province Natural Science Foundation(2020J01168);Fujian Provincial Department of Education (Rank A)(JAT210053)

Abstract:

The mainstream methods in the field of facial attribute manipulation had the following two defects due to data and model architecture limitations.First, the bottleneck structure of the autoencoder model results in the loss of feature information, and the traditional method of continuously injected styles to the source domain features during the decoding process makes the generated image too referential to the target domain while losing the identity information and fine-grained details.Second, differences in facial attributes composition between images, such as gender, ethnicity, or age can cause variations in frequency domain information.And the current unsupervised training methods do not automatically adjust the proportion of source and target domain information in the style injection stage, resulting in artifacts in generated images.A facial gender forgery model based on generative adversarial networks and image-to-image translation techniques, namely fused wavelet shortcut connection generative adversarial network (WscGAN), was proposed to address the these issues.Shortcut connections were added to the autoencoder structure, and the outputs of different encoding stages were decomposed at the feature level by wavelet transform.Attention mechanism was employed to process them one by one, to dynamically change the proportion of source domain features at different frequencies in the decoding process.This model could complete forgery of facial images in terms of gender attributes.To verify the effectiveness of the model, it was conducted on the CelebA-HQ dataset and the FFHQ dataset.Compared with the existing optimal models, the method improves the FID and LPIPS indices by 5.4% and 11.2%, and by 1.8% and 6.7%, respectively.Furthermore, the effectiveness of the proposed method in improving the gender attribute conversion of facial images is fully demonstrated by the results based on qualitative visual comparisons.

Key words: image generation, generative adversarial network, image-to-image translation, facial attribute manipulation, wavelet transform

CLC Number: 

No Suggested Reading articles found!