智能科学与技术学报 ›› 2021, Vol. 3 ›› Issue (1): 76-84.doi: 10.11959/j.issn.2096-6652.202108

• 专题:情感脑机接口 • 上一篇    下一篇

基于3D分层卷积融合的多模态生理信号情绪识别

凌文芬1,2, 陈思含1,2, 彭勇1,2, 孔万增1,2   

  1. 1 杭州电子科技大学计算机学院,浙江 杭州 310018
    2 浙江省脑机协同智能重点实验室,浙江 杭州 310018
  • 修回日期:2021-02-05 出版日期:2021-03-15 发布日期:2021-03-01
  • 作者简介:凌文芬(1995- ),女,杭州电子科技大学计算机学院硕士生,主要研究方向为情绪生理信号、特征提取、情绪识别。
    陈思含(1997- ),女,杭州电子科技大学计算机学院硕士生,主要研究方向为情绪生理信号、特征提取、情绪识别。
    彭勇(1985- ),男,博士,杭州电子科技大学计算机学院副教授,主要研究方向为机器学习、模式识别与脑机交互算法及应用。
    孔万增(1980- ),男,博士,杭州电子科技大学计算机学院认知与智能计算研究所教授、博士生导师,杭州电子科技大学研究生院副院长,浙江省脑机协同智能重点实验室主任,主要研究方向为人工智能与模式识别、嵌入式可穿戴计算、脑机交互与认知计算等。
  • 基金资助:
    国家重点研发计划基金资助项目(2017YFE0116800);国家自然科学基金资助项目(U1909202);浙江省科技计划项目(2018C04012);浙江省“脑机协同智能”重点实验室开放基金项目(20200E10010)

Multi-modal physiological signal emotion recognition based on 3D hierarchical convolution fusion

Wenfen LING1,2, Sihan CHEN1,2, Yong PENG1,2, Wanzeng KONG1,2   

  1. 1 College of Computer Science and Techonology, Hangzhou Dianzi University, Hangzhou 310018, China
    2 Key Laboratory of Brain Machine Collaborative Intelligence of Zhejiang Province, Hangzhou 310018, China
  • Revised:2021-02-05 Online:2021-03-15 Published:2021-03-01
  • Supported by:
    The National Key Research and Development Program of China(2017YFE0116800);The National Natural Science Foundation of China(U1909202);Science and Technology Program of Zhejiang Province(2018C04012);Key Laboratory of Brain Machine Collaborative Intelligence of Zhejiang Province(20200E10010)

摘要:

近年来,脑电等生理信号由于能客观体现真实情绪已逐渐成为情绪识别研究的热门对象。然而,单模态的脑电信号存在情绪信息特征不完备问题,多模态生理信号存在情绪信息交互不充分问题。针对这些问题,提出基于3D分层卷积的多模态特征融合模型,旨在充分挖掘多模态交互关系,更准确地刻画情感信息。首先分别通过深度可分离卷积网络提取脑电、眼电和肌电3种模态的生理信号的多模态初级情绪特征信息,再对得到的多模态初级情绪特征信息进行3D卷积融合操作,实现两两模态间的局部交互以及所有模态间的全局交互,获取包含不同生理信号情绪特征的多模态融合特征。实验结果表明,提出的模型在DEAP数据集的效价、唤醒度的二分类和四分类任务中达到了98%的平均准确率。

关键词: 生理信号, 情绪识别, 3D分层卷积, 多模态交互

Abstract:

In recent years, physiological signals such as electroencephalograhpy (EEG) have gradually become popular objects of emotion recognition research because they can objectively reflect true emotions.However, the single-modal EEG signal has the problem of incomplete emotional information representation, and the multi-modal physiological signal has the problem of insufficient emotional information interaction.Therefore, a 3D hierarchical convolutional fusion model was proposed, which aimed to fully explore multi-modal interaction relationships and more accurately describe emotional information.The method first extracted the primary emotional representation information of EEG , electro-oculogram (EOG) and electromyography (EMG) by depthwise separable convolution network, and then performed 3D convolution fusion operation on the obtained multi-modal primary emotional representation information to realize the pairwise mode local interactions between states and global interactions among all modalities, so as to obtain multi-modal fusion representations containing emotional characteristics of different physiological signals.The results show that the accuracy in the valence and arousal of the two-class and four-class tasks on DEAP dataset are both 98% by the proposed model.

Key words: physiological signal, emotion recognition, 3D hierarchical convolutional, multi-modal interaction

中图分类号: 

No Suggested Reading articles found!