[1] |
韩文静, 李海峰, 阮华斌 , 等. 语音情感识别研究进展综述[J]. 软件学报, 2014, 25(1): 37-50.
|
|
HAN W J , LI H F , RUAN H B , et al. A review of research progress in speech emotion recognition[J]. Journal of Software, 2014, 25(1): 37-50.
|
[2] |
王海坤, 潘嘉, 刘聪 . 语音识别技术的研究进展与展望[J]. 电信科学, 2018, 34(2): 1-11.
|
|
WANG H K , PAN J , LIU C . Research progress and prospect of speech recognition technology[J]. Telecommunications Science, 2018, 34(2): 1-11.
|
[3] |
YAMADA T , HASHIMOTO H , TOSA N . Pattern recognition of emotion with neural network[C]// The 1995 IEEE IECON 21st International Conference on Industrial Electronics,Control,and Instrumentation,Nov 6-10,1995,Orlando,FL,USA. Piscataway:IEEE Press, 1995: 183-187.
|
[4] |
TENG Z , JI W . Speech emotion recognition with i-vector feature and rnn model[C]// 2015 IEEE China Summit and International Conference on Signal and Information Processing (China SIP),July 12-15,2015,Chengdu,China. Piscataway:IEEE Press, 2015: 524-528.
|
[5] |
BASU A , CHAKRABORTY J , AFTABUDDIN M . Emotion recognition from speech using convolutional neural network with recurrent neural network architecture[C]// 2016 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA),Dec 13-16,2016,Jeju,South Korea. Piscataway:IEEE Press, 2017: 333-336.
|
[6] |
SHI B , BAI X , YAO C . An end-to-end trainable neural network for image-based sequence recognition and its application to scene text recognition[J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2017, 39(11).
|
[7] |
ZAZO R , LOZANO-DIEZ A , GONZALEZ D J , et al. Language identification in short utterances using long short-term memory (LSTM)[J]. Recurrent Neural Networks, 2016(1).
|
[8] |
GELLY G , GAUVAIN J L , LE V , et al. A divide-and-conquer approach for language identification based on recurrent neural networks[Z]. 2016.
|
[9] |
LOZANO-DIEZ A , ZAZO C R , GONZLEZ D J , et al. An end-to-end approach to language identification in short utterances using convolutional neural networks[J]. 2015.
|
[10] |
ZHANG X R , SONG P , ZHA C , et al. Auditory attention model based on Chirplet for cross-corpus speech emotion recognition[J]. Journal of Southeast University, 2016, 32(4): 402-407.
|
[11] |
纪滨, 杨盼盼, 申元霞 . 基于改进ITTI模型及粒子群优化算法的白细胞区域提取[J]. 安徽工业大学学报, 2016, 33(3): 284-288.
|
|
JI B , YANG P P , SHEN Y X . Leukocyte region extraction based on improved ITTI model and particle swarm optimization algorithm[J]. Journal of Anhui University of Technology, 2016, 33(3): 284-288.
|
[12] |
刘兵, 霍键亮 . 基于灰度概率统计的视觉注意改进算法[J]. 电子设计工程, 2013, 21(5): 54-57.
|
|
LIU B , HUO J L . Improved visual attention algorithm based on gray probability statistics[J]. Electronic Design Engineering, 2013, 21(5): 54-57.
|
[13] |
KALINLI O , CHEN R . Speech syllable/vowel /phone boundary detection using auditory attention cues: US20120253812[P].2016-02-02.
|
[14] |
STEVENS C , HARN B , CHARD D J , et al. Examining the role of attention and instruction in at-risk kind ergarteners electrophysiological measures of selective auditory attention before and after an early literacy intervention[J]. Journal of Learning Disabilities, 2013, 46(1): 73-86.
|
[15] |
张欣然, 巨晓正, 宋鹏 , 等. 用于垮库语音情感识别的 DBN特征融合方法[J]. 信号处理, 2017, 33(5): 649-650.
|
|
ZHANG X R , JU X Z , SONG P , et al. DBN feature fusion method for voice emotion recognition in library[J]. Signal Processing, 2017, 33(5): 649-650.
|
[16] |
HU J , SHEN L , SUN G . Squeeze-and-excitation networks[J]. arXiv: 1709.01507, 2017.
|
[17] |
EYBEN F , WOLLMER M , SCHULLER B . openSMILE—the Munich versatile and fast open-source audio feature extractor[C]// The 18th ACM International Conference on Multimedia,October 25-29,2010,Firenze,Italy. New York:ACM Press, 2010: 1459-1462.
|
[18] |
BARTZ C , HEROLD T , HAOJIN Y , et al. Language identification using deep convolutional recurrent neural networks[J]. arXiv: 1708.04811v1, 2017.
|