通信学报 ›› 2018, Vol. 39 ›› Issue (2): 53-64.doi: 10.11959/j.issn.1000-436x.2018024
修回日期:
2017-12-08
出版日期:
2018-02-01
发布日期:
2018-03-28
作者简介:
王衡军(1973-),男,湖南衡阳人,解放军信息工程大学副教授、硕士生导师,主要研究方向为机器学习、自然语言处理和信息安全。|司念文(1992-),男,湖北襄阳人,解放军信息工程大学硕士生,主要研究方向为机器学习、自然语言处理。|宋玉龙(1995-),男,安徽阜阳人,73671部队助理工程师,主要研究方向为网络与信息安全。|单义栋(1988-),男,山东乳山人,解放军信息工程大学硕士生,主要研究方向为自然语言处理。
Hengjun WANG1,Nianwen SI1(),Yulong SONG2,Yidong SHAN1
Revised:
2017-12-08
Online:
2018-02-01
Published:
2018-03-28
摘要:
利用时序型长短时记忆(LSTM,long short term memory)网络和分片池化的卷积神经网络(CNN,convolutional neural network),分别提取词向量特征和全局向量特征,将2类特征结合输入前馈网络中进行训练;模型训练中,采用基于概率的训练方法。与改进前的模型相比,该模型能够更多地关注句子的全局特征;相较于最大化间隔训练算法,所提训练方法更充分地利用所有可能的依存句法树进行参数更新。为了验证该模型的性能,在宾州中文树库(CTB5,Chinese Penn Treebank 5)上进行实验,结果表明,与已有的仅使用LSTM或CNN的句法分析模型相比,该模型在保证一定效率的同时,能够有效提升依存分析准确率。
中图分类号:
王衡军,司念文,宋玉龙,单义栋. 结合全局向量特征的神经网络依存句法分析模型[J]. 通信学报, 2018, 39(2): 53-64.
Hengjun WANG,Nianwen SI,Yulong SONG,Yidong SHAN. Neural network model for dependency parsing incorporating global vector feature[J]. Journal on Communications, 2018, 39(2): 53-64.
表3
预训练词向量维度数值"
词语 | 维度1 | 维度2 | 维度3 | 维度4 | 维度5 | 维度6 | 维度7 | 维度8 | … |
北京 | 0.784 532 | 0.126 508 | ?0.057 374 | ?0.154 533 | 0.332 767 | ?0.097 764 | 0.326 162 | ?0.050 384 | … |
是 | 0.266 844 | 0.302 665 | ?0.264 444 | ?0.226 088 | 0.441 961 | 0.016 561 | 0.301 905 | 0.502 168 | … |
中国 | 0.219 274 | ?0.014 053 | 0.131 701 | ?0.286 423 | 0.335 406 | ?0.358 501 | 0.673 590 | ?0.042 960 | … |
的 | 0.217 868 | 0.127 214 | ?0.149 943 | ?0.321 670 | 0.492 081 | 0.059 336 | 0.125 756 | 0.118 209 | … |
首都 | ?0.026 037 | ?0.481 110 | 0.242 895 | 0.234 439 | 0.002 181 | 0.174 378 | 0.376 564 | 0.373 155 | … |
。 | 0.472 912 | 0.190 043 | ?0.269 932 | ?0.252 637 | 0.368 207 | ?0.009 401 | 0.306 351 | 0.124 188 | … |
上海 | 0.730 383 | ?0.120 522 | 0.057 258 | ?0.139 031 | 0.066 892 | ?0.064 185 | 0.207 783 | ?0.374 539 | … |
浦东 | 0.484 598 | ?0.354 264 | ?0.023 650 | 0.244 754 | ?0.167 892 | 0.134 869 | 0.174 820 | ?0.593 116 | … |
开发 | 0.095 341 | ?0.730 478 | ?0.233 886 | ?0.270 567 | 0.182 623 | 0.313 196 | ?0.008 622 | ?0.193 032 | … |
与 | ?0.206 168 | ?0.178 119 | 0.087 402 | ?0.033 061 | 0.410 609 | ?0.186 066 | 0.368 784 | 0.011 376 | … |
法制 | ?0.176 999 | 0.472 356 | ?0.179 480 | ?0.312 853 | 0.651 259 | ?0.345 816 | ?0.471 174 | ?0.225 371 | … |
建设 | 0.003 175 | ?0.034 065 | ?0.223 146 | 0.021 553 | 0.255 378 | 0.221 632 | ?0.279 153 | 0.244 122 | … |
同步 | ?0.144 631 | ?0.027 353 | ?0.243 820 | ?0.205 473 | 0.200 052 | 0.040 610 | 0.059 072 | ?0.277 645 | … |
[1] | MCDONALD R , CRAMMER K , PEREIRA F . Online large-margin training of dependency parsers[C]// The 43rd Annual Meeting on Association for Computational Linguistics. 2005: 91-98. |
[2] | EISNER J M . Three new probabilistic models for dependency parsing:an exploration[J]. Computer Science, 1997: 340-345. |
[3] | MCDONALD R T , PEREIRA F C N . Online learning of approximate dependency parsing algorithms[C]// The 11th Conference of the European Chapter of the Association for Computational Linguistics. 2006: 81-88. |
[4] | CARRERAS X , . Experiments with a higher-order projective dependency parser[C]// The 2007 Joint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning. 2007: 957-961. |
[5] | KOO T , COLLINS M . Efficient third-order dependency parsers[C]// The 48th Annual Meeting of the Association for Computational Linguistics. 2010: 1-11. |
[6] | 马学喆 . 依存句法分析的若干关键问题的研究[D]. 上海:上海交通大学, 2013. |
MA X Z . Research on key issues of dependency parsing[D]. Shanghai:Shanghai Jiaotong University, 2013. | |
[7] | PEI W Z , GE T , CHANG B B . An effective neural network model for graph-based dependency parsing[C]// The 53rd Annual Meeting of the Association for Computational Linguistics. 2015: 313-322. |
[8] | WANG W H , CHANG B B . Graph-based dependency parsing with bidirectional LSTM[C]// The 54th Annual Meeting of the Association for Computational Linguistics. 2016: 2306-2315. |
[9] | KIPERWASSER E , GOLDBERG Y . Simple and accurate dependency parsing using bidirectional LSTM feature representations[J]. Transactions of the Association for Computational Linguistics, 2016(4): 313-327. |
[10] | ZHANG Z S , ZHAO H , QIN L H . Probabilistic graph-based dependency parsing with convolutional neural network[C]// The 54th Annual Meeting of the Association for Computational Linguistics. 2016: 1382-1392. |
[11] | MA X Z , HOVY E . Neural probabilistic model for non-projective MST parsing[J].arXiv:arXiv,2017. arXiv:arXiv:1701.00874 2017. |
[12] | ZHANG J J , ZHANG D K , HAO J . Local translation prediction with global sentence representation[C]// The 24th International Joint Conference on Artificial Intelligence. 2015: 1398-1404. |
[13] | KALCHBRENNER N , GREFENSTETTE E , BLUNSOM P . A convolutional neural network for modelling sentences[J]. Eprint Arxiv, 2014(1). |
[14] | COLLOBERT R , WESTON J , BOTTOU L ,et al. Natural language processing (almost) from scratch[J]. Journal of Machine Learning Research, 2011(12): 2493-2537. |
[15] | JASON C , ERIC N . Named entity recognition with bidirectional lstm-cnns[J]. Transactions of the Association for Computational Linguistics, 2016(4): 357-370. |
[16] | 曾谁飞, 张笑燕, 杜晓峰 ,等. 基于神经网络的文本表示模型新方法[J]. 通信学报, 2017,38(4): 86-98. |
ZENG S F , ZHANG X Y , DU X F ,et al. New method of text representation model based on neural network[J]. Journal on Communications, 2017,38(4): 86-98. | |
[17] | CHEN D Q , MANNING C . A fast and accurate dependency parser using neural networks[C]// Conference on Empirical Methods in Natural Language Processing. 2014: 740-750. |
[18] | WEISS D , ALBERTI C , COLLINS M ,et al. Structured training for neural network transition-based parsing[C]// The 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing. 2015: 323-333. |
[19] | ANDOR D , ALBERTI C , WEISS D ,et al. Globally normalized transition-based neural networks[C]// The 54rd Annual Meeting of the Association for Computational Linguistics. 2016: 2442-2452. |
[20] | DYER C , BALLESTEROS M , WANG L ,et al. Transition-based dependency parsing with stack long short-term memory[J]. ComputerScience, 2015,37(2): 321-332. |
[21] | CHENG H , FANG H , HE X D ,et al. Bi-directional attention with agreement for dependency parsing[C]// Conference on Empirical Methods in Natural Language Processing. 2016: 2204-2214. |
[22] | DOZAT T , MANNING C D . Deep biaffine attention for neural dependency parsing[J].arXiv:arXiv 1611.01734,2016. arXiv:arXiv 1611,01734, 2016. |
[23] | NEUBIG G , DYER C , GOLDBERG Y ,et al. DyNet:the dynamic neural network toolkit[J]. arXiv:arXiv 1701.03980, 2017. |
[24] | ZHANG Y , NIVRE J . Transition-based dependency parsing with rich non-local features[C]// The 49th Annual Meeting of the Association for Computational Linguistics. 2011: 188-193. |
[25] | ZHANG H , MCDONALD R . Enforcing structural diversity in cube-pruned dependency parsing[C]// The 52nd Annual Meeting of the Association for Computational Linguistics. 2014: 656-666. |
[1] | 胡斌, 谈潇, 王森鹏. 基于分治策略的SAT差分自动化搜索算法及其应用[J]. 通信学报, 2023, 44(4): 137-144. |
[2] | 江沸菠, 彭于波, 董莉. 面向6G的深度图像语义通信模型[J]. 通信学报, 2023, 44(3): 198-208. |
[3] | 仪双燕, 梁永生, 陆晶晶, 柳伟, 胡涛, 何震宇. 联合低秩重构和投影重构的稳健特征选择方法[J]. 通信学报, 2023, 44(3): 209-219. |
[4] | 刘建勋, 丁领航, 康国胜, 曹步清, 肖勇. 基于特征深度融合的Web服务QoS联合预测[J]. 通信学报, 2022, 43(7): 215-226. |
[5] | 李昂, 陈建新, 魏昕, 周亮. 面向6G的跨模态信号重建技术[J]. 通信学报, 2022, 43(6): 28-40. |
[6] | 兰巨龙, 朱棣, 李丹. 面向多模态网络业务切片的虚拟网络功能资源容量智能预测方法[J]. 通信学报, 2022, 43(6): 143-155. |
[7] | 王晓丹, 李京泰, 宋亚飞. DDAC:面向卷积神经网络图像隐写分析模型的特征提取方法[J]. 通信学报, 2022, 43(5): 68-81. |
[8] | 廖育荣, 王海宁, 林存宝, 李阳, 方宇强, 倪淑燕. 基于深度学习的光学遥感图像目标检测研究进展[J]. 通信学报, 2022, 43(5): 190-203. |
[9] | 王欣, 付威. 基于推土机距离的证据冲突强度量方法[J]. 通信学报, 2022, 43(5): 204-213. |
[10] | 张帆, 黄赟, 方子茁, 郭威. 卷积神经网络的损失最小训练后参数量化方法[J]. 通信学报, 2022, 43(4): 114-122. |
[11] | 言洪萍, 周强, 王世豪, 姚旺, 何刘坤, 王良民. 基于SDN的实际网络流中Tor网页复合特征提取方法[J]. 通信学报, 2022, 43(3): 76-87. |
[12] | 朱政宇, 侯庚旺, 黄崇文, 孙钢灿, 郝万明, 梁静. 基于并行CNN的RIS辅助D2D保密通信系统资源分配算法[J]. 通信学报, 2022, 43(3): 172-179. |
[13] | 王泽宇, 布树辉, 黄伟, 郑远攀, 吴庆岗, 常化文, 张旭. 基于多模态特征的无监督领域自适应多级对抗语义分割网络[J]. 通信学报, 2022, 43(12): 157-171. |
[14] | 谢丽霞, 李雪鸥, 杨宏宇, 张良, 成翔. 基于样本特征强化的APT攻击多阶段检测方法[J]. 通信学报, 2022, 43(12): 66-76. |
[15] | 段雪源, 付钰, 王坤, 李彬. 基于简单统计特征的LDoS攻击检测方法[J]. 通信学报, 2022, 43(11): 53-64. |
阅读次数 | ||||||
全文 |
|
|||||
摘要 |
|
|||||
|