HyperED: A hierarchy‐aware network based on hyperbolic geometry for event detection

事件(粒子物理) 计算机科学 人工智能 杠杆(统计) 等级制度 模式识别(心理学) 机器学习 数据挖掘 市场经济 量子力学 物理 经济
作者
Meng Zhang,Zhiwen Xie,Jin Liu,Xiao Liu,Xiao Yu,Bo Huang
出处
期刊:Computational Intelligence [Wiley]
卷期号:40 (1)
标识
DOI:10.1111/coin.12627
摘要

Computational IntelligenceVolume 40, Issue 1 e12627 SPECIAL ISSUE ARTICLE HyperED: A hierarchy-aware network based on hyperbolic geometry for event detection Meng Zhang, Meng Zhang orcid.org/0000-0002-3649-0940 School of Computer Science, Wuhan University, Wuhan, ChinaSearch for more papers by this authorZhiwen Xie, Zhiwen Xie School of Computer Science, Central China Normal University, Wuhan, ChinaSearch for more papers by this authorJin Liu, Corresponding Author Jin Liu [email protected] School of Computer Science, Wuhan University, Wuhan, China Correspondence Jin Liu, School of Computer Science, Wuhan University. 430072 Wuhan China. Email: [email protected]Search for more papers by this authorXiao Liu, Xiao Liu School of Information Technology, Deakin University, Geelong, AustraliaSearch for more papers by this authorXiao Yu, Xiao Yu School of Computer Science and Artificial Intelligence, Wuhan University of Technology, Wuhan, ChinaSearch for more papers by this authorBo Huang, Bo Huang School of Electronic and Electrical Engineering, Shanghai University of Engineering Science, Shanghai, ChinaSearch for more papers by this author Meng Zhang, Meng Zhang orcid.org/0000-0002-3649-0940 School of Computer Science, Wuhan University, Wuhan, ChinaSearch for more papers by this authorZhiwen Xie, Zhiwen Xie School of Computer Science, Central China Normal University, Wuhan, ChinaSearch for more papers by this authorJin Liu, Corresponding Author Jin Liu [email protected] School of Computer Science, Wuhan University, Wuhan, China Correspondence Jin Liu, School of Computer Science, Wuhan University. 430072 Wuhan China. Email: [email protected]Search for more papers by this authorXiao Liu, Xiao Liu School of Information Technology, Deakin University, Geelong, AustraliaSearch for more papers by this authorXiao Yu, Xiao Yu School of Computer Science and Artificial Intelligence, Wuhan University of Technology, Wuhan, ChinaSearch for more papers by this authorBo Huang, Bo Huang School of Electronic and Electrical Engineering, Shanghai University of Engineering Science, Shanghai, ChinaSearch for more papers by this author First published: 04 January 2024 https://doi.org/10.1111/coin.12627Read the full textAboutPDF ToolsRequest permissionExport citationAdd to favoritesTrack citation ShareShare Give accessShare full text accessShare full-text accessPlease review our Terms and Conditions of Use and check box below to share full-text version of article.I have read and accept the Wiley Online Library Terms and Conditions of UseShareable LinkUse the link below to share a full-text version of this article with your friends and colleagues. Learn more.Copy URL Share a linkShare onEmailFacebookTwitterLinkedInRedditWechat Abstract Event detection plays an essential role in the task of event extraction. It aims at identifying event trigger words in a sentence and classifying event types. Generally, multiple event types are usually well-organized with a hierarchical structure in real-world scenarios, and hierarchical correlations between event types can be used to enhance event detection performance. However, such kind of hierarchical information has received insufficient attention which can lead to misclassification between multiple event types. In addition, the most existing methods perform event detection in Euclidean space, which cannot adequately represent hierarchical relationships. To address these issues, we propose a novel event detection network HyperED which embeds the event context and types in Poincaré ball of hyperbolic geometry to help learn hierarchical features between events. Specifically, for the event detection context, we first leverage the pre-trained BERT or BiLSTM in Euclidean space to learn the semantic features of ED sentences. Meanwhile, to make full use of the dependency knowledge, a GNN-based model is applied when encoding event types to learn the correlations between events. Then we use a simple neural-based transformation to project the embeddings into the Poincaré ball to capture hierarchical features, and a distance score in hyperbolic space is computed for prediction. The experiments on MAVEN and ACE 2005 datasets indicate the effectiveness of the HyperED model and prove the natural advantages of hyperbolic spaces in expressing hierarchies in an intuitive way. CONFLICT OF INTEREST STATEMENT The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper. Open Research DATA AVAILABILITY STATEMENT The data that support the findings of this study are available from the corresponding author upon reasonable request. REFERENCES 1Buyko E, Faessler E, Wermter J, Hahn U. Syntactic simplification and semantic enrichment–trimming dependency graphs for event extraction. Comput Intell. 2011; 27(4): 610-644. 10.1111/j.1467-8640.2011.00402.x Web of Science®Google Scholar 2Liu J, Chen Y, Liu K, Bi W, Liu X. Event extraction as machine reading comprehension. Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP). Association for Computational Linguistics; 2020: 1641-1651. 10.18653/v1/2020.emnlp-main.128 Google Scholar 3Li F, Peng W, Chen Y, et al. Event extraction as multi-turn question answering. Findings of the Association for Computational Linguistics (EMNLP). Vol 2020. Association for Computational Linguistics; 2020: 829-838. Google Scholar 4Ahn D. The stages of event extraction. Proceedings of the Workshop on Annotating and Reasoning about Time and Events. Association for Computational Linguistics; 2006: 1-8. 10.3115/1629235.1629236 Google Scholar 5Wang X, Wang Z, Han X, et al. MAVEN: A Massive General Domain Event Detection Dataset. Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP). Association for Computational Linguistics; 2020: 1652-1671. 10.18653/v1/2020.emnlp-main.129 Google Scholar 6Walker C, Strassel S, Medero J, Maeda K. ACE 2005 multilingual training corpus ldc2006t06, 2006. https://catalog.ldc.upenn.edu/LDC2006T06 Google Scholar 7Liu S, Chen Y, He S, Liu K, Zhao J. Leveraging framenet to improve automatic event detection. Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics. The Association for Computer Linguistics; 2016: 2134-2143. 10.18653/v1/P16-1201 Google Scholar 8Chen Y, Xu L, Liu K, Zeng D, Zhao J. Event extraction via dynamic multi-pooling convolutional neural networks. Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing. The Association for Computer Linguistics; 2015: 167-176. Google Scholar 9Wang X, Han X, Liu Z, Sun M, Li P. Adversarial training for weakly supervised event detection. Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. The Association for Computer Linguistics; 2019: 998-1008. 10.18653/v1/N19-1105 Google Scholar 10Lafferty J, McCallum A, Pereira FC. Conditional random fields: probabilistic models for segmenting and labeling sequence data. 2001. Google Scholar 11Ji H, Grishman R. Refining event extraction through cross-document inference. Proceedings of the 46th Annual Meeting of the Association for Computational Linguistics. The Association for Computer Linguistics; 2008: 254-262. Google Scholar 12Liao S, Grishman R. Using document level cross-event inference to improve event extraction. Proceedings of the 48th Annual Meeting of the Association for Computational Linguistics. The Association for Computer Linguistics; 2010: 789-797. Google Scholar 13Nguyen TH, Cho K, Grishman R. Joint event extraction via recurrent neural networks. Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. The Association for Computational Linguistics; 2016: 300-309. 10.18653/v1/N16-1034 Google Scholar 14Liu S, Chen Y, Liu K, Zhao J. Exploiting argument information to improve event detection via supervised attention mechanisms. Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics; 2017: 1789-1798. 10.18653/v1/P17-1164 Google Scholar 15Nguyen T, Grishman R. Graph convolutional networks with argument-aware pooling for event detection. Proceedings of the AAAI Conference on Artificial Intelligence. AAAI Press; 2018. Google Scholar 16Chen Y, Yang H, Liu K, Zhao J, Jia Y. Collective event detection via a hierarchical and bias tagging networks with gated multi-level attention mechanisms. Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics; 2018: 1267-1276. 10.18653/v1/D18-1158 Google Scholar 17Nickel M, Kiela D. Poincaré embeddings for learning hierarchical representations. Adv Neural Inform Process Syst. 2017; 30:6338-6347. Google Scholar 18López F, Heinzerling B, Strube M. Fine-Grained Entity Typing in Hyperbolic Space. Proceedings of the 4th Workshop on Representation Learning for NLP (RepL4NLP-2019), Association for Computational Linguistics; 2019: 169-180. 10.18653/v1/W19-4319 Google Scholar 19Wu Z, Pan S, Chen F, Long G, Zhang C, Philip SY. A comprehensive survey on graph neural networks. IEEE Trans Neural Netw Learn Syst. 2020; 32(1): 4-24. 10.1109/TNNLS.2020.2978386 Web of Science®Google Scholar 20Liu S, Liu K, He S, Zhao J. A probabilistic soft logic based approach to exploiting latent and global information in event classification. Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence. AAAI Press; 2016. 10.1609/aaai.v30i1.10375 Google Scholar 21Veyseh APB, Van Nguyen M, Trung NN, Min B, Nguyen TH. Modeling document-level context for event detection via important context selection. Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics; 2021: 5403-5413. 10.18653/v1/2021.emnlp-main.439 Google Scholar 22Lou D, Liao Z, Deng S, Zhang N, Chen H. MLBiNet: A Cross-Sentence Collective Event Detection Network. Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing. Association for Computational Linguistics; 2021: 4829-4839. 10.18653/v1/2021.acl-long.373 Google Scholar 23Xu S, Li P, Zhu Q. Improving Event Coreference Resolution Using Document-level and Topic-level Information. Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics; 2022: 6765-6775. 10.18653/v1/2022.emnlp-main.454 Google Scholar 24Liang Y, Jiang Z, Yin D, Ren B. RAAT: Relation-Augmented Attention Transformer for Relation Modeling in Document-Level Event Extraction. Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. Association for Computational Linguistics; 2022: 4985-4997. 10.18653/v1/2022.naacl-main.367 Google Scholar 25Rumelhart DE, Hinton GE, Williams RJ. Learning representations by back-propagating errors. nature. 1986; 323(6088): 533-536. 10.1038/323533a0 Web of Science®Google Scholar 26Kenton JDMWC, Toutanova LK. BERT: pre-training of deep bidirectional transformers for language understanding. Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. Association for Computational Linguistics; 2019: 4171-4186. Google Scholar 27Yan H, Jin X, Meng X, Guo J, Cheng X. Event detection with multi-order graph convolution and aggregated attention. Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP). Association for Computational Linguistics; 2019: 5766-5770. 10.18653/v1/D19-1582 Google Scholar 28Yao L, Mao C, Luo Y. Graph convolutional networks for text classification. Proceedings of the AAAI conference on artificial intelligence. AAAI Press; 2019: 7370-7377. 10.1609/aaai.v33i01.33017370 Google Scholar 29Cui S, Yu B, Liu T, Zhang Z, Wang X, Shi J. Edge-Enhanced Graph Convolution Networks for Event Detection with Syntactic Relation. Findings of the Association for Computational Linguistics: EMNLP 2020. Association for Computational Linguistics; 2020: 2329-2339. 10.18653/v1/2020.findings-emnlp.211 Google Scholar 30Hsu I, Huang K, Boschee E, et al. DEGREE: a data-efficient generation-based event extraction model. Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL 2022). Association for Computational Linguistics; 2022: 1890-1908. Google Scholar 31Liu X, Huang H, Shi G, Wang B. Dynamic Prefix-Tuning for Generative Template-based Event Extraction. Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). Vol 2022. ACL; 2022: 5216-5228. 10.18653/v1/2022.acl-long.358 Google Scholar 32Ma MD, Taylor A, Wang W, Peng N. DICE: Data-Efficient Clinical Event Extraction with Generative Models. Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). Vol 2023. ACL; 2023: 15898-15917. 10.18653/v1/2023.acl-long.886 Google Scholar 33López F, Strube M. A fully hyperbolic neural model for hierarchical multi-class classification. Findings of the Association for Computational Linguistics (EMNLP). Vol 2020. Association for Computational Linguistics; 2020: 460-475. Google Scholar 34Dhingra B, Shallue C, Norouzi M, Dai A, Dahl G. Embedding Text in Hyperbolic Spaces. Proceedings of the Twelfth Workshop on Graph-Based Methods for Natural Language Processing (TextGraphs-12). Association for Computational Linguistics; 2018: 59-69. 10.18653/v1/W18-1708 Google Scholar 35Hamann M. On the tree-likeness of hyperbolic spaces. Mathematical Proceedings of the Cambridge Philosophical Society. Cambridge University Press; 2018: 345-361. Google Scholar 36Peng W, Varanka T, Mostafa A, Shi H, Zhao G. Hyperbolic Deep Neural Networks: A Survey. IEEE Trans Pattern Anal Mach Intell. 2021;44(12):10023-10044. Google Scholar 37Ganea O, Bécigneul G, Hofmann T. Hyperbolic neural networks. Adv Neural Inform Process Syst. 2018; 31:5350-5360. Google Scholar 38Liu Q, Nickel M, Kiela D. Hyperbolic graph neural networks. Adv Neural Inform Process Syst. 2019; 32:8228-8239. PubMedGoogle Scholar 39Gulcehre C, Denil M, Malinowski M, et al. Hyperbolic attention networks. arXiv:1805.09786 2018. Google Scholar 40Chen B, Huang X, Xiao L, Cai Z, Jing L. Hyperbolic interaction model for hierarchical multi-label classification. Proceedings of the AAAI Conference on Artificial Intelligence. AAAI Press; 2020: 7496-7503. Google Scholar 41Vaswani A, Shazeer N, Parmar N, et al. Attention is all you need. Adv Neural Inform Process Syst. 2017; 30:5998-6008. Google Scholar 42Cannon JW, Floyd WJ, Kenyon R, Parry WR. Hyperbolic geometry. Flavors Geomet. 1997; 31(59–115): 2. Google Scholar 43Chen B, Fu Y, Xu G, et al. Probing BERT in hyperbolic spaces. 2020. Google Scholar 44Zeb A, Haq AU, Chen J, Lei Z, Zhang D. Learning hyperbolic attention-based embeddings for link prediction in knowledge graphs. Knowl-Based Syst. 2021; 229:107369. 10.1016/j.knosys.2021.107369 Web of Science®Google Scholar 45Hinton G, Vinyals O, Dean J. Distilling the knowledge in a neural network. arXiv:1503.02531 2015. Google Scholar 46Bonnabel S. Stochastic Gradient Descent on Riemannian Manifolds. IEEE Trans Autom Contr. 2013; 58(9): 2217-2229. 10.1109/TAC.2013.2254619 Web of Science®Google Scholar 47Kochurov M, Karimov R, Kozlukov S. Geoopt: Riemannian optimization in pytorch. arXiv:2005.02819 2020. Google Scholar 48Loshchilov I, Hutter F. Decoupled weight decay regularization. arXiv:1711.05101 2017. Google Scholar 49Baker CF, Fillmore CJ, Lowe JB. The berkeley framenet project. COLING 1998 Volume 1: The 17th International Conference on Computational Linguistics. Association for Computational Linguistics; 1998. Google Scholar 50Pennington J, Socher R, Manning CD. Glove: global vectors for word representation. Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP). Association for Computational Linguistics; 2014: 1532-1543. 10.3115/v1/D14-1162 Google Scholar 51Sheng J, Sun R, Guo S, et al. CorED: incorporating type-level and instance-level correlations for fine-grained event detection. Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval. Association for Computing Machinery; 2022: 1122-1132. 10.1145/3477495.3531956 Google Scholar 52Schmidhuber J, Hochreiter S, et al. Long short-term memory. Neural Comput. 1997; 9(8): 1735-1780. 10.1162/neco.1997.9.8.1735 PubMedWeb of Science®Google Scholar 53Veličković P, Cucurull G, Casanova A, Romero A, Liò P, Bengio Y. Graph attention networks. International Conference on Learning Representations; 2018. Google Scholar 54Ma J, Ballesteros M, Doss S, et al. Label semantics for few shot named entity recognition. Findings of the Association for Computational Linguistics (ACL). Vol 2022. Association for Computational Linguistics; 2022: 1956-1971. 10.18653/v1/2022.findings-acl.155 Google Scholar 55Zhao K, Liu J, Xu Z, et al. Graph4Web: a relation-aware graph attention network for web service classification. J Syst Softw. 2022; 190:111324. 10.1016/j.jss.2022.111324 Web of Science®Google Scholar 56Yu J, Zhao K, Liu J, Liu X, Xu Z, Wang X. Exploiting gated graph neural network for detecting and explaining self-admitted technical debts. J Syst Softw. 2022; 187:111219. 10.1016/j.jss.2022.111219 Web of Science®Google Scholar 57Xie Z, Zhu R, Zhao K, Liu J, Zhou G, Huang JX. Dual gated graph attention networks with dynamic iterative training for cross-lingual entity alignment. ACM Trans Inform Syst. 2021; 40(3): 1-30. Web of Science®Google Scholar 58Schölkopf B, Smola A, Müller KR. Nonlinear component analysis as a kernel eigenvalue problem. Neural Comput. 1998; 10(5): 1299-1319. 10.1162/089976698300017467 Web of Science®Google Scholar 59Choi E, Levy O, Choi Y, Zettlemoyer L. Ultra-fine entity typing. Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics; 2018: 87-96. Google Scholar Volume40, Issue1February 2024e12627 ReferencesRelatedInformation
最长约 10秒,即可获得该文献文件

科研通智能强力驱动
Strongly Powered by AbleSci AI
更新
PDF的下载单位、IP信息已删除 (2025-6-4)

科研通是完全免费的文献互助平台,具备全网最快的应助速度,最高的求助完成率。 对每一个文献求助,科研通都将尽心尽力,给求助人一个满意的交代。
实时播报
西南楚留香完成签到,获得积分10
3秒前
Newky发布了新的文献求助10
4秒前
6秒前
6秒前
王路飞完成签到,获得积分10
7秒前
10秒前
mmm发布了新的文献求助10
11秒前
boltos发布了新的文献求助10
13秒前
ASA完成签到,获得积分10
13秒前
abin发布了新的文献求助10
14秒前
Newky完成签到,获得积分10
14秒前
卿君完成签到,获得积分10
18秒前
18秒前
Jasper应助乔佳怡采纳,获得10
20秒前
Akim应助hwezhu采纳,获得10
20秒前
21秒前
alpv完成签到,获得积分10
21秒前
22秒前
南冥发布了新的文献求助10
25秒前
25秒前
26秒前
迷路的绿藻头完成签到 ,获得积分10
26秒前
量子星尘发布了新的文献求助10
27秒前
哈哈完成签到,获得积分10
27秒前
A阿澍发布了新的文献求助10
28秒前
abin完成签到,获得积分10
28秒前
30秒前
30秒前
30秒前
hwezhu发布了新的文献求助10
31秒前
33秒前
jyy应助科研通管家采纳,获得10
35秒前
czh应助科研通管家采纳,获得20
35秒前
烟花应助科研通管家采纳,获得10
36秒前
充电宝应助科研通管家采纳,获得10
36秒前
小马甲应助科研通管家采纳,获得10
36秒前
36秒前
今后应助科研通管家采纳,获得10
36秒前
36秒前
无花果应助科研通管家采纳,获得10
36秒前
高分求助中
A new approach to the extrapolation of accelerated life test data 1000
ACSM’s Guidelines for Exercise Testing and Prescription, 12th edition 500
‘Unruly’ Children: Historical Fieldnotes and Learning Morality in a Taiwan Village (New Departures in Anthropology) 400
Indomethacinのヒトにおける経皮吸収 400
Phylogenetic study of the order Polydesmida (Myriapoda: Diplopoda) 370
基于可调谐半导体激光吸收光谱技术泄漏气体检测系统的研究 350
Robot-supported joining of reinforcement textiles with one-sided sewing heads 320
热门求助领域 (近24小时)
化学 材料科学 医学 生物 工程类 有机化学 生物化学 物理 内科学 纳米技术 计算机科学 化学工程 复合材料 遗传学 基因 物理化学 催化作用 冶金 细胞生物学 免疫学
热门帖子
关注 科研通微信公众号,转发送积分 3989263
求助须知:如何正确求助?哪些是违规求助? 3531418
关于积分的说明 11253814
捐赠科研通 3270066
什么是DOI,文献DOI怎么找? 1804884
邀请新用户注册赠送积分活动 882084
科研通“疑难数据库(出版商)”最低求助积分说明 809136