skip to main content
short-paper

Chinese Event Extraction via Graph Attention Network

Authors Info & Claims
Published:19 January 2022Publication History
Skip Abstract Section

Abstract

Event extraction plays an important role in natural language processing (NLP) applications, including question answering and information retrieval. Most of the previous state-of-the-art methods were lack of ability in capturing features in long range. Recent methods applied dependency tree via dependency-bridge and attention-based graph. However, most of the automatic processing tools used in those methods show poor performance on Chinese texts due to mismatching between word segmentation and labels, which results in error propagation. In this article, we propose a novel character-level Chinese event extraction framework via graph attention network (CAEE). We build our model upon the sequence labeling model, but enhance it with word information by incorporating the word lexicon into the character representations. We further exploit the inter-dependencies between event triggers and argument by building a word-character-based graph network via syntactic shortcut arcs with dependency-parsing. The architecture of the graph minimizes error propagation, which is the result of the error detection of the word boundaries in the processing of Chinese texts. To demonstrate the effectiveness of our work, we build a large-scale real-world corpus consisting of announcements of Chinese financial news without golden entities. Experiments on the corpus show that our approach achieves competitive results compared with previous work in the field of Chinese texts.

REFERENCES

  1. [1] Cai Deng and Lam Wai. 2020. Graph transformer for graph-to-sequence learning. In AAAI. 74647471.Google ScholarGoogle Scholar
  2. [2] Chen Chen and Ng Vincent. 2012. Joint modeling for Chinese event extraction with rich linguistic features. In COLING. 529544.Google ScholarGoogle Scholar
  3. [3] Chen Yubo, Xu Liheng, Liu Kang, Zeng Daojian, and Zhao Jun. 2015. Event extraction via dynamic multi-pooling convolutional neural networks. In ACL. The Association for Computer Linguistics, 167176.Google ScholarGoogle Scholar
  4. [4] Cho Kyunghyun, Merriënboer Bart Van, Gulcehre Caglar, Bahdanau Dzmitry, Bougares Fethi, Schwenk Holger, and Bengio Yoshua. 2014. Learning phrase representations using RNN encoder-decoder for statistical machine translation. arXiv preprint arXiv:1406.1078 (2014).Google ScholarGoogle Scholar
  5. [5] Devlin Jacob, Chang Ming-Wei, Lee Kenton, and Toutanova Kristina. 2018. BERT: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018).Google ScholarGoogle Scholar
  6. [6] Gui Tao, Zou Yicheng, Zhang Qi, Peng Minlong, Fu Jinlan, Wei Zhongyu, and Huang Xuan-Jing. 2019. A lexicon-based graph neural network for Chinese NER. In EMNLP-IJCNLP. 10391049.Google ScholarGoogle Scholar
  7. [7] Hochreiter Sepp and Schmidhuber Jürgen. 1997. Long short-term memory. Neural Computat. 9, 8 (1997), 17351780. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. [8] Kipf Thomas N. and Welling Max. 2016. Semi-supervised classification with graph convolutional networks. arXiv preprint arXiv:1609.02907 (2016).Google ScholarGoogle Scholar
  9. [9] Li Shen, Zhao Zhe, Hu Renfen, Li Wensi, Liu Tao, and Du Xiaoyong. 2018. Analogical reasoning on Chinese morphological and semantic relations. In ACL. Association for Computational Linguistics, 138143. Retrieved from http://aclweb.org/anthology/P18-2023.Google ScholarGoogle Scholar
  10. [10] Lin Hongyu, Lu Yaojie, Han Xianpei, and Sun Le. 2018. Nugget proposal networks for Chinese event detection. arXiv preprint arXiv:1805.00249 (2018).Google ScholarGoogle Scholar
  11. [11] Liu Shulin, Chen Yubo, He Shizhu, Liu Kang, and Zhao Jun. 2016. Leveraging FrameNet to improve automatic event detection. In ACL. The Association for Computer Linguistics.Google ScholarGoogle Scholar
  12. [12] Liu Shulin, Liu Kang, He Shizhu, and Zhao Jun. 2016. A probabilistic soft logic based approach to exploiting latent and global information in event classification. In AAAI. AAAI Press, 29932999. Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. [13] Liu Xiao, Luo Zhunchen, and Huang Heyan. 2018. Jointly multiple events extraction via attention-based graph information aggregation. In EMNLP. Association for Computational Linguistics, 12471256.Google ScholarGoogle Scholar
  14. [14] Ma Ruotian, Peng Minlong, Zhang Qi, Wei Zhongyu, and Huang Xuanjing. 2020. Simplify the usage of lexicon in Chinese NER. In ACL. Association for Computational Linguistics, 59515960.Google ScholarGoogle Scholar
  15. [15] Mikolov Tomas, Chen Kai, Corrado Greg, and Dean Jeffrey. 2013. Efficient estimation of word representations in vector space. arXiv preprint arXiv:1301.3781 (2013).Google ScholarGoogle Scholar
  16. [16] Nguyen Thien Huu, Cho Kyunghyun, and Grishman Ralph. 2016. Joint event extraction via recurrent neural networks. In HLT-NAACL. The Association for Computational Linguistics, 300309.Google ScholarGoogle Scholar
  17. [17] Nguyen Thien Huu and Grishman Ralph. 2015. Event detection and domain adaptation with convolutional neural networks. In ACL/IJCNLP. Association for Computational Linguistics, Beijing, China, 365371. DOI: https://doi.org/10.3115/v1/P15-2060Google ScholarGoogle Scholar
  18. [18] Sha Lei, Qian Feng, Chang Baobao, and Sui Zhifang. 2018. Jointly extracting event triggers and arguments by dependency-bridge RNN and tensor-based argument interaction. In AAAI. AAAI Press, 59165923. Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. [19] Song Linfeng, Zhang Yue, Wang Zhiguo, and Gildea Daniel. 2018. N-ary relation extraction using graph state LSTM. arXiv preprint arXiv:1808.09101 (2018).Google ScholarGoogle Scholar
  20. [20] Sui Dianbo, Chen Yubo, Liu Kang, Zhao Jun, and Liu Shengping. 2019. Leverage lexical knowledge for Chinese named entity recognition via collaborative graph network. In EMNLP/IJCNLP. Association for Computational Linguistics, 38283838.Google ScholarGoogle Scholar
  21. [21] Veličković Petar, Cucurull Guillem, Casanova Arantxa, Romero Adriana, Lio Pietro, and Bengio Yoshua. 2017. Graph attention networks. arXiv preprint arXiv:1710.10903 (2017).Google ScholarGoogle Scholar
  22. [22] Yang Bishan and Mitchell Tom M.. 2016. Joint extraction of events and entities within a document context. In HLT-NAACL. The Association for Computational Linguistics, 289299.Google ScholarGoogle Scholar
  23. [23] Zeng Ying, Yang Honghui, Feng Yansong, Wang Zheng, and Zhao Dongyan. 2016. A convolution BiLSTM neural network model for Chinese event extraction. In Natural Language Understanding and Intelligent Applications. Springer, 275287.Google ScholarGoogle Scholar
  24. [24] Zhang Shuai, Wang Lijie, Sun Ke, and Xiao Xinyan. 2020. A Practical Chinese Dependency Parser Based on a Large-scale Dataset. arxiv:cs.CL/2009.00901.Google ScholarGoogle Scholar
  25. [25] Zhang Yue and Yang Jie. 2018. Chinese NER using lattice LSTM. In ACL. Association for Computational Linguistics, 15541564.Google ScholarGoogle Scholar
  26. [26] Zheng Shun, Cao Wei, Xu Wei, and Bian Jiang. 2019. Doc2EDAG: An end-to-end document-level framework for Chinese financial event extraction. In EMNLP/IJCNLP. Association for Computational Linguistics, 337346.Google ScholarGoogle Scholar

Index Terms

  1. Chinese Event Extraction via Graph Attention Network

      Recommendations

      Comments

      Login options

      Check if you have access through your login credentials or your institution to get full access on this article.

      Sign in

      Full Access

      • Published in

        cover image ACM Transactions on Asian and Low-Resource Language Information Processing
        ACM Transactions on Asian and Low-Resource Language Information Processing  Volume 21, Issue 4
        July 2022
        464 pages
        ISSN:2375-4699
        EISSN:2375-4702
        DOI:10.1145/3511099
        Issue’s Table of Contents

        Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

        Publisher

        Association for Computing Machinery

        New York, NY, United States

        Publication History

        • Published: 19 January 2022
        • Accepted: 1 October 2021
        • Revised: 1 September 2021
        • Received: 1 January 2021
        Published in tallip Volume 21, Issue 4

        Permissions

        Request permissions about this article.

        Request Permissions

        Check for updates

        Qualifiers

        • short-paper
        • Refereed

      PDF Format

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader

      Full Text

      View this article in Full Text.

      View Full Text

      HTML Format

      View this article in HTML Format .

      View HTML Format
      About Cookies On This Site

      We use cookies to ensure that we give you the best experience on our website.

      Learn more

      Got it!