skip to main content
research-article

Attention-Gated Graph Convolutions for Extracting Drug Interaction Information from Drug Labels

Authors Info & Claims
Published:04 January 2021Publication History
Skip Abstract Section

Abstract

Preventable adverse events as a result of medical errors present a growing concern in the healthcare system. As drug-drug interactions (DDIs) may lead to preventable adverse events, being able to extract DDIs from drug labels into a machine-processable form is an important step toward effective dissemination of drug safety information. Herein, we tackle the problem of jointly extracting mentions of drugs and their interactions, including interaction outcome, from drug labels. Our deep learning approach entails composing various intermediate representations, including graph-based context derived using graph convolutions (GCs) with a novel attention-based gating mechanism (holistically called GCA), which are combined in meaningful ways to predict on all subtasks jointly. Our model is trained and evaluated on the 2018 TAC DDI corpus. Our GCA model in conjunction with transfer learning performs at 39.20% F1 and 26.09% F1 on entity recognition (ER) and relation extraction (RE), respectively, on the first official test set and at 45.30% F1 and 27.87% F1 on ER and RE, respectively, on the second official test set. These updated results lead to improvements over our prior best by up to 6 absolute F1 points. After controlling for available training data, the proposed model exhibits state-of-the-art performance for this task.

References

  1. Daniel Andor, Chris Alberti, David Weiss, Aliaksei Severyn, Alessandro Presta, Kuzman Ganchev, Slav Petrov, and Michael Collins. 2016. Globally normalized transition-based neural networks. arXiv preprint arXiv:1603.06042 (2016).Google ScholarGoogle Scholar
  2. Masaki Asada, Makoto Miwa, and Yutaka Sasaki. 2018. Enhancing drug-drug interaction extraction from texts by molecular structure information. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers). 680--685.Google ScholarGoogle Scholar
  3. Dzmitry Bahdanau, Kyunghyun Cho, and Yoshua Bengio. 2015. Neural machine translation by jointly learning to align and translate. In Proceedings of 3th International Conference on Learning Representations (ICLR’15).Google ScholarGoogle Scholar
  4. Jason P. C. Chiu and Eric Nichols. 2016. Named entity recognition with bidirectional LSTM-CNNs. Transactions of the Association for Computational Linguistics 4 (2016), 357--370.Google ScholarGoogle ScholarCross RefCross Ref
  5. Bharath Dandala, Diwakar Mahajan, and Ananya Poddar. 2018. IBM research system at TAC 2018: Deep learning architectures for drug-drug interaction extraction from structured product labels. In Proceedings of the 2018 Text Analysis Conference (TAC’18).Google ScholarGoogle Scholar
  6. Dina Demner-Fushman, Kin Wah Fung, Phong Do, Richard D. Boyce, and Travis Goodwin. 2018. Overview of the TAC 2018 drug-drug interaction extraction from drug labels track. In Proceedings of the 2018 Text Analysis Conference (TAC’18).Google ScholarGoogle Scholar
  7. Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2018. Bert: Pre-training of deep bidirectional transformers for language understanding. Proceedings of the 17th Annual Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies.Google ScholarGoogle Scholar
  8. Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. 2016. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 770--778.Google ScholarGoogle ScholarCross RefCross Ref
  9. María Herrero-Zazo, Isabel Segura-Bedmar, Paloma Martínez, and Thierry Declerck. 2013. The DDI corpus: An annotated corpus with pharmacological substances and drug--drug interactions. Journal of Biomedical Informatics 46, 5 (2013), 914--920.Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. Rosie Jones, Andrew McCallum, Kamal Nigam, and Ellen Riloff. 1999. Bootstrapping for text learning tasks. In IJCAI-99 Workshop on Text Mining: Foundations, Techniques and Applications, Vol. 1.Google ScholarGoogle Scholar
  11. Ramakanth Kavuluru, Anthony Rios, and Tung Tran. 2017. Extracting drug-drug interactions with word and character-level recurrent neural networks. In 5th IEEE International Conference on Healthcare Informatics (ICHI’17). IEEE, 5--12.Google ScholarGoogle Scholar
  12. J.-D. Kim, Tomoko Ohta, Yuka Tateisi, and Jun’ichi Tsujii. 2003. GENIA corpus—A semantically annotated corpus for bio-textmining. Bioinformatics 19, Suppl_1 (2003), i180--i182.Google ScholarGoogle ScholarCross RefCross Ref
  13. Yoon Kim. 2014. Convolutional neural networks for sentence classification. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP’14). Association for Computational Linguistics, Doha, Qatar, 1746--1751. http://www.aclweb.org/anthology/D14-1181.Google ScholarGoogle ScholarCross RefCross Ref
  14. Linda T. Kohn, Janet M. Corrigan, and Molla S. Donaldson. 2000. To Err Is Human: Building a Safer Health System. Vol. 6. National Academies Press.Google ScholarGoogle Scholar
  15. Jinhyuk Lee, Wonjin Yoon, Sungdong Kim, Donghyeon Kim, Sunkyu Kim, Chan Ho So, and Jaewoo Kang. 2020. BioBERT: A pre-trained biomedical language representation model for biomedical text mining. Bioinformatics 36, 4 (2020), 1234--1240.Google ScholarGoogle Scholar
  16. Daniel R. Levinson. 2010. Adverse events in hospitals: National incidence among Medicare beneficiaries. Department of Health and Human Services Office of the Inspector General.Google ScholarGoogle Scholar
  17. Fei Li, Meishan Zhang, Guohong Fu, and Donghong Ji. 2017. A neural joint model for entity and relation extraction from biomedical text. BMC Bioinformatics 18, 1 (2017), 198.Google ScholarGoogle ScholarCross RefCross Ref
  18. Sangrak Lim, Kyubum Lee, and Jaewoo Kang. 2018. Drug drug interaction extraction from the literature using a recursive neural network. PloS One 13, 1 (2018), e0190926.Google ScholarGoogle ScholarCross RefCross Ref
  19. Shengyu Liu, Kai Chen, Qingcai Chen, and Buzhou Tang. 2016. Dependency-based convolutional neural network for drug-drug interaction extraction. In 2016 IEEE International Conference on Bioinformatics and Biomedicine (BIBM’16). IEEE, 1074--1080.Google ScholarGoogle Scholar
  20. Shengyu Liu, Buzhou Tang, Qingcai Chen, and Xiaolong Wang. 2016. Drug-drug interaction extraction via convolutional neural networks. Computational and Mathematical Methods in Medicine 2016 (2016), 1--8.Google ScholarGoogle ScholarCross RefCross Ref
  21. Yuan Luo, Özlem Uzuner, and Peter Szolovits. 2016. Bridging semantics and syntax with graph algorithms—State-of-the-art of extracting biomedical relations. Briefings in Bioinformatics 18, 1 (2016), 160--178.Google ScholarGoogle ScholarCross RefCross Ref
  22. Sampo Pyysalo, Filip Ginter, Hans Moen, Tapio Salakoski, and Sophia Ananiadou. 2013. Distributional semantics resources for biomedical text processing. In Proceedings of 5th International Symposium on Languages in Biology and Medicine. 39--44.Google ScholarGoogle Scholar
  23. Lev Ratinov and Dan Roth. 2009. Design challenges and misconceptions in named entity recognition. In Proceedings of the 13th Conference on Computational Natural Language Learning. Association for Computational Linguistics, 147--155.Google ScholarGoogle ScholarCross RefCross Ref
  24. Sunil Kumar Sahu and Ashish Anand. 2018. Drug-drug interaction extraction from biomedical texts using long short-term memory network. Journal of Biomedical Informatics 86 (2018), 15--24.Google ScholarGoogle ScholarCross RefCross Ref
  25. Burr Settles. 2012. Active learning. Synthesis Lectures on Artificial Intelligence and Machine Learning 6, 1 (2012), 1--114.Google ScholarGoogle ScholarDigital LibraryDigital Library
  26. Víctor Suárez-Paniagua, Isabel Segura-Bedmar, and Paloma Martínez. 2017. Exploring convolutional neural networks for drug--drug interaction extraction. Database 2017 (2017), 1--15.Google ScholarGoogle Scholar
  27. Xia Sun, Ke Dong, Long Ma, Richard Sutcliffe, Feijuan He, Sushing Chen, and Jun Feng. 2019. Drug-drug interaction extraction via recurrent hybrid convolutional neural networks with an improved focal loss. Entropy 21, 1 (2019), 37.Google ScholarGoogle Scholar
  28. Siliang Tang, Qi Zhang, Tianpeng Zheng, Mengdi Zhou, Zhan Chen, Lixing Shen, Xiang Ren, Yueting Zhuang, Shiliang Pu, and Fei Wu Wu. 2018. Two step joint model for drug drug interaction extraction. In Proceedings of the 2018 Text Analysis Conference (TAC’18).Google ScholarGoogle Scholar
  29. Tung Tran, Ramakanth Kavuluru, and Halil Kilicoglu. 2018. A multi-task learning framework for extracting drugs and their interactions from drug labels. In Proceedings of the 2018 Text Analysis Conference (TAC’18).Google ScholarGoogle Scholar
  30. Andrew Viterbi. 1967. Error bounds for convolutional codes and an asymptotically optimum decoding algorithm. IEEE Transactions on Information Theory 13, 2 (1967), 260--269.Google ScholarGoogle ScholarDigital LibraryDigital Library
  31. Yuhao Zhang, Peng Qi, and Christopher D. Manning. 2018. Graph convolution over pruned dependency trees improves relation extraction. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing.Google ScholarGoogle Scholar
  32. Zhehuan Zhao, Zhihao Yang, Ling Luo, Hongfei Lin, and Jian Wang. 2016. Drug drug interaction extraction from biomedical literature using syntax convolutional neural network. Bioinformatics 32, 22 (2016), 3444--3453.Google ScholarGoogle Scholar

Index Terms

  1. Attention-Gated Graph Convolutions for Extracting Drug Interaction Information from Drug Labels

        Recommendations

        Comments

        Login options

        Check if you have access through your login credentials or your institution to get full access on this article.

        Sign in

        Full Access

        PDF Format

        View or Download as a PDF file.

        PDF

        eReader

        View online with eReader.

        eReader

        HTML Format

        View this article in HTML Format .

        View HTML Format
        About Cookies On This Site

        We use cookies to ensure that we give you the best experience on our website.

        Learn more

        Got it!