skip to main content
note

Building a Closed-Domain Question Answering System for a Low-Resource Language

Published:10 March 2023Publication History
Skip Abstract Section

Abstract

In recent years, the Question Answering System (QAS) has been widely used to develop many systems, such as conversation systems, chatbots, and intelligent search. Depending on the amount of information or knowledge that the system processes, the system can be applied in answering the questions in an open domain or closed domain. There are many approaches to solving the QA problem, but the neural network models have yielded impressive and promising results, especially the Machine Reading Comprehension approach. In this article, we build a closed-domain QAS for a low-resource language, Vietnamese—specifically, “The Postgraduate Admission of Ho Chi Minh City University of Food Industry, Vietnam.” In addition, we have created two datasets to serve our QAS: vi-SQuAD v1.1, which is automatically translated and edited from SQuAD (Stanford University Question Answering Dataset), and HUFI-PostGrad, which is manually collected. We use two main models for the system, including the Intent Classification model and the Machine Reading Comprehension model. Experimental results initially show that our QAS gives encouraging results.

REFERENCES

  1. [1] Lee Kyungjae, Yoon Kyoungho, Park Sunghyun, and Hwang Seung-Won. 2018. Semi-supervised training data generation for multilingual question answering. In Proceedings of the Eleventh International Conference on Language Resources and Evaluation (LREC’18), Miyazaki, Japan, Calzolari Nicoletta, Choukri Khalid, Cieri Christopher, Declerck Thierry, Goggi Sara, Hasida Kôiti, Isahara Hitoshi, et al. (Eds.). European Language Resources Association. http://www.lrec-conf.org/proceedings/lrec2018/summaries/711.html.Google ScholarGoogle Scholar
  2. [2] Rajpurkar Pranav, Zhang Jian, Lopyrev Konstantin, and Liang Percy. 2016. SQuAD: 100,000+ questions for machine comprehension of text. CoRR abs/1606.05250 (2016).Google ScholarGoogle Scholar
  3. [3] Jurafsky Dan and Martin James H.. 2009. Speech and Language Processing: An Introduction to Natural Language Processing, Computational Linguistics, and Speech Recognition (2nd ed.) Prentice Hall, Pearson Education International. https://www.worldcat.org/oclc/315913020.Google ScholarGoogle Scholar
  4. [4] Gur Izzeddin, Hewlett Daniel, Lacoste Alexandre, and Jones Llion. 2017. Accurate supervised and semi-supervised machine reading for long documents. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing (EMNLP’17), Copenhagen, Denmark, Palmer Martha, Hwa Rebecca, and Riedel Sebastian (Eds.). Association for Computational Linguistics, 20112020. DOI:Google ScholarGoogle ScholarCross RefCross Ref
  5. [5] Reddy Siva, Chen Danqi, and Manning Christopher D.. 2019. CoQA: A conversational question answering challenge. Trans. Assoc. Comput. Linguistics 7 (2019), 249266. https://transacl.org/ojs/index.php/tacl/article/view/1572.Google ScholarGoogle ScholarCross RefCross Ref
  6. [6] Cui Lei, Huang Shaohan, Wei Furu, Tan Chuanqi, Duan Chaoqun, and Zhou Ming. 2017. SuperAgent: A customer service chatbot for e-commerce websites. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (ACL’17), Vancouver, Canada, System Demonstrations, Bansal Mohit and Ji Heng (Eds.). Association for Computational Linguistics, 97102. DOI:Google ScholarGoogle ScholarCross RefCross Ref
  7. [7] Pan Sinno Jialin and Yang Qiang. 2010. A survey on transfer learning. IEEE Trans. Knowl. Data Eng. 22, 10 (2010), 13451359. DOI:Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. [8] Zhao Helen Jiahe and Liu Jiamou. 2018. Finding answers from the word of God: Domain adaptation for neural networks in biblical question answering. CoRR abs/1810.12118 (2018).Google ScholarGoogle Scholar
  9. [9] Jones Karen Spärck. 2004. A statistical interpretation of term specificity and its application in retrieval. J. Documentation 60, 5 (2004), 493502. DOI:Google ScholarGoogle ScholarCross RefCross Ref
  10. [10] Mikolov Tomás, Sutskever Ilya, Chen Kai, Corrado Gregory S., and Dean Jeffrey. 2013. Distributed representations of words and phrases and their compositionality. In Advances in Neural Information Processing Systems 26: 27th Annual Conference on Neural Information Processing Systems 2013. Proceedings of a Meeting Held December 5–8, 2013, Lake Tahoe, Nevada, United States, Burges Christopher J. C., Bottou Léon, Ghahramani Zoubin, and Weinberger Kilian Q. (Eds.). Curran Associates, 31113119. https://proceedings.neurips.cc/paper/2013/hash/9aa42b31882ec039965f3c4923ce901b-Abstract.html.Google ScholarGoogle Scholar
  11. [11] Pennington Jeffrey, Socher Richard, and Manning Christopher D.. 2014. GloVe: Global vectors for word representation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing, EMNLP 2014, October 25–29, 2014, Doha, Qatar, a Meeting of SIGDAT, a Special Interest Group of the ACL, Moschitti Alessandro, Pang Bo, and Daelemans Walter (Eds.). Association for Computational Linguistics, 15321543. DOI:Google ScholarGoogle ScholarCross RefCross Ref
  12. [12] Peters Matthew E., Neumann Mark, Iyyer Mohit, Gardner Matt, Clark Christopher, Lee Kenton, and Zettlemoyer Luke. 2018. Deep contextualized word representations. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2018, New Orleans, Louisiana, USA, June 1–6, 2018, Volume 1 (Long Papers), Walker Marilyn A., Ji Heng, and Stent Amanda (Eds.). Association for Computational Linguistics, 22272237. DOI:Google ScholarGoogle ScholarCross RefCross Ref
  13. [13] Hochreiter Sepp and Schmidhuber Jürgen. 1997. Long short-term memory. Neural Comput. 9, 8 (1997), 17351780. DOI:Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. [14] LeCun Yann and Bengio Yoshua. 1998. Convolutional Networks for Images, Speech, and Time Series. MIT Press, Cambridge, MA, 255258. Google ScholarGoogle Scholar
  15. [15] Vaswani Ashish, Shazeer Noam, Parmar Niki, Uszkoreit Jakob, Jones Llion, Gomez Aidan N., Kaiser Lukasz, and Polosukhin Illia. 2017. Attention is all you need. In Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, December 4–9, 2017, Long Beach, CA, USA, Guyon Isabelle, Luxburg Ulrike von, Bengio Samy, Wallach Hanna M., Fergus Rob, Vishwanathan S. V. N., and Garnett Roman (Eds.). Curran Associates, 5998–6008. https://proceedings.neurips.cc/paper/2017/hash/3f5ee243547dee91fbd053c1c4a845aa-Abstract.html.Google ScholarGoogle Scholar
  16. [16] Harris Zellig S.. 1954. Distributional structure. WORD 10, 2–3 (1954), 146162. DOI:Google ScholarGoogle ScholarCross RefCross Ref
  17. [17] Luhn Hans Peter. 1957. A statistical approach to mechanized encoding and searching of literary information. IBM J. Res. Dev. 1, 4 (1957), 309317. DOI:Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. [18] Seo Min Joon, Kembhavi Aniruddha, Farhadi Ali, and Hajishirzi Hannaneh. 2017. Bidirectional attention flow for machine comprehension. In Proceedings of the 5th International Conference on Learning Representations (ICLR’17). https://openreview.net/forum?id=HJ0UKP9ge.Google ScholarGoogle Scholar
  19. [19] Devlin Jacob, Chang Ming-Wei, Lee Kenton, and Toutanova Kristina. 2019. BERT: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2019, Minneapolis, MN, USA, June 2–7, 2019, Volume 1 (Long and Short Papers), Burstein Jill, Doran Christy, and Solorio Thamar (Eds.). Association for Computational Linguistics, 41714186. DOI:Google ScholarGoogle ScholarCross RefCross Ref
  20. [20] Conneau Alexis, Khandelwal Kartikay, Goyal Naman, Chaudhary Vishrav, Wenzek Guillaume, Guzmán Francisco, Grave Edouard, Ott Myle, Zettlemoyer Luke, and Stoyanov Veselin. 2020. Unsupervised cross-lingual representation learning at scale. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, ACL 2020, Online, July 5–10, 2020, Jurafsky Dan, Chai Joyce, Schluter Natalie, and Tetreault Joel R. (Eds.). Association for Computational Linguistics, 84408451. DOI:Google ScholarGoogle ScholarCross RefCross Ref
  21. [21] Lewis Patrick S. H., Oguz Barlas, Rinott Ruty, Riedel Sebastian, and Schwenk Holger. 2020. MLQA: Evaluating cross-lingual extractive question answering. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, ACL 2020, Online, July 5–10, 2020, Jurafsky Dan, Chai Joyce, Schluter Natalie, and Tetreault Joel R. (Eds.). Association for Computational Linguistics, 73157330. DOI:Google ScholarGoogle ScholarCross RefCross Ref
  22. [22] Rajpurkar Pranav, Jia Robin, and Liang Percy. 2018. Know what you don’t know: Unanswerable questions for SQuAD. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics, ACL 2018, Melbourne, Australia, July 15–20, 2018, Volume 2: Short Papers, Gurevych Iryna and Miyao Yusuke (Eds.). Association for Computational Linguistics, 784789. DOI:Google ScholarGoogle ScholarCross RefCross Ref

Index Terms

  1. Building a Closed-Domain Question Answering System for a Low-Resource Language

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in

    Full Access

    • Published in

      cover image ACM Transactions on Asian and Low-Resource Language Information Processing
      ACM Transactions on Asian and Low-Resource Language Information Processing  Volume 22, Issue 3
      March 2023
      570 pages
      ISSN:2375-4699
      EISSN:2375-4702
      DOI:10.1145/3579816
      Issue’s Table of Contents

      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 10 March 2023
      • Online AM: 12 October 2022
      • Accepted: 3 October 2022
      • Revised: 2 August 2022
      • Received: 18 May 2021
      Published in tallip Volume 22, Issue 3

      Permissions

      Request permissions about this article.

      Request Permissions

      Check for updates

      Qualifiers

      • note
    • Article Metrics

      • Downloads (Last 12 months)316
      • Downloads (Last 6 weeks)35

      Other Metrics

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Full Text

    View this article in Full Text.

    View Full Text

    HTML Format

    View this article in HTML Format .

    View HTML Format
    About Cookies On This Site

    We use cookies to ensure that we give you the best experience on our website.

    Learn more

    Got it!