skip to main content
research-article

Learning Syllables Using Conv-LSTM Model for Swahili Word Representation and Part-of-speech Tagging

Authors Info & Claims
Published:26 May 2021Publication History
Skip Abstract Section

Abstract

The need to capture intra-word information in natural language processing (NLP) tasks has inspired research in learning various word representations at word, character, or morpheme levels, but little attention has been given to syllables from a syllabic alphabet. Motivated by the success of compositional models in morphological languages, we present a Convolutional-long short term memory (Conv-LSTM) model for constructing Swahili word representation vectors from syllables. The unified architecture addresses the word agglutination and polysemous nature of Swahili by extracting high-level syllable features using a convolutional neural network (CNN) and then composes quality word embeddings with a long short term memory (LSTM). The word embeddings are then validated using a syllable-aware language model (31.267) and a part-of-speech (POS) tagging task (98.78), both yielding very competitive results to the state-of-art models in their respective domains. We further validate the language model using Xhosa and Shona, which are syllabic-based languages. The novelty of the study is in its capability to construct quality word embeddings from syllables using a hybrid model that does not use max-over-pool common in CNN and then the exploitation of these embeddings in POS tagging. Therefore, the study plays a crucial role in the processing of agglutinative and syllabic-based languages by contributing quality word embeddings from syllable embeddings, a robust Conv–LSTM model that learns syllables for not only language modeling and POS tagging, but also for other downstream NLP tasks.

References

  1. A. Adedjouma Sèmiyou, John O. R. Aoga, and Mamoud A. Igue. 2012. Part-of-speech tagging of Yoruba standard, language of Niger-Congo family. Research Journal of Computer and Information Technology Sciences 1 (2013), 2.Google ScholarGoogle Scholar
  2. Alan Akbik, Duncan Blythe, and Roland Vollgraf. 2018. Contextual string embeddings for sequence labeling. In Proceedings of the 27th International conference on Computational Linguistics (COLING’18). 1638–1649.Google ScholarGoogle Scholar
  3. Assibi Apatewon Amidu. 1995. Kiswahili: People, language, literature and lingua franca. Nordic J. Afric. Stud. 4, 1 (1995), 104–123.Google ScholarGoogle Scholar
  4. H. Arvi. 1999. Swahili language manager-SALAMA. Nordic J. Afric. Stud. 8, 2 (1999), 139–157.Google ScholarGoogle Scholar
  5. Ethel O. Ashton. 1947. Kiswahili Grammar, Longmans.Google ScholarGoogle Scholar
  6. Zhenisbek Assylbekov, Rustem Takhanov, Bagdat Myrzakhmetov, and Jonathan N. Washington. 2017. Syllable-aware neural language models: A failure to beat character-aware ones. In Proceedings of the Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics, 1866–1872. DOI:https://doi.org/10.18653/v1/D17-1199Google ScholarGoogle Scholar
  7. Michael Auli, Michel Galley, Chris Quirk, and Geoffrey Zweig. 2013. Joint language and translation modeling with recurrent neural networks. In Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing. 1044--1054.Google ScholarGoogle Scholar
  8. Kaspars Balodis and Daiga Deksne. 2019. FastText-based intent detection for inflected languages. Information 10, 5 (2019), 161.Google ScholarGoogle ScholarCross RefCross Ref
  9. Yoshua Bengio, Réjean Ducharme, Pascal Vincent, and Christian Jauvin. 2003. A neural probabilistic language model. J. Mach. Learn. Res. 3, Feb. (2003), 1137–1155. Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. Piotr Bojanowski, Edouard Grave, Armand Joulin, and Tomas Mikolov. 2017. Enriching word vectors with subword information. Trans. Assoc. Comput. Ling. 5 (2017), 135–146.Google ScholarGoogle ScholarCross RefCross Ref
  11. Jan Botha and Phil Blunsom. 2014. Compositional morphology for word representations and language modelling. In Proceedings of the International Conference on Machine Learning. 1899–1907. Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. Kris Cao and Marek Rei. 2016. A joint model for word embedding and word morphology. In Proceedings of the 1st Workshop on Representation Learning for NLP. Association for Computational Linguistics, 18–26. DOI:https://doi.org/10.18653/v1/W16-1603Google ScholarGoogle ScholarCross RefCross Ref
  13. Xinxiong Chen, Lei Xu, Zhiyuan Liu, Maosong Sun, and Huanbo Luan. 2015. Joint learning of character and word embeddings. In Proceedings of the 24th International Joint Conference on Artificial Intelligence. Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. Kyunghyun Cho, Bart Van Merriënboer, Caglar Gulcehre, Dzmitry Bahdanau, Fethi Bougares, Holger Schwenk, and Yoshua Bengio. 2014. Learning phrase representations using RNN encoder-decoder for statistical machine translation. In Proceedings of the Conference on Empirical Methods in Natural Language Processing. 1724–1734.Google ScholarGoogle ScholarCross RefCross Ref
  15. Sanghyuk Choi, Taeuk Kim, Jinseok Seol, and Sang-goo Lee. 2017. A syllable-based technique for word embeddings of Korean words. In Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP’17). 36.Google ScholarGoogle ScholarCross RefCross Ref
  16. Ronan Collobert and Jason Weston. 2008. A unified architecture for natural language processing: Deep neural networks with multitask learning. In Proceedings of the 25th International Conference on Machine Learning. ACM, 160–167. Google ScholarGoogle ScholarDigital LibraryDigital Library
  17. Ronan Collobert, Jason Weston, Léon Bottou, Michael Karlen, Koray Kavukcuoglu, and Pavel Kuksa. 2011. Natural language processing (almost) from scratch. J. Mach. Learn. Res. 12, Aug. (2011), 2493–2537. Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. Ryan Cotterell and Hinrich Schütze. 2015. Morphological word-embeddings. In Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. 1287–1292.Google ScholarGoogle ScholarCross RefCross Ref
  19. Mathias Creutz and Krista Lagus. 2005. Unsupervised Morpheme Segmentation and Morphology Induction from Text Corpora Using Morfessor 1.0. Helsinki University of Technology Helsinki.Google ScholarGoogle Scholar
  20. Guy De Pauw and Gilles-Maurice De Schryver. 2008. Improving the computational morphological analysis of a Swahili corpus for lexicographic purposes. Lexikos 18, 1 (2008).Google ScholarGoogle Scholar
  21. Guy De Pauw, Gilles-Maurice de Schryver, and Janneke van de Loo. 2012. Resource-light Bantu part-of-speech tagging. In Proceedings of the Workshop on Language Technology for Normalisation of Less-Resourced Languages (SALTMIL 8-AFLAT’12). European Language Resources Association, 85–92.Google ScholarGoogle Scholar
  22. Guy De Pauw, Gilles-Maurice De Schryver, and Peter W Wagacha. 2006. Data-driven part-of-speech tagging of Kiswahili. In Proceedings of the International Conference on Text, Speech and Dialogue. Springer, 197–204. Google ScholarGoogle ScholarDigital LibraryDigital Library
  23. Guy De Pauw, Peter Waiganjo Wagacha, and Gilles-Maurice de Schryver. 2011. Towards English-Swahili machine translation. In Research Workshop of the Israel Science Foundation.Google ScholarGoogle Scholar
  24. Kamil Ud Deen Salah Ud Deen. 2002. The acquisition of Nairobi Swahili: The morphosyntax of inflectional prefixes and subjects. Ph.D. Dissertation. University of California, Los Angeles.Google ScholarGoogle Scholar
  25. Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL-HLT’19).Google ScholarGoogle Scholar
  26. Jiachen Du, Lin Gui, Yulan He, Ruifeng Xu, and Xuan Wang. 2019. Convolution-based neural attention with applications to sentiment classification. IEEE Access 7 (2019), 27983--27992.Google ScholarGoogle ScholarCross RefCross Ref
  27. Jeffrey L. Elman. 1990. Finding structure in time. Cog. Sci. 14, 2 (1990), 179–211.Google ScholarGoogle ScholarCross RefCross Ref
  28. R. Elwell. 2008. Using syllables as features in Morpheme tagging in Swahili. In Proceedings of the 5th Midwest Computational Linguistics Colloquium.Google ScholarGoogle Scholar
  29. G. David Forney. 1973. The Viterbi algorithm. Proc. IEEE 61, 3 (1973), 268–278.Google ScholarGoogle ScholarCross RefCross Ref
  30. Hadrien Gelas, Laurent Besacier, and Francois Pellegrino. 2012. Developments of Swahili resources for an automatic speech recognition system. In Proceedings of the Workshop on Spoken Language Technologies for Under-Resourced Languages, Cape-Town, South Africa (SLTU’12). Retrieved from http://hal.inria.fr/hal-00954048.Google ScholarGoogle Scholar
  31. John Goldsmith, Yu Hu, Irina Matveeva, and Colin Sprague. 2005. A Heuristic for Morpheme Discovery Based on String Edit Distance. Technical Report of Computer Science Department, University of Chicago.Google ScholarGoogle Scholar
  32. Edouard Grave, Piotr Bojanowski, Prakhar Gupta, Armand Joulin, and Tomas Mikolov. 2018. Learning word vectors for 157 languages. In Proceedings of the 11th International Conference on Language Resources and Evaluation (LREC’18). European Languages Resources Association (ELRA), 3483–3487. Retrieved from https://www.aclweb.org/anthology/L18-1550.Google ScholarGoogle Scholar
  33. Prakhar Gupta, Matteo Pagliardini, and Martin Jaggi. 2019. Better word embeddings by disentangling contextual n-gram information. In Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies.Google ScholarGoogle ScholarCross RefCross Ref
  34. Abdalraouf Hassan and Ausif Mahmood. 2018. Convolutional recurrent deep learning model for sentence classification. IEEE Access 6 (2018), 13949–13957.Google ScholarGoogle ScholarCross RefCross Ref
  35. Georg Heigold, Guenter Neumann, and Josef van Genabith. 2017. An extensive empirical evaluation of character-based morphological tagging for 14 languages. In Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics. 505–513.Google ScholarGoogle ScholarCross RefCross Ref
  36. Sepp Hochreiter and Jürgen Schmidhuber. 1997. Long short-term memory. Neural Comput. 9, 8 (1997), 1735–1780. Google ScholarGoogle ScholarDigital LibraryDigital Library
  37. Zhiheng Huang, Wei Xu, and Kai Yu. 2015. Bidirectional LSTM-CRF models for sequence tagging. arXiv preprint arXiv:1508.01991 (2015).Google ScholarGoogle Scholar
  38. Arvi Hurskainen. 1996. Disambiguation of morphological analysis in Bantu languages. In Proceedings of the 16th Conference on Computational Linguistics. Association for Computational Linguistics, 568–573. Google ScholarGoogle ScholarDigital LibraryDigital Library
  39. A. Hurskainen. 2016. Helsinki Corpus of Swahili 2.0 (HCS 2.0) Annotated Version. http://urn.fi/urn:nbn:fi:lb-201608301.Google ScholarGoogle Scholar
  40. Go Inoue, Hiroyuki Shindo, and Yuji Matsumoto. 2017. Joint prediction of morphosyntactic categories for fine-grained Arabic part-of-speech tagging exploiting tag dictionary information. In Proceedings of the 21st Conference on Computational Natural Language Learning (CoNLL’17). 421–431.Google ScholarGoogle ScholarCross RefCross Ref
  41. Ishani Joshi, Purvi Koringa, and Suman Mitra. 2019. Word embeddings in low resource Gujarati language. In Proceedings of the International Conference on Document Analysis and Recognition Workshops (ICDARW’19). IEEE, 110–115.Google ScholarGoogle ScholarCross RefCross Ref
  42. Jurgita Kapočiūtė-Dzikienė, Robertas Damaševičius, and Marcin Woźniak. 2019. Sentiment analysis of Lithuanian texts using traditional and deep learning approaches. Computers 8, 1 (2019), 4.Google ScholarGoogle ScholarCross RefCross Ref
  43. Andrej Karpathy and Li Fei-Fei. 2015. Deep visual-semantic alignments for generating image descriptions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 3128–3137.Google ScholarGoogle ScholarCross RefCross Ref
  44. Yoon Kim, Yacine Jernite, David Sontag, and Alexander M. Rush. 2016. Character-aware neural language models. In Proceedings of the 30th AAAI Conference on Artificial Intelligence. Google ScholarGoogle ScholarDigital LibraryDigital Library
  45. Alex Krizhevsky, Ilya Sutskever, and Geoffrey E. Hinton. 2012. Imagenet classification with deep convolutional neural networks. In Proceedings of the International Conference on Advances in Neural Information Processing Systems. 1097–1105. Google ScholarGoogle ScholarDigital LibraryDigital Library
  46. Siwei Lai, Liheng Xu, Kang Liu, and Jun Zhao. 2015. Recurrent convolutional neural networks for text classification. In Proceedings of the 29th AAAI Conference on Artificial Intelligence. Google ScholarGoogle ScholarDigital LibraryDigital Library
  47. Angeliki Lazaridou, Marco Marelli, Roberto Zamparelli, and Marco Baroni. 2013. Compositionally derived representations of morphologically complex words in distributional semantics. In Proceedings of the 51st Meeting of the Association for Computational Linguistics. 1517–1526.Google ScholarGoogle Scholar
  48. Yann LeCun and Yoshua Bengio. 1995. Convolutional networks for images, speech, and time series. Handb. Brain Theor Neural Netw. 3361, 10 (1995).Google ScholarGoogle Scholar
  49. Yann LeCun, Yoshua Bengio, and Geoffrey Hinton. 2015. Deep learning. Nature 521, 7553 (2015), 436.Google ScholarGoogle Scholar
  50. Wang Ling, Chris Dyer, Alan W. Black, Isabel Trancoso, Ramón Fermandez, Silvio Amir, Luís Marujo, and Tiago Luís. 2015. Finding function in form: Compositional character models for open vocabulary word representation. In Proceedings of the Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics, 1520–1530. DOI:https://doi.org/10.18653/v1/D15-1176Google ScholarGoogle ScholarCross RefCross Ref
  51. Qihe Liu, Xiaonan Hu, Mao Ye, Xianqiong Cheng, and Fan Li. 2015. Gas recognition under sensor drift by using deep learning. Int. J. Intell. Syst. 30, 8 (2015), 907–922. Google ScholarGoogle ScholarDigital LibraryDigital Library
  52. Thang Luong, Richard Socher, and Christopher Manning. 2013. Better word representations with recursive neural networks for morphology. In Proceedings of the 17th Conference on Computational Natural Language Learning. 104–113.Google ScholarGoogle Scholar
  53. Christopher D. Manning. 2011. Part-of-speech tagging from 97% to 100%: Is it time for some linguistics? In Proceedings of the International Conference on Intelligent Text Processing and Computational Linguistics. Springer, 171–189. Google ScholarGoogle ScholarDigital LibraryDigital Library
  54. Mitchell Marcus, Beatrice Santorini, and Mary Ann Marcinkiewicz. 1993. Building a large annotated corpus of English: The Penn Treebank. Comput. Ling. 19, 2 (1993), 313--330. Google ScholarGoogle ScholarDigital LibraryDigital Library
  55. Giuseppe Marra, Andrea Zugarini, Stefano Melacci, and Marco Maggini. 2018. An unsupervised character-aware neural approach to word and context representation learning. In Proceedings of the International Conference on Artificial Neural Networks. Springer, 126–136.Google ScholarGoogle ScholarCross RefCross Ref
  56. Innocent Jacodah Masengo. 2018. Cross-Linguistic Influence in Third Language Production among Kiswahili Learners. Ph.D. Dissertation. Makerere University, Uganda.Google ScholarGoogle Scholar
  57. Bryan McCann, James Bradbury, Caiming Xiong, and Richard Socher. 2017. Learned in translation: Contextualized word vectors. In Proceedings of the International Conference on Advances in Neural Information Processing Systems. 6294–6305. Google ScholarGoogle ScholarDigital LibraryDigital Library
  58. Warren S. McCulloch and Walter Pitts. 1943. A logical calculus of the ideas immanent in nervous activity. Bull. Math. Biophys. 5, 4 (1943), 115–133.Google ScholarGoogle ScholarCross RefCross Ref
  59. Fandong Meng, Zhengdong Lu, Mingxuan Wang, Hang Li, Wenbin Jiang, and Qun Liu. 2015. Encoding source language with convolutional neural network for machine translation. arXiv preprint arXiv:1503.01838 (2015).Google ScholarGoogle Scholar
  60. Tomáš Mikolov, Kai Chen, Greg Corrado, and Jeffrey Dean. 2013. Efficient estimation of word representations in vector space. In Proceedings of the 1st International Conference on Learning Representations, (ICLR’13).Google ScholarGoogle Scholar
  61. Tomáš Mikolov, Martin Karafiát, Lukáš Burget, Jan Černockỳ, and Sanjeev Khudanpur. 2010. Recurrent neural network based language model. In Proceedings of the 11th Conference of the International Speech Communication Association. Google ScholarGoogle ScholarDigital LibraryDigital Library
  62. Tomáš Mikolov, Ilya Sutskever, Kai Chen, Greg S. Corrado, and Jeff Dean. 2013. Distributed representations of words and phrases and their compositionality. In Proceedings of the International Conference on Advances in Neural Information Processing Systems. 3111–3119. Google ScholarGoogle ScholarDigital LibraryDigital Library
  63. Tomáš Mikolov, Ilya Sutskever, Anoop Deoras, Hai-Son Le, Stefan Kombrink, and Jan Cernocky. 2012. Subword Language Modeling with Neural Networks. Faculty of Information Technology, Brno University of Technology.Google ScholarGoogle Scholar
  64. Seonwoo Min, Byunghan Lee, and Sungroh Yoon. 2017. Deep learning in bioinformatics. Brief. Bioinf. 18, 5 (2017), 851–869.Google ScholarGoogle Scholar
  65. Wanjiku Ng’ang’a. 2003. Semantic analysis of Kiswahili words using the self organizing map. Nordic J. Afric. Stud. 12, 3 (2003), 407–425.Google ScholarGoogle Scholar
  66. Jeffrey Pennington, Richard Socher, and Christopher Manning. 2014. GloVe: Global vectors for word representation. In Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP’14). 1532–1543.Google ScholarGoogle ScholarCross RefCross Ref
  67. Matthew E. Peters, Mark Neumann, Mohit Iyyer, Matt Gardner, Christopher Clark, Kenton Lee, and Luke Zettlemoyer. 2018. Deep contextualized word representations. arXiv preprint arXiv:1802.05365 (2018).Google ScholarGoogle Scholar
  68. Edgar C. Polomé. 1967. Swahili Language Handbook. ERIC.Google ScholarGoogle Scholar
  69. Siyu Qiu, Qing Cui, Jiang Bian, Bin Gao, and Tie-Yan Liu. 2014. Co-learning of word representations and morpheme representations. In Proceedings of, the 25th International Conference on Computational Linguistics (COLING’14). 141–150.Google ScholarGoogle Scholar
  70. Cicero D. Santos and Bianca Zadrozny. 2014. Learning character-level representations for part-of-speech tagging. In Proceedings of the 31st International Conference on Machine Learning (ICML’14). 1818–1826. Google ScholarGoogle ScholarDigital LibraryDigital Library
  71. Mike Schuster and Kuldip K. Paliwal. 1997. Bidirectional recurrent neural networks. IEEE Trans. Sig. Proc. 45, 11 (1997), 2673–2681. Google ScholarGoogle ScholarDigital LibraryDigital Library
  72. Devashish Shankar, Sujay Narumanchi, H. A. Ananya, Pramod Kompalli, and Krishnendu Chaudhury. 2017. Deep learning based large scale visual recommendation and search for e-commerce. arXiv preprint arXiv:1703.02344 (2017).Google ScholarGoogle Scholar
  73. Yelong Shen, Xiaodong He, Jianfeng Gao, Li Deng, and Grégoire Mesnil. 2014. Learning semantic representations using convolutional neural networks for web search. In Proceedings of the 23rd International Conference on World Wide Web. ACM, 373–374. Google ScholarGoogle ScholarDigital LibraryDigital Library
  74. Casper S. Shikali, Zhou Sijie, Liu Qihe, and Refuoe Mokhosi. 2019. Better word representation vectors using syllabic alphabet: A case study of Swahili. Appl. Sci. 9, 18 (2019), 3648.Google ScholarGoogle ScholarCross RefCross Ref
  75. Yi Mon Shwe Sin and Khin Mar Soe. 2019. Attention-based syllable level neural machine translation system for Myanmar to English language pair. International Journal on Natural Language Computing (IJNLC) 8, 2 (2019).Google ScholarGoogle Scholar
  76. Anders Søgaard. 2011. Semisupervised condensed nearest neighbor for part-of-speech tagging. In Proceedings of the 49th Meeting of the Association for Computational Linguistics: Human Language Technologies. Association for Computational Linguistics, 48–52. Google ScholarGoogle ScholarDigital LibraryDigital Library
  77. Rupesh Kumar Srivastava, Klaus Greff, and Jürgen Schmidhuber. 2015. Highway networks. Proceedings of the Deep Learning Workshop International Conference on Machine Learning. 1504–1515.Google ScholarGoogle Scholar
  78. Rupesh K. Srivastava, Klaus Greff, and Jürgen Schmidhuber. 2015. Training very deep networks. In Proceedings of the International Conference on Advances in Neural Information Processing Systems. 2377–2385. Google ScholarGoogle ScholarDigital LibraryDigital Library
  79. Chi Sun, Xipeng Qiu, and Xuan-Jing Huang. 2019. VCWE: Visual character-enhanced word embeddings. In Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. 2710–2719.Google ScholarGoogle Scholar
  80. Ilya Sutskever, James Martens, and Geoffrey E. Hinton. 2011. Generating text with recurrent neural networks. In Proceedings of the 28th International Conference on Machine Learning (ICML’11). 1017–1024. Google ScholarGoogle ScholarDigital LibraryDigital Library
  81. Ahmet Üstün, Murathan Kurfalı, and Burcu Can. 2018. Characters or morphemes: How to represent words? In Proceedings of the 3rd Workshop on Representation Learning for NLP. 144–153.Google ScholarGoogle ScholarCross RefCross Ref
  82. Clara Vania and Adam Lopez. 2017. From characters to words to in between: Do we capture morphology? In Proceedings of the 55th Meeting of the Association for Computational Linguistics. Association for Computational Linguistics, 2016–2027. DOI:https://doi.org/10.18653/v1/P17-1184Google ScholarGoogle ScholarCross RefCross Ref
  83. Csaba Veres and Paul Kapustin. 2020. Enhancing subword embeddings with open N-grams. In Proceedings of the International Conference on Applications of Natural Language to Information Systems. Springer, 3–15.Google ScholarGoogle ScholarCross RefCross Ref
  84. John Wieting, Mohit Bansal, Kevin Gimpel, and Karen Livescu. 2016. Charagram: Embedding words and sentences via character n-grams. In Proceedings of the Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics, 1504–1515. DOI:https://doi.org/10.18653/v1/D16-1157Google ScholarGoogle ScholarCross RefCross Ref
  85. Huijia Wu, Jiajun Zhang, and Chengqing Zong. 2017. A dynamic window neural network for CCG supertagging. In Proceedings of the 31st AAAI Conference on Artificial Intelligence. 3337–3343. Google ScholarGoogle ScholarDigital LibraryDigital Library
  86. Yonghui Wu, Mike Schuster, Zhifeng Chen, Quoc V. Le, Mohammad Norouzi, Wolfgang Macherey, Maxim Krikun, Yuan Cao, Qin Gao, Klaus Macherey, et al. 2016. Google’s neural machine translation system: Bridging the gap between human and machine translation. arXiv preprint arXiv:1609.08144 (2016).Google ScholarGoogle Scholar
  87. Wen-tau Yih, Xiaodong He, and Christopher Meek. 2014. Semantic parsing for single-relation question answering. In Proceedings of the 52nd Meeting of the Association for Computational Linguistics. 643–648.Google ScholarGoogle Scholar
  88. Tom Young, Devamanyu Hazarika, Soujanya Poria, and Erik Cambria. 2018. Recent trends in deep learning based natural language processing. IEEE Comput. Intell. Mag. 13, 3 (2018), 55–75.Google ScholarGoogle ScholarCross RefCross Ref
  89. Seunghak Yu, Nilesh Kulkarni, Haejun Lee, and Jihie Kim. 2017. Syllable-level neural language model for agglutinative language. In Proceedings of the 1st Workshop on Subword and Character Level Models in NLP. Association for Computational Linguistics, 92–96. DOI:https://doi.org/10.18653/v1/W17-4113Google ScholarGoogle ScholarCross RefCross Ref
  90. Wojciech Zaremba, Ilya Sutskever, and Oriol Vinyals. 2014. Recurrent neural network regularization. In Proceedings of the International Conference on Learning Representations. 1–8.Google ScholarGoogle Scholar
  91. Jinman Zhao, Sidharth Mudgal, and Yingyu Liang. 2018. Generalizing word embeddings using bag of subwords. In Proceedings of the Conference on Empirical Methods in Natural Language Processing. 601–606.Google ScholarGoogle ScholarCross RefCross Ref
  92. Chunting Zhou, Chonglin Sun, Zhiyuan Liu, and Francis Lau. 2015. A C-LSTM neural network for text classification. arXiv preprint arXiv:1511.08630 (2015).Google ScholarGoogle Scholar
  93. Andrea Zugarini, Stefano Melacci, and Marco Maggini. 2019. Neural poetry: Learning to generate poems using syllables. In Proceedings of the International Conference on Artificial Neural Networks. Springer, 313–325.Google ScholarGoogle ScholarCross RefCross Ref

Index Terms

  1. Learning Syllables Using Conv-LSTM Model for Swahili Word Representation and Part-of-speech Tagging

          Recommendations

          Comments

          Login options

          Check if you have access through your login credentials or your institution to get full access on this article.

          Sign in

          Full Access

          PDF Format

          View or Download as a PDF file.

          PDF

          eReader

          View online with eReader.

          eReader

          HTML Format

          View this article in HTML Format .

          View HTML Format
          About Cookies On This Site

          We use cookies to ensure that we give you the best experience on our website.

          Learn more

          Got it!