skip to main content
research-article

A Supplementary Feature Set for Sentiment Analysis in Japanese Dialogues

Authors Info & Claims
Published:07 May 2019Publication History
Skip Abstract Section

Abstract

Recently, real-time affect-awareness has been applied in several commercial systems, such as dialogue systems and computer games. Real-time recognition of affective states, however, requires the application of costly feature extraction methods and/or labor-intensive annotation of large datasets, especially in the case of Asian languages where large annotated datasets are seldom available. To improve recognition accuracy, we propose the use of cognitive context in the form of “emotion-sensitive” intentions. Intentions are often represented through dialogue acts and, as an emotion-sensitive model of dialogue acts, a tagset of interpersonal-relations-directing interpersonal acts (the IA model) is proposed. The model's adequacy is assessed using a sentiment classification task in comparison with two well-known dialogue act models, the SWBD-DAMSL and the DIT++. For the assessment, five Japanese in-game dialogues were annotated with labels of sentiments and the tags of all three dialogue act models which were used to enhance a baseline sentiment classifier system. The adequacy of the IA tagset is demonstrated by a 9% improvement to the baseline sentiment classifier's recognition accuracy, outperforming the other two models by more than 5%.

References

  1. O. Abdel-Hamid, L. Deng, and D. Yu. 2013. Exploring convolutional neural network structures and optimization techniques for speech recognition. In Proceedings of 14th Annual Conference of the International Speech Communication Association (INTERSPEECH'13).Google ScholarGoogle Scholar
  2. J. Ang, R. Dhillon, A. Krupski, E. Shriberg, and A. Stolcke. 2002. Prosody-based automatic detection of annoyance and frustration in human-computer dialog. In Proceedings of the 7th International Conference on Spoken Language Processing (INTERSPEECH'02).Google ScholarGoogle Scholar
  3. Y. Arimoto and H. Mori. 2017. Emotion category mapping to emotional space by cross-corpus emotion labeling. In Proceedings of the International Conference on Situated Interaction (INTERSPEECH'17).Google ScholarGoogle Scholar
  4. A. Batliner, K. Fischer, R. Huber, J. Spilker, and E. Noth. 2003. How to find trouble in communication. Speech Communication 40, 1 117--143. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. P. Brown and S. C. Levinson. 1987. Politeness: Some Universals in Language Usage (Vol. 4), Cambridge University Press, Cambridge.Google ScholarGoogle ScholarCross RefCross Ref
  6. H. Bunt. 2009. The DIT++ taxonomy for functional dialogue markup. In Proceedings of AAMAS 2009 Work. 13--24.Google ScholarGoogle Scholar
  7. H. Bunt. 2011. Multifunctionality in dialogue. Comput. Speech Lang. 25 222--245. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. H. Bunt, J. Alexandersson, J. Choe, A. C. Fang, K. Hasida, V. Petukhova, A. Popescu-Belis, and D. Traum. 2012. ISO 24617-2: A semantically-based standard for dialogue annotation. In Proceedings of LREC 2012. 430--437.Google ScholarGoogle Scholar
  9. H. Bunt, V. Petukhova, D. Traum, and J. Alexandersson. 2017. Dialogue Act Annotation with the ISO 24617-2 Standard. Multimodal Interaction with W3C Standards: Towards Natural User Interfaces to Everything, Deborah Dahl (Ed.). Springer, Berlin, 109--135.Google ScholarGoogle Scholar
  10. J. Chung, C. Gulcehre, K. Cho, and Y. Bengio. 2014. Empirical evaluation of gated recurrent neural networks on sequence modeling. ArXiv1--9.Google ScholarGoogle Scholar
  11. D. Duncan, G. Shine, and C. English. 2016. Facial Emotion Recognition in Real Time. Report, Stanford.Google ScholarGoogle Scholar
  12. F. Eyben, M. Wolmiller, and B. Schuller. 2009. OpenEAR—Introducing the Munich open-source emotion and affect recognition toolkit. In Proceedings of the International Conference on Affective Computing and Intelligent Interaction. IEEE, 1--6.Google ScholarGoogle Scholar
  13. H. M. Fayek, M. Lech, and L. Cavedon. 2015. Towards real-time speech emotion recognition using deep neural networks. In Proceedings of the Conference on Signal Processing and Communication Systems (ICSPCS). IEEE, 1--5.Google ScholarGoogle Scholar
  14. N. H. Frijda. 1987. Emotion, cognitive structure, and action tendency. Cogn. Emot. 1 115--143.Google ScholarGoogle Scholar
  15. P. L. Ihasz, T. H. Van, and V. V. Kryssanov. 2015. A computational model for conversational Japanese. In Proceedings of 2015 International Conference on Culture and Computing. 64--71. Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. D. Jurafsky, E. Shriberg, and D. Biasca. 1997. Switchboard SWBD-DAMSL Shallow Discourse-Function Annotation (Coders Manual, Draft 13). Technical Report 97-02. University of Colorado, Institute of Cognitive Science, Colorado.Google ScholarGoogle Scholar
  17. D. P. Kingma and J. Ba. 2014. Adam: A Method for Stochastic Optimization. ArXiv Preprint ArXiv:1412.6980.Google ScholarGoogle Scholar
  18. C. M. Lee and S. S. Narayanan. 2005. Toward detecting emotions in spoken dialogs. IEEE Transactions on Speech and Audio Processing 13, 2, 293--303.Google ScholarGoogle ScholarCross RefCross Ref
  19. J. Liscombe, G. Riccardi, and D. Hakkani-Tur. 2005. Using context to improve emotion detection in spoken dialog systems. In Proceedings of the 9th European Conference on Speech Communication and Technology (INTERSPEECH'05).Google ScholarGoogle Scholar
  20. M. Mateas and A. Stern. 2005. Structuring content in the façade interactive drama architecture. In Proceedings of the 1st Artificial Intelligence and Interactive Digital Entertainment Conference. 93--98. Google ScholarGoogle ScholarDigital LibraryDigital Library
  21. Y. Matsumoto. 1988. Reexamination of the universality of face: Politeness phenomena in Japanese. Journal of Pragmatics 12, 4 (1988), 403--426.Google ScholarGoogle ScholarCross RefCross Ref
  22. M. Obaid, C. Han, and M. Billinghurst. 2008. Feed the fish: An affect-aware game. In Proceedings of the 5th Australasian Conference on Interactive Entertainment. ACM. Google ScholarGoogle ScholarDigital LibraryDigital Library
  23. D. W. Opitz and R. Maclin. 1999. Popular ensemble methods: An empirical study. Journal of Artificial Intelligence Resources 11, 169--198. Google ScholarGoogle ScholarDigital LibraryDigital Library
  24. J. Pennington, R. Socher, and C. D. Manning. 2014. GloVe: Global vectors for word representation. In Proceedings of the 2014 Conference on Empirical Methods of Natural Language Processes. 532--543.Google ScholarGoogle Scholar
  25. S. Planet and I. Iriondo. 2012. Comparison between decision-level and feature-level fusion of acoustic and linguistic features for spontaneous emotion recognition. In Proceedings of the 7th Iberian Conference on Information Systems and Technologies (CISTI). IEEE, 1--6.Google ScholarGoogle Scholar
  26. R. Plutchik. 2001. The nature of emotions. American Scientist 89, 4, 344--350.Google ScholarGoogle ScholarCross RefCross Ref
  27. A. Popescu-Belis. 2008. Dimensionality of dialogue act tagsets: An empirical analysis of large corpora. Language Res. Eval. 42, 99--107.Google ScholarGoogle ScholarCross RefCross Ref
  28. J. A. Russel. 1980. A circumplex model of affect. Journal of Personality and Social Psychology 39, 6, 1161.Google ScholarGoogle Scholar
  29. M. Szwoch and W. Szwoch. 2014. Emotion recognition for affect aware video games. Image Processing 8 Communications Challenges 6, 313, 227.Google ScholarGoogle Scholar
  30. L. Tian, J. D. Moore, and C. Lai. 2015. Emotion recognition in spontaneous and acted dialogues. In Proceedings of the 2015 International Conference on Affective Computing and Intelligent Interaction. IEEE, 698--704. Google ScholarGoogle ScholarDigital LibraryDigital Library
  31. T. Vogt, E. Andre, and N. Bee. 2008. Emovoice—A framework for online recognition of emotions from voice. In Proceedings of the 4th Tutorial and Research Workshop on Perception in Multimodal Dialogue Systems. IEEE, 188--199. Google ScholarGoogle ScholarDigital LibraryDigital Library
  32. Wikimedia Project Editors. 2017. Wikimedia database dump of the Japanese Wikipedia on July 20, 2016, https://archive.org/details/jawiki-20160720, Last accessed: 2017/09/04.Google ScholarGoogle Scholar
  33. H. Yoon, S. Park, Y. K. Lee, and J. H. Jang. 2013. Emotion recognition of serious game players using a simple brain computer interface. In Proceedings of ICT Convergence (ICTC). IEEE, 783--786.Google ScholarGoogle Scholar

Index Terms

  1. A Supplementary Feature Set for Sentiment Analysis in Japanese Dialogues

          Recommendations

          Comments

          Login options

          Check if you have access through your login credentials or your institution to get full access on this article.

          Sign in

          Full Access

          • Published in

            cover image ACM Transactions on Asian and Low-Resource Language Information Processing
            ACM Transactions on Asian and Low-Resource Language Information Processing  Volume 18, Issue 4
            December 2019
            305 pages
            ISSN:2375-4699
            EISSN:2375-4702
            DOI:10.1145/3327969
            Issue’s Table of Contents

            Copyright © 2019 ACM

            Publisher

            Association for Computing Machinery

            New York, NY, United States

            Publication History

            • Published: 7 May 2019
            • Revised: 1 January 2019
            • Accepted: 1 January 2019
            • Received: 1 February 2018
            Published in tallip Volume 18, Issue 4

            Permissions

            Request permissions about this article.

            Request Permissions

            Check for updates

            Qualifiers

            • research-article
            • Research
            • Refereed

          PDF Format

          View or Download as a PDF file.

          PDF

          eReader

          View online with eReader.

          eReader

          HTML Format

          View this article in HTML Format .

          View HTML Format
          About Cookies On This Site

          We use cookies to ensure that we give you the best experience on our website.

          Learn more

          Got it!