Abstract
Chinese poetry generation has been a challenging part of natural language processing due to the unique literariness and aesthetics of poetry. In most cases, the content of poetry is topic related. In other words, specific thoughts or emotions are usually expressed regarding given topics. However, topic information is rarely taken into consideration in current studies about poetry generation models. In this article, we propose a topic-enhanced Chinese poetry generation model called TPoet in which the topic model is integrated into the Transformer-based auto-regressive text generation model. By feeding topic information to the input layer and heterogeneous attention mechanism, TPoet can implicitly learn the latent information of topic distribution. In addition, by setting multiple identifiers such as segment, rhyme, and tone, the model can explicitly learn the constraints of generated poems. Extensive experimental results show that the quality of TPoet-generated poems outperforms the current advanced models or systems, and the topic consistency and diversity in generated poems have been significantly improved as well.
- [1] . 2003. Latent Dirichlet allocation. Journal of Machine Learning Research 3 (2003), 993–1022.Google Scholar
Digital Library
- [2] . 2019. Sentiment-controllable Chinese poetry generation. In Proceedings of the 28th International Joint Conference on Artificial Intelligence. 4925–4931.Google Scholar
Cross Ref
- [3] . 2020. An iterative polishing framework based on quality aware masked language model for Chinese poetry generation. In Proceedings of the 24th AAAI Conference on Artificial Intelligence. 7643–7650. https://aaai.org/ojs/index.php/AAAI/article/view/6265.Google Scholar
Cross Ref
- [4] . 2019. BERT: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers). 4171–4186. Google Scholar
Cross Ref
- [5] . 2004. Chinese new rhyme (the fourteen rhymes). Chinese Poetry 5 (2004), 38–51.Google Scholar
- [6] . 2018. Hierarchical neural story generation. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). 889–898. Google Scholar
Cross Ref
- [7] . 2012. Generating Chinese classical poems with statistical machine translation models. In Proceedings of the 26th AAAI Conference on Artificial Intelligence. 1650–1656.Google Scholar
Digital Library
- [8] . 2014. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014).Google Scholar
- [9] . 2016. A diversity-promoting objective function for neural conversation models. In Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. 110–119. Google Scholar
Cross Ref
- [10] . 2018. Generating classical Chinese poems via conditional variational autoencoder and adversarial training. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing. 3890–3900.Google Scholar
Cross Ref
- [11] . 2020. Rigid formats controlled text generation. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. 742–751. Google Scholar
Cross Ref
- [12] . 2018. Improving Language Understanding by Generative Pre-Training.
Technical Report . OpenAI.Google Scholar - [13] . 2013. i, Poet: Automatic Chinese poetry composition through a generative summarization framework under constrained optimization. In Proceedings of the 22nd International Joint Conference on Artificial Intelligence. 2197–2203.Google Scholar
- [14] . 2017. Attention is all you need. In Proceedings of the 31st Conference on Neural Information Processing Systems (NIPS’17). 5998–6008. Google Scholar
Cross Ref
- [15] . 2016. Chinese poetry generation with planning based neural network. In Proceedings of the 26th International Conference on Computational Linguistics: Technical Papers. 1051–1060.Google Scholar
- [16] . 2018. Generating thematic Chinese poetry using conditional variational autoencoders with hybrid decoders. In Proceedings of the 27th International Joint Conference on Artificial Intelligence. 4539–4545.Google Scholar
Digital Library
- [17] . 2016. Generating Chinese classical poems with RNN encoder-decoder. arXiv preprint arXiv:1604.01537 (2016).Google Scholar
- [18] . 2018. Automatic poetry generation with mutual reinforcement learning. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing. 3143–3153.Google Scholar
Cross Ref
- [19] . 2017. Flexible and creative Chinese poetry generation using neural memory. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics. 1364–1373.Google Scholar
Cross Ref
- [20] . 2014. Chinese poetry generation with recurrent neural networks. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing. 670–680.Google Scholar
Cross Ref
- [21] . 2010. Genetic algorithm and its implementation of automatic generation of Chinese SONGCI. Journal of Software 21, 3 (2010), 427–437. Google Scholar
Cross Ref
Index Terms
TPoet: Topic-Enhanced Chinese Poetry Generation
Recommendations
Topic-driven reader comments summarization
CIKM '12: Proceedings of the 21st ACM international conference on Information and knowledge managementReaders of a news article often read its comments contributed by other readers. By reading comments, readers obtain not only complementary information about this news article but also the opinions from other readers. However, the existing ranking ...
Topic sentiment change analysis
MLDM'11: Proceedings of the 7th international conference on Machine learning and data mining in pattern recognitionPublic opinions on a topic may change over time. Topic Sentiment change analysis is a new research problem consisting of two main components: (a) mining opinions on a certain topic, and (b) detect significant changes of sentiment of the opinions on the ...
Research on Multi-document Summarization Based on LDA Topic Model
IHMSC '14: Proceedings of the 2014 Sixth International Conference on Intelligent Human-Machine Systems and Cybernetics - Volume 02Compared with VSM (Vector Space Model) and graph-ranking models, LDA (Latent Dirichlet Allocation) Model can discover latent topics in the corpus and latent topics are beneficial to use sentence-ranking mechanisms to form a good summary. In the paper, ...






Comments