TY - GEN
T1 - A Novel Distributed Reinforcement Learning Method for Classical Chinese Poetry Generation
AU - Ma, Liangliang
AU - Shen, Hong
AU - Liang, Shangsong
N1 - Publisher Copyright:
© 2021, Springer Nature Switzerland AG.
PY - 2021
Y1 - 2021
N2 - Poetry generation has been a classic natural language generation task recently. But so far the methods for this topic mainly imitate and reproduce the poems on the training data set, which indicates that they either have not much connotation or overfit too much like plagiarism of the existing poems. To solve this problem, unlike previous work, instead of tuning the trade-off between connotation and innovation, we propose a distributed reinforcement learning framework, which consists of two stages of training, to generate creative and meaningful poetry. At the first stage we train a model in parallel on a large poetry corpus at word level to master how poets write poems. At the second stage we train the model with a distributed architecture to learn how connotation is developed in human literary art works at sentence level and force the model to imitate itself when it composes some ‘good poems’ to further improve performance. Experiments on generating classical Chinese poetry demonstrate that the proposed model is able to achieve better performance and the high efficiency of training compared to the state-of-the-art.
AB - Poetry generation has been a classic natural language generation task recently. But so far the methods for this topic mainly imitate and reproduce the poems on the training data set, which indicates that they either have not much connotation or overfit too much like plagiarism of the existing poems. To solve this problem, unlike previous work, instead of tuning the trade-off between connotation and innovation, we propose a distributed reinforcement learning framework, which consists of two stages of training, to generate creative and meaningful poetry. At the first stage we train a model in parallel on a large poetry corpus at word level to master how poets write poems. At the second stage we train the model with a distributed architecture to learn how connotation is developed in human literary art works at sentence level and force the model to imitate itself when it composes some ‘good poems’ to further improve performance. Experiments on generating classical Chinese poetry demonstrate that the proposed model is able to achieve better performance and the high efficiency of training compared to the state-of-the-art.
KW - Distribution
KW - Natural Language Generation
KW - Reinforcement learning
UR - http://www.scopus.com/inward/record.url?scp=85104371685&partnerID=8YFLogxK
U2 - 10.1007/978-3-030-69244-5_3
DO - 10.1007/978-3-030-69244-5_3
M3 - Conference contribution
AN - SCOPUS:85104371685
SN - 9783030692438
T3 - Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
SP - 30
EP - 42
BT - Parallel and Distributed Computing, Applications and Technologies - 21st International Conference, PDCAT 2020, Proceedings
A2 - Zhang, Yong
A2 - Xu, Yicheng
A2 - Tian, Hui
PB - Springer Science and Business Media Deutschland GmbH
T2 - 21st International Conference on Parallel and Distributed Computing, Applications, and Technologies, PDCAT 2020
Y2 - 28 December 2020 through 30 December 2020
ER -