Non-Parallel Text Style Transfer with Self-Parallel Supervision

Ruibo Liu · CHONGYANG GAO · Chenyan Jia · Guangxuan Xu · Soroush Vosoughi

Keywords: [ language models ] [ style transfer ] [ imitation learning ]

[ Abstract ]
[ Visit Poster at Spot C0 in Virtual World ] [ OpenReview
Mon 25 Apr 2:30 a.m. PDT — 4:30 a.m. PDT


The performance of existing text style transfer models is severely limited by the non-parallel datasets on which the models are trained. In non-parallel datasets, no direct mapping exists between sentences of the source and target style; the style transfer models thus only receive weak supervision of the target sentences during training, which often leads the model to discard too much style-independent information, or utterly fail to transfer the style.In this work, we propose LaMer, a novel text style transfer framework based on large-scale language models. LaMer first mines the roughly parallel expressions in the non-parallel datasets with scene graphs, and then employs MLE training, followed by imitation learning refinement, to leverage the intrinsic parallelism within the data. On two benchmark tasks (sentiment & formality transfer) and a newly proposed challenging task (political stance transfer), our model achieves qualitative advances in transfer accuracy, content preservation, and fluency. Further empirical and human evaluations demonstrate that our model not only makes training more efficient, but also generates more readable and diverse expressions than previous models.

Chat is not available.