3.8 Proceedings Paper

latent-GLAT: Glancing at Latent Variables for Parallel Text Generation

出版社

ASSOC COMPUTATIONAL LINGUISTICS-ACL

关键词

-

资金

  1. National Science Foundation of China [U1836221, 6217020152]

向作者/读者索取更多资源

This paper introduces a method called latent-GLAT, which uses discrete latent variables to capture word categorical information and employs curriculum learning technique to alleviate the multi-modality problem. Experimental results show that this method outperforms strong baselines without the help of an autoregressive model, further broadening the application scenarios of the parallel decoding paradigm.
Recently, parallel text generation has received widespread attention due to its success in generation efficiency. Although many advanced techniques are proposed to improve its generation quality, they still need the help of an autoregressive model for training to overcome the one-to-many multi-modal phenomenon in the dataset, limiting their applications. In this paper, we propose latent-GLAT, which employs the discrete latent variables to capture word categorical information and invoke an advanced curriculum learning technique, alleviating the multi-modality problem. Experiment results show that our method outperforms strong baselines without the help of an autoregressive model, which further broadens the application scenarios of the parallel decoding paradigm. (double dagger)

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

3.8
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据