ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2010.05994
22
17

Improving Text Generation with Student-Forcing Optimal Transport

12 October 2020
Guoyin Wang
Chunyuan Li
Jianqiao Li
Hao Fu
Yuh-Chen Lin
Liqun Chen
Yizhe Zhang
Chenyang Tao
Ruiyi Zhang
Wenlin Wang
Dinghan Shen
Qian Yang
Lawrence Carin
    OT
ArXivPDFHTML
Abstract

Neural language models are often trained with maximum likelihood estimation (MLE), where the next word is generated conditioned on the ground-truth word tokens. During testing, however, the model is instead conditioned on previously generated tokens, resulting in what is termed exposure bias. To reduce this gap between training and testing, we propose using optimal transport (OT) to match the sequences generated in these two modes. An extension is further proposed to improve the OT learning, based on the structural and contextual information of the text sequences. The effectiveness of the proposed method is validated on machine translation, text summarization, and text generation tasks.

View on arXiv
Comments on this paper