arXiv Analytics

Sign in

arXiv:1910.00292 [cs.LG]AbstractReferencesReviewsResources

Generalization in Generation: A closer look at Exposure Bias

Florian Schmidt

Published 2019-10-01Version 1

Exposure bias refers to the train-test discrepancy that seemingly arises when an autoregressive generative model uses only ground-truth contexts at training time but generated ones at test time. We separate the contributions of the model and the learning framework to clarify the debate on consequences and review proposed counter-measures. In this light, we argue that generalization is the underlying property to address and propose unconditional generation as its fundamental benchmark. Finally, we combine latent variable modeling with a recent formulation of exploration in reinforcement learning to obtain a rigorous handling of true and generated contexts. Results on language modeling and variational sentence auto-encoding confirm the model's generalization capability.

Related articles: Most relevant | Search more
arXiv:2302.00077 [cs.LG] (Published 2023-01-31)
Personalized Privacy Auditing and Optimization at Test Time
arXiv:1712.06646 [cs.LG] (Published 2017-12-18)
When Not to Classify: Anomaly Detection of Attacks (ADA) on DNN Classifiers at Test Time
arXiv:2501.08821 [cs.LG] (Published 2025-01-15)
A Closer Look at the Learnability of Out-of-Distribution (OOD) Detection