Historical Text Normalization with Delayed Rewards

Research output: Chapter in Book/Report/Conference proceedingArticle in proceedingsResearchpeer-review

Documents

Training neural sequence-to-sequence models with simple token-level log-likelihood is now a standard approach to historical text normalization, albeit often outperformed by phrase-based models. Policy gradient training enables direct optimization for exact matches, and while the small datasets in historical text normalization are prohibitive of from-scratch reinforcement learning, we show that policy gradient fine-tuning leads to significant improvements across the board. Policy gradient training, in particular, leads to more accurate normalizations for long or unseen words
Original languageEnglish
Title of host publicationProceedings of the 57th Annual Meeting of the Association for Computational Linguistics
PublisherAssociation for Computational Linguistics
Publication date2019
Pages1614-1619
DOIs
Publication statusPublished - 2019
Event57th Annual Meeting of the Association for Computational Linguistics - Florence, Italy
Duration: 1 Jul 20191 Jul 2019

Conference

Conference57th Annual Meeting of the Association for Computational Linguistics
LandItaly
ByFlorence,
Periode01/07/201901/07/2019

Number of downloads are based on statistics from Google Scholar and www.ku.dk


No data available

ID: 239617712