Home > Research > Publications & Outputs > Sequence-to-Sequence Models for Data-to-Text Na...

Links

View graph of relations

Sequence-to-Sequence Models for Data-to-Text Natural Language Generation: Word- vs. Character-based Processing and Output Diversity

Research output: Contribution in Book/Report/Proceedings - With ISBN/ISSNConference contribution/Paperpeer-review

Published
Close
Publication date11/2018
Host publicationProceedings of the 11th International Natural Language Generation Conference
PublisherAssociation for Computational Linguistics
Pages221-232
Number of pages12
ISBN (print)9781948087865
<mark>Original language</mark>English
Event11th International Conference on Natural Language Generation (INLG) - Tilburg, Netherlands
Duration: 5/11/20188/11/2018

Conference

Conference11th International Conference on Natural Language Generation (INLG)
Country/TerritoryNetherlands
CityTilburg
Period5/11/188/11/18

Conference

Conference11th International Conference on Natural Language Generation (INLG)
Country/TerritoryNetherlands
CityTilburg
Period5/11/188/11/18

Abstract

We present a comparison of word-based and character-based sequence-to sequence models for data-to-text natural language generation, which generate natural language descriptions for structured inputs. On the datasets of two recent generation challenges, our models achieve comparable or better automatic evaluation results than the best challenge submissions.
Subsequent detailed statistical and human analyses shed light on the differences
between the two input representations and the diversity of the generated texts. In a controlled experiment with synthetic training data generated from templates, we demonstrate the ability of neural models to learn novel combinations of the templates and thereby generalize beyond the linguistic structures they were trained on.