Self-Training for Compositional Neural NLG in Task-Oriented Dialogue

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

Neural approaches to natural language generation in task-oriented dialogue have typically required large amounts of annotated training data to achieve satisfactory performance, especially when generating from compositional inputs. To address this issue, we show that self-training enhanced with constrained decoding yields large gains in data efficiency on a conversational weather dataset that employs compositional meaning representations. In particular, our experiments indicate that self-training with constrained decoding can enable sequence-to-sequence models to achieve satisfactory quality using vanilla decoding with five to ten times less data than with ordinary supervised baseline; moreover, by leveraging pretrained models, data efficiency can be increased further to fifty times. We confirm the main automatic results with human evaluations and show that they extend to an enhanced, compositional version of the E2E dataset. The end result is an approach that makes it possible to achieve acceptable performance on compositional NLG tasks using hundreds rather than tens of thousands of training samples.

Original languageEnglish (US)
Title of host publicationINLG 2021 - 14th International Conference on Natural Language Generation, Proceedings
EditorsAnya Belz, Angela Fan, Ehud Reiter, Yaji Sripada
PublisherAssociation for Computational Linguistics (ACL)
Pages87-102
Number of pages16
ISBN (Electronic)9781954085510
DOIs
StatePublished - 2021
Externally publishedYes
Event14th International Conference on Natural Language Generation, INLG 2021 - Virtual, Online, United Kingdom
Duration: Sep 20 2021Sep 24 2021

Publication series

NameINLG 2021 - 14th International Conference on Natural Language Generation, Proceedings

Conference

Conference14th International Conference on Natural Language Generation, INLG 2021
Country/TerritoryUnited Kingdom
CityVirtual, Online
Period9/20/219/24/21

ASJC Scopus subject areas

  • Computer Science Applications
  • Information Systems
  • Computational Theory and Mathematics
  • Software

Fingerprint

Dive into the research topics of 'Self-Training for Compositional Neural NLG in Task-Oriented Dialogue'. Together they form a unique fingerprint.

Cite this