Adversarial domain adaptation for variational neural language generation in dialogue systems
conference contribution
posted on 2018-01-01, 00:00authored byVan Khanh Tran, Le-Minh Nguyen
Domain Adaptation arises when we aim at learning from source domain a model that can perform acceptably well on a different target domain. It is especially crucial for Natural Language
Generation (NLG) in Spoken Dialogue Systems when there are sufficient annotated data in the
source domain, but there is a limited labeled data in the target domain. How to effectively utilize
as much of existing abilities from source domains is a crucial issue in domain adaptation. In this
paper, we propose an adversarial training procedure to train a Variational encoder-decoder based
language generator via multiple adaptation steps. In this procedure, a model is first trained on
a source domain data and then fine-tuned on a small set of target domain utterances under the
guidance of two proposed critics. Experimental results show that the proposed method can effectively leverage the existing knowledge in the source domain to adapt to another related domain
by using only a small amount of in-domain data.
History
Pagination
1205-1217
Location
Santa Fe, N.M.
Start date
2018-08-20
End date
2018-08-26
ISBN-13
978-1-948087-50-6
Language
eng
Publication classification
E1.1 Full written paper - refereed
Copyright notice
2018, Association for Computational Linguistics
Editor/Contributor(s)
[Unknown]
Title of proceedings
COLING 2018 : Proceedings of the 27th International Conference on Computational Linguistics
Event
Association for Computational Linguistics. Conference (27th : 2018 : Santa Fe, N.M.)
Publisher
Association for Computational Linguistics
Place of publication
Stroudsburg, Pa.
Series
Association for Computational Linguistics Conference