@article { , title = {Domain transfer for deep natural language generation from abstract meaning representations}, abstract = {Stochastic natural language generation systems that are trained from labelled datasets are often domainspecific in their annotation and in their mapping from semantic input representations to lexical-syntactic outputs. As a result, learnt models fail to generalize across domains, heavily restricting their usability beyond single applications. In this article, we focus on the problem of domain adaptation for natural language generation. We show how linguistic knowledge from a source domain, for which labelled data is available, can be adapted to a target domain by reusing training data across domains. As a key to this, we propose to employ abstract meaning representations as a common semantic representation across domains. We model natural language generation as a long short-term memory recurrent neural network encoderdecoder, in which one recurrent neural network learns a latent representation of a semantic input, and a second recurrent neural network learns to decode it to a sequence of words. We show that the learnt representations can be transferred across domains and can be leveraged effectively to improve training on new unseen domains. Experiments in three different domains and with six datasets demonstrate that the lexical-syntactic constructions learnt in one domain can be transferred to new domains and achieve up to 75-100\% of the performance of in-domain training. This is based on objective metrics such as BLEU and semantic error rate and a subjective human rating study. Training a policy from prior knowledge from a different domain is consistently better than pure in-domain training by up to 10\%.}, doi = {10.1109/mci.2017.2708558}, issn = {1556-603X}, issue = {3}, journal = {IEEE computational intelligence magazine}, pages = {18-28}, publicationstatus = {Published}, publisher = {Institute of Electrical and Electronics Engineers}, url = {https://hull-repository.worktribe.com/output/453764}, volume = {12}, keyword = {High Performance Computing (Hull Viper), Dependable Intelligent Systems, Specialist Research - Other, Semantics, Natural languages, Training, Stochastic processes, Pragmatics, Adaptation models, Machine learning, Natural language processing}, year = {2017}, author = {Dethlefs, Nina} }