论文标题
关于跨NLU任务的多语言原型模型的跨语性可传递性
On the cross-lingual transferability of multilingual prototypical models across NLU tasks
论文作者
论文摘要
有监督的基于深度学习的方法已应用于以任务为导向的对话框,并在有足够数量的培训示例可用时对有限的域和语言应用有效。在实践中,这些方法遭受了域驱动设计和资源不足的语言的缺点。域和语言模型应该随着问题空间的发展而增长和变化。一方面,对转移学习的研究证明了基于多语言变压器模型学习语义丰富的表示的跨语性能力。另一方面,除了上述方法外,元学习还可以开发任务和语言学习算法,能够具有很大的概括。在这种情况下,本文提出了使用典型的神经网络和基于多语言变压器的模型来研究使用协同少的学习的跨语性可传递性。自然语言的实验理解多亚提斯++语料库上的任务表明,我们的方法显着改善了低资源和高资源语言之间观察到的转移学习表现。更普遍地说,我们的方法证实,可以将具有给定语言的有意义的潜在空间推广到使用元学习的情况下看不见和资源不足的潜在空间。
Supervised deep learning-based approaches have been applied to task-oriented dialog and have proven to be effective for limited domain and language applications when a sufficient number of training examples are available. In practice, these approaches suffer from the drawbacks of domain-driven design and under-resourced languages. Domain and language models are supposed to grow and change as the problem space evolves. On one hand, research on transfer learning has demonstrated the cross-lingual ability of multilingual Transformers-based models to learn semantically rich representations. On the other, in addition to the above approaches, meta-learning have enabled the development of task and language learning algorithms capable of far generalization. Through this context, this article proposes to investigate the cross-lingual transferability of using synergistically few-shot learning with prototypical neural networks and multilingual Transformers-based models. Experiments in natural language understanding tasks on MultiATIS++ corpus shows that our approach substantially improves the observed transfer learning performances between the low and the high resource languages. More generally our approach confirms that the meaningful latent space learned in a given language can be can be generalized to unseen and under-resourced ones using meta-learning.