December 4, 2018
One of the first steps in the utterance interpretation pipeline of many task-oriented conversational AI systems is to identify user intents and the corresponding slots. Neural sequence labeling models have achieved very high accuracy on these tasks when trained on large amounts of training data. However, collecting this data is very time-consuming and therefore it is unfeasible to collect large amounts of data for many languages. For this reason, it is desirable to make use of existing data in a high-resource language to train models in low-resource languages. In this paper, we investigate the performance of three different methods for cross-lingual transfer learning, namely (1) translating the training data, (2) using cross-lingual pre-trained embeddings, and (3) a novel method of using a multilingual machine translation encoder as contextual word representations. We find that given several hundred training examples in the the target language, the latter two methods outperform translating the training data. Further, in very low-resource settings, we find that multilingual contextual word representations give better results than using cross-lingual static embeddings. We release the new dataset and plan to release our implementation of the NLU models in the near future.
October 16, 2019
Awni Hannun, Adrien Dufraux, Matthijs Douze, Armelle Brun, Emmanuel Vincent
October 16, 2019
July 27, 2019
Pierre-Emmanuel Mazaré, Antoine Bordes, Jason Weston, Braden Hancock
July 27, 2019
June 03, 2019
Adriana Romero Soriano, Dave Meger, Edward Smith, Scott Fujimoto
June 03, 2019
May 29, 2019
Douwe Kiela, Abi See, Jason Weston, Stephen Roller
May 29, 2019
December 04, 2018
Sebastian Schuster, Sonal Gupta, Rushin Shah, Mike Lewis
December 04, 2018
July 28, 2019
Abigail See, Stephen Roller, Douwe Kiela, Jason Weston
July 28, 2019
November 05, 2019
Shane Moon, Pararth Shah, Anuj Kumar, Rajen Subba
November 05, 2019
October 31, 2018
Prasanna Parthasarathi, Joelle Pineau
October 31, 2018
Foundational models
Latest news
Foundational models