Latent Multi-Task Architecture Learning

Research output: Chapter in Book/Report/Conference proceedingArticle in proceedingsResearchpeer-review

Multi-task learning (MTL) allows deep neural networks to learn from related tasks by sharing parameters with other networks. In practice, however, MTL involves searching an enormous space of possible parameter sharing architectures to find (a) the layers or subspaces that benefit from sharing, (b) the appropriate amount of sharing, and (c) the appropriate relative weights of the different task losses. Recent work has addressed each of the above problems in isolation. In this work we present an approach that learns a latent multi-task architecture that jointly addresses (a)–(c). We present experiments on synthetic data and data from OntoNotes 5.0, including four different tasks and seven different domains. Our extension consistently outperforms previous approaches to learning latent architectures for multi-task problems and achieves up to 15% average error reductions over common approaches to MTL.
Original languageEnglish
Title of host publicationProceedings of 33nd AAAI Conference on Artificial Intelligence, AAAI 2019
PublisherAAAI Press
Publication date2019
Pages4822-4829
ISBN (Electronic)978-1-57735-809-1
DOIs
Publication statusPublished - 2019
Event33rd AAAI Conference on Artificial Intelligence - AAAI 2019 - Honolulu, United States
Duration: 27 Jan 20191 Feb 2019

Conference

Conference33rd AAAI Conference on Artificial Intelligence - AAAI 2019
LandUnited States
ByHonolulu
Periode27/01/201901/02/2019

Links

ID: 240627841