Open Peer Review. Open Publishing. Open Access. Open Discussion. Open Directory. Open Recommendations. Open API. Open Source.
Deep Multi-task Representation Learning: A Tensor Factorisation Approach
Yongxin Yang, Timothy M. Hospedales
Nov 04, 2016 (modified: Feb 16, 2017)ICLR 2017 conference submissionreaders: everyone
Abstract:Most contemporary multi-task learning methods assume linear models. This setting is considered shallow in the era of deep learning. In this paper, we present a new deep multi-task representation learning framework that learns cross-task sharing structure at every layer in a deep network. Our approach is based on generalising the matrix factorisation techniques explicitly or implicitly used by many conventional MTL algorithms to tensor factorisation, to realise automatic learning of end-to-end knowledge sharing in deep networks. This is in contrast to existing deep learning approaches that need a user-defined multi-task sharing strategy. Our approach applies to both homogeneous and heterogeneous MTL. Experiments demonstrate the efficacy of our deep multi-task representation learning in terms of both higher accuracy and fewer design choices.
TL;DR:A multi-task representation learning framework that learns cross-task sharing structure at every layer in a deep network.
Enter your feedback below and we'll get back to you as soon as possible.