Abstract

We propose a novel multi-task learning architecture, which allows learning of task-specific feature-level attention. Our design, the Multi-Task Attention Network (MTAN), consists of a single shared network containing a global feature pool, together with a soft-attention module for each task. These modules allow for learning of task-specific features from the global features, whilst simultaneously allowing for features to be shared across different tasks. The architecture can be trained end-to-end and can be built upon any feed-forward neural network, is simple to implement, and is parameter efficient. We evaluate our approach on a variety of datasets, across both image-to-image predictions and image classification tasks. We show that our architecture is state-of-the-art in multi-task learning compared to existing methods, and is also less sensitive to various weighting schemes in the multi-task loss function. Code is available at https://github.com/lorenmt/mtan.

Keywords

Computer scienceTask (project management)WeightingFeature (linguistics)Artificial intelligenceCode (set theory)Artificial neural networkEnd-to-end principleMulti-task learningNetwork architectureArchitectureTask analysisFunction (biology)Feature extractionMachine learningEngineeringSet (abstract data type)

Affiliated Institutions

Related Publications

Publication Info

Year
2019
Type
article
Pages
1871-1880
Citations
1042
Access
Closed

External Links

Social Impact

Social media, news, blog, policy document mentions

Citation Metrics

1042
OpenAlex

Cite This

Shikun Liu, Edward Johns, Andrew J. Davison (2019). End-To-End Multi-Task Learning With Attention. , 1871-1880. https://doi.org/10.1109/cvpr.2019.00197

Identifiers

DOI
10.1109/cvpr.2019.00197