Abstract

Feature fusion, the combination of features from different layers or branches, is an omnipresent part of modern network architectures. It is often implemented via simple operations, such as summation or concatenation, but this might not be the best choice. In this work, we propose a uniform and general scheme, namely attentional feature fusion, which is applicable for most common scenarios, including feature fusion induced by short and long skip connections as well as within Inception layers. To better fuse features of inconsistent semantics and scales, we propose a multiscale channel attention module, which addresses issues that arise when fusing features given at different scales. We also demonstrate that the initial integration of feature maps can become a bottleneck and that this issue can be alleviated by adding another level of attention, which we refer to as iterative attentional feature fusion. With fewer layers or parameters, our models outperform state-of-the-art networks on both CIFAR-100 and ImageNet datasets, which suggests that more sophisticated attention mechanisms for feature fusion hold great potential to consistently yield better results compared to their direct counterparts. Our codes and trained models are available online <sup xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">1</sup> .

Keywords

Concatenation (mathematics)Feature (linguistics)Computer scienceBottleneckFuse (electrical)Artificial intelligenceSemantics (computer science)FusionScheme (mathematics)Pattern recognition (psychology)Machine learningMathematicsEngineering

Affiliated Institutions

Related Publications

Deep Layer Aggregation

Visual recognition requires rich representations that span levels from low to high, scales from small to large, and resolutions from fine to coarse. Even with the depth of featu...

2018 2018 IEEE/CVF Conference on Computer ... 1501 citations

Publication Info

Year
2021
Type
article
Pages
3559-3568
Citations
997
Access
Closed

External Links

Social Impact

Altmetric

Social media, news, blog, policy document mentions

Citation Metrics

997
OpenAlex

Cite This

Yimian Dai, Fabian Gieseke, Stefan Oehmcke et al. (2021). Attentional Feature Fusion. , 3559-3568. https://doi.org/10.1109/wacv48630.2021.00360

Identifiers

DOI
10.1109/wacv48630.2021.00360