Abstract:
Multi-scale representation ability is one of key criteria for measuring convolutional neural networks (CNNs) effectiveness. Recent studies have shown that multi-scale fea...Show MoreMetadata
Abstract:
Multi-scale representation ability is one of key criteria for measuring convolutional neural networks (CNNs) effectiveness. Recent studies have shown that multi-scale features can represent different semantic information of original images, and a combination of them would have positive influence on vision tasks. Many researchers are investigated in extract the multi-scale features in a layerwise manner and equipped with relatively inflexible receptive field. In this paper, we propose a multi-scale attention (MSA) module for CNNs, namely MSANet, where the residual block comprises hierarchical attention connections and skip connections. The MSANet improves the multi-scale representation power of the network by adaptively enriching the receptive fields of each convolutional branch. We insert the proposed MSANet block into several backbone CNN models and achieve consistent improvement over backbone models on CIFAR-100 dataset. To better verify the effectiveness of MSANet, the experimental results on major network details, i.e., scale, depth, further demonstrate the superiority of the MSANet over the Res2Net methods.
Published in: 2019 IEEE 14th International Conference on Intelligent Systems and Knowledge Engineering (ISKE)
Date of Conference: 14-16 November 2019
Date Added to IEEE Xplore: 18 August 2020
ISBN Information: