Back to List

CBAM: Convolutional Block Attention Module

Jongchan Park1 et al. — ECCV 2018

We propose Convolutional Block Attention Module (CBAM), a simple and effective attention module that can be integrated with any feed-forward convolutional neural networks. Given an intermediate feature map, our module sequentially infers attention maps along two separate dimensions, channel and spatial, then the attention maps are multiplied to the input feature map for adaptive feature refinement. Because CBAM is a lightweight and general module, it can be integrated into any CNN architecture seamlessly with negligible overheads. Our module is end-to-end trainable along with base CNNs. We validate our CBAM through extensive experiments on ImageNet-1K, MS COCO detection, and VOC 2007 detection datasets. Our experiments show consistent improvements on classification and detection performances with various models, demonstrating the wide applicability of CBAM. The code and models will be publicly available.

전체 내용 보기

Jongchan Park1, Sanghyun Woo2, Joon-Young Lee3 and In So Kweon2

1Lunit Inc., 2KAIST, 3Adobe Research

ECCV 2018

Read more