What is: Attention Gate?
Source | Attention U-Net: Learning Where to Look for the Pancreas |
Year | 2000 |
Data Source | CC BY-SA - https://paperswithcode.com |
Attention gate focuses on targeted regions while suppressing feature activations in irrelevant regions. Given the input feature map and the gating signal which is collected at a coarse scale and contains contextual information, the attention gate uses additive attention to obtain the gating coefficient. Both the input and the gating signal are first linearly mapped to an dimensional space, and then the output is squeezed in the channel domain to produce a spatial attention weight map . The overall process can be written as \begin{align} S &= \sigma(\varphi(\delta(\phi_x(X)+\phi_g(G)))) \end{align} \begin{align} Y &= S X \end{align} where , and are linear transformations implemented as convolutions.
The attention gate guides the model's attention to important regions while suppressing feature activation in unrelated areas. It substantially enhances the representational power of the model without a significant increase in computing cost or number of model parameters due to its lightweight design. It is general and modular, making it simple to use in various CNN models.