What is: Filter Response Normalization?
Source | Filter Response Normalization Layer: Eliminating Batch Dependence in the Training of Deep Neural Networks |
Year | 2000 |
Data Source | CC BY-SA - https://paperswithcode.com |
Filter Response Normalization (FRN) is a type of normalization that combines normalization and an activation function, which can be used as a replacement for other normalizations and activations. It operates on each activation channel of each batch element independently, eliminating the dependency on other batch elements.
To demonstrate, assume we are dealing with the feed-forward convolutional neural network. We follow the usual convention that the filter responses (activation maps) produced after a convolution operation are a 4D tensor with shape , where is the mini-batch size, are the spatial extents of the map, and is the number of filters used in convolution. is also referred to as output channels. Let , where , be the vector of filter responses for the filter for the batch point. Let , be the mean squared norm of .
Then Filter Response Normalization is defined as the following:
where is a small positive constant to prevent division by zero.
A lack of mean centering in FRN can lead to activations having an arbitrary bias away from zero. Such a bias in conjunction with ReLU can have a detrimental effect on learning and lead to poor performance and dead units. To address this the authors augment ReLU with a learned threshold to yield:
Since , the effect of this activation is the same as having a shared bias before and after ReLU.