What is: Shake-Shake Regularization?
Source | Shake-Shake regularization |
Year | 2000 |
Data Source | CC BY-SA - https://paperswithcode.com |
Shake-Shake Regularization aims to improve the generalization ability of multi-branch networks by replacing the standard summation of parallel branches with a stochastic affine combination. A typical pre-activation ResNet with 2 residual branches would follow this equation:
Shake-shake regularization introduces a random variable following a uniform distribution between 0 and 1 during training:
Following the same logic as for dropout, all are set to the expected value of at test time.