What is: Collapsing Linear Unit?
Source | Deeper Learning with CoLU Activation |
Year | 2000 |
Data Source | CC BY-SA - https://paperswithcode.com |
CoLU is an activation function similar to Swish and Mish in properties. It is defined as: It is smooth, continuously differentiable, unbounded above, bounded below, non-saturating, and non-monotonic. Based on experiments done with CoLU with different activation functions, it is observed that CoLU usually performs better than other functions on deeper neural networks.