What is: Kernel Activation Function?
Source | Kafnets: kernel-based non-parametric activation functions for neural networks |
Year | 2000 |
Data Source | CC BY-SA - https://paperswithcode.com |
A Kernel Activation Function is a non-parametric activation function defined as a one-dimensional kernel approximator:
where:
- The dictionary of the kernel elements is fixed by sampling the -axis with a uniform step around 0.
- The user selects the kernel function (e.g., Gaussian, ReLU, Softplus) and the number of kernel elements as a hyper-parameter. A larger dictionary leads to more expressive activation functions and a larger number of trainable parameters.
- The linear coefficients are adapted independently at every neuron via standard back-propagation.
In addition, the linear coefficients can be initialized using kernel ridge regression to behave similarly to a known function in the beginning of the optimization process.