What is: Model-Agnostic Meta-Learning?
Source | Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks |
Year | 2000 |
Data Source | CC BY-SA - https://paperswithcode.com |
MAML, or Model-Agnostic Meta-Learning, is a model and task-agnostic algorithm for meta-learning that trains a model’s parameters such that a small number of gradient updates will lead to fast learning on a new task.
Consider a model represented by a parametrized function with parameters . When adapting to a new task , the model’s parameters become . With MAML, the updated parameter vector is computed using one or more gradient descent updates on task . For example, when using one gradient update,
The step size may be fixed as a hyperparameter or metalearned. The model parameters are trained by optimizing for the performance of with respect to across tasks sampled from . More concretely the meta-objective is as follows:
Note that the meta-optimization is performed over the model parameters , whereas the objective is computed using the updated model parameters . In effect MAML aims to optimize the model parameters such that one or a small number of gradient steps on a new task will produce maximally effective behavior on that task. The meta-optimization across tasks is performed via stochastic gradient descent (SGD), such that the model parameters are updated as follows:
where is the meta step size.