元学习——MAML、Reptile与ANIL

作者:凯鲁嘎吉 - 博客园 http://www.cnblogs.com/kailugaji/

        之前介绍过元学习——从MAML到MAML++,这次在此基础上进一步探讨,深入了解MAML的本质,引出MAML高效学习的原因究竟是快速学习,学到一个很厉害的初始化参数,还是特征重用,初始化参数与最终结果很接近?因此得到ANIL(Almost No Inner Loop),随后我们阅读了Reptile——On first-order meta-learning algorithms,另一种元学习方法,并比较了MAML、Reptile与模型预训练之间的区别。

1. Meta Learning vs Machine Learning

2. MAML vs Model Pre-Training

3. MAML——Feature Reuse

4. MAML vs ANIL

5. Reptile: On First-Order Meta-Learning Algorithms

6. MAML, Model Pre-Training, and Reptile

7. 参考文献

[1] GitHub - Fafa-DL/Lhy_Machine_Learning: 李宏毅2021春季机器学习课程课件及作业 https://github.com/Fafa-DL/Lhy_Machine_Learning

[2] Finn, C., Abbeel, P. & Levine, S. Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks. ICML 2017.

[3] Aniruddh Raghu, Maithra Raghu, Samy Bengio, Oriol Vinyals, Rapid Learning or Feature Reuse? Towards Understanding the Effectiveness of MAML, ICLR, 2020.

[4] Nichol A, Achiam J, Schulman J. On first-order meta-learning algorithms. arXiv preprint arXiv:1803.02999, 2018.

[5] Nichol A, Schulman J. Reptile: a scalable metalearning algorithm. arXiv preprint arXiv:1803.02999, 2018, 2(3): 4.

[6] Reptile: A Scalable Meta-Learning Algorithm. OpenAI. https://openai.com/blog/reptile/

[7] 一文入门元学习(Meta-Learning)(附代码) - 知乎 https://zhuanlan.zhihu.com/p/136975128

内容来源于网络如有侵权请私信删除

文章来源: 博客园

原文链接: https://www.cnblogs.com/kailugaji/p/15156806.html

你还没有登录,请先登录注册
  • 还没有人评论,欢迎说说您的想法!