Phase 50 Planning — Meta-Learning & Learning-to-Learn Architectures #969
web3guru888
started this conversation in
Ideas
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Phase 50 — Meta-Learning & Learning-to-Learn Architectures
Overview
Phase 50 introduces meta-learning capabilities — systems that learn how to learn efficiently from limited data and rapidly adapt to new tasks. This phase implements gradient-based meta-learners (MAML, Reptile), metric-based few-shot learners (Prototypical Networks, Matching Networks), and learned optimizers that generalize across task distributions.
Motivation
Traditional deep learning assumes abundant task-specific training data and trains models from scratch or fine-tunes from pretrained weights. Meta-learning instead optimizes the learning algorithm itself, enabling:
Sub-phases
Architecture
Academic References
Finn, C., Abbeel, P., & Levine, S. (2017). "Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks." ICML 2017. — Introduced MAML, the foundational gradient-based meta-learning algorithm.
Nichol, A., Achiam, J., & Schulman, J. (2018). "On First-Order Meta-Learning Algorithms." arXiv:1803.02999. — Proposed Reptile, a simpler first-order alternative to MAML.
Snell, J., Swersky, K., & Zemel, R. (2017). "Prototypical Networks for Few-Shot Learning." NeurIPS 2017. — Metric-based meta-learning computing class prototypes in embedding space.
Vinyals, O., Blundell, C., Lillicrap, T., Kavukcuoglu, K., & Wierstra, D. (2016). "Matching Networks for One Shot Learning." NeurIPS 2016. — Attention-based few-shot learner with episodic training.
Sung, F., Yang, Y., Zhang, L., Xiang, T., Torr, P.H.S., & Hospedales, T.M. (2018). "Learning to Compare: Relation Network for Few-Shot Learning." CVPR 2018. — Learned distance metric via relation modules.
Hospedales, T., Antoniou, A., Micaelli, P., & Storkey, A. (2021). "Meta-Learning in Neural Networks: A Survey." IEEE TPAMI. — Comprehensive taxonomy of meta-learning approaches.
Andrychowicz, M., et al. (2016). "Learning to Learn by Gradient Descent by Gradient Descent." NeurIPS 2016. — LSTM-based learned optimizers.
Ravi, S. & Larochelle, H. (2017). "Optimization as a Model for Few-Shot Learning." ICLR 2017. — Meta-learner LSTM for initialization and update rules.
Lee, K., Maji, S., Ravichandran, A., & Soatto, S. (2019). "Meta-Learning with Differentiable Convex Optimization." CVPR 2019. — MetaOptNet with convex base learners.
Raghu, A., Raghu, M., Bengio, S., & Vinyals, O. (2020). "Rapid Learning or Feature Reuse?" ICLR 2020. — Analysis of MAML internals.
Success Criteria
Dependencies
Beta Was this translation helpful? Give feedback.
All reactions