Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
144 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Meta-Learning: A Survey (1810.03548v1)

Published 8 Oct 2018 in cs.LG and stat.ML

Abstract: Meta-learning, or learning to learn, is the science of systematically observing how different machine learning approaches perform on a wide range of learning tasks, and then learning from this experience, or meta-data, to learn new tasks much faster than otherwise possible. Not only does this dramatically speed up and improve the design of machine learning pipelines or neural architectures, it also allows us to replace hand-engineered algorithms with novel approaches learned in a data-driven way. In this chapter, we provide an overview of the state of the art in this fascinating and continuously evolving field.

Citations (731)

Summary

  • The paper provides an extensive review of meta-learning methods that extract insights from model evaluations, task properties, and prior models to improve learning efficiency.
  • It illustrates how learning from previous tasks guides algorithm configuration and transfer learning strategies, aiding rapid adaptation in new scenarios.
  • The paper also explores applications beyond supervised learning, including reinforcement and active learning, to automate and optimize model design.

Overview

Meta-learning, also known as "learning to learn", is concerned with the paper and design of systems that can improve at their tasks over time as a result of experience. It entails observing various machine learning approaches across multiple tasks, and learning from this how to enhance the design and execution of algorithms. Meta-learning not only accelerates the creation of machine learning models but can also advance the automation of this process by replacing traditional hand-engineered techniques with data-driven solutions. The surveyed field of meta-learning has evolved rapidly and spans various approaches and applications.

Learning from Model Evaluations

A fundamental concept in meta-learning involves extracting valuable insights from the past performance of machine learning models across multiple tasks. These insights can guide the identification of generally useful configurations, transfer knowledge between similar tasks, and optimize models for new problems. This area discusses methods such as using model evaluations to provide task-independent recommendations, designing better search spaces for configurations, transferring configuration information to similar tasks, and different strategies for combining evaluations of past tasks to infer a good starting point for new tasks.

Learning from Task Properties

Another method to bolster model performance is to characterize tasks based on their meta-features, such as the number of instances or features in a dataset. These characterizations help determine tasks' similarities and guide the transfer of information between tasks. By leveraging these meta-features, we can train meta-learners that can predict or rank configurations for new tasks, thus jump-starting the optimization process. Furthermore, task properties can be exploited to synthesize entire machine learning pipelines, decide whether certain algorithm parameters are worth tuning for a specific task, and guide model selection.

Learning from Prior Models

Meta-learning also extends to leveraging prior trained models, particularly in the context of transfer learning and few-shot learning. Transfer learning approaches aim to utilize models trained on one set of tasks to bootstrap models for new, but related, tasks. Few-shot learning challenges meta-learning systems to train accurate models with very limited data, capitalizing on knowledge from similar, well-represented tasks. These approaches include optimizing initial model parameters, learning update rules, and developing model-agnostic methods for rapid adaptation to new data.

Beyond Supervised Learning

The applicability of meta-learning is not limited to supervised learning tasks; it can be applied to various domains such as reinforcement learning, active learning, density estimation, and beyond. For example, meta-reinforcement learning uses experiences from related tasks to learn better policies, while meta-active learning determines the most informative instances to label for a new task.

Conclusion

Meta-learning presents vast opportunities for improving the efficiency of machine learning systems across a wide array of tasks. It focuses on the adaptive aspect of algorithms, striving to use the knowledge and experience gained from previous tasks to tackle new challenges more effectively. As the scope of meta-learning widens, future developments are directed towards increasingly efficient and sophisticated systems that can readily generalize to a broad set of applications.