Advances and Challenges in Meta-Learning: A Technical Review (2307.04722v1)
Abstract: Meta-learning empowers learning systems with the ability to acquire knowledge from multiple tasks, enabling faster adaptation and generalization to new tasks. This review provides a comprehensive technical overview of meta-learning, emphasizing its importance in real-world applications where data may be scarce or expensive to obtain. The paper covers the state-of-the-art meta-learning approaches and explores the relationship between meta-learning and multi-task learning, transfer learning, domain adaptation and generalization, self-supervised learning, personalized federated learning, and continual learning. By highlighting the synergies between these topics and the field of meta-learning, the paper demonstrates how advancements in one area can benefit the field as a whole, while avoiding unnecessary duplication of efforts. Additionally, the paper delves into advanced meta-learning topics such as learning from complex multi-modal task distributions, unsupervised meta-learning, learning to efficiently adapt to data distribution shifts, and continual meta-learning. Lastly, the paper highlights open problems and challenges for future research in the field. By synthesizing the latest research developments, this paper provides a thorough understanding of meta-learning and its potential impact on various machine learning applications. We believe that this technical overview will contribute to the advancement of meta-learning and its practical implications in addressing real-world problems.
- “Automl to date and beyond: Challenges and opportunities” In ACM Computing Surveys (CSUR) 54.8 ACM New York, NY, 2021, pp. 1–36
- Joaquin Vanschoren “Meta-Learning” In Automated Machine Learning: Methods, Systems, Challenges Springer International Publishing, 2019, pp. 35–61
- “Meta-learning in neural networks: A survey” In IEEE transactions on pattern analysis and machine intelligence 44.9 IEEE, 2021, pp. 5149–5169
- “A perspective view and survey of meta-learning” In Artificial intelligence review 18 Springer, 2002, pp. 77–95
- Mike Huisman, Jan N Van Rijn and Aske Plaat “A survey of deep meta-learning” In Artificial Intelligence Review 54.6 Springer, 2021, pp. 4483–4541
- Oriol Vinyals “Talk: Model vs Optimization Meta Learning”, 2017 Neural Information Processing Systems (NIPS’17) URL: https://evolution.ml/pdf/vinyals.pdf
- Jake Snell, Kevin Swersky and Richard Zemel “Prototypical networks for few-shot learning” In Advances in neural information processing systems 30, 2017
- Chelsea Finn, Pieter Abbeel and Sergey Levine “Model-agnostic meta-learning for fast adaptation of deep networks” In International conference on machine learning, 2017, pp. 1126–1135 PMLR
- “Induction Networks for Few-Shot Text Classification” In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), 2019, pp. 3904–3913
- “Few-shot aspect category sentiment analysis via meta-learning” In ACM Transactions on Information Systems 41.1 ACM New York, NY, 2023, pp. 1–31
- “Meta-learning for low-resource neural machine translation” In 2018 Conference on Empirical Methods in Natural Language Processing, EMNLP 2018, 2020, pp. 3622–3631 Association for Computational Linguistics
- “Personalizing dialogue agents via meta-learning” In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, 2019, pp. 5454–5459
- “Domain Adaptive Dialog Generation via Meta Learning” In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, 2019, pp. 2639–2649
- “Meta-learning for low-resource natural language generation in task-oriented dialogue systems” In Proceedings of the 28th International Joint Conference on Artificial Intelligence, 2019, pp. 3151–3157
- “One-shot visual imitation learning via meta-learning” In Conference on robot learning, 2017, pp. 357–368 PMLR
- “ProtoTransformer: A meta-learning approach to providing student feedback” In arXiv preprint arXiv:2107.14035, 2021
- “Multi-task learning as multi-objective optimization” In Advances in neural information processing systems 31, 2018
- “Gradnorm: Gradient normalization for adaptive loss balancing in deep multitask networks” In International conference on machine learning, 2018, pp. 794–803 PMLR
- Alex Kendall, Yarin Gal and Roberto Cipolla “Multi-task learning using uncertainty to weigh losses for scene geometry and semantics” In Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 7482–7491
- “Cross-stitch networks for multi-task learning” In Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 3994–4003
- “Latent multi-task architecture learning” In Proceedings of the AAAI Conference on Artificial Intelligence 33.01, 2019, pp. 4822–4829
- “Nddr-cnn: Layerwise feature fusing in multi-task cnns by neural discriminative dimensionality reduction” In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2019, pp. 3205–3214
- “Feature-wise transformations” https://distill.pub/2018/feature-wise-transformations In Distill, 2018 DOI: 10.23915/distill.00011
- Shikun Liu, Edward Johns and Andrew J Davison “End-to-end multi-task learning with attention” In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2019, pp. 1871–1880
- “Learning multiple tasks with multilinear relationship networks” In Advances in neural information processing systems 30, 2017
- “Perceiver IO: A General Architecture for Structured Inputs & Outputs” In International Conference on Learning Representations
- “Efficiently identifying task groupings for multi-task learning” In Advances in Neural Information Processing Systems 34, 2021, pp. 27503–27516
- “A survey on multi-task learning” In IEEE Transactions on Knowledge and Data Engineering 34.12 IEEE, 2021, pp. 5586–5609
- Michael Crawshaw “Multi-task learning with deep neural networks: A survey” In arXiv preprint arXiv:2009.09796, 2020
- Minyoung Huh, Pulkit Agrawal and Alexei A Efros “What makes ImageNet good for transfer learning?” In arXiv preprint arXiv:1608.08614, 2016
- “Bert: Pre-training of deep bidirectional transformers for language understanding” In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2019, Minneapolis, MN, USA, June 2-7, 2019, Volume 1 (Long and Short Papers) Association for Computational Linguistics, 2019, pp. 4171–4186
- “Palm: Scaling language modeling with pathways” In arXiv preprint arXiv:2204.02311, 2022
- “Llama: Open and efficient foundation language models” In arXiv preprint arXiv:2302.13971, 2023
- OpenAI “GPT-4 Technical Report” In ArXiv abs/2303.08774, 2023
- “Fine-Tuning can Distort Pretrained Features and Underperform Out-of-Distribution” In International Conference on Learning Representations
- “Surgical Fine-Tuning Improves Adaptation to Distribution Shifts” In NeurIPS 2022 Workshop on Distribution Shifts: Connecting Methods and Applications
- Jason Phang, Thibault Févry and Samuel R Bowman “Sentence encoders on stilts: Supplementary training on intermediate labeled-data tasks” In arXiv preprint arXiv:1811.01088, 2018
- “Universal Language Model Fine-tuning for Text Classification” In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2018, pp. 328–339
- “Meta-learning with memory-augmented neural networks” In International conference on machine learning, 2016, pp. 1842–1850 PMLR
- “A simple neural attentive meta-learner” In International Conference on Learning Representations, 2018
- “Meta networks” In International conference on machine learning, 2017, pp. 2554–2563 PMLR
- “Conditional neural processes” In International conference on machine learning, 2018, pp. 1704–1713 PMLR
- “Optimization as a model for few-shot learning” In International conference on learning representations, 2017
- “Learning to learn by gradient descent by gradient descent” In Advances in neural information processing systems 29, 2016
- “Learning to Optimize” In International Conference on Learning Representations, 2017
- “Learned optimizers that scale and generalize” In International Conference on Machine Learning, 2017, pp. 3751–3760 PMLR
- “Meta architecture search” In Advances in Neural Information Processing Systems 32, 2019
- “Towards fast adaptation of neural architectures with meta learning” In International Conference on Learning Representations, 2019
- “Bilevel programming for hyperparameter optimization and meta-learning” In International conference on machine learning, 2018, pp. 1568–1577 PMLR
- “Meta-learning and universality: Deep representations and gradient descent can approximate any learning algorithm” In International Conference on Learning Representations, 2018
- “Meta-sgd: Learning to learn quickly for few-shot learning” In arXiv preprint arXiv:1707.09835, 2017
- Harkirat Singh Behl, Atılım Güneş Baydin and Philip HS Torr “Alpha maml: Adaptive model-agnostic meta-learning” In 6th ICML Workshop on Automated Machine Learning, Thirty-Sixth International Conference on Machine Learning (ICML), 2019
- Fengwei Zhou, Bin Wu and Zhenguo Li “Deep meta-learning: Learning to learn in the concept space” In arXiv preprint arXiv:1802.03596, 2018
- “Rapid learning or feature reuse? towards understanding the effectiveness of maml” In International conference on learning representations, 2023
- “Boil: Towards representation change for few-shot learning” In The International Conference on Learning Representations (ICLR), 2021
- Antreas Antoniou, Harrison Edwards and Amos Storkey “How to train your MAML” In International Conference on Learning Representations, 2018
- “Fast context adaptation via meta-learning” In International Conference on Machine Learning, 2019, pp. 7693–7702 PMLR
- Markus Hiller, Mehrtash Harandi and Tom Drummond “On Enforcing Better Conditioned Meta-Learning for Rapid Few-Shot Adaptation” In Advances in Neural Information Processing Systems, 2022
- Ian Goodfellow, Yoshua Bengio and Aaron Courville “Deep learning” MIT press, 2016
- Alex Nichol, Joshua Achiam and John Schulman “On first-order meta-learning algorithms” In arXiv preprint arXiv:1803.02999, 2018
- “Meta-learning with differentiable closed-form solvers” In International Conference on Learning Representations (ICLR), 2019, 2019
- “Meta-learning with differentiable convex optimization” In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2019, pp. 10657–10665
- “Meta-learning with implicit gradients” In Advances in neural information processing systems 32, 2019
- “Nearest neighbor pattern classification” In IEEE Transactions on Information Theory 13.1, 1967, pp. 21–27 DOI: 10.1109/TIT.1967.1053964
- “The unreasonable effectiveness of deep features as a perceptual metric” In Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 586–595
- Gregory Koch, Richard Zemel and Ruslan Salakhutdinov “Siamese neural networks for one-shot image recognition” In ICML deep learning workshop 2.1, 2015 Lille
- “Matching networks for one shot learning” In Advances in neural information processing systems 29, 2016
- “On episodes, prototypical networks, and few-shot learning” In Advances in Neural Information Processing Systems 34, 2021, pp. 24581–24592
- “Learning to compare: Relation network for few-shot learning” In Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 1199–1208
- “Few-shot learning with graph neural networks” In International Conference on Learning Representations, 2018
- “Infinite mixture prototypes for few-shot learning” In International Conference on Machine Learning, 2019, pp. 232–241 PMLR
- “Learning to learn with conditional class dependencies” In International conference on learning representations, 2019
- “Meta-learning with latent embedding optimization” In International Conference on Learning Representations, 2019
- “Meta-dataset: A dataset of datasets for learning to learn from few examples” In International Conference on Learning Representations, 2020
- “A hybrid approach with optimization-based and metric-based meta-learner for few-shot learning” In Neurocomputing 349 Elsevier, 2019, pp. 202–211
- “Rethinking few-shot image classification: a good embedding is all you need?” In Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part XIV 16, 2020, pp. 266–282 Springer
- “Meta-Baseline: Exploring Simple Meta-Learning for Few-Shot Learning” In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), 2021, pp. 9062–9071
- “A broader study of cross-domain few-shot learning” In Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part XXVII 16, 2020, pp. 124–141 Springer
- “A Closer Look at Few-shot Classification” In International Conference on Learning Representations, 2019
- “Multimodal model-agnostic meta-learning via task-aware modulation” In Advances in neural information processing systems 32, 2019
- Giulia Denevi, Massimiliano Pontil and Carlo Ciliberto “The advantage of conditional meta-learning for biased regularization and fine tuning” In Advances in Neural Information Processing Systems 33, 2020, pp. 964–974
- “Hierarchically structured meta-learning” In International Conference on Machine Learning, 2019, pp. 7045–7054 PMLR
- Weisen Jiang, James Kwok and Yu Zhang “Subspace learning for effective meta-learning” In International Conference on Machine Learning, 2022, pp. 10177–10194 PMLR
- “Reconciling meta-learning and continual learning with online mixtures of tasks” In Advances in Neural Information Processing Systems 32, 2019
- “Task similarity aware meta learning: Theory-inspired improvement on maml” In Uncertainty in Artificial Intelligence, 2021, pp. 23–33 PMLR
- “LGM-Net: Learning to generate matching networks for few-shot learning” In International conference on machine learning, 2019, pp. 3825–3834 PMLR
- “A Universal Representation Transformer Layer for Few-Shot Image Classification” In International Conference on Learning Representations, 2020
- Wei-Hong Li, Xialei Liu and Hakan Bilen “Universal representation learning from multiple domains for few-shot classification” In Proceedings of the IEEE/CVF International Conference on Computer Vision, 2021, pp. 9526–9535
- “Fast and flexible multi-task classification using conditional neural adaptive processes” In Advances in Neural Information Processing Systems 32, 2019
- Nikita Dvornik, Cordelia Schmid and Julien Mairal “Selecting relevant features from a multi-domain representation for few-shot classification” In Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part X 16, 2020, pp. 769–786 Springer
- “Learning a universal template for few-shot dataset generalization” In International Conference on Machine Learning, 2021, pp. 10424–10433 PMLR
- Wei-Yu Lee, Jheng-Yu Wang and Yu-Chiang Frank Wang “Domain-Agnostic Meta-Learning for Cross-Domain Few-Shot Classification” In ICASSP 2022-2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2022, pp. 1715–1719 IEEE
- “Transferable Meta Learning Across Domains.” In UAI, 2018, pp. 177–187
- “Semi-supervised meta-learning for cross-domain few-shot intent classification” In Proceedings of the 1st Workshop on Meta Learning and Its Applications to Natural Language Processing, 2021, pp. 67–75
- “Task-level Self-supervision for Cross-domain Few-shot Learning” In Proceedings of the AAAI Conference on Artificial Intelligence 36.3, 2022, pp. 3215–3223
- “Self-supervised meta-learning for few-shot natural language classification tasks” In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), 2020
- “Communication-efficient learning of deep networks from decentralized data” In Artificial intelligence and statistics, 2017, pp. 1273–1282 PMLR
- “Federated optimization in heterogeneous networks” In Proceedings of Machine learning and systems 2, 2020, pp. 429–450
- Qinbin Li, Bingsheng He and Dawn Song “Model-contrastive federated learning” In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 10713–10722
- “Scaffold: Stochastic controlled averaging for federated learning” In International Conference on Machine Learning, 2020, pp. 5132–5143 PMLR
- “Federated learning of a mixture of global and local models” In arXiv preprint arXiv:2002.05516, 2020
- Canh T Dinh, Nguyen Tran and Josh Nguyen “Personalized federated learning with moreau envelopes” In Advances in Neural Information Processing Systems 33, 2020, pp. 21394–21405
- “Ditto: Fair and robust federated learning through personalization” In International Conference on Machine Learning, 2021, pp. 6357–6368 PMLR
- Jian Xu, Xinyi Tong and Huang Shao-Lun “Personalized Federated Learning with Feature Alignment and Classifier Collaboration” In International conference on learning representations, 2023
- “An efficient framework for clustered federated learning” In Advances in Neural Information Processing Systems 33, 2020, pp. 19586–19597
- “Flexible clustered federated learning for client-level data distribution shift” In IEEE Transactions on Parallel and Distributed Systems 33.11 IEEE, 2021, pp. 2661–2674
- Felix Sattler, Klaus-Robert Müller and Wojciech Samek “Clustered federated learning: Model-agnostic distributed multitask optimization under privacy constraints” In IEEE transactions on neural networks and learning systems 32.8 IEEE, 2020, pp. 3710–3722
- “Personalized federated learning on non-IID data via group-based meta-learning” In ACM Transactions on Knowledge Discovery from Data 17.4 ACM New York, NY, 2023, pp. 1–20
- “Improving federated learning personalization via model agnostic meta learning” In arXiv preprint arXiv:1909.12488, 2019
- Alireza Fallah, Aryan Mokhtari and Asuman Ozdaglar “Personalized federated learning with theoretical guarantees: A model-agnostic meta-learning approach” In Advances in Neural Information Processing Systems 33, 2020, pp. 3557–3568
- “Federated meta-learning with fast convergence and efficient communication” In arXiv preprint arXiv:1802.07876, 2018
- Mikhail Khodak, Maria-Florina F Balcan and Ameet S Talwalkar “Adaptive gradient-based meta-learning methods” In Advances in Neural Information Processing Systems 32, 2019
- “A simple framework for contrastive learning of visual representations” In International conference on machine learning, 2020, pp. 1597–1607 PMLR
- “Momentum contrast for unsupervised visual representation learning” In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2020, pp. 9729–9738
- “Bootstrap your own latent-a new approach to self-supervised learning” In Advances in neural information processing systems 33, 2020, pp. 21271–21284
- Aaron van den Oord, Yazhe Li and Oriol Vinyals “Representation learning with contrastive predictive coding” In arXiv preprint arXiv:1807.03748, 2018
- Kyle Hsu, Sergey Levine and Chelsea Finn “Unsupervised learning via meta-learning” In International Conference on Learning Representations, 2019
- Jeff Donahue, Philipp Krähenbühl and Trevor Darrell “Adversarial feature learning” In International Conference on Learning Representations, 2017
- “Large scale adversarial representation learning” In Advances in neural information processing systems 32, 2019
- “Deep clustering for unsupervised learning of visual features” In Proceedings of the European conference on computer vision (ECCV), 2018, pp. 132–149
- Siavash Khodadadeh, Ladislau Boloni and Mubarak Shah “Unsupervised meta-learning for few-shot image classification” In Advances in neural information processing systems 32, 2019
- “Unsupervised meta-learning through latent-space interpolation in generative models” In International Conference on Learning Representations, 2021
- “Autoaugment: Learning augmentation strategies from data” In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2019, pp. 113–123
- “R3m: A universal visual representation for robot manipulation” In Conference on Robot Learning, 2023, pp. 892–909 PMLR
- Alex Tamkin, Mike Wu and Noah Goodman “Viewmaker networks: Learning views for unsupervised representation learning” In International Conference on Learning Representations, 2020
- “Meta-gmvae: Mixture of gaussian vae for unsupervised meta-learning” In International Conference on Learning Representations, 2021
- Deqian Kong, Bo Pang and Ying Nian Wu “Unsupervised Meta-Learning via Latent Space Energy-based Model of Symbol Vector Coupling” In Fifth Workshop on Meta-Learning at the Conference on Neural Information Processing Systems, 2021
- Diederik P Kingma and Max Welling “Auto-encoding variational bayes” In International Conference on Learning Representations, 2014
- “Energy-based models for sparse overcomplete representations” In Journal of Machine Learning Research 4.Dec, 2003, pp. 1235–1260
- “The close relationship between contrastive learning and meta-learning” In International Conference on Learning Representations, 2021
- Zhanyuan Yang, Jinghua Wang and Yingying Zhu “Few-shot classification with contrastive learning” In Computer Vision–ECCV 2022: 17th European Conference, Tel Aviv, Israel, October 23–27, 2022, Proceedings, Part XX, 2022, pp. 293–309 Springer
- “Self-Supervised Set Representation Learning for Unsupervised Meta-Learning” In International Conference on Learning Representations, 2023
- Huiwon Jang, Hankook Lee and Jinwoo Shin “Unsupervised Meta-learning via Few-shot Pseudo-supervised Contrastive Learning” In International Conference on Learning Representations, 2023
- “Language models are few-shot learners” In Advances in neural information processing systems 33, 2020, pp. 1877–1901
- Chuhan Wu, Fangzhao Wu and Yongfeng Huang “One teacher is enough? pre-trained language model distillation from multiple teachers” In Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021, 2021, pp. 4408–4413
- Trapit Bansal, Rishikesh Jha and Andrew McCallum “Learning to few-shot learn across diverse natural language classification tasks” In Proceedings of the 28th International Conference on Computational Linguistics, 2020, pp. 5108–5123
- “Deep domain confusion: Maximizing for domain invariance” In arXiv preprint arXiv:1412.3474, 2014
- “Domain-adversarial training of neural networks” In The journal of machine learning research 17.1 JMLR. org, 2016, pp. 2096–2030
- “Unpaired image-to-image translation using cycle-consistent adversarial networks” In Proceedings of the IEEE international conference on computer vision, 2017, pp. 2223–2232
- “Rl-cyclegan: Reinforcement learning aware simulation-to-real” In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 11157–11166
- “Avid: Learning multi-stage tasks via pixel-level translation of human videos” In arXiv preprint arXiv:1912.04443, 2019
- “Cycada: Cycle-consistent adversarial domain adaptation” In International conference on machine learning, 2018, pp. 1989–1998 Pmlr
- “Adversarial multiple source domain adaptation” In Advances in neural information processing systems 31, 2018
- “Unsupervised multi-target domain adaptation through knowledge distillation” In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, 2021, pp. 1339–1347
- “Blending-target domain adaptation by adversarial meta-adaptation networks” In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2019, pp. 2248–2257
- “Unsupervised multi-target domain adaptation: An information theoretic approach” In IEEE Transactions on Image Processing 29 IEEE, 2020, pp. 3993–4002
- “Adaptive risk minimization: Learning to adapt to domain shift” In Advances in Neural Information Processing Systems 34, 2021, pp. 23664–23678
- “Few-Shot Unsupervised Domain Adaptation via Meta Learning” In 2022 IEEE International Conference on Multimedia and Expo (ICME), 2022, pp. 1–6 IEEE
- “Similarity-based meta-learning network with adversarial domain adaptation for cross-domain fault identification” In Knowledge-Based Systems 217 Elsevier, 2021, pp. 106829
- Anthony Sicilia, Xingchen Zhao and Seong Jae Hwang “Domain adversarial neural networks for domain generalization: When it works and how to improve” In Machine Learning Springer, 2023, pp. 1–37
- “Deep coral: Correlation alignment for deep domain adaptation” In Computer Vision–ECCV 2016 Workshops: Amsterdam, The Netherlands, October 8-10 and 15-16, 2016, Proceedings, Part III 14, 2016, pp. 443–450 Springer
- “mixup: Beyond empirical risk minimization” In International Conference on Learning Representations, 2018
- “Manifold mixup: Better representations by interpolating hidden states” In International conference on machine learning, 2019, pp. 6438–6447 PMLR
- “Improving out-of-distribution robustness via selective augmentation” In International Conference on Machine Learning, 2022, pp. 25407–25437 PMLR
- “Domain generalization via model-agnostic learning of semantic features” In Advances in Neural Information Processing Systems 32, 2019
- “Episodic training for domain generalization” In Proceedings of the IEEE/CVF International Conference on Computer Vision, 2019, pp. 1446–1455
- “Feature-critic networks for heterogeneous domain generalization” In International Conference on Machine Learning, 2019, pp. 3915–3924 PMLR
- “Learning to generalize: Meta-learning for domain generalization” In Proceedings of the AAAI conference on artificial intelligence 32.1, 2018
- Yogesh Balaji, Swami Sankaranarayanan and Rama Chellappa “Metareg: Towards domain generalization using meta-regularization” In Advances in neural information processing systems 31, 2018
- “Open domain generalization with domain-augmented meta-learning” In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 9624–9633
- Keyu Chen, Di Zhuang and J Morris Chang “Discriminative adversarial domain generalization with meta-learning based cross-domain validation” In Neurocomputing 467 Elsevier, 2022, pp. 418–426
- Gido M Ven, Tinne Tuytelaars and Andreas S Tolias “Three types of incremental learning” In Nature Machine Intelligence 4.12 Nature Publishing Group UK London, 2022, pp. 1185–1197
- “Gradient episodic memory for continual learning” In Advances in neural information processing systems 30, 2017
- “Efficient lifelong learning with a-gem” In International Conference on Learning Representations, 2019
- “icarl: Incremental classifier and representation learning” In Proceedings of the IEEE conference on Computer Vision and Pattern Recognition, 2017, pp. 2001–2010
- Rahaf Aljundi, Marcus Rohrbach and Tinne Tuytelaars “Selfless sequential learning” In International Conference on Learning Representations, 2019
- “Overcoming catastrophic forgetting in neural networks” In Proceedings of the national academy of sciences 114.13 National Acad Sciences, 2017, pp. 3521–3526
- “Overcoming catastrophic forgetting with hard attention to the task” In International Conference on Machine Learning, 2018, pp. 4548–4557 PMLR
- “Learn to grow: A continual structure learning framework for overcoming catastrophic forgetting” In International Conference on Machine Learning, 2019, pp. 3925–3934 PMLR
- “Contextual transformation networks for online continual learning” In International Conference on Learning Representations, 2021
- “Progressive neural networks” In arXiv preprint arXiv:1606.04671, 2016
- “Learning to continually learn” In 24th European Conference on Artificial Intelligence, 2020
- “Learning to learn without forgetting by maximizing transfer and minimizing interference” In International Conference on Learning Representations, 2019
- “Meta-learning representations for continual learning” In Advances in neural information processing systems 32, 2019
- Gunshi Gupta, Karmesh Yadav and Liam Paull “Look-ahead meta learning for continual learning” In Advances in Neural Information Processing Systems 33, 2020, pp. 11588–11598
- “Few-shot learning for dermatological disease diagnosis” In Machine Learning for Healthcare Conference, 2019, pp. 532–552 PMLR
- “Cross-modal generalization: Learning in low resource modalities via meta-alignment” In Proceedings of the 29th ACM International Conference on Multimedia, 2021, pp. 2680–2689
- “Flamingo: a visual language model for few-shot learning” In Advances in Neural Information Processing Systems 35, 2022, pp. 23716–23736
- “A generalist agent” In Transactions on Machine Learning Research, 2022
- “Meta-learning for few-shot land cover classification” In Proceedings of the ieee/cvf conference on computer vision and pattern recognition workshops, 2020, pp. 200–201
- “One-shot imitation from observing humans via domain-adaptive meta-learning” In Robotics: Science and Systems XIV Robotics: ScienceSystems Foundation, 2018
- Cuong Q Nguyen, Constantine Kreatsoulas and Kim M Branson “Meta-learning gnn initializations for low-resource molecular property prediction” In 4th Lifelong Machine Learning Workshop at ICML 2020, 2020
- “Few-shot human motion prediction via meta-learning” In Proceedings of the European Conference on Computer Vision (ECCV), 2018, pp. 432–450
- “Meta-album: Multi-domain meta-dataset for few-shot image classification” In Advances in Neural Information Processing Systems 35, 2022, pp. 3232–3247
- “NEVIS’22: A Stream of 100 Tasks Sampled from 30 Years of Computer Vision Research” In arXiv preprint arXiv:2211.11747, 2022
- “Meta-world: A benchmark and evaluation for multi-task and meta reinforcement learning” In Conference on robot learning, 2020, pp. 1094–1100 PMLR
- “The visual task adaptation benchmark”, 2019
- “Taskonomy: Disentangling task transfer learning” In Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 3712–3722
- “Value: A multi-task benchmark for video-and-language understanding evaluation” In Thirty-fifth Conference on Neural Information Processing Systems Datasets and Benchmarks Track (Round 1), 2021
- “Beyond the imitation game: Quantifying and extrapolating the capabilities of language models” In Transactions on Machine Learning Research, 2023
- “Tasks, stability, architecture, and compute: Training more effective learned optimizers, and using them to train themselves” In arXiv preprint arXiv:2009.11243, 2020
- Anna Vettoruzzo (4 papers)
- Mohamed-Rafik Bouguelia (5 papers)
- Joaquin Vanschoren (68 papers)
- Thorsteinn Rögnvaldsson (16 papers)
- KC Santosh (7 papers)