Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
194 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
46 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Active Representation Learning for General Task Space with Applications in Robotics (2306.08942v1)

Published 15 Jun 2023 in cs.LG and cs.RO

Abstract: Representation learning based on multi-task pretraining has become a powerful approach in many domains. In particular, task-aware representation learning aims to learn an optimal representation for a specific target task by sampling data from a set of source tasks, while task-agnostic representation learning seeks to learn a universal representation for a class of tasks. In this paper, we propose a general and versatile algorithmic and theoretic framework for \textit{active representation learning}, where the learner optimally chooses which source tasks to sample from. This framework, along with a tractable meta algorithm, allows most arbitrary target and source task spaces (from discrete to continuous), covers both task-aware and task-agnostic settings, and is compatible with deep representation learning practices. We provide several instantiations under this framework, from bilinear and feature-based nonlinear to general nonlinear cases. In the bilinear case, by leveraging the non-uniform spectrum of the task representation and the calibrated source-target relevance, we prove that the sample complexity to achieve $\varepsilon$-excess risk on target scales with $ (k*)2 |v*|_22 \varepsilon{-2}$ where $k*$ is the effective dimension of the target and $|v*|_22 \in (0,1]$ represents the connection between source and target space. Compared to the passive one, this can save up to $\frac{1}{d_W}$ of sample complexity, where $d_W$ is the task space dimension. Finally, we demonstrate different instantiations of our meta algorithm in synthetic datasets and robotics problems, from pendulum simulations to real-world drone flight datasets. On average, our algorithms outperform baselines by $20\%-70\%$.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (23)
  1. Neural lander: Stable drone landing control using learned dynamics. In 2019 International Conference on Robotics and Automation (ICRA), pages 9784–9790. IEEE, 2019.
  2. Learning quadrupedal locomotion over challenging terrain. Science robotics, 5(47):eabc5986, 2020.
  3. Learning transferable visual models from natural language supervision. arXiv preprint arXiv:2103.00020, 2021.
  4. Language models are few-shot learners. arXiv preprint arXiv:2005.14165, 2020.
  5. Coca: Contrastive captioners are image-text foundation models. arXiv preprint arXiv:2205.01917, 2022.
  6. Flamingo: a visual language model for few-shot learning. Advances in Neural Information Processing Systems, 35:23716–23736, 2022.
  7. Neural-fly enables rapid learning for agile flight in strong winds. Science Robotics, 7(66):eabm6597, 2022.
  8. Active multi-task representation learning. In International Conference on Machine Learning, pages 3271–3298. PMLR, 2022.
  9. Few-shot learning via learning the representation, provably, 2021.
  10. On the theory of transfer learning: The importance of task diversity, 2020.
  11. Provable meta-learning of linear representations. In International Conference on Machine Learning, pages 10434–10443. PMLR, 2021.
  12. Sample efficient linear meta-learning by alternating minimization. arXiv preprint arXiv:2105.08306, 2021.
  13. Representation learning beyond linear prediction functions. Advances in Neural Information Processing Systems, 34:4792–4804, 2021.
  14. Maml and anil provably learn representations. In International Conference on Machine Learning, pages 4238–4310. PMLR, 2022.
  15. Uniform approximation of functions with random bases. In 2008 46th annual allerton conference on communication, control, and computing, pages 555–561. IEEE, 2008.
  16. Michael J Todd. Minimum-volume ellipsoids: Theory and algorithms. SIAM, 2016.
  17. Theoretical foundation for cma-es from information geometry perspective. Algorithmica, 64:698–716, 2012.
  18. Rapid learning or feature reuse? towards understanding the effectiveness of maml. arXiv preprint arXiv:1909.09157, 2019.
  19. Reptile: a scalable metalearning algorithm. arXiv preprint arXiv:1803.02999, 2(3):4, 2018.
  20. How to train your maml. arXiv preprint arXiv:1810.09502, 2018.
  21. Meta-learning in neural networks: A survey. IEEE transactions on pattern analysis and machine intelligence, 44(9):5149–5169, 2021.
  22. Weighted training for cross-task learning, 2021.
  23. Meta-adaptive nonlinear control: Theory and algorithms. Advances in Neural Information Processing Systems, 34:10013–10025, 2021.
Citations (3)

Summary

We haven't generated a summary for this paper yet.