Transductive Active Learning: Theory and Applications
Abstract: We study a generalization of classical active learning to real-world settings with concrete prediction targets where sampling is restricted to an accessible region of the domain, while prediction targets may lie outside this region. We analyze a family of decision rules that sample adaptively to minimize uncertainty about prediction targets. We are the first to show, under general regularity assumptions, that such decision rules converge uniformly to the smallest possible uncertainty obtainable from the accessible data. We demonstrate their strong sample efficiency in two key applications: active fine-tuning of large neural networks and safe Bayesian optimization, where they achieve state-of-the-art performance.
- Abbasi-Yadkori, Y. Online learning for linearly parametrized control problems. PhD thesis, University of Alberta, 2013.
- Adapting the linearised laplace model evidence for modern deep learning. In International Conference on Machine Learning, pp. 796–821. PMLR, 2022.
- On exact computation with an infinitely wide neural net. NeurIPS, 32, 2019.
- k-means++: The advantages of careful seeding. In SODA, volume 7, 2007.
- Gone fishing: Neural active learning with fisher embeddings. NeurIPS, 34, 2021.
- Deep batch active learning by diverse, uncertain gradient lower bounds. ICLR, 2020.
- Best arm identification in multi-armed bandits. In COLT, 2010.
- Ann-benchmarks: A benchmarking tool for approximate nearest neighbor algorithms. Information Systems, 87, 2020.
- A cookbook of self-supervised learning. arXiv preprint arXiv:2304.12210, 2023.
- Barrett, A. B. Exploration of synergistic and redundant information sharing in static and dynamical gaussian systems. Physical Review E, 91(5), 2015.
- Gosafe: Globally optimal safe robot learning. In ICRA, 2021.
- Curriculum learning. In ICML, volume 26, 2009.
- Feature selection via mutual information: New theoretical insights. In IJCNN, 2019.
- Safe controller optimization for quadrotors with gaussian processes. In ICRA, 2016.
- Bayesian optimization with safety constraints: safe and automatic parameter tuning in robotics. Machine Learning, 2021.
- Weight uncertainty in neural network. In ICML, 2015.
- Truncated variance reduction: A unified approach to bayesian optimization and level-set estimation. NeurIPS, 29, 2016.
- Information-theoretic safe exploration with gaussian processes. NeurIPS, 35, 2022.
- Information-theoretic safe bayesian optimization. arXiv preprint arXiv:2402.15347, 2024.
- Pure exploration in multi-armed bandits problems. In ALT, volume 20, 2009.
- Bayesian experimental design: A review. Statistical Science, 1995.
- Chandra, B. Quadrotor simulation, 2023. URL https://github.com/Bharath2/Quadrotor-Simulation.
- Near-optimal batch mode active learning and adaptive submodular optimization. In ICML, 2013.
- On kernelized multi-armed bandits. In ICML, 2017.
- Cohn, D. Neural network exploration using optimal experiment design. Advances in neural information processing systems, 6, 1993.
- Similarity search for efficient active learning and search of rare concepts. In AAAI, volume 36, 2022.
- Multidimensional bayesian estimation for deep brain stimulation using the safeopt algorithm. medRxiv, 2022.
- Cover, T. M. Elements of information theory. John Wiley & Sons, 1999.
- Algorithms for subset selection in linear regression. In STOC, volume 40, 2008.
- Approximate submodularity and its applications: Subset selection, sparse approximation and dictionary selection. JMLR, 19(1), 2018.
- Laplace redux-effortless bayesian deep learning. NeurIPS, 34, 2021.
- Imagenet: A large-scale hierarchical image database. In CVPR, 2009.
- BERT: Pre-training of deep bidirectional transformers for language understanding. In NAACL, 2019.
- Likelihood ratio confidence sets for sequential decision making. NeurIPS, 37, 2023.
- Adaptivity in adaptive submodularity. In COLT, 2021.
- Sequential experimental design for transductive linear bandits. NeurIPS, 32, 2019.
- Consistency-based semi-supervised active learning: Towards minimizing labeling cost. In ECCV, 2020.
- Bayesian optimization with inequality constraints. In ICML, volume 2014, 2014.
- Deep active learning over the long tail. arXiv preprint arXiv:1711.00941, 2017.
- Adaptive submodularity: Theory and applications in active learning and stochastic optimization. JAIR, 42, 2011.
- Semi-supervised learning by entropy minimization. NeurIPS, 17, 2004.
- Automated curriculum learning for neural networks. In ICML, 2017.
- Graybill, F. A. An introduction to linear statistical models. Literary Licensing, LLC, 1961.
- Accelerating large-scale inference with anisotropic vector quantization. In ICML, 2020.
- Optimistic active-learning using mutual information. In IJCAI, volume 7, 2007.
- Bayesian deep ensembles via the neural tangent kernel. NeurIPS, 33, 2020.
- Benchmarking neural network robustness to common corruptions and perturbations. ICLR, 2019.
- A baseline for detecting misclassified and out-of-distribution examples in neural networks. ICLR, 2017.
- Entropy search for information-efficient global optimization. JMLR, 13(6), 2012.
- Predictive entropy search for efficient global optimization of black-box functions. NeurIPS, 27, 2014.
- Output-space predictive entropy search for flexible global optimization. In NeurIPS workshop on Bayesian Optimization, 2015.
- A framework and benchmark for deep batch active learning for regression. JMLR, 24(164), 2023.
- Hopfield, J. J. Neural networks and physical systems with emergent collective computational abilities. Proceedings of the national academy of sciences, 79(8), 1982.
- Bayesian active learning for classification and preference learning. CoRR, 2011.
- Universal language model fine-tuning for text classification. In ACL, 2018.
- Active few-shot fine-tuning. arXiv preprint arXiv:2402.15441, 2024.
- Neural tangent kernel: Convergence and generalization in neural networks. NeurIPS, 31, 2018.
- Billion-scale similarity search with gpus. IEEE Transactions on Big Data, 7(3), 2019.
- Probabilistic active meta-learning. NeurIPS, 33, 2020.
- Neural contextual bandits without regret. In AISTATS, 2022.
- Approximate inference turns deep networks into gaussian processes. NeurIPS, 32, 2019.
- Scalable greedy feature selection via weak submodularity. In AISTATS, 2017.
- Adam: A method for stochastic optimization. In ICLR, 2014.
- Batchbald: Efficient and diverse batch acquisition for deep bayesian active learning. NeurIPS, 32, 2019.
- Adaptive and safe bayesian optimization in high dimensions via one-dimensional subspaces. In ICML, 2019.
- Wilds: A benchmark of in-the-wild distribution shifts. In ICML, 2021.
- Do better imagenet models transfer better? In CVPR, 2019.
- Submodular function maximization. Tractability, 3, 2014.
- Nonmyopic active learning of gaussian processes: an exploration-exploitation approach. In ICML, volume 24, 2007.
- Near-optimal sensor placements in gaussian processes: Theory, efficient algorithms and empirical studies. JMLR, 9(2), 2008.
- Learning multiple layers of features from tiny images. Technical report, University of Toronto, 2009.
- The mnist database of handwritten digits. http://yann.lecun.com/exdb/mnist/, 1998.
- Deep neural networks as gaussian processes. ICLR, 2018.
- Wide neural networks of any depth evolve as linear models under gradient descent. NeurIPS, 32, 2019.
- Surgical fine-tuning improves adaptation to distribution shifts. NeurIPS workshop on Distribution Shifts, 2022.
- A sequential algorithmfor training text classifiers. In SIGIR, 1994.
- Heterogeneous uncertainty sampling for supervised learning. In Machine learning proceedings 1994. Elsevier, 1994.
- Sampling from gaussian process posteriors using stochastic gradient descent. NeurIPS, 37, 2023.
- A simple baseline for bayesian uncertainty in deep learning. NeurIPS, 32, 2019.
- A kernel-based view of language model fine-tuning. In ICML, 2023.
- Optimizing neural networks with kronecker-factored approximate curvature. In ICML, 2015.
- Murphy, K. P. Probabilistic machine learning: Advanced topics. MIT Press, 2023.
- Experimental design for linear functionals in reproducing kernel hilbert spaces. NeurIPS, 35, 2022.
- An analysis of approximations for maximizing submodular set functions—i. Mathematical programming, 14, 1978.
- The effectiveness of lloyd-type methods for the k-means problem. JACM, 2013.
- Experiment planning with function approximation. NeurIPS, 37, 2024.
- Feature selection based on mutual information criteria of max-dependency, max-relevance, and min-redundancy. IEEE Transactions on pattern analysis and machine intelligence, 27(8), 2005.
- Random features for large-scale kernel machines. NeurIPS, 20, 2007.
- Do imagenet classifiers generalize to imagenet? In ICML, 2019.
- Meta-learning priors for safe bayesian optimization. In COLT, 2023.
- A tutorial on thompson sampling. Foundations and Trends® in Machine Learning, 11(1), 2018.
- Active hidden markov models for information extraction. In IDA, 2001.
- Safe exploration for active learning with gaussian processes. In ECML PKDD, 2015.
- Active learning for convolutional neural networks: A core-set approach. ICLR, 2017.
- Gaussian process regression: Active data selection and test point rejection. In Mustererkennung 2000. Springer, 2000.
- Settles, B. Active learning literature survey. Technical report, University of Wisconsin-Madison Department of Computer Sciences, 2009.
- An analysis of active learning strategies for sequence labeling tasks. In EMNLP, 2008.
- Partial is better than all: revisiting fine-tuning strategy for few-shot learning. In AAAI, volume 35, 2021.
- Experimental design for overparameterized learning with application to single shot deep active learning. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2023.
- To compress or not to compress–self-supervised learning and information theory: A review. arXiv preprint arXiv:2304.09355, 2023.
- Towards foundation models and few-shot parameter-efficient fine-tuning for volumetric organ segmentation. In MICCAI, 2023.
- Curriculum learning: A survey. IJCV, 2022.
- Gaussian process optimization in the bandit setting: No regret and experimental design. In ICML, volume 27, 2009.
- Safe exploration for optimization with gaussian processes. In ICML, 2015.
- Gosafeopt: Scalable safe exploration for global optimization of dynamical systems. Artificial Intelligence, 2023.
- Active learning helps pretrained models learn the intended task. NeurIPS, 35, 2022.
- Efficientnet: Rethinking model scaling for convolutional neural networks. In ICML, 2019.
- JAX: A python library for differentiable optimal control on accelerators, 2023. URL http://github.com/google/trajax.
- Safe exploration for interactive machine learning. NeurIPS, 32, 2019.
- On information gain and regret bounds in gaussian process bandits. In AISTATS, 2021.
- Vapnik, V. Estimation of dependences based on empirical data. Springer Science & Business Media, 2006.
- A review of feature selection methods based on mutual information. Neural computing and applications, 24, 2014.
- Matching networks for one shot learning. NeurIPS, 29, 2016.
- Wainwright, M. J. High-dimensional statistics: A non-asymptotic viewpoint, volume 48. Cambridge university press, 2019.
- Max-value entropy search for efficient bayesian optimization. In ICML, 2017.
- More than a toy: Random matrix models predict how real-world neural representations generalize. In ICML, 2022.
- Tuning legged locomotion controllers via safe bayesian optimization. CORL, 2023.
- Wilks, S. S. Certain generalizations in the analysis of variance. Biometrika, 1932.
- Gaussian processes for machine learning, volume 2. MIT press Cambridge, MA, 2006.
- A model-free algorithm to safely approach the handling limit of an autonomous racecar. In ICCVE, 2019.
- Passive sampling for regression. In ICDM, 2010.
- Active learning via transductive experimental design. In ICML, volume 23, 2006.
- Design of experiments for stochastic contextual linear bandits. NeurIPS, 34, 2021.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.