Active Learning for Regression based on Wasserstein distance and GroupSort Neural Networks (2403.15108v1)
Abstract: This paper addresses a new active learning strategy for regression problems. The presented Wasserstein active regression model is based on the principles of distribution-matching to measure the representativeness of the labeled dataset. The Wasserstein distance is computed using GroupSort Neural Networks. The use of such networks provides theoretical foundations giving a way to quantify errors with explicit bounds for their size and depth. This solution is combined with another uncertainty-based approach that is more outlier-tolerant to complete the query strategy. Finally, this method is compared with other classical and recent solutions. The study empirically shows the pertinence of such a representativity-uncertainty approach, which provides good estimation all along the query procedure. Moreover, the Wasserstein active regression often achieves more precise estimations and tends to improve accuracy faster than other models.
- (2007). Modeling slump flow of concrete using second-order regressions and artificial neural networks. Cement and Concrete Composites, 29(6):474–480.
- (2021). Pool-based unsupervised active learning for regression using iterative representativeness-diversity maximization (irdm). Pattern Recognition Letters, 142:11–19.
- (2023). Active learning for regression by inverse distance weighting. Information Sciences, 626:275–292.
- Sorting out lipschitz function approximation.
- A theory of learning from different domains. Machine Learning, 79:151–175.
- Curriculum learning. International Conference on Machine Learning.
- Some Theoretical Properties of GANs. Annals of Statistics, 48(3):1539–1566.
- Some theoretical insights into wasserstein gans. 22(1).
- An iterative algorithm for computing the best estimate of an orthogonal matrix. SIAM Journal on Numerical Analysis, 8(2):358–364.
- Active learning for regression based on query by committee. In Yin, H., Tino, P., Corchado, E., Byrne, W., and Yao, X., editors, Intelligent Data Engineering and Automated Learning - IDEAL 2007, pages 209–218, Berlin, Heidelberg. Springer Berlin Heidelberg.
- Cédric, V. (right 2009). Optimal transport : old and new / Villani Cédric. Grundlehren der mathematischen Wissenschaften. Springer, Berlin.
- Dudley, R. M. (1969). The Speed of Mean Glivenko-Cantelli Convergence. The Annals of Mathematical Statistics, 40(1):40 – 50.
- On the rate of convergence in Wasserstein distance of the empirical measure. Probability Theory and Related Fields, 162(3-4):707.
- Geometry, resistance and stability of the delft systematic yacht hull series. International shipbuilding progress, 28:276–297.
- Generative adversarial nets. In Ghahramani, Z., Welling, M., Cortes, C., Lawrence, N., and Weinberger, K., editors, Advances in Neural Information Processing Systems, volume 27. Curran Associates, Inc.
- Regularisation of neural networks by enforcing lipschitz continuity. Machine Learning, 110(2):393–416.
- Improved training of wasserstein gans. CoRR, abs/1704.00028.
- Hedonic housing prices and the demand for clean air. Journal of Environmental Economics and Management, 5(1):81–102.
- Adam: A method for stochastic optimization. International Conference on Learning Representations.
- A sequential algorithm for training text classifiers. In Croft, B. W. and van Rijsbergen, C. J., editors, SIGIR ’94, pages 3–12, London. Springer London.
- One-shot learning of object categories. IEEE Trans. Pattern Anal. Mach. Intell., 28(4):594–611.
- An invitation to statistics in wasserstein space.
- A survey of deep active learning. arXiv preprint arXiv:2009.00236.
- Toward optimal active learning through sampling estimation of error reduction. In International Conference on Machine Learning.
- Settles, B. (2012). Active Learning. Synthesis Lectures on Artificial Intelligence and Machine Learning Series. Morgan & Claypool.
- Shannon, C. E. (2001). A mathematical theory of communication. SIGMOBILE Mob. Comput. Commun. Rev., 5(1):3–55.
- Deep active learning: Unified and principled method for query and training. In International Conference on Artificial Intelligence and Statistics, pages 1308–1318. PMLR.
- Curriculum learning: A survey. Int. J. Comput. Vision, 130(6):1526–1565.
- Optimal 1-wasserstein distance for wgans. ArXiv, abs/2201.02824.
- Approximating lipschitz continuous functions with groupsort neural networks. ArXiv, abs/2006.05254.
- Airfoil self-noise and prediction. Technical report, NASA RP-1218.
- Effective data augmentation with diffusion models.
- Accurate quantitative estimation of energy performance of residential buildings using statistical machine learning tools. Energy and Buildings, 49:560–567.
- Active learning for regression using greedy sampling. Information Sciences, 474:90–105.
- A probability analysis on the value of unlabeled data for classification problems. In Proc. 17th International Conf. on Machine Learning, pages 1191–1198. Morgan Kaufmann, San Francisco, CA.