Outlier-robust Estimation of a Sparse Linear Model Using Invexity
Abstract: In this paper, we study problem of estimating a sparse regression vector with correct support in the presence of outlier samples. The inconsistency of lasso-type methods is well known in this scenario. We propose a combinatorial version of outlier-robust lasso which also identifies clean samples. Subsequently, we use these clean samples to make a good estimation. We also provide a novel invex relaxation for the combinatorial problem and provide provable theoretical guarantees for this relaxation. Finally, we conduct experiments to validate our theory and compare our results against standard lasso.
- Fair sparse regression with clustering: An invex relaxation for a combinatorial problem. Advances in Neural Information Processing Systems, 34:23245–23257.
- Sparse mixed linear regression with guarantees: Taming an intractable problem with invex relaxation. International Conference on Machine Learning.
- What is invexity? The ANZIAM Journal, 28(1):1–9.
- Robust sparse regression under adversarial corruption. In International conference on machine learning, pages 774–782. PMLR.
- A convex formulation for mixed regression with two components: Minimax optimal rates. In Conference on Learning Theory, pages 560–604. PMLR.
- Outlier-robust estimation of a sparse linear model using l1-penalized huber’s m-estimator. Advances in neural information processing systems, 32.
- Estimating Diffusion Network Structures: Recovery Conditions, Sample Complexity & Soft-Thresholding Algorithm. In International Conference on Machine Learning, pages 793–801.
- Consistent estimation for pca and sparse regression with oblivious outliers. Advances in Neural Information Processing Systems, 34:25427–25438.
- Efficient online and batch learning using forward backward splitting. The Journal of Machine Learning Research, 10:2899–2934.
- Variable selection via nonconcave penalized likelihood and its oracle properties. Journal of the American statistical Association, 96(456):1348–1360.
- Sensitivity analysis in linear regression. John Wiley & Sons.
- Robust statistics: the approach based on influence functions. Wiley-Interscience; New York.
- Hanson, M. A. (1981). On sufficiency of the kuhn-tucker conditions. Journal of Mathematical Analysis and Applications, 80(2):545–550.
- Haynsworth, E. V. (1968). Determination of the inertia of a partitioned hermitian matrix. Linear algebra and its applications, 1(1):73–81.
- Matrix analysis. Cambridge university press.
- A tail inequality for quadratic forms of subgaussian random vectors. Electronic Communications in Probability, 17.
- Huber, P. J. (2011). Robust statistics. In International encyclopedia of statistical science, pages 1248–1251. Springer.
- Robust regression through the huber’s criterion and adaptive lasso penalty. Electronic Journal of Statistics, 5:1015–1053.
- High dimensional robust sparse regression. In International Conference on Artificial Intelligence and Statistics, pages 411–421. PMLR.
- Minimum distance lasso for robust high-dimensional regression. Electronic Journal of Statistics, 10(1):1296–1340.
- High-dimensional graphs and variable selection with the lasso. The annals of statistics, 34(3):1436–1462.
- Improved imaging by invex regularizers with global optima guarantees. arXiv preprint arXiv:2211.10112.
- Spam: Sparse Additive Models. In Proceedings of the 20th International Conference on Neural Information Processing Systems, pages 1201–1208. Curran Associates Inc.
- High-dimensional Ising Model Selection Using L1-Regularized Logistic Regression. The Annals of Statistics, 38(3):1287–1319.
- Adaptive huber regression. Journal of the American Statistical Association, 115(529):254–265.
- Vershynin, R. (2012). How close is the sample covariance matrix to the actual covariance matrix? Journal of Theoretical Probability, 25(3):655–686.
- Wainwright, M. J. (2009). Sharp Thresholds for High-Dimensional and Noisy Sparsity Recovery Using L1-Constrained Quadratic Programming (Lasso). IEEE transactions on information theory, 55(5):2183–2202.
- Wainwright, M. J. (2019). High-dimensional statistics: A non-asymptotic viewpoint, volume 48. Cambridge University Press.
- Alternating minimization for mixed linear regression. In International Conference on Machine Learning, pages 613–621. PMLR.
- Zou, H. (2006). The adaptive lasso and its oracle properties. Journal of the American statistical association, 101(476):1418–1429.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.