Papers
Topics
Authors
Recent
AI Research Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 80 tok/s
Gemini 2.5 Pro 49 tok/s Pro
GPT-5 Medium 26 tok/s Pro
GPT-5 High 32 tok/s Pro
GPT-4o 92 tok/s Pro
Kimi K2 182 tok/s Pro
GPT OSS 120B 438 tok/s Pro
Claude Sonnet 4 38 tok/s Pro
2000 character limit reached

Tuning parameter selection rules for nuclear norm regularized multivariate linear regression (1901.06478v1)

Published 19 Jan 2019 in math.ST and stat.TH

Abstract: We consider the tuning parameter selection rules for nuclear norm regularized multivariate linear regression (NMLR) in high-dimensional setting. High-dimensional multivariate linear regression is widely used in statistics and machine learning, and regularization technique is commonly applied to deal with the special structures in high-dimensional data. As we know, how to select the tuning parameter is an essential issue for regularization approach and it directly affects the model estimation performance. To the best of our knowledge, there are no rules about the tuning parameter selection for NMLR from the point of view of optimization. In order to establish such rules, we study the duality theory of NMLR. Then, we claim the choice of tuning parameter for NMLR is based on the sample data and the solution of NMLR dual problem, which is a projection on a nonempty, closed and convex set. Moreover, based on the (firm) nonexpansiveness and the idempotence of the projection operator, we build four tuning parameter selection rules PSR, PSRi, PSRfn and PSR+. Furthermore, we give a sequence of tuning parameters and the corresponding intervals for every rule, which states that the rank of the estimation coefficient matrix is no more than a fixed number for the tuning parameter in the given interval. The relationships between these rules are also discussed and PSR+ is the most efficient one to select the tuning parameter. Finally, the numerical results are reported on simulation and real data, which show that these four tuning parameter selection rules are valuable.

Summary

We haven't generated a summary for this paper yet.

Lightbulb On Streamline Icon: https://streamlinehq.com

Continue Learning

We haven't generated follow-up questions for this paper yet.

Authors (2)

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Don't miss out on important new AI/ML research

See which papers are being discussed right now on X, Reddit, and more:

“Emergent Mind helps me see which AI papers have caught fire online.”

Philip

Philip

Creator, AI Explained on YouTube