Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Non-asymptotic oracle inequalities for the Lasso in high-dimensional mixture of experts (2009.10622v7)

Published 22 Sep 2020 in math.ST, cs.AI, cs.LG, stat.ME, stat.ML, and stat.TH

Abstract: We investigate the estimation properties of the mixture of experts (MoE) model in a high-dimensional setting, where the number of predictors is much larger than the sample size, and for which the literature is particularly lacking in theoretical results. We consider the class of softmax-gated Gaussian MoE (SGMoE) models, defined as MoE models with softmax gating functions and Gaussian experts, and focus on the theoretical properties of their $l_1$-regularized estimation via the Lasso. To the best of our knowledge, we are the first to investigate the $l_1$-regularization properties of SGMoE models from a non-asymptotic perspective, under the mildest assumptions, namely the boundedness of the parameter space. We provide a lower bound on the regularization parameter of the Lasso penalty that ensures non-asymptotic theoretical control of the Kullback--Leibler loss of the Lasso estimator for SGMoE models. Finally, we carry out a simulation study to empirically validate our theoretical findings.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. TrungTin Nguyen (17 papers)
  2. Hien D Nguyen (26 papers)
  3. Geoffrey J McLachlan (24 papers)
  4. Faicel Chamroukhi (35 papers)
Citations (1)

Summary

We haven't generated a summary for this paper yet.