Eliminating polynomial dependence on rank d in learning low logit rank models
Establish whether there exists an algorithm for learning approximately low logit rank language models under the same setting as Theorem 7 (the main learning result with logit queries) that only requires the final total variation error bound ε* to be lower bounded by a polynomial in the average logit misspecification ε_avg, the sequence length T, the alphabet size |Σ|, the boundedness parameter α, and 1/δ, but with no polynomial dependence on the logit rank d.
Sponsor
References
We are not even sure if it is possible to avoid polynomial dependence on $d$ altogether: Is there an algorithm for learning low logit rank models in the setting of \cref{thm:approx-main} that only requires $\ep\st \geq \Omega(\poly(\epavg, T, |\Sigma|, \alpha,1/\delta))$, i.e., without any polynomial dependence on the rank $d$?