Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Finer-Grained Hardness of Kernel Density Estimation (2407.02372v1)

Published 2 Jul 2024 in cs.DS, cs.NA, and math.NA

Abstract: In batch Kernel Density Estimation (KDE) for a kernel function $f$, we are given as input $2n$ points $x{(1)}, \cdots, x{(n)}, y{(1)}, \cdots, y{(n)}$ in dimension $m$, as well as a vector $v \in \mathbb{R}n$. These inputs implicitly define the $n \times n$ kernel matrix $K$ given by $K[i,j] = f(x{(i)}, y{(j)})$. The goal is to compute a vector $v$ which approximates $K w$ with $|| Kw - v||_\infty < \varepsilon ||w||_1$. A recent line of work has proved fine-grained lower bounds conditioned on SETH. Backurs et al. first showed the hardness of KDE for Gaussian-like kernels with high dimension $m = \Omega(\log n)$ and large scale $B = \Omega(\log n)$. Alman et al. later developed new reductions in roughly this same parameter regime, leading to lower bounds for more general kernels, but only for very small error $\varepsilon < 2{- \log{\Omega(1)} (n)}$. In this paper, we refine the approach of Alman et al. to show new lower bounds in all parameter regimes, closing gaps between the known algorithms and lower bounds. In the setting where $m = C\log n$ and $B = o(\log n)$, we prove Gaussian KDE requires $n{2-o(1)}$ time to achieve additive error $\varepsilon < \Omega(m/B){-m}$, matching the performance of the polynomial method up to low-order terms. In the low dimensional setting $m = o(\log n)$, we show that Gaussian KDE requires $n{2-o(1)}$ time to achieve $\varepsilon$ such that $\log \log (\varepsilon{-1}) > \tilde \Omega ((\log n)/m)$, matching the error bound achievable by FMM up to low-order terms. To our knowledge, no nontrivial lower bound was previously known in this regime. Our new lower bounds make use of an intricate analysis of a special case of the kernel matrix -- the `counting matrix'. As a key technical lemma, we give a novel approach to bounding the entries of its inverse by using Schur polynomials from algebraic combinatorics.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (30)
  1. Optimal-Degree Polynomial Approximations for Exponentials and Gaussian Kernel Density Estimation. In 37th Computational Complexity Conference (CCC 2022), 2022.
  2. Algorithms and hardness for linear algebra on geometric graphs. In 2020 IEEE 61st Annual Symposium on Foundations of Computer Science (FOCS), 2020.
  3. Fast attention requires bounded entries. In NeurIPS, 2023.
  4. Probabilistic polynomials and hamming nearest neighbors. 2015 IEEE 56th Annual Symposium on Foundations of Computer Science, Oct 2015.
  5. Efficient density evaluation for smooth kernels. In 2018 IEEE 59th Annual Symposium on Foundations of Computer Science (FOCS), pages 615–626, 2018.
  6. Lectures on integrable probability, 2015.
  7. On the fine-grained complexity of empirical risk minimization: Kernel methods and neural networks. In I. Guyon, U. Von Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett, editors, Advances in Neural Information Processing Systems, volume 30. Curran Associates, Inc., 2017.
  8. Space and time efficient kernel density estimation in high dimensions. In Proceedings of the 33rd International Conference on Neural Information Processing Systems, Red Hook, NY, USA, 2019. Curran Associates Inc.
  9. Yen-Chi Chen. A tutorial on kernel density estimation and recent advances. Biostatistics & Epidemiology, 1(1):161–187, 2017.
  10. Lijie Chen. On the hardness of approximate and exact (bichromatic) maximum inner product. Theory of Computing, 16(4):1–50, 2020.
  11. Kernel density estimation through density constrained near neighbor search. In 2020 IEEE 61st Annual Symposium on Foundations of Computer Science (FOCS), pages 172–183, 2020.
  12. A quasi-monte carlo data structure for smooth kernel evaluations. In Proceedings of the 2024 Annual ACM-SIAM Symposium on Discrete Algorithms (SODA), pages 5118–5144, 2024.
  13. Hashing-based-estimators for kernel density in high dimensions. In 2017 IEEE 58th Annual Symposium on Foundations of Computer Science (FOCS), pages 1032–1043, 2017.
  14. Multi-resolution hashing for fast pairwise summations. In 2019 IEEE 60th Annual Symposium on Foundations of Computer Science (FOCS), pages 769–792, 2019.
  15. L Greengard and V Rokhlin. A fast algorithm for particle simulations. Journal of Computational Physics, 73(2):325–348, 1987.
  16. The fast gauss transform. SIAM J. Sci. Comput., 12:79–94, 1991.
  17. Normal subgroup reconstruction and quantum computation using group representations. In Proceedings of the Thirty-Second Annual ACM Symposium on Theory of Computing, STOC ’00, page 627–635, New York, NY, USA, 2000. Association for Computing Machinery.
  18. On the complexity of k-sat. Journal of Computer and System Sciences, 62(2):367–375, 2001.
  19. Rectangular kronecker coefficients and plethysms in geometric complexity theory. In 2016 IEEE 57th Annual Symposium on Foundations of Computer Science (FOCS), pages 396–405, 2016.
  20. On the sign patterns of entrywise positivity preservers in fixed dimension. American Journal of Mathematics, 143:1863 – 1929, 2021.
  21. Kernel mean embedding of distributions: A review and beyond. Foundations and Trends® in Machine Learning, 10(1-2):1–141, 2017.
  22. Andrei Okounkov. Infinite wedge and random partitions. Selecta Mathematica, 7:57–81, 1999.
  23. Quantum spectrum testing. In Proceedings of the Forty-Seventh Annual ACM Symposium on Theory of Computing, STOC ’15, page 529–538, New York, NY, USA, 2015. Association for Computing Machinery.
  24. Near-optimal coresets of kernel density estimates. Discrete Comput. Geom., 63(4):867–887, jun 2020.
  25. Aviad Rubinstein. Hardness of approximate nearest neighbor search. Proceedings of the 50th Annual ACM SIGACT Symposium on Theory of Computing, Jun 2018.
  26. Learning with kernels: support vector machines, regularization, optimization, and beyond. MIT press, 2002.
  27. Richard Stanley. Enumerative Combinatorics. Cambridge Studies in Advanced Mathematics. Cambridge University Press, 2 edition, 2023.
  28. Paraskevas Syminelakis. Fast kernel evaluation in high dimensions: Importance sampling and near neighbor search. PhD thesis, Stanford University, 2019.
  29. Ryan Williams. A new algorithm for optimal 2-constraint satisfaction and its implications. Theoretical Computer Science, 348(2-3):357–365, 2005.
  30. Ryan Williams. On the difference between closest, furthest, and orthogonal pairs: Nearly-linear vs barely-subquadratic complexity. In Proceedings of the Twenty-Ninth Annual ACM-SIAM Symposium on Discrete Algorithms, 2018.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (2)
  1. Josh Alman (36 papers)
  2. Yunfeng Guan (12 papers)
Citations (1)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com