Papers
Topics
Authors
Recent
Search
2000 character limit reached

KABC: Bandit Clustering & Kaczmarz Acceleration

Updated 19 January 2026
  • KABC algorithm is a dual-framework method that applies kernel-based adaptive bandit clustering and clustering-accelerated Kaczmarz to enhance statistical and computational efficiency.
  • It employs kernel mean embedding and maximum mean discrepancy tests to achieve precise, variance-aware, nonparametric clustering of bandit arms under PAC guarantees.
  • In accelerating Kaczmarz, clustering techniques like Cluster-JL and Cluster-Block reduce iteration counts and computational costs by exploiting structured matrix row similarities.

The acronym KABC refers to distinct algorithms in two research domains: (1) Kernel-based Adaptive Bandit Clustering for nonparametric clustering with bandit feedback, and (2) Kaczmarz Algorithm Based on Clustering for accelerating iterative solvers for linear systems. Both methods exploit clustering structure—either of probability distributions or matrix rows—to improve statistical or computational efficiency. The following sections distinguish these two frameworks, detail their methodologies, formal properties, and elucidate core contributions.

1. Problem Setting and Motivation

1.1 Nonparametric Kernel Clustering with Bandit Feedback

The task is to partition the arms of an NN-armed stochastic bandit according to their underlying distributions, grouping two arms if and only if they share the same distribution. At each round tt, the learner selects arm At[N]A_t \in [N] and observes XtPAtX_t \sim P_{A_t}. The goal is to output a clustering Cτ\mathcal{C}_\tau after τ\tau samples such that, with probability at least 1δ1-\delta, Cτ\mathcal{C}_\tau equals the true partition C\mathcal{C}^* (the δ\delta-PAC guarantee). The objective is to minimize the total number of samples τ\tau while making no parametric assumptions on the PiP_i (Thuot et al., 12 Jan 2026).

1.2 Accelerating Kaczmarz for Linear Systems via Clustering

Given a consistent linear system Ax=bAx = b with ARn×pA \in \mathbb{R}^{n \times p}, the randomized Kaczmarz algorithm (RKA) projects the iterate xkx_k onto hyperplanes defined by randomly selected equations. When AA's rows cluster directionally, exploiting this structure can accelerate convergence. KABC clusters rows and uses this partition to improve both block Kaczmarz procedures and single-row selection strategies, with demonstrated speedups in high-dimensional Gaussian data (Li et al., 2015).

2. Algorithmic Methodology

2.1 Kernel Mean Embedding and MMD for Bandit Clustering

A characteristic, bounded kernel g:X×XRg: \mathcal{X}\times \mathcal{X} \rightarrow \mathbb{R} induces a reproducing kernel Hilbert space (RKHS) H\mathcal{H}. The kernel mean embedding (KME) for each arm ii is μi=EXPi[φ(X)]H\mu_i = \mathbb{E}_{X \sim P_i}[\varphi(X)] \in \mathcal{H}. The maximum mean discrepancy (MMD) between Pi,PjP_i, P_j is μiμjH\|\mu_i - \mu_j\|_{\mathcal{H}}, with Pi=PjP_i = P_j iff μi=μj\mu_i = \mu_j for characteristic kernels. Thus, clustering PiP_i's reduces to clustering KMEs in H\mathcal{H} (Thuot et al., 12 Jan 2026).

2.2 Signal-to-Noise Ratio and Thresholding

For reliable clustering, within-cluster variance and between-cluster separation must be balanced. Define arm-wise RKHS variance as Vi=EXPi[φ(X)μi2]\mathcal{V}_i^* = \mathbb{E}_{X \sim P_i}[\|\varphi(X) - \mu_i\|^2] and the minimal signal-to-noise ratio as

Δ=minij,μiμjMMD(Pi,Pj)max{i,j}V.\Delta = \min_{i\neq j,\, \mu_i \neq \mu_j} \frac{\mathrm{MMD}(P_i, P_j)}{ \max_{\ell \in \{i,j\}} \sqrt{\mathcal{V}_\ell^*} }.

Empirical quantities for sampling-based confidence bounds are assembled into variance-aware thresholds for two-sample testing (Thuot et al., 12 Jan 2026).

2.3 Clustering in Kaczmarz Methods

KABC for linear systems begins with clustering (e.g., k-means) of the rows AiA_i into cc clusters C1,,CcC_1, \dots, C_c. Two implementation variants are developed:

  • Cluster-JL: Within each iteration, the furthest cluster center (in residual sense) is identified, rows are sampled from this cluster using Johnson–Lindenstrauss (JL) projections to estimate high-residual candidates, and the update proceeds along the maximally violating row.
  • Cluster-Block: Blocks are formed by selecting one row per cluster. Each block is used as an update direction, and blocks are chosen uniformly at random (Li et al., 2015).

3. Formal Algorithmic Description

3.1 KABC for Nonparametric Bandit Clustering

The KABC routine employs an adaptive, round-based schedule:

  • At each round kk, per-arm sample size nk=2kln(8(N2N)/δk)n_k = \lceil 2^k \ln(8(N^2-N)/\delta_k) \rceil and error tolerance δk=δ/(4k2)\delta_k = \delta/(4k^2).
  • For each pair (i,j)(i, j), empirical KMEs and MMDs are computed, and a variance-aware threshold Bij(nk,δk)B_{ij}(n_k, \delta_k) is derived.
  • Edges are placed in a similarity graph based on DijBijD_{ij} \leq B_{ij}; connected components provide a clustering.
  • The process terminates when exactly KK clusters are found (Thuot et al., 12 Jan 2026).

3.2 KABC for Accelerated Kaczmarz

Pseudocode:

Cluster-JL:

  1. Cluster {Ai}\{A_i\} into C1,,CcC_1,\ldots,C_c.
  2. For each iteration, identify the most violating center, sample ss rows from the associated cluster using projected norms via JL mapping, and update xkx_k with the row of maximal projected residual.

Cluster-Block:

  1. Cluster {Ai}\{A_i\}, sample blocks with one row from each cluster.
  2. Iterate updates xk+1=xk+Aτ(bτAτxk)x_{k+1} = x_k + A_\tau^{\dagger}(b_\tau - A_\tau x_k), with block index τ\tau chosen uniformly at random (Li et al., 2015).

4. Theoretical Properties

4.1 Correctness and Sample Complexity for Bandit Clustering

KABC is δ\delta-PAC: with probability at least 1δ1-\delta, it outputs the correct clustering Cτ=C\mathcal{C}_\tau = \mathcal{C}^*. The sample complexity τ\tau satisfies

τ8N[128s21]ln(32(N2N)k2/δ)\tau \leq 8N \left[ \frac{128}{s_*^2} \vee 1 \right]\ln\left(32(N^2-N)k_*^2/\delta\right)

with s2s_*^2 a variance-aware squared SNR and k=log2(128/s2)k_* = \lceil \log_2(128/s_*^2)\rceil. The method is adaptive: no prior estimate of Δ\Delta or {Vi}\{\mathcal{V}_i^*\} is needed (Thuot et al., 12 Jan 2026).

4.2 Acceleration and Convergence for Kaczmarz Variants

In the block Kaczmarz context, the contraction per iteration improves upon random partitions. Specifically, block spectral norm β1+cϵ\beta\leq 1 + c\epsilon and minimum eigenvalue α1ϵ\alpha \geq 1 - \epsilon for clustering-based blocks, where ϵ\epsilon controls within-block coherence. The contraction factor per iteration 1σmin2(A)/((1+ϵc)m)1-\sigma_{\min}^2(A)/((1+\epsilon c)m) is strictly smaller than in randomly formed blocks, leading to faster convergence. Empirically, Cluster-JL and Cluster-Block variants halved iteration counts or improved CPU time by 30–40% in synthetic Gaussian problems (Li et al., 2015).

5. Computational Complexity and Implementation

5.1 Nonparametric Clustering

Time per round for KABC is O(Nnk)O(N n_k) for sampling and O(N2nk)O(N^2 n_k) for pairwise distance calculations; dominant cost arises for kkk\approx k_*, giving total time O(N2/s2)O(N^2/s_*^2). Memory requirements are O(N2)O(N^2), mainly for dot-products and statistics; streaming architectures are feasible. Translation-invariant kernels such as Gaussian RBF meet the required criteria (Thuot et al., 12 Jan 2026).

5.2 Kaczmarz with Clustering

Clustering cost is O(npc)O(np c) for nn rows and pp dimensions. Each Cluster-JL or Cluster-Block iteration matches the per-iteration costs of standard Kaczmarz but achieves lower per-iteration contraction by exploiting cluster structure, markedly reducing total iterations for well-clustered data (Li et al., 2015).

6. Applications, Limitations, and Practical Remarks

6.1 Applications

For kernel-based bandit clustering, target domains include recommendation systems (clustering user distributions), adaptive grouping in medical trials, and other settings where cluster membership is defined by identity of distributions and sampling is expensive (Thuot et al., 12 Jan 2026).

Clustering-accelerated Kaczmarz methods target high-dimensional linear systems arising in compressed sensing, large-scale regression, or applications where matrix rows exhibit structural redundancy or coherence (Li et al., 2015).

6.2 Limitations and Assumptions

  • For kernel-based bandit clustering, the number of clusters KK must be known, and kernel gg must be bounded and characteristic. If between-distribution MMD separation Δ\Delta is arbitrarily small, no finite-time clustering is possible.
  • For Kaczmarz with clustering, acceleration critically depends on the presence of cluster structure; otherwise, benefits may be nominal. The analysis assumes block spectral norms and within-block incoherence can be bounded via clustering (Thuot et al., 12 Jan 2026, Li et al., 2015).

7. Summary Table: Distinctions Between KABC Algorithms

Domain Clustering Object Main Technique
Nonparametric bandit clustering (Thuot et al., 12 Jan 2026) Arms/distributions via kernels Kernel mean embedding, adaptive MMD tests
Accelerated linear solvers (Li et al., 2015) Rows of matrix AA Clustering-guided block/sampling for Kaczmarz

Both methods demonstrate that pre-clustering and adaptively exploiting cluster structure yield provable and empirical benefits for high-dimensional inference and optimization.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (2)

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to KABC Algorithm.