KABC: Bandit Clustering & Kaczmarz Acceleration
- KABC algorithm is a dual-framework method that applies kernel-based adaptive bandit clustering and clustering-accelerated Kaczmarz to enhance statistical and computational efficiency.
- It employs kernel mean embedding and maximum mean discrepancy tests to achieve precise, variance-aware, nonparametric clustering of bandit arms under PAC guarantees.
- In accelerating Kaczmarz, clustering techniques like Cluster-JL and Cluster-Block reduce iteration counts and computational costs by exploiting structured matrix row similarities.
The acronym KABC refers to distinct algorithms in two research domains: (1) Kernel-based Adaptive Bandit Clustering for nonparametric clustering with bandit feedback, and (2) Kaczmarz Algorithm Based on Clustering for accelerating iterative solvers for linear systems. Both methods exploit clustering structure—either of probability distributions or matrix rows—to improve statistical or computational efficiency. The following sections distinguish these two frameworks, detail their methodologies, formal properties, and elucidate core contributions.
1. Problem Setting and Motivation
1.1 Nonparametric Kernel Clustering with Bandit Feedback
The task is to partition the arms of an -armed stochastic bandit according to their underlying distributions, grouping two arms if and only if they share the same distribution. At each round , the learner selects arm and observes . The goal is to output a clustering after samples such that, with probability at least , equals the true partition (the -PAC guarantee). The objective is to minimize the total number of samples while making no parametric assumptions on the (Thuot et al., 12 Jan 2026).
1.2 Accelerating Kaczmarz for Linear Systems via Clustering
Given a consistent linear system with , the randomized Kaczmarz algorithm (RKA) projects the iterate onto hyperplanes defined by randomly selected equations. When 's rows cluster directionally, exploiting this structure can accelerate convergence. KABC clusters rows and uses this partition to improve both block Kaczmarz procedures and single-row selection strategies, with demonstrated speedups in high-dimensional Gaussian data (Li et al., 2015).
2. Algorithmic Methodology
2.1 Kernel Mean Embedding and MMD for Bandit Clustering
A characteristic, bounded kernel induces a reproducing kernel Hilbert space (RKHS) . The kernel mean embedding (KME) for each arm is . The maximum mean discrepancy (MMD) between is , with iff for characteristic kernels. Thus, clustering 's reduces to clustering KMEs in (Thuot et al., 12 Jan 2026).
2.2 Signal-to-Noise Ratio and Thresholding
For reliable clustering, within-cluster variance and between-cluster separation must be balanced. Define arm-wise RKHS variance as and the minimal signal-to-noise ratio as
Empirical quantities for sampling-based confidence bounds are assembled into variance-aware thresholds for two-sample testing (Thuot et al., 12 Jan 2026).
2.3 Clustering in Kaczmarz Methods
KABC for linear systems begins with clustering (e.g., k-means) of the rows into clusters . Two implementation variants are developed:
- Cluster-JL: Within each iteration, the furthest cluster center (in residual sense) is identified, rows are sampled from this cluster using Johnson–Lindenstrauss (JL) projections to estimate high-residual candidates, and the update proceeds along the maximally violating row.
- Cluster-Block: Blocks are formed by selecting one row per cluster. Each block is used as an update direction, and blocks are chosen uniformly at random (Li et al., 2015).
3. Formal Algorithmic Description
3.1 KABC for Nonparametric Bandit Clustering
The KABC routine employs an adaptive, round-based schedule:
- At each round , per-arm sample size and error tolerance .
- For each pair , empirical KMEs and MMDs are computed, and a variance-aware threshold is derived.
- Edges are placed in a similarity graph based on ; connected components provide a clustering.
- The process terminates when exactly clusters are found (Thuot et al., 12 Jan 2026).
3.2 KABC for Accelerated Kaczmarz
Pseudocode:
Cluster-JL:
- Cluster into .
- For each iteration, identify the most violating center, sample rows from the associated cluster using projected norms via JL mapping, and update with the row of maximal projected residual.
Cluster-Block:
- Cluster , sample blocks with one row from each cluster.
- Iterate updates , with block index chosen uniformly at random (Li et al., 2015).
4. Theoretical Properties
4.1 Correctness and Sample Complexity for Bandit Clustering
KABC is -PAC: with probability at least , it outputs the correct clustering . The sample complexity satisfies
with a variance-aware squared SNR and . The method is adaptive: no prior estimate of or is needed (Thuot et al., 12 Jan 2026).
4.2 Acceleration and Convergence for Kaczmarz Variants
In the block Kaczmarz context, the contraction per iteration improves upon random partitions. Specifically, block spectral norm and minimum eigenvalue for clustering-based blocks, where controls within-block coherence. The contraction factor per iteration is strictly smaller than in randomly formed blocks, leading to faster convergence. Empirically, Cluster-JL and Cluster-Block variants halved iteration counts or improved CPU time by 30–40% in synthetic Gaussian problems (Li et al., 2015).
5. Computational Complexity and Implementation
5.1 Nonparametric Clustering
Time per round for KABC is for sampling and for pairwise distance calculations; dominant cost arises for , giving total time . Memory requirements are , mainly for dot-products and statistics; streaming architectures are feasible. Translation-invariant kernels such as Gaussian RBF meet the required criteria (Thuot et al., 12 Jan 2026).
5.2 Kaczmarz with Clustering
Clustering cost is for rows and dimensions. Each Cluster-JL or Cluster-Block iteration matches the per-iteration costs of standard Kaczmarz but achieves lower per-iteration contraction by exploiting cluster structure, markedly reducing total iterations for well-clustered data (Li et al., 2015).
6. Applications, Limitations, and Practical Remarks
6.1 Applications
For kernel-based bandit clustering, target domains include recommendation systems (clustering user distributions), adaptive grouping in medical trials, and other settings where cluster membership is defined by identity of distributions and sampling is expensive (Thuot et al., 12 Jan 2026).
Clustering-accelerated Kaczmarz methods target high-dimensional linear systems arising in compressed sensing, large-scale regression, or applications where matrix rows exhibit structural redundancy or coherence (Li et al., 2015).
6.2 Limitations and Assumptions
- For kernel-based bandit clustering, the number of clusters must be known, and kernel must be bounded and characteristic. If between-distribution MMD separation is arbitrarily small, no finite-time clustering is possible.
- For Kaczmarz with clustering, acceleration critically depends on the presence of cluster structure; otherwise, benefits may be nominal. The analysis assumes block spectral norms and within-block incoherence can be bounded via clustering (Thuot et al., 12 Jan 2026, Li et al., 2015).
7. Summary Table: Distinctions Between KABC Algorithms
| Domain | Clustering Object | Main Technique |
|---|---|---|
| Nonparametric bandit clustering (Thuot et al., 12 Jan 2026) | Arms/distributions via kernels | Kernel mean embedding, adaptive MMD tests |
| Accelerated linear solvers (Li et al., 2015) | Rows of matrix | Clustering-guided block/sampling for Kaczmarz |
Both methods demonstrate that pre-clustering and adaptively exploiting cluster structure yield provable and empirical benefits for high-dimensional inference and optimization.