Papers
Topics
Authors
Recent
Search
2000 character limit reached

Sparse Broad Learning System (S-BLS)

Updated 29 November 2025
  • S-BLS is a model framework that improves robustness and interpretability of Broad Learning Systems by integrating Sequential Threshold Least-Squares.
  • It employs a sparse output weight learning mechanism to prune noisy nodes while maintaining predictive performance.
  • Experimental evaluations show a 20–35% RMSE reduction and significant node pruning in both nonlinear system identification and CSTR benchmarks.

The Sparse Broad Learning System (S-BLS) is a model framework developed to enhance the robustness and interpretability of Broad Learning Systems in the presence of sensor noise and outliers, which are common in industrial and nonlinear system identification tasks. S-BLS integrates a Sequential Threshold Least-Squares (STLS) mechanism for learning sparse output weights, offering significant improvements in generalization, efficiency, and noise resilience compared to standard ridge-regression-based BLS solutions (Li, 22 Nov 2025).

1. Structure of the Broad Learning System

The conventional Broad Learning System (BLS) is a shallow network architecture designed for efficiently mapping input data to target outputs with reduced computational complexity. For input data XRN×DX \in \mathbb{R}^{N \times D}, BLS constructs two primary layers before linear output regression:

  • Feature Mapping Layer: Input samples are transformed via random weights and nonlinearity ϕ\phi into nn feature groups, each of size kk,

Zi=ϕ(XWei+βei)RN×k,i=1,,n,Z_i = \phi(X W_{e_i} + \beta_{e_i}) \in \mathbb{R}^{N \times k},\quad i = 1, \dots, n,

which are concatenated to form Zn=[Z1,...,Zn]RN×(nk)Z^n = [Z_1, ..., Z_n] \in \mathbb{R}^{N \times (nk)}.

  • Enhancement Layer: The mapped features ZnZ^n are further expanded via mm enhancement groups (size qq) with another nonlinearity ξ\xi,

Hj=ξ(ZnWhj+βhj)RN×q,j=1,...,m,H_j = \xi(Z^n W_{h_j} + \beta_{h_j}) \in \mathbb{R}^{N \times q},\quad j = 1, ..., m,

and concatenated to yield HmRN×(mq)H^m \in \mathbb{R}^{N \times (mq)}.

The system input for regression, A=[ZnHm]RN×LA = [Z^n \mid H^m] \in \mathbb{R}^{N \times L}, with L=nk+mqL = nk + mq, feeds a linear read-out Y=AWoutY = A W_{\text{out}} to match the targets YRN×CY \in \mathbb{R}^{N \times C}.

Standard BLS computes WoutW_{\text{out}} by ridge-regularized least squares (pseudoinverse), leading to a solution that is typically dense, with each node contributing to the output, thus propagating noise and risking overfitting in non-ideal measurement environments.

2. Sparse Output-Weight Learning via Sequential Threshold Least-Squares

S-BLS modifies the output-weight learning by introducing direct sparsity in WoutW_{\text{out}} using an L0L_0-penalized regression objective,

minW  12YAWF2+αW0.\min_{W}\;\frac12\,\|Y - A\,W\|_F^2 + \alpha\,\|W\|_0.

Since obtaining the exact L0L_0 solution is NP-hard, S-BLS adopts the Sequential Threshold Least-Squares (STLS) strategy, which alternates between elementwise hard-thresholding and least-squares projection on active variables.

STLS Steps:

  1. Hard Thresholding: Zero out weights w<λ|w| < \lambda via

Tλ(w)={0,w<λ, w,wλ.\mathcal{T}_{\lambda}(w) = \begin{cases} 0, & |w| < \lambda, \ w, & |w| \geq \lambda. \end{cases}

  1. Least-Squares Projection: Restrict AA to active columns S\mathcal{S} where Wi,d0W_{i,d} \neq 0. Compute restricted least squares,

WS=(ASTAS)1ASTY,W_{\mathcal{S}} = (A_{\mathcal{S}}^{T}A_{\mathcal{S}})^{-1}A_{\mathcal{S}}^{T}Y,

with all other weights fixed at zero.

  1. Iteration: Start with W(0)=AYW^{(0)} = A^\dagger Y and apply thresholding and projection for a fixed number TT (typically $5$–$10$) of iterations.

Following this procedure results in a sparse WoutW_{\text{out}} that selectively removes noise-influenced nodes while maintaining predictive performance.

3. Training Procedure and Implementation

The S-BLS training process follows these steps:

  1. Random Layer Construction: Sample Wei,βei,Whj,βhjW_{e_i}, \beta_{e_i}, W_{h_j}, \beta_{h_j} for all feature and enhancement nodes.
  2. Layer Computation: Compute all ZiZ_i, HjH_j, and concatenate to form AA.
  3. Initialization: Compute W(0)=AYW^{(0)} = A^\dagger Y.
  4. Iterative STLS: For t=1,,Tt = 1,\dots,T,
    • Hard-threshold W(t1)W^{(t-1)},
    • For each output dd, re-solve least squares on active set Sd\mathcal{S}_d.
  5. Termination: Return the final sparse output matrix W(T)W^{(T)}.

This approach preserves the efficient analytical nature of classical BLS while achieving node-level sparsity that benefits downstream interpretability.

4. Experimental Evaluation

S-BLS has been extensively evaluated in two settings:

4.1 Nonlinear System Identification

  • Model: A nonlinear difference equation,

y(n)=y(n1)y(n2)(y(n1)+2.5)1+y2(n1)+y2(n2)+u(n1)y(n) = \frac{y(n-1)\,y(n-2)\,(y(n-1)+2.5)}{1 + y^2(n-1)+y^2(n-2)}+u(n-1)

with u(n)u(n) stochastic in training and sinusoidal in test.

  • Noise: Gaussian/uniform, with levels γ=0.1\gamma = 0.1 to $0.4$.
  • Metrics: Test RMSE and sparsity ratio (active/L).
Noise RMSE (BLS) RMSE (S-BLS) Active/L
0.1 0.3370 0.2588 201/401 (50.1%)
0.2 0.2222 0.1673 201/401
0.3 0.2361 0.1648 201/401
0.4 0.2528 0.1632 201/401

S-BLS achieves 20–35 % RMSE reduction and prunes roughly 50 % of nodes.

4.2 CSTR Benchmark

  • System: Nonlinear continuous stirred tank reactor (CSTR) model under noisy, outlier-corrupted conditions.
  • Samples: 2,000.
  • Noise levels: γ=0.2,0.3,0.4\gamma = 0.2,0.3,0.4.
Noise RMSE (BLS) RMSE (S-BLS) Active/201
0.2 0.0644 0.0590 61 (30%)
0.3 0.0886 0.0809 61
0.4 0.1111 0.1031 61

S-BLS consistently achieves lower RMSE and prunes approximately 70% of nodes.

5. Computational Complexity and Comparison

Let NN denote the number of samples, LL the number of initial nodes, LaL_a the number remaining after pruning.

  • Standard BLS: One pseudoinverse, O(NL2+L3)\mathcal{O}(NL^2 + L^3).
  • Lasso-BLS (e.g. ISTA/ADMM): Per iteration O(NL)\mathcal{O}(NL), but typically requires many iterations.
  • S-BLS: Initialization is the same as standard BLS. Each STLS iteration (fixed T10T \approx 10):
    • Thresholding: O(LC)\mathcal{O}(LC), negligible.
    • Least squares on LaL_a columns: O(NLa2+La3)\mathcal{O}(NL_a^2 + L_a^3).
    • Total: O(NL2+L3)+TO(NLa2+La3)\mathcal{O}(NL^2 + L^3) + T\mathcal{O}(NL_a^2 + L_a^3).

Because LaLL_a \ll L after a few steps, the added cost is marginal relative to standard BLS and much lower than Lasso-based solutions. The sparsity also yields faster inference and improved generalization.

6. Practical Considerations and Applicability

  • Threshold Selection (λ\lambda): Should be aligned with the standard deviation of the noise. Cross-validation on the training set is recommended. Overly small λ\lambda retains noise; overly large values remove informative nodes.
  • Regularization Parameter (α\alpha): Linked to λ\lambda (via KKT conditions), but S-BLS optimizes λ\lambda directly.
  • Iterations (TT): Empirically, T=5T=5–$10$ suffices, as the active support stabilizes rapidly.
  • Parameterization of Nodes (n,k,m,qn,k,m,q): Moderate over-parameterization is beneficial; initial redundancy is pruned by STLS.
  • Robustness: S-BLS employs hard rather than soft thresholding (as in Lasso), making it particularly adept at rejecting noise-dominated components and handling outlier or non-Gaussian corruption.
  • Domains of Application: S-BLS is well-suited for system identification, control-oriented modeling, and any scenario prioritizing interpretability and computational speed.
  • Limitations: No global L0L_0-optimality guarantee. Performance relies on a quality initial overcomplete basis and careful hyperparameter selection.

In summary, S-BLS achieves a balance between analytical efficiency, model sparsity, and robustness to noise by integrating the STLS algorithm into the BLS framework, offering superior performance in both synthetic and real-world system identification benchmarks, without substantial computational overhead relative to standard BLS (Li, 22 Nov 2025).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Sparse Broad Learning System (S-BLS).