Sparse Broad Learning System (S-BLS)
- S-BLS is a model framework that improves robustness and interpretability of Broad Learning Systems by integrating Sequential Threshold Least-Squares.
- It employs a sparse output weight learning mechanism to prune noisy nodes while maintaining predictive performance.
- Experimental evaluations show a 20–35% RMSE reduction and significant node pruning in both nonlinear system identification and CSTR benchmarks.
The Sparse Broad Learning System (S-BLS) is a model framework developed to enhance the robustness and interpretability of Broad Learning Systems in the presence of sensor noise and outliers, which are common in industrial and nonlinear system identification tasks. S-BLS integrates a Sequential Threshold Least-Squares (STLS) mechanism for learning sparse output weights, offering significant improvements in generalization, efficiency, and noise resilience compared to standard ridge-regression-based BLS solutions (Li, 22 Nov 2025).
1. Structure of the Broad Learning System
The conventional Broad Learning System (BLS) is a shallow network architecture designed for efficiently mapping input data to target outputs with reduced computational complexity. For input data , BLS constructs two primary layers before linear output regression:
- Feature Mapping Layer: Input samples are transformed via random weights and nonlinearity into feature groups, each of size ,
which are concatenated to form .
- Enhancement Layer: The mapped features are further expanded via enhancement groups (size ) with another nonlinearity ,
and concatenated to yield .
The system input for regression, , with , feeds a linear read-out to match the targets .
Standard BLS computes by ridge-regularized least squares (pseudoinverse), leading to a solution that is typically dense, with each node contributing to the output, thus propagating noise and risking overfitting in non-ideal measurement environments.
2. Sparse Output-Weight Learning via Sequential Threshold Least-Squares
S-BLS modifies the output-weight learning by introducing direct sparsity in using an -penalized regression objective,
Since obtaining the exact solution is NP-hard, S-BLS adopts the Sequential Threshold Least-Squares (STLS) strategy, which alternates between elementwise hard-thresholding and least-squares projection on active variables.
STLS Steps:
- Hard Thresholding: Zero out weights via
- Least-Squares Projection: Restrict to active columns where . Compute restricted least squares,
with all other weights fixed at zero.
- Iteration: Start with and apply thresholding and projection for a fixed number (typically $5$–$10$) of iterations.
Following this procedure results in a sparse that selectively removes noise-influenced nodes while maintaining predictive performance.
3. Training Procedure and Implementation
The S-BLS training process follows these steps:
- Random Layer Construction: Sample for all feature and enhancement nodes.
- Layer Computation: Compute all , , and concatenate to form .
- Initialization: Compute .
- Iterative STLS: For ,
- Hard-threshold ,
- For each output , re-solve least squares on active set .
- Termination: Return the final sparse output matrix .
This approach preserves the efficient analytical nature of classical BLS while achieving node-level sparsity that benefits downstream interpretability.
4. Experimental Evaluation
S-BLS has been extensively evaluated in two settings:
4.1 Nonlinear System Identification
- Model: A nonlinear difference equation,
with stochastic in training and sinusoidal in test.
- Noise: Gaussian/uniform, with levels to $0.4$.
- Metrics: Test RMSE and sparsity ratio (active/L).
| Noise | RMSE (BLS) | RMSE (S-BLS) | Active/L |
|---|---|---|---|
| 0.1 | 0.3370 | 0.2588 | 201/401 (50.1%) |
| 0.2 | 0.2222 | 0.1673 | 201/401 |
| 0.3 | 0.2361 | 0.1648 | 201/401 |
| 0.4 | 0.2528 | 0.1632 | 201/401 |
S-BLS achieves 20–35 % RMSE reduction and prunes roughly 50 % of nodes.
4.2 CSTR Benchmark
- System: Nonlinear continuous stirred tank reactor (CSTR) model under noisy, outlier-corrupted conditions.
- Samples: 2,000.
- Noise levels: .
| Noise | RMSE (BLS) | RMSE (S-BLS) | Active/201 |
|---|---|---|---|
| 0.2 | 0.0644 | 0.0590 | 61 (30%) |
| 0.3 | 0.0886 | 0.0809 | 61 |
| 0.4 | 0.1111 | 0.1031 | 61 |
S-BLS consistently achieves lower RMSE and prunes approximately 70% of nodes.
5. Computational Complexity and Comparison
Let denote the number of samples, the number of initial nodes, the number remaining after pruning.
- Standard BLS: One pseudoinverse, .
- Lasso-BLS (e.g. ISTA/ADMM): Per iteration , but typically requires many iterations.
- S-BLS: Initialization is the same as standard BLS. Each STLS iteration (fixed ):
- Thresholding: , negligible.
- Least squares on columns: .
- Total: .
Because after a few steps, the added cost is marginal relative to standard BLS and much lower than Lasso-based solutions. The sparsity also yields faster inference and improved generalization.
6. Practical Considerations and Applicability
- Threshold Selection (): Should be aligned with the standard deviation of the noise. Cross-validation on the training set is recommended. Overly small retains noise; overly large values remove informative nodes.
- Regularization Parameter (): Linked to (via KKT conditions), but S-BLS optimizes directly.
- Iterations (): Empirically, –$10$ suffices, as the active support stabilizes rapidly.
- Parameterization of Nodes (): Moderate over-parameterization is beneficial; initial redundancy is pruned by STLS.
- Robustness: S-BLS employs hard rather than soft thresholding (as in Lasso), making it particularly adept at rejecting noise-dominated components and handling outlier or non-Gaussian corruption.
- Domains of Application: S-BLS is well-suited for system identification, control-oriented modeling, and any scenario prioritizing interpretability and computational speed.
- Limitations: No global -optimality guarantee. Performance relies on a quality initial overcomplete basis and careful hyperparameter selection.
In summary, S-BLS achieves a balance between analytical efficiency, model sparsity, and robustness to noise by integrating the STLS algorithm into the BLS framework, offering superior performance in both synthetic and real-world system identification benchmarks, without substantial computational overhead relative to standard BLS (Li, 22 Nov 2025).