Next Generation Reservoir Computing (2106.07688v2)
Abstract: Reservoir computing is a best-in-class machine learning algorithm for processing information generated by dynamical systems using observed time-series data. Importantly, it requires very small training data sets, uses linear optimization, and thus requires minimal computing resources. However, the algorithm uses randomly sampled matrices to define the underlying recurrent neural network and has a multitude of metaparameters that must be optimized. Recent results demonstrate the equivalence of reservoir computing to nonlinear vector autoregression, which requires no random matrices, fewer metaparameters, and provides interpretable results. Here, we demonstrate that nonlinear vector autoregression excels at reservoir computing benchmark tasks and requires even shorter training data sets and training time, heralding the next generation of reservoir computing.
Summary
- The paper introduces NG-RC, a nonlinear vector autoregression method that replaces complex random reservoirs with engineered time-delayed feature vectors for modeling dynamical systems.
- It achieves efficient forecasting with up to 33-162x faster training speeds and state-of-the-art performance on chaotic systems using as few as 400 training points.
- The method trains only the output weights via regularized linear regression, significantly reducing metaparameter tuning and eliminating randomness issues in traditional reservoir computing.
This paper, "Next Generation Reservoir Computing" (Next Generation Reservoir Computing, 2021), introduces Nonlinear Vector Autoregression (NVAR), termed the Next Generation Reservoir Computer (NG-RC), as a practical and more efficient alternative to traditional Reservoir Computing (RC) for modeling and forecasting dynamical systems from time-series data.
Traditional RC, while effective for dynamical systems tasks and less data-hungry than some deep learning methods, relies on randomly initialized and fixed reservoir connections and input weights. This introduces variability in performance depending on the random matrices used and necessitates extensive tuning of numerous metaparameters. The NG-RC addresses these issues by demonstrating that the functional capability of a traditional RC with a linear reservoir and nonlinear output layer is mathematically equivalent to an NVAR model.
The core idea of the NG-RC is to replace the complex, random recurrent neural network (the reservoir) with a simple feature vector constructed directly from time-delayed observations of the input data and nonlinear functions of these observations. This feature vector, Ototal,i, for a given time step i, is typically composed of:
- A constant term (c).
- Linear features (Olin,i): Concatenation of the input vector X at the current time ti and at k−1 previous time steps, spaced by s. If Xi is d-dimensional, Olin,i=Xi⊕Xi−s⊕Xi−2s⊕…⊕Xi−(k−1)s. This results in a feature vector of size d×k.
- Nonlinear features (Ononlin,i): Nonlinear functions of the linear features, typically monomials (polynomials). For a quadratic feature vector Ononlinear(2), it includes all unique quadratic terms of the elements in Olin,i. A p-order polynomial feature vector Ononlinear(p) consists of all unique monomials of order up to p of the elements in Olin,i.
The total feature vector is Ototal=c⊕Olin⊕Ononlin.
The output Yi+1 is then computed as a linear transformation of this feature vector:
Yi+1=WoutOtotal,i+1
Instead of directly predicting Yi+1, the paper shows improved performance by training the NG-RC to predict the difference between consecutive time steps, effectively learning the "flow" of the dynamical system:
Yi+1=Yi+WoutOtotal,i+1
The output weights Wout are the only parameters that need training. This is done using supervised learning via regularized linear least squares (specifically, Tikhonov regularization or ridge regression). Given a training dataset of input-output pairs (Ototal,Ytarget), the optimal Wout is found by minimizing a cost function that includes both the mean squared error and a regularization term penalizing large weights:
Wout=YtargetOtotalT(OtotalOtotalT+αI)−1
Here, α is the regularization parameter controlling the strength of the penalty, and I is the identity matrix. This linear optimization is computationally efficient and avoids complex backpropagation or reservoir state optimization.
Practical Implementation Aspects and Advantages:
- Reduced Data Requirement: NG-RC requires significantly less training data compared to traditional RC or deep learning methods. The authors demonstrate state-of-the-art performance on benchmark chaotic systems (Lorenz63 and a double-scroll circuit) using only 400 training data points, a drastically smaller amount than typical requirements for traditional RCs (thousands to millions of points). The minimum data needed is hypothesized to be related to the number of unknown fit parameters (Ntotal×d), plus some overhead for generalization.
- Faster Training: Training involves constructing the feature matrix from the time series and performing a single linear regression. This is computationally much cheaper than training the weights of a recurrent neural network. The paper estimates speed-ups of 33-162x compared to an efficient traditional RC and over 10⁶x for a high-accuracy traditional RC for the Lorenz63 task.
- Fewer Metaparameters: The primary metaparameters to tune are k (number of delays), s (delay spacing), the polynomial order p, and the regularization parameter α. This is a much smaller and more interpretable set than the large number of parameters in traditional RC reservoirs (spectral radius, connectivity sparsity, input scaling, etc.). The authors find small values of k and low polynomial orders (p=2,3) are often sufficient.
- Shorter Warm-up Period: Traditional RCs require a "warm-up" phase to allow the reservoir state to become independent of initial conditions. NG-RC only needs s×k steps to populate the initial feature vector, which is typically very short (e.g., 2 steps for k=2,s=1).
- Avoids Randomness Issues: Since there is no randomly generated reservoir matrix, the performance is deterministic given the hyperparameters, removing the need to generate and test multiple random instances.
- Interpretability: Although not deeply explored in the paper, the trained output weights Wout directly map combinations of past states (Ototal) to future changes. Analyzing Wout can potentially provide insights into which past states and nonlinear interactions are most predictive of the system's dynamics.
Application Examples Demonstrated in the Paper:
- Forecasting Chaotic Dynamics: NG-RC successfully forecasts the dynamics of Lorenz63 and the double-scroll circuit for several Lyapunov times and reconstructs their strange attractors, matching the long-term statistical properties ('climate'). This is achieved with minimal training data and computational cost.
- Inferring Unseen Dynamics: NG-RC can infer missing variables of a dynamical system. For Lorenz63, it was trained on all three variables (x,y,z) but tested by inferring z given only x and y. This is relevant for applications where only partial state information is available.
Implementation Details & Code:
The paper notes that the implementation uses standard Python libraries like NumPy and SciPy. The core steps for implementation would be:
- Data Collection: Obtain time-series data X(t) from the dynamical system. Discretize it with a suitable time step dt.
- Feature Engineering: For each time step i in the training data, construct the feature vector Ototal,i by collecting delayed inputs (Xi,Xi−s,…,Xi−(k−1)s) and their chosen nonlinear combinations (e.g., quadratic or cubic monomials). Include a constant term if desired.
- Target Variable Preparation: Determine the target output Ytarget,i. For forecasting the flow, this would be Ytarget,i=Xi+1−Xi. For direct prediction, it would be Ytarget,i=Xi+1.
- Training: Stack the feature vectors Ototal,i row-wise into a matrix Ototal and the target variables Ytarget,i row-wise into a matrix Ytarget. Compute the output weight matrix Wout using the Tikhonov regularization formula:
1 2 3 4 5 6 7 8
# Using NumPy for linear algebra # O_total is the matrix where each row is a feature vector # Y_target is the matrix where each row is the target output # alpha is the regularization parameter identity_matrix = np.eye(O_total.shape[1]) # W_out = Y_target @ O_total.T @ np.linalg.inv(O_total @ O_total.T + alpha * identity_matrix) # Potentially unstable # More stable using least squares solver W_out = np.linalg.solve(O_total.T @ O_total + alpha * identity_matrix, O_total.T @ Y_target).T
- Prediction/Forecasting:
- Start with the last known state Xlast.
- Construct the feature vector Ototal,last using Xlast and required past states Xlast−s,….
- Predict the change (or next state): ΔX=WoutOtotal,last (or Xnext=WoutOtotal,last).
- Update the state: Xnext=Xlast+ΔX (or Xnext=Xnext).
- Repeat, using the predicted Xnext as the input for the next step's feature vector calculation.
The authors provide their code publicly on GitHub (link provided in the paper), which can serve as a direct reference for implementation.
Computational Considerations:
The dominant computational cost during training is solving the linear system for Wout, which is O(MtrainNtotal2+Ntotal3) where Mtrain is the number of training points and Ntotal is the number of features. During forecasting, the cost per time step is simply multiplying the feature vector by Wout, which is O(Ntotal×dout) where dout is the dimension of the output. Since Mtrain and Ntotal are typically small in NG-RC compared to traditional methods, the overall cost is low. The number of features Ntotal grows polynomially with d and k, and with the polynomial order p. Choosing small values for these parameters is key to computational efficiency.
Robustness and Limitations:
The paper demonstrates robustness to noise by showing that increasing the regularization parameter α allows the NG-RC to learn the underlying deterministic system even in the presence of significant noise. While the paper focuses on low-dimensional systems, the theoretical equivalence suggests applicability to high-dimensional systems where traditional RC has been successful. Future work includes exploring better feature selection methods (like LASSO) or kernel methods, especially when Ntotal becomes large.
In summary, the NG-RC offers a computationally efficient, data-thrifty, and conceptually simpler method for modeling and forecasting dynamical systems by leveraging the power of nonlinear regression on time-delayed input features, providing a strong practical alternative to traditional reservoir computing approaches.
Related Papers
- Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification (2015)
- Hybridizing Traditional and Next-Generation Reservoir Computing to Accurately and Efficiently Forecast Dynamical Systems (2024)
- Optical next generation reservoir computing (2024)
- Machine Learning: a Lecture Note (2025)
- SLOT: Sample-specific Language Model Optimization at Test-time (2025)