PAST: Primary-Auxiliary Spatio-Temporal Network
- The paper demonstrates that decomposing spatio-temporal data into primary internal dependencies and auxiliary contextual influences enables robust imputation under diverse missingness regimes.
- It leverages a Graph-Integrated Module to learn dynamic node relations and a Cross-Gated Module to fuse external timestamps and node metadata effectively.
- Empirical results reveal up to 26.2% RMSE and 31.6% MAE improvements over state-of-the-art baselines, highlighting its practical significance in traffic data reconstruction.
The Primary-Auxiliary Spatio-Temporal Network (PAST) is a neural framework designed for traffic time series imputation under severe and heterogeneous missing data conditions. Unlike conventional methods that separately or statically model spatial and temporal patterns, PAST explicitly decomposes dependencies into primary patterns—internal relationships between time series nodes—and auxiliary patterns—influences from external features such as timestamps or node metadata. Its architecture integrates a Graph-Integrated Module (@@@@3@@@@) for learning primary spatio-temporal dependencies and a Cross-Gated Module (CGM) for embedding and modulating auxiliary contextual information, enabling robust estimation of missing values across random, block, and fiber-type missingness regimes (Hu et al., 17 Nov 2025).
1. Conceptual Foundations
PAST formalizes the imputation challenge by distinguishing between primary and auxiliary spatio-temporal patterns. Primary patterns refer to intrinsic sensor or node relations inferred directly from available measurements, while auxiliary patterns encode exogenous covariates, including temporal (e.g., time-of-day, day-of-week) or static node characteristics. This dichotomy addresses limitations in prior imputation models that use static, hand-crafted graphs or simply concatenate auxiliary information, which impedes adaptability to dynamic and irregular missingness.
The model is structured into two interacting modules:
- The Graph-Integrated Module (GIM) is responsible for dynamic inference of node-to-node relations and propagation of hidden representations capturing primary patterns.
- The Cross-Gated Module (CGM) injects auxiliary features via learnable gating and bidirectional fusion with the shared representation.
Configuring the model in this manner allows PAST to adaptively reconstruct missing data even when intrinsic or extrinsic dependencies are inconsistently observable in the input.
2. Graph-Integrated Module (GIM): Dynamic Spatio-Temporal Modeling
GIM operationalizes primary pattern extraction by constructing a time-varying, fully-connected node adjacency matrix at each time step, computed directly from the shared hidden state . This is achieved using a learnable projection :
where is a temperature parameter. This adjacency is interpreted as encoding the current, context-dependent affinities between all node pairs.
An alternative attention-based variant employs a LeakyReLU attention:
Crucially, GIM does not assume prior spatial proximity but learns emergent, possibly non-local dependencies at each time step.
3. Interval-Aware Dropout: Regularization under Missingness
To prevent overfitting to direct node connections and to model the disruptions induced by missing data, GIM integrates an interval-aware dropout mechanism that masks edges in probabilistically based on the history of observed co-occurrence for each node pair. For each pair:
where is the time since the last observed co-occurrence, is a base dropout rate, and controls decay. This mechanism compels the model to utilize multi-hop, indirect interactions—especially when strong direct evidence is unavailable.
In ablation experiments, removing dynamic adjacency (switching to static/geographic ) increased RMSE by 9.4% and MAE by 8.7% under 50% random missing; removing interval-aware dropout increased RMSE by 4.1% (Hu et al., 17 Nov 2025).
4. Multi-Order Graph Convolutions
The information propagation backbone of GIM leverages layered graph convolutions of up to orders:
where is the diagonal node degree matrix. The -th layer update is: Here, controls the receptive field, enabling non-local aggregation as increases. Ablation results demonstrate that reducing from 3 to 1 increased RMSE by 6.8%, emphasizing the importance of multi-hop message passing in settings with extensive missingness.
5. Interaction with the Cross-Gated Module (CGM) and Fusion Mechanism
At each time , GIM generates an updated hidden state encoding primary dependencies; CGM computes based on auxiliary inputs. These are fused via a gated network: consists of sigmoid gates and linear projections, determining for each node and timepoint the mixing proportion of primary and auxiliary signals. This fusion enables the unified hidden state to simultaneously propagate internal and external information through backpropagation.
6. Self-Supervised Training and View Ensemble Strategy
PAST is trained via a masked reconstruction objective: at each time , a random subset of the traffic matrix is masked, and the model seeks to reconstruct these entries: where indicates unmasked entries.
Each training minibatch ensembles different random masking patterns, forwarding versions of input and averaging reconstructions: Loss is computed relative to to mitigate mask variance and promote robustness.
7. Hyperparameters, Computational Aspects, and Empirical Performance
Notable hyperparameters and implementation settings in PAST include:
- Number of GIM layers
- Max graph order or 3
- Hidden dimensionality
- Interval-aware dropout base rate , decay
- Number of masked views per minibatch
Computational complexity per step is for adjacency calculation and for graph convolutions (with in dense graphs but reduced after sparsification). Sparsification is achieved by keeping only the top- edges per node, and high-order normalized adjacencies are precomputed by recursive sparse-matrix multiplication.
Empirically, PAST outperforms seven state-of-the-art baselines by up to 26.2% in RMSE and 31.6% in MAE across three datasets and 27 missing data conditions, especially excelling under large-scale and non-random missing regimes (Hu et al., 17 Nov 2025). Its dynamic graph construction, interval-aware dropout, and multi-order propagation enable it to adaptively recover missing values where static or shallow models fail.
8. Comparative Analysis and Practical Significance
The central innovation of PAST lies in the synergy between dynamic, order- graph modeling for primary dependencies and learnable cross-gating with auxiliary signals. Unlike fixed graph approaches, PAST’s GIM adapts to shifting spatial-temporal correlations and routes information via alternative paths in the presence of missingness. Interval-aware dropout further regularizes the propagation process, forcing the model to plan for multi-hop contingencies. GIM’s multi-order convolutions grant a tunable receptive field, crucial for long-range dependency modeling under extended block or fiber missing.
A plausible implication is that the primary-auxiliary decomposition in PAST may generalize to other temporal graph imputation settings where external contextual information is available and internal dependencies are non-stationary across time. This architectural separation and integration mechanism addresses the frequently observed brittleness of previous methods when confronted with complex missingness patterns in large spatio-temporal systems (Hu et al., 17 Nov 2025).