Dual Relation Fusion Network (DRFN)
- The paper introduces DRFN, a neural framework that fuses long-term static relations with short-term market dynamics to improve stock prediction.
- It employs cross-attention and recurrent fusion modules to integrate multimodal inputs and capture both persistent and transient market trends.
- Empirical evaluations show DRFN achieves lower RMSE and MAE with higher Pearson correlations in forecasting next-day returns on US and Chinese markets.
A Dual Relation Fusion Network (DRFN) is a neural framework designed to jointly model heterogeneous relational structures by recursively integrating static (long-term) and dynamic (short-term) relations. In financial prediction, DRFN leverages temporal multimodal inputs—such as quantitative indicators and qualitative news—alongside prior knowledge of business connections to forecast stock returns with enhanced sensitivity to co-movement and abrupt market shifts (Chen et al., 12 Oct 2025).
1. Foundational Motivation and Conceptual Design
The DRFN architecture is motivated by the observed complementarity between static relational structures—capturing persistent inter-stock connections—and dynamic relations that reflect rapid changes in market conditions. Existing methods typically operate with either fixed structural priors or exclusively learn short-term interactions, thus neglecting the interaction between stable and volatile network dynamics. DRFN is specifically constructed to address this, enabling robust operation under both long-term market regularities and sudden informational shocks.
The static relation matrix () encodes predefined business interdependencies, often sourced from structured knowledge graphs such as Wikidata. Its limitation is the absence of ultra-short-term market information, notably overnight effects. DRFN resolves this via a relative static relation module, fusing with the dynamic relation matrix from the previous trading session () using learnable weights and balancing coefficients.
2. Static Relation Component
The relative static relation module updates the conventional static relation matrix to incorporate time-varying patterns and overnight influences. The module realizes this integration by applying the following computation:
where is a learnable parameter tensor modulating static importance, denotes Hadamard (element-wise) multiplication, and is a learnable scalar balancing the contributions of static and dynamic sources. This recurrent update mechanism enables the static topology to reflect not only enterprise embeddedness but also temporally conditioned relational strengths.
A plausible implication is that networks using only will underestimate short-term volatility, while DRFN’s fusion exploits the directional flow of information from recent market events into the otherwise stationary relational graph.
3. Dynamic Relation Mechanism
The dynamic relation module of DRFN captures evolving short-term interdependencies through a distance-aware cross-attention mechanism. Temporal multimodal features—news () and market indicators ()—are first fused:
Resulting hidden states are processed to obtain positive (synergistic) and negative (adversarial) relation matrices:
where and denotes the hidden dimension. The L₁ distance between stock representations,
directly modulates the initial attention score via exponential weighting:
A cross-attention strategy is employed to synthesize positive and negative relations:
The final dynamic relation matrix is projected via a GELU activation:
This process enables DRFN to respond to and encode rapid market transitions in the latent inter-stock relationship graph.
4. Recurrent Fusion of Static and Dynamic Relations
The key innovation in DRFN lies in recurrent bidirectional cross-attention fusion of static and dynamic relational views. After parallel formation of and , these are integrated:
with the standard scaled dot-product attention:
This bidirectional integration allows the long-term relational graph to dynamically evolve by recurrently incorporating new short-term evidence, while retaining the stabilizing anchors of business structure.
This suggests the DRFN recurrent architecture is uniquely equipped to maintain sensitivity to evolving interdependencies, preserving temporal continuity without omitting abrupt changes.
5. Performance Benchmarking and Sensitivity
Empirical evaluation of DRFN on datasets spanning US and Chinese stock markets demonstrates superior performance relative to LSTM, GRU, GCN, GAT, MASTER, SAMBA, and LSR-IGRU (Chen et al., 12 Oct 2025). Notably, DRFN achieves lower RMSE and MAE in next-day return forecasting, validating the utility of dual relation fusion.
Ablation studies confirm that either static-only or dynamic-only modules yield inferior accuracy. The comprehensive fusion, including multimodal feature alignment and adaptive residual connections, produces higher Pearson correlation coefficients between predicted relational strength changes and actual price movements. In case comparisons (e.g., AAPL, AMZN), DRFN more accurately tracks both direction and magnitude of price changes, evidencing improved responsiveness.
6. Mathematical Formulations and Predictive Output
A summary of principal mathematical components applied in DRFN is tabulated below:
| Module | Formulation Example | Role |
|---|---|---|
| Static Relation Update | Fusing prior dynamic/static relations | |
| Dynamic Relation Scoring | Measure synergistic/adversarial links | |
| Distance Weighting | Enhances attention by proximity | |
| Fusion Attention | Combines static/dynamic views | |
| Output Prediction | Forecast next-day stock price |
Final predictions are produced via an adaptive residual fusion layer, accounting for stock-specific weights and the contributions of both news and market features.
7. Practical Implications and Interpretability
DRFN’s dual relation approach facilitates robust forecasting across different market regimes, benefiting applications requiring both rapid responsiveness to market news and structural stability provided by long-established relations. Its multimodal integration of qualitative and quantitative signals enhances versatility and transferability. The modular architecture, with explicit attention pathways and recurrent updates, delivers interpretability regarding which relations and modalities primarily drive predictive accuracy.
A plausible implication is that DRFN’s explicit separation and recurrent fusion of static and dynamic relations lays the groundwork for extending similar architectures to domains where evolving multi-modal and multi-relational data are prevalent (e.g., multi-agent systems, industrial supply chains, macroeconomic indicators). The attention-based modules facilitate post-hoc analysis of relational influence—enabling targeted interventions in algorithmic trading or risk assessment workflows.