Quantum Temporal CNN (QTCNN)
- Quantum Temporal Convolutional Neural Network (QTCNN) is a hybrid architecture that combines classical multi-scale temporal encoders with quantum convolution circuits to predict equity returns.
- It employs efficient angle embedding and hierarchical, parameter-sharing quantum layers to harness superposition and entanglement for robust sequential modeling in noisy financial environments.
- Empirical evaluations on the JPX Tokyo Stock Exchange dataset demonstrate superior out-of-sample Sharpe ratios, highlighting its performance advantage over classical models.
The Quantum Temporal Convolutional Neural Network (QTCNN) is a hybrid quantum–classical architecture engineered for cross-sectional equity return prediction in noisy, nonstationary financial environments. QTCNN integrates a classical temporal encoder with hierarchical, parameter-efficient quantum convolutional circuits, designed to leverage quantum features such as superposition and entanglement for more robust and generalizable sequential modeling. The primary modeling pipeline consists of a classical temporal encoder ("TConv") that extracts multi-scale temporal features, followed by quantum angle embedding and convolutional processing, and culminating in a hybrid classification head. Empirical evaluation on the JPX Tokyo Stock Exchange dataset demonstrates superior out-of-sample Sharpe ratios compared to both classical and other quantum deep learning baselines (Chen et al., 7 Dec 2025).
1. Architectural Overview
QTCNN is composed of three major modules: (1) a classical temporal CNN encoder (TConv), (2) a quantum convolutional circuit, and (3) a hybrid classification stage. Input data , representing multi-day sequences of technical indicators per equity, is processed through these stages as follows:
- Classical Temporal Encoder: Extracts temporally compact feature .
- Quantum Convolutional Circuit: Embeds via angle embedding onto qubits, follows with hierarchical quantum convolution and pooling layers, and outputs a single-qubit expectation value .
- Hybrid Classification Head: Concatenates and , propagates through an MLP, and outputs probability prediction via sigmoid activation.
The canonical instantiation sets the qubit count , sequence length , and utilizes dozens of feature channels derived from engineered technical indicators.
2. Classical Temporal Encoder Design
Given a sequential input , the TConv encoder applies a stack of one-dimensional temporal convolutions:
- Convolutional Layer:
- Nonlinear Activation: ReLU or variant.
- Global Average Pooling (GAP):
- MLP Projection:
This encoder ensures efficient multi-scale temporal pattern extraction, including capturing momentum and volatility fluctuations. The projection to yields an -dimensional latent representation suitable for quantum encoding.
3. Quantum Convolutional Module
3.1 Data Encoding (“Angle Embedding”)
The feature vector is mapped to a quantum state using rotations independently applied to each qubit:
where .
3.2 Hierarchical Quantum Convolution and Pooling
Each convolutional layer applies a shared six-parameter unitary to adjacent pairs :
- Parameters are shared across all pairs within layer , reducing trainable quantum parameters.
- After each layer, pooling discards alternate qubits (odd indices), halving the register size. This is recursively applied over layers, resulting in depth and a final single-qubit state.
3.3 Quantum Measurement
The final state’s single qubit is measured in the basis to obtain output :
4. Hybrid Classification and Training
The quantum output is concatenated with classical embedding , forming . A two-layer MLP propagates using ReLU activations, with intermediate dimensionalities , and applies a sigmoid to yield probability prediction .
4.1 Loss Function and Optimization
Classification targets are binary labels , identifying extreme long/short equity candidates. The binary cross entropy objective is:
QTCNN uses AdamW optimization (, batch size 128, 50 epochs).
4.2 Quantum Gradient Evaluation
For quantum circuit parameters , observables’ gradients are computed via the parameter-shift rule:
These gradients interface with classical optimization frameworks using PennyLane’s autodifferentiation.
5. Benchmarking and Empirical Performance
5.1 Dataset, Feature Engineering, and Labeling
The principal evaluation utilizes daily JPX Tokyo Stock Exchange data (2017–2022), focusing on Universe0 stocks (2,000 equities). Feature engineering comprises momentum, volatility, and volume signals, all z-scored and winsorized cross-sectionally. Labels are assigned with weak supervision, selecting the top/bottom equities per day as positive/negative classes.
- Training set: First 80% temporal partition
- Testing set: Last 20% (out-of-sample, OOS)
- Subsampling with stride : 100 train days, 22 OOS days per run
- Evaluation metric: Annualized Sharpe ratio of long/short portfolios ( stocks per side), with bootstrap 95% confidence intervals
5.2 Comparative Results
| Model | OOS Sharpe Ratio | 95% CI |
|---|---|---|
| QTCNN | 0.538 | ±0.042 |
| QNN | 0.467 | ±0.038 |
| QCNN | 0.361 | ±0.298 |
| QLSTM | 0.333 | ±0.051 |
| Transformer | 0.313 | ±0.036 |
| LSTM | 0.044 | ±0.029 |
QTCNN achieves a Sharpe ratio of $0.538$, outperforming the best classical baseline (Transformer) by approximately , and surpassing quantum convolutional variants lacking temporal encoders by . This suggests a critical role for the temporal encoder and parameter-efficient quantum convolution structure in enhancing predictive robustness and generalization.
6. Quantum Circuit Implementation and Complexity
- Qubit count:
- Quantum depth: (determined by pooling hierarchy)
- Entanglement layout: Nearest-neighbor ring CNOTs, NISQ-compatible
- Simulation framework: PennyLane, using default.qubit/lightning backends
- Per-iteration wall time: QTCNN 0.25 s, LSTM 0.002 s (hardware: RTX3090 + i9 CPU)
The implementation leverages parameter sharing and pooling to optimize sample efficiency and scalability with limited qubit resources.
7. Limitations and Prospective Directions
Key limitations include the modest OOS sample (22 days) and extensive subsampling (9% of available days), leading to reduced statistical power. All experiments utilize idealized (noiseless) qubit simulation; future work must address robustness to quantum noise (e.g., depolarizing, amplitude damping channels) and physical device constraints. Feature engineering is restricted to price-volume technical indicators; incorporating sentiment, macroeconomic, or order-book data could potentially enrich the representational scope. Generalizability to other global equity markets and asset classes remains unverified. Additionally, theoretical investigation into expressivity, barren-plateau phenomena, and conditions for quantum advantage in financial time series models is an open avenue.
In summary, QTCNN represents a hybrid paradigm in quantum machine learning, synthesizing classical multi-scale temporal encoding with hierarchical, parameter sharing quantum circuits to address predictive challenges in complex financial data. Empirical results indicate substantial outperformance relative to canonical deep learning and hybrid quantum approaches, underscoring the practical relevance of quantum-enhanced forecasting for quantitative finance (Chen et al., 7 Dec 2025).