fMRI-to-Gesture Reconstruction Networks
- The paper presents an end-to-end framework that maps raw fMRI volumes to 2D gaze coordinates without relying on template-based co-registration.
- It utilizes advanced techniques like morphological eyeball isolation, 3D Retina-Net localization, and residual network regression to enhance computational efficiency and accuracy.
- Results indicate improved metrics including lower mean absolute error and higher Pearson correlation, enabling robust, real-time gaze tracking across varied cognitive tasks.
fMRI2GES refers to the automated reconstruction of eye gaze points from functional magnetic resonance imaging (fMRI) data, as operationalized in the MRGazer framework. This approach dispenses with traditional co-registration to a standard template and instead achieves gaze estimation in “individual space” through end-to-end learning, with substantial advances in both computational efficiency and accuracy compared to prior art that relies on co-registered eyeball masks. The method is characterized by an initial step of eyeball localization—either via morphological analysis or a trainable 3D Retina-Net—followed by residual network-based gaze regression. fMRI2GES establishes a robust, template-free pipeline for camera-less gaze tracking in cognitive neuroscience experiments (Wu et al., 2023).
1. Formal Problem Setting
The core task is to learn a mapping from raw or minimally processed fMRI volumes to the corresponding 2D eye gaze point. Given an fMRI sample (or for time sequences), the goal is to predict , the gaze coordinates on the stimulus screen. The mapping is implemented by a parameterized function , learning to minimize the mean squared error: where . This direct regression framing enables applicability across a diverse array of tasks, from controlled fixation to naturalistic viewing and saccade paradigms.
2. Eyeball Localization and Preprocessing
Accurate eyeball localization is fundamental, as downstream gaze inference exploits local fMRI signals from the ocular globes. Two main strategies are defined:
- Morphological Eyeball Isolation (Offline): After standard motion correction, a rough region of interest (ROI) is binarized at , where and . A morphological opening operation disconnects brain and ocular signals, and connected-component analysis isolates the two largest non-brain clusters as eyeball candidates. Axis-aligned bounding boxes (usually voxels) are extracted per eye.
- 3D Retina-Net Eyeball Detector (Online): A trainable pipeline uses a full-brain input (common: ) and processes through 3D Res-Net blocks. An FPN fuses multiscale features, and per-level heads predict "eye" versus "background" via focal loss and refine bounding boxes through regression. This module achieves operation at 0.01 s/volume in inference.
These approaches eliminate the need for template-based co-registration and promote efficient processing at scale (Wu et al., 2023).
3. Residual Network-Based Gaze Regression
Following extraction, each cropped eyeball volume serves as input to a residual network regressor. The standard architecture involves two identical 12-layer 3D Res-Nets (one per spatial direction):
- Initial 3D convolution: (kernel , 64 filters)
- Three stacked residual blocks: each consists of two applied convolutions with skip connections
- Global average pooling, flattening
- Sequential fully connected layers (FC(512) ReLU FC(2048) ReLU FC(512) ReLU)
- Final linear output (no activation) to yield the predicted gaze coordinate
When outlier detection is required, the backbone remains but with binary classification heads and a cross-entropy loss. No geometric augmentations are applied by default, but random left/right flips are feasible.
4. Training Protocols, Datasets, and Metrics
fMRI2GES is validated across multiple datasets and gaze tasks:
- Datasets:
- Fixation (HBN Biobank Peer1–3): hundreds of volumes per subject, labeled for 27 discrete screen locations.
- Naturalistic viewing: e.g., “The Present” (∼214 volumes, participants for ground-truth eye tracker average).
- Pro-/anti-saccade (OpenNeuro ds000120/119): both training and held-out testing for generalization.
- Splits: Across-scan (train on some Peer datasets, test on others), across-individual (random split with held-out subjects), and fivefold (for saccade data).
- Optimization: Adam; regression lr , batch size 128, early-stop at 60 epochs.
- Performance Metrics:
- Mean absolute error (MAE, in visual degrees)
- Pearson correlation for each direction
- Euclidean error per volume
- Saccade: pixel-distance error, precision, recall, F1
| Task/Method | MAE (°) | MAE (°) | ||
|---|---|---|---|---|
| MRGazer (ResNet12) | 1.11±0.69 | 0.91±0.10 | 1.12±0.48 | 0.87±0.11 |
Top quartile outlier removal further refines error metrics. Naturalistic viewing yields against averaged eye tracking. Runtime is 0.02 s/volume (morphological extraction + Retina-Net + regression), markedly faster than template-based DeepMReye (0.3 s/volume).
5. Task-Specific Results and Comparative Analysis
In controlled fixation, MRGazer surpasses DeepMReye on Pearson- and achieves MAE, MAE (Wu et al., 2023). Across-individual test splits confirm generalizability (MAE, ). In naturalistic viewing, correlation with eye tracker ground truth is high. For pro-/anti-saccade, accuracy is competitive: anti-saccade "visual guide" achieves F1 of 88.5%.
The pipeline is robust to low SNR but sensitivity to anatomical variance or head motion remains. Outlier detection via classification achieves F1 of 88.2%.
A plausible implication is that individualized preprocessing and learning-based spatial localization are critical for robust, generalizable fMRI2GES performance across heterogeneous acquisition settings.
6. Technical Implementation and Environment
MRGazer is implemented under a standard modern deep learning stack. Key details:
- Hardware: NVIDIA Tesla V100 (16 GB), Intel Xeon Gold 6230R, 32 GB RAM.
- Software: Ubuntu 18.04, Python 3.8, PyTorch 1.8 + CUDA 10.2, torchvision, scikit-image, nibabel, numpy, scipy, scikit-learn.
- Modular codebase:
1. fmri_preproc.py: motion correction, intensity normalization 2. eye_extract.py: ROI / bounding box logic 3. retina_net.py: Retina-Net training/inference 4. resnet_gaze.py: 3D Res-Net regression/outlier detection 5. train.py / eval.py: orchestration
Total pipeline achieves real-time (50 Hz) operation for single-volume inference.
7. Limitations and Prospects
- Eyeball Extraction Robustness: Morphological methods may fail under low-SNR or anatomical variability. Parameter tuning and alternative segmenters (e.g., 3D U-Net) are advised.
- Regression Heads: Separate networks for x/y could be unified to improve parameter efficiency and reduce overfitting.
- Generalization: Screen calibration parameters are assumed constant; cross-scanner adaptation requires new strategies.
- Head Motion and Demographics: Performance is degraded in high-motion or young subjects. Integrating movement estimates or demographic covariates may mitigate this.
- Future Extensions: Performance on blinks or pursuit, higher temporal resolution, and leveraging multimodal (e.g., anatomical, head-pose) inputs are active areas of investigation.
By combining bespoke spatial preprocessing, lightweight end-to-end learning, and carefully benchmarked evaluation, fMRI2GES establishes the state-of-the-art for non-invasive, template-free eye tracking from fMRI, and offers a modular platform for further exploration of visual attention and oculomotor control (Wu et al., 2023).