Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
92 tokens/sec
Gemini 2.5 Pro Premium
51 tokens/sec
GPT-5 Medium
24 tokens/sec
GPT-5 High Premium
17 tokens/sec
GPT-4o
97 tokens/sec
DeepSeek R1 via Azure Premium
92 tokens/sec
GPT OSS 120B via Groq Premium
458 tokens/sec
Kimi K2 via Groq Premium
222 tokens/sec
2000 character limit reached

BrainWave: A Brain Signal Foundation Model for Clinical Applications (2402.10251v6)

Published 15 Feb 2024 in q-bio.NC, cs.AI, cs.LG, and eess.SP

Abstract: Neural electrical activity is fundamental to brain function, underlying a range of cognitive and behavioral processes, including movement, perception, decision-making, and consciousness. Abnormal patterns of neural signaling often indicate the presence of underlying brain diseases. The variability among individuals, the diverse array of clinical symptoms from various brain disorders, and the limited availability of diagnostic classifications, have posed significant barriers to formulating reliable model of neural signals for diverse application contexts. Here, we present BrainWave, the first foundation model for both invasive and non-invasive neural recordings, pretrained on more than 40,000 hours of electrical brain recordings (13.79 TB of data) from approximately 16,000 individuals. Our analysis show that BrainWave outperforms all other competing models and consistently achieves state-of-the-art performance in the diagnosis and identification of neurological disorders. We also demonstrate robust capabilities of BrainWave in enabling zero-shot transfer learning across varying recording conditions and brain diseases, as well as few-shot classification without fine-tuning, suggesting that BrainWave learns highly generalizable representations of neural signals. We hence believe that open-sourcing BrainWave will facilitate a wide range of clinical applications in medicine, paving the way for AI-driven approaches to investigate brain disorders and advance neuroscience research.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (61)
  1. EEG Signal Analysis for Diagnosing Neurological Disorders Using Discrete Wavelet Transform and Intelligent Techniques. Sensors 20 (2020).
  2. Diego Alvarez-Estevez and Roselyne Rijsman. 2022. Haaglanden Medisch Centrum sleep staging database (version 1.1). https://doi.org/10.13026/t79q-fr32.
  3. Diego Alvarez-Estevez and Roselyne M Rijsman. 2021. Inter-database validation of a deep learning approach for automatic sleep scoring. PloS one 16 (2021).
  4. Sequential Modeling Enables Scalable Learning for Large Vision Models. arXiv:2312.00785 [cs.CV]
  5. AASM scoring manual updates for 2017 (version 2.4).
  6. On the Opportunities and Risks of Foundation Models. arXiv:2108.07258 [cs.LG]
  7. MBrain: A Multi-Channel Self-Supervised Learning Framework for Brain Signals. In Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining.
  8. EEG/SEEG Signal Modelling using Frequency and Fractal Analysis.. In BIOSIGNALS.
  9. BrainNet: Epileptic Wave Detection from SEEG with Hierarchical Graph Diffusion Learning. In Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining.
  10. Previous, current, and future stereotactic EEG techniques for localising epileptic foci. Expert Review of Medical Devices 19 (2022), 571–580.
  11. Paolo Detti. 2020. Siena Scalp EEG Database (version 1.0.0). https://doi.org/10.13026/5d4a-j060.
  12. EEG Synchronization Analysis for Seizure Prediction: A Study on Data of Noninvasive Recordings. Processes 8 (2020).
  13. Improved manual annotation of EEG signals through convolutional neural network guidance. Eneuro 9 (2022).
  14. SimMTM: A Simple Pre-Training Framework for Masked Time-Series Modeling. In Thirty-seventh Conference on Neural Information Processing Systems.
  15. Differential entropy feature for EEG-based emotion classification. In 6th International IEEE/EMBS Conference on Neural Engineering (NER).
  16. Stereotactic EEG Practices: A Survey of United States Tertiary Referral Epilepsy Centers. Journal of Clinical Neurophysiology 39 (2022).
  17. PhysioBank, PhysioToolkit, and PhysioNet: Components of a New Research Resource for Complex Physiologic Signals. Circulation 101 (2000).
  18. John Guttag. 2010. CHB-MIT Scalp EEG Database. PhysioNet. https://doi.org/10.13026/C2K01R
  19. The TUH EEG CORPUS: A big data resource for automated EEG interpretation. In 2014 IEEE signal processing in medicine and biology symposium (SPMB).
  20. Exploiting Interactivity and Heterogeneity for Sleep Stage Classification Via Heterogeneous Graph Neural Network. In ICASSP 2023-2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
  21. Analysis of EEG Signal for the Detection of Brain Abnormalities.
  22. Kranti Kamble and Joydeep Sengupta. 2023. A comprehensive survey on emotion recognition based on electroencephalograph (EEG) signals. Multimedia Tools and Applications (2023), 1–36.
  23. Analysis of a sleep-dependent neuronal feedback loop: the slow-wave microcontinuity of the EEG. IEEE Transactions on Biomedical Engineering 47 (2000).
  24. Cost-effectiveness analysis of invasive EEG monitoring in drug-resistant epilepsy. Epilepsy & Behavior 114 (2021).
  25. Beyond Scale: the Diversity Coefficient as a Data Quality Metric Demonstrates LLMs are Pre-trained on Formally Diverse Data. arXiv:2306.13840 [cs.CL]
  26. From Quantity to Quality: Boosting LLM Performance with Self-Guided Data Selection for Instruction Tuning. arXiv:2308.12032 [cs.CL]
  27. Removing artefacts and periodically retraining improve performance of neural network-based seizure prediction models. Scientific Reports (2023).
  28. Ilya Loshchilov and Frank Hutter. 2017. Decoupled weight decay regularization. arXiv preprint arXiv:1711.05101 abs/1711.05101 (2017).
  29. Zhengqing Miao and Meirong Zhao. 2023a. Time-space-frequency feature Fusion for 3-channel motor imagery classification. arXiv preprint arXiv:2304.01461 abs/2304.01461 (2023).
  30. Zhengqing Miao and Meirong Zhao. 2023b. Time-space-frequency feature Fusion for 3-channel motor imagery classification. arXiv:2304.01461 [cs.LG]
  31. Santiago Morales and Maureen Bowers. 2022. Time-frequency analysis methods and their application in developmental EEG data. Developmental Cognitive Neuroscience 54 (2022).
  32. Multicenter intracranial EEG dataset for classification of graphoelements and artifactual signals. Scientific data 7 (2020).
  33. A Time Series is Worth 64 Words: Long-term Forecasting with Transformers. In The Eleventh International Conference on Learning Representations.
  34. XSleepNet: Multi-view sequential model for automatic sleep staging. IEEE Transactions on Pattern Analysis and Machine Intelligence 44 (2021).
  35. Huy Phan and Kaare Mikkelsen. 2022. Automatic sleep staging of EEG signals: recent development, challenges, and future directions. Physiological Measurement 43 (2022).
  36. L-SeqSleepNet: Whole-cycle Long Sequence Modeling for Automatic Sleep Staging. IEEE Journal of Biomedical and Health Informatics 27 (2023).
  37. Language models are unsupervised multitask learners. OpenAI blog 1 (2019).
  38. Searching for Activation Functions. CoRR abs/1710.05941 (2017).
  39. BCI2000: A General-Purpose Brain-Computer Interface (BCI) System. IEEE Transactions on Biomedical Engineering 51 (2004).
  40. Ali Hossam Shoeb. 2009. Application of machine learning to epileptic seizure onset detection and treatment. Ph. D. Dissertation. Massachusetts Institute of Technology.
  41. EEG Emotion Recognition Using Dynamical Graph Convolutional Neural Networks. IEEE Transactions on Affective Computing 11 (2020).
  42. EEG Conformer: Convolutional Transformer for EEG Decoding and Visualization. IEEE Transactions on Neural Systems and Rehabilitation Engineering 31 (2023).
  43. Akara Supratak and Yike Guo. 2020. TinySleepNet: An efficient deep learning model for sleep stage scoring based on raw single-channel EEG. In 2020 42nd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC).
  44. LLaMA: Open and Efficient Foundation Language Models. arXiv:2302.13971 [cs.CL]
  45. Llama 2: Open Foundation and Fine-Tuned Chat Models. arXiv:2307.09288 [cs.CL]
  46. Attention Is All You Need. arXiv:1706.03762 [cs.CL]
  47. BrainBERT: Self-supervised representation learning for intracranial recordings. In The Eleventh International Conference on Learning Representations.
  48. InternImage: Exploring Large-Scale Vision Foundation Models With Deformable Convolutions. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
  49. TimesNet: Temporal 2D-Variation Modeling for General Time Series Analysis. In The Eleventh International Conference on Learning Representations.
  50. Baichuan 2: Open Large-scale Language Models. arXiv:2309.10305 [cs.CL]
  51. Learning Topology-Agnostic EEG Representations with Geometry-Aware Modeling. In Thirty-seventh Conference on Neural Information Processing Systems.
  52. An overview of power spectral density (PSD) calculations. In Optical Manufacturing and Testing VI.
  53. Florence: A New Foundation Model for Computer Vision. CoRR abs/2111.11432 (2021).
  54. PPi: Pretraining Brain Signal Model for Patient-independent Seizure Detection. In Thirty-seventh Conference on Neural Information Processing Systems.
  55. Biao Zhang and Rico Sennrich. 2019. Root Mean Square Layer Normalization. CoRR abs/1910.07467 (2019).
  56. Brant: Foundation Model for Intracranial Neural Signal. In Thirty-seventh Conference on Neural Information Processing Systems.
  57. A survey on deep learning-based non-invasive brain signals: recent advances and new frontiers. Journal of Neural Engineering 18 (2021).
  58. Self-supervised contrastive pre-training for time series via time-frequency consistency. Advances in Neural Information Processing Systems 35 (2022).
  59. ScatterFormer: Locally-Invariant Scattering Transformer for Patient-Independent Multispectral Detection of Epileptiform Discharges. arXiv:2304.14919 [eess.SP]
  60. Wei-Long Zheng and Bao-Liang Lu. 2015. Investigating Critical Frequency Bands and Channels for EEG-based Emotion Recognition with Deep Neural Networks. IEEE Transactions on Autonomous Mental Development 7, 3 (2015).
  61. One Fits All: Power General Time Series Analysis by Pretrained LM. In Thirty-seventh Conference on Neural Information Processing Systems.
Citations (3)

Summary

  • The paper introduces Brant-2, a foundation model featuring 1 billion parameters and trained on 4TB of mixed EEG and SEEG data to advance brain signal analysis.
  • It employs temporal and spatial encoders within a multi-FFN Transformer block and uses dual pretraining tasks like mask prediction and forecasting to enhance performance.
  • Empirical evaluations demonstrate robust results in seizure detection, sleep stage classification, emotion recognition, and motor imagery, even in low-resource clinical scenarios.

Advancing the Frontier of Brain Signal Analysis with Brant-2: A Comprehensive Foundation Model

Introduction

The burgeoning interest in brain signal analysis is motivated by its potential to unlock novel insights into brain function, disease mechanisms, and innovative applications in neurology, sleep studies, and even human-computer interaction. The landscape of brain signal research, straddling both invasive (e.g., SEEG) and non-invasive (e.g., EEG) techniques, presents unique challenges, notably the massive diversity in data and the prohibitive costs of data annotation. This context sets the stage for the significant contribution of foundation models—models pretrained on large, unlabeled datasets that can be fine-tuned for specific tasks with minimal labeled data. In this discourse, we explore Brant-2, the latest in foundation models for brain signals, marking a significant leap in addressing the multifaceted challenges of this domain.

Model Overview

Brant-2 emerges as a progressive iteration over its predecessor, Brant, with substantial advancements in its pre-training corpus, robustness to data variations, and adaptability across a diverse range of tasks. Featuring a staggering 1 billion parameters and trained on nearly 4 TB of mixed SEEG and EEG data from over 15k subjects, Brant-2 embodies a strategic fusion of volume and diversity in its training regime.

Architectural Innovations

Key to Brant-2's architecture is the incorporation of temporal and spatial encoders within a multi-FFN (Feedforward Neural Network) Transformer block, enabling the model to master both the spatial correlations between electrodes and the temporal dependencies across signals. This layered architectural strategy is complemented by a sophisticated data augmentation module, enhancing the model's resilience to data variations and modeling scales. Additionally, two-pretraining tasks—mask-prediction and forecasting—equip Brant-2 with the ability to extract rich semantic information, setting a solid foundation for its generalization capabilities.

Empirical Evaluation

Extensive evaluations underscore Brant-2’s superior performance across a spectrum of downstream tasks, including but not limited to seizure detection and prediction, sleep stage classification, emotion recognition, and motor imagery classification. Brant-2 consistently outperforms prior models and achieves competitive results against task-specific methods, demonstrating its broad applicability and robust generalization prowess.

Scalability and Efficiency in Low-Resource Scenarios

Our scalability analysis reveals a proportional improvement in model performance with an increase in parameters, affirming the potential of Brant-2 to scale adeptly. Furthermore, in scenarios with scarce labels, Brant-2's performance exhibits remarkable stability in comparison to fully supervised methods. This characteristic is particularly beneficial in clinical settings, where obtaining extensive labeled data is often impractical, thus highlighting the model's practical utility and the significant reduction in dependency on labeled data.

Conclusion and Future Directions

Brant-2 stands as a testament to the power of foundation models in the field of brain signal analysis, setting a new benchmark for both academic research and practical applications. Looking ahead, the ambition is to further enrich Brant-2’s pre-training corpus, incorporating a wider array of data sources to extend its applicability across more research areas and real-world scenarios. The promising results of Brant-2 not only fortify the foundation model paradigm in brain signal analysis but also pave the way for new avenues of exploration and innovation in understanding the enigmatic workings of the human brain.

Dice Question Streamline Icon: https://streamlinehq.com

Follow-up Questions

We haven't generated follow-up questions for this paper yet.

Don't miss out on important new AI/ML research

See which papers are being discussed right now on X, Reddit, and more:

“Emergent Mind helps me see which AI papers have caught fire online.”

Philip

Philip

Creator, AI Explained on YouTube