Depression Detection Based on Electroencephalography Using a Hybrid Deep Neural Network CNN-GRU and MRMR Feature Selection
Abstract: This study investigates the detection and classification of depressive and non-depressive states using deep learning approaches. Depression is a prevalent mental health disorder that substantially affects quality of life, and early diagnosis can greatly enhance treatment effectiveness and patient care. However, conventional diagnostic methods rely heavily on self-reported assessments, which are often subjective and may lack reliability. Consequently, there is a strong need for objective and accurate techniques to identify depressive states. In this work, a deep learning based framework is proposed for the early detection of depression using EEG signals. EEG data, which capture underlying brain activity and are not influenced by external behavioral factors, can reveal subtle neural changes associated with depression. The proposed approach combines convolutional neural networks (CNNs) and gated recurrent units (GRUs) to jointly extract spatial and temporal features from EEG recordings. The minimum redundancy maximum relevance (MRMR) algorithm is then applied to select the most informative features, followed by classification using a fully connected neural network. The results demonstrate that the proposed model achieves high performance in accurately identifying depressive states, with an overall accuracy of 98.74%. By effectively integrating temporal and spatial information and employing optimized feature selection, this method shows strong potential as a reliable tool for clinical applications. Overall, the proposed framework not only enables accurate early detection of depression but also has the potential to support improved treatment strategies and patient outcomes.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Explain it Like I'm 14
What is this paper about?
This paper is about using brainwave recordings (called EEG) and a smart computer program (deep learning) to tell whether a person is likely depressed or not. The goal is to create a fast, objective test that doesn’t depend only on interviews or questionnaires, which can sometimes be subjective.
What questions did the researchers ask?
- Can we detect depression from EEG brain signals using a computer model?
- If we combine two kinds of neural networks—one good at recognizing patterns across space (where on the head the signals come from) and one good at following patterns over time—does the detection get better?
- Does picking only the most important pieces of information (features) make the model more accurate and efficient?
How did they do it?
EEG in simple terms
EEG (electroencephalography) records tiny electrical signals from the brain using small sensors placed on the scalp. Think of it like placing microphones on your head that “listen” to the brain’s activity. In this study, a simple wearable device with three sensors was used.
Before analyzing, the team cleaned the signals:
- They removed noise (like static on a radio).
- They filtered out unhelpful frequencies to keep the parts related to brain activity that matter for depression.
The hybrid network: CNN + GRU
The model uses two parts that work together:
- CNN (Convolutional Neural Network): Imagine looking at a photo to spot shapes and patterns. The CNN looks across the different EEG sensors to find “where” patterns show up on the scalp. That’s the spatial part.
- GRU (Gated Recurrent Unit): Imagine watching a short video and remembering what happened from one moment to the next. The GRU follows how the brain signals change over time. That’s the temporal (time-based) part.
These two sets of patterns—“where on the head” and “how it changes over time”—are then combined into one set of features.
Choosing the best features: MRMR
The model creates many features (pieces of information). MRMR (Minimum Redundancy Maximum Relevance) is a smart filter. Think of it like picking a team: you want players who are each very useful (maximum relevance) but who don’t all do the same job (minimum redundancy). MRMR keeps the most helpful, least repetitive features and throws out the rest, making the model faster and often more accurate.
Training and testing the model
- Data: They used an open dataset (MODMA) with EEG from 53 people: 24 with major depressive disorder and 29 healthy controls, aged 16–52.
- Each recording was split into 10 short segments, creating 530 total segments.
- The computer learned from 70% of the segments (training) and was tested on the remaining 30% (testing) it hadn’t seen before.
- Performance was checked using:
- Accuracy: How many predictions were correct overall.
- Precision: When the model says “depressed,” how often is it right?
- Recall: How many of the truly depressed cases did it find?
- F1 score: A balanced score that combines precision and recall.
What did they find and why it matters?
The combined CNN+GRU model with MRMR feature selection worked very well.
- High accuracy: Around 98–99% (one test showed 98.74%; the average over many runs was 98.42%).
- It correctly identified almost all depressed cases (high recall) and avoided wrongly labeling healthy people as depressed (high precision).
- Extra checks (like ROC/AUC) also showed strong performance, meaning the model separates depressed vs. healthy signals reliably.
Why this matters:
- It shows EEG can provide objective clues about depression.
- A tool like this could help doctors screen patients earlier and more consistently.
- Faster, more reliable detection can lead to earlier treatment and better outcomes.
What does this mean for the future?
- Potential clinical helper: This kind of model could become a support tool for doctors—especially in clinics where quick, objective screening is useful. It wouldn’t replace a professional diagnosis but could make it more accurate and efficient.
- Tracking treatment: Because EEG measures brain activity directly, this approach might also help monitor how someone responds to treatments over time.
- Next steps and limits:
- The study used a relatively small dataset and only three EEG sensors. Larger studies with more people and more sensors are needed to prove it works widely.
- Testing in real hospitals and in real time would show how well it performs in everyday conditions.
In short, the study shows that mixing “where” brain patterns happen (CNN), “how they change over time” (GRU), and smart feature picking (MRMR) can make a powerful, objective tool for spotting depression from EEG signals—promising progress toward faster, fairer mental health care.
Knowledge Gaps
Knowledge gaps, limitations, and open questions
The paper leaves several concrete knowledge gaps, limitations, and open questions that future work could address:
- Subject-level data leakage risk: Training/testing splits were performed at the epoch (segment) level, not subject level, which likely places segments from the same participant in both sets. Evaluate with subject-wise splits (e.g., leave-one-subject-out or grouped cross-validation) to assess true generalization.
- Small and homogeneous sample: Only 53 recordings (24 MDD, 29 controls) from a single open dataset with a wearable three-electrode device. Validate on larger, multi-site cohorts and diverse populations to assess generalizability.
- Electrode montage and acquisition details missing: Electrode positions, sampling rate, reference scheme, epoch length, and stimulation/rest protocols are unspecified. Provide these to enable reproducibility and to examine montage-dependent performance.
- Preprocessing pipeline insufficiently described: “Denoising” is not specified (e.g., ICA, EOG/EMG artifact removal, re-referencing, notch filtering for line noise). Perform and report standardized artifact correction steps and evaluate their impact on performance.
- Bandpass selection rationale unclear: A 4.5–45 Hz filter excludes delta bands that can be informative in depression. Conduct sensitivity analyses across frequency bands and justify the chosen ranges.
- Architecture transparency lacking: CNN kernel sizes/strides, number of layers, GRU units, activation functions, optimizer, learning rate, batch size, regularization (dropout/weight decay), and early stopping criteria are not reported. Provide full configurations for reproducibility.
- End-to-end vs. feature-selection pipeline justification: Applying MRMR to deep features is atypical. Quantify the incremental contribution of MRMR via ablation (CNN-GRU only vs. CNN-GRU+MRMR) and compare to fully end-to-end models without post-hoc feature selection.
- Deep feature provenance unclear: Specify from which layers the “20 spatial” and “100 temporal” features are extracted, whether they are pooled activations or learned embeddings, and justify why these layers were chosen.
- Interpretability absent: Identify which channels, time windows, and feature groups drive decisions (e.g., saliency maps, Grad-CAM, SHAP, attention weights). Link selected features to neurophysiological markers of depression.
- External validation not performed: Test the trained model on independent EEG datasets to evaluate domain shift and cross-dataset generalization.
- Comparisons may be non-equivalent: Table 1 likely aggregates results from models trained on different datasets or protocols. Re-implement baselines on the same dataset and standardized splits to ensure fair benchmarking.
- Metrics inconsistency and lack of uncertainty: Reported accuracies (e.g., 96.74% vs. 98.42% vs. 98.74%) are inconsistent. Provide confidence intervals/standard deviations across runs, conduct statistical tests, and reconcile metric discrepancies.
- Subject-level performance absent: Present per-subject metrics (accuracy, sensitivity, specificity) and summarize performance at the participant level, not just per-epoch.
- Class imbalance handling unclear: Describe whether stratified splitting was used, and report additional metrics (balanced accuracy, MCC, specificity) and calibration (Brier score) to better characterize performance.
- Robustness to noise/artifacts untested: Quantify resilience to motion, ocular/muscle artifacts, and low SNR; perform stress tests and report performance degradation under realistic clinical noise conditions.
- Probability calibration and clinical utility: Assess calibration (reliability curves), choose clinically relevant operating points, and perform decision-curve analysis or report PPV/NPV at realistic prevalence.
- Covariate effects unexplored: Analyze performance stratified by age, sex, medication status, comorbidities, and severity to detect potential biases and differential accuracy.
- Rest vs. stimulation condition effects: Specify which conditions were used for training/testing and evaluate performance across tasks; determine whether task-specific models are needed.
- Real-time feasibility not demonstrated: Estimate latency, memory footprint, and computational load; evaluate on-device inference and deployment constraints for clinical screening.
- Longitudinal stability unknown: Assess test–retest reliability, session-to-session variability, and model stability over time within individuals.
- Beyond binary classification: Examine severity estimation (e.g., correlation with PHQ-9/clinician ratings), subtypes of depression, and differential diagnosis (e.g., anxiety vs. depression).
- Ethical and clinical integration: Define workflows for integrating model outputs into clinical practice, handling false positives/negatives, and communicating results to patients and clinicians.
- Reproducibility resources missing: Release code, detailed preprocessing scripts, trained model weights, and exact data splits to enable independent verification.
- Use of regression plot in classification context: Clarify the purpose of the “regression analysis” (Figure 1) for a binary classifier; if assessing calibration or continuous risk scores, use appropriate calibration plots and report AUC confidence intervals.
- Hyperparameter sensitivity: Systematically vary key hyperparameters (feature counts, MRMR selection size, epoch length) and report sensitivity analyses to identify stable operating regimes.
- Electrode optimization and channel reduction: Investigate which electrode placements yield maximal performance and whether fewer/more channels improve accuracy and robustness.
- Personalization and transfer learning: Explore subject-specific adaptation (fine-tuning, few-shot learning) and domain adaptation across devices/sites to improve individual-level performance.
- Multimodal integration: Evaluate whether combining EEG with complementary modalities (e.g., actigraphy, text, clinical variables) yields more robust and generalizable depression detection.
Practical Applications
Immediate Applications
Below are specific, actionable uses that can be piloted or deployed now, drawing directly from the paper’s methods (hybrid CNN-GRU, MRMR feature selection) and its reported performance. Each item notes relevant sectors, potential tools/workflows, and key assumptions or dependencies.
- Clinical adjunct screening in research settings
- Sectors: healthcare, medical devices
- What: Use a three-electrode wearable EEG and the CNN-GRU-MRMR model as an objective adjunct to intake assessments (e.g., alongside PHQ-9) to flag probable depressive states during outpatient visits.
- Tools/workflows: Research-mode software plugin for existing EEG systems; cloud or on-prem inference service; standardized preprocessing (denoising, 4.5–45 Hz bandpass), 5–10-minute recording; summary report integrated into EMR.
- Assumptions/dependencies: IRB/ethics approval; trained personnel for electrode placement; local validation on the clinic’s patient population; acceptance that outputs are adjunctive (not standalone diagnosis); model recalibration for the site’s EEG hardware and protocols.
- Treatment monitoring dashboards for neuromodulation and pharmacotherapy
- Sectors: healthcare (psychiatry, neuromodulation), pharma (clinical research)
- What: Track EEG-derived depression risk scores longitudinally to quantify neural changes during TMS/tDCS or antidepressant treatment.
- Tools/workflows: Session-based EEG capture, automated preprocessing, model inference, visualization of trajectories; alerts for non-response patterns.
- Assumptions/dependencies: Consistent recording conditions; evidence that EEG markers correlate with symptom change; clinic buy-in; basic MLOps to maintain model versions.
- Telepsychiatry augmentation with mailed/supervised EEG kits (research use only)
- Sectors: telehealth, digital health
- What: Remote EEG acquisition (three-electrode headband) with guided setup; cloud inference provides objective signals to augment tele-visits.
- Tools/workflows: Patient app with setup instructions, artifact checks; secure upload; clinician dashboard with risk score and confidence.
- Assumptions/dependencies: Sufficient data quality outside clinic; patient training; privacy/security; adherence to research-only disclaimers pending regulatory clearance.
- Academic replication and benchmarking pipeline
- Sectors: academia (neuroscience, biomedical engineering, machine learning)
- What: Adopt the paper’s end-to-end pipeline (CNN-GRU feature extraction + MRMR + dense classifier) as a reproducible baseline on public EEG datasets (e.g., MODMA).
- Tools/workflows: Open-source code release, standardized preprocessing notebooks, cross-validation protocols, ablation studies (with/without MRMR).
- Assumptions/dependencies: Access to datasets and labels; compute resources; transparent reporting of hyperparameters and segmentation.
- Extension to other EEG-based conditions (baseline architecture)
- Sectors: academia, healthcare (AD, Parkinson’s, epilepsy, ADHD)
- What: Reuse the hybrid spatial–temporal architecture with MRMR for classification tasks in other neurological/psychiatric EEG datasets.
- Tools/workflows: Condition-specific preprocessing and labeling; performance benchmarking against task-specific baselines.
- Assumptions/dependencies: Task-relevant features exist in EEG bands; adequate datasets; domain-specific validation.
- Research-mode productization by EEG vendors
- Sectors: medical devices
- What: Integrate “depression risk index” as a research-only module in EEG acquisition software, enabling customers to run pilots in clinical or academic labs.
- Tools/workflows: Embedded preprocessing, model inference on device or connected workstation; operator training; documentation and warnings.
- Assumptions/dependencies: Clear non-diagnostic labeling; device-specific calibration; post-market surveillance for research customers.
- Clinical trial biomarker (secondary endpoint)
- Sectors: pharma, CROs, healthcare
- What: Use model outputs to quantify brain-based changes as secondary endpoints in antidepressant or neuromodulation trials.
- Tools/workflows: Pre-specified EEG acquisition schedule; harmonized preprocessing; blinded analysis; data capture integrated with trial EDC.
- Assumptions/dependencies: Regulatory acceptance as exploratory endpoints; sensor protocol standardization; predefined statistical analysis plans.
- Standardized EEG preprocessing toolkit
- Sectors: software, academia, healthcare
- What: Package denoising, artifact repair, normalization, and bandpass filtering (4.5–45 Hz) into a reusable library to improve data quality for clinical and research EEG pipelines.
- Tools/workflows: Python/MATLAB packages; automated QC reports; plug-ins for common EEG platforms.
- Assumptions/dependencies: Compatibility with diverse EEG formats; validation across hardware vendors.
- Education and workforce training
- Sectors: education, academia, healthcare
- What: Develop course modules and hands-on labs on hybrid deep learning for EEG (CNN-GRU), feature selection (MRMR), and mental health applications.
- Tools/workflows: Jupyter notebooks, lecture materials, synthetic datasets; capstone projects replicating the paper’s pipeline.
- Assumptions/dependencies: Institutional adoption; access to anonymized datasets or simulators.
- EEG lab QA and artifact detection
- Sectors: healthcare, research labs
- What: Use learned spatial–temporal signatures to flag poor electrode contact, motion artifacts, or non-physiological signals before clinical interpretation.
- Tools/workflows: Real-time artifact scoring during acquisition; operator prompts to re-seat electrodes.
- Assumptions/dependencies: Model retraining for artifact detection; distinct labeling of artifacts vs. pathology.
- EMR integration for structured reporting (pilot)
- Sectors: healthcare IT
- What: Append standardized “EEG depression risk” summaries to patient records via FHIR/HL7 for internal clinical audits and research.
- Tools/workflows: Interoperability adapters; clinician-friendly PDFs/JSON; audit trails.
- Assumptions/dependencies: IT governance approval; clear disclaimers; data security controls.
Long-Term Applications
These opportunities require further validation, scaling, regulatory approval, or technical development before routine deployment.
- FDA/CE-cleared EEG-based depression diagnostic or screening device
- Sectors: healthcare, policy/regulation, medical devices
- What: A clinically validated, multi-site, multi-demographic device+software system providing objective depression screening/diagnosis.
- Tools/workflows: Prospective trials; rigorous generalization studies; post-market surveillance; human factors and usability validation.
- Assumptions/dependencies: Large, diverse datasets; consistent performance across devices and settings; regulatory approval; acceptable PPV/NPV in population screening.
- Primary care screening protocol integrating EEG + PHQ-9
- Sectors: healthcare (primary care), policy
- What: A brief, standardized EEG test to complement questionnaires, improving early detection and triage.
- Tools/workflows: 5-minute capture workflow; nurse-led setup; embedded decision support; referral pathways.
- Assumptions/dependencies: Training and staffing; reimbursement pathways; cost-effectiveness evidence; minimal disruption to clinic flow.
- Closed-loop neuromodulation personalized by EEG biomarkers
- Sectors: neuromodulation, robotics (medical), software
- What: Real-time EEG state estimation guides TMS/tDCS parameters (site, polarity, intensity) to maximize treatment responsiveness.
- Tools/workflows: Low-latency inference; safety controllers; adaptive protocols; RCTs comparing closed- vs. open-loop.
- Assumptions/dependencies: Causal linkage between EEG features and treatment outcomes; real-time artifact resilience; stringent safety/regulatory approvals.
- Precision psychiatry: treatment selection and response prediction
- Sectors: healthcare, pharma
- What: Stratify patients by EEG phenotypes to predict response to SSRIs/SNRIs, psychotherapy, or neuromodulation, reducing time-to-effective treatment.
- Tools/workflows: Multimodal models combining EEG with demographics/clinical history; decision support integrated into care pathways.
- Assumptions/dependencies: Large labeled cohorts linking EEG to outcomes; fairness audits; clinician acceptance.
- Consumer-grade wellness devices for mood tracking (non-diagnostic)
- Sectors: consumer health, software
- What: Headbands/apps offering trend monitoring of EEG-based mood indices with coaching and lifestyle recommendations.
- Tools/workflows: Edge inference on wearables; privacy-first data handling; user feedback loops.
- Assumptions/dependencies: Clear marketing as non-medical; harm-minimization safeguards; robust artifact handling in everyday environments.
- Population-level mental health surveillance (opt-in, privacy-preserving)
- Sectors: public health policy, academia
- What: Aggregated, anonymized EEG-derived indicators to track community-level trends and allocate resources.
- Tools/workflows: Federated analytics; differential privacy; dashboards for health departments.
- Assumptions/dependencies: Voluntary participation; strong governance; societal acceptance; bias mitigation.
- Multimodal fusion platforms (EEG + actigraphy + text/EHR)
- Sectors: healthcare IT, software, digital health
- What: Integrated risk scoring leveraging EEG, wearables, and clinical notes to enhance sensitivity/specificity.
- Tools/workflows: Data pipelines; feature fusion models; clinician-facing explanations.
- Assumptions/dependencies: Data interoperability; consent management; explainability and transparency.
- High-density EEG expansion for improved granularity
- Sectors: academia, medical devices
- What: Adapt the architecture to 32–64+ channels to capture richer spatial patterns and improve detection across subtypes of depression.
- Tools/workflows: Scalable training; channel selection strategies; MRMR or alternative feature selection tailored to high-dimensional data.
- Assumptions/dependencies: Cost and complexity of high-density systems; standardized cap layouts; computational scaling.
- Fairness, bias, and cross-cultural generalization auditing tools
- Sectors: policy, academia, healthcare
- What: Toolkits to evaluate performance across age, sex, ethnicity, comorbidities, and medications, with remediation strategies.
- Tools/workflows: Bias dashboards; domain adaptation; stratified reporting; governance checklists.
- Assumptions/dependencies: Diverse datasets; stakeholder oversight; evolving regulatory guidance on AI fairness.
- Standards and consortia for EEG mental health benchmarking
- Sectors: academia, policy, industry
- What: Open challenges, shared datasets, and harmonized protocols for acquisition, labeling, and evaluation in depression detection.
- Tools/workflows: Data commons; reference implementations; pre-registered evaluation plans.
- Assumptions/dependencies: Funding and multi-institution collaboration; legal/data-sharing frameworks.
- Scalable, secure cloud/edge inference platforms
- Sectors: software, digital health
- What: HIPAA/GDPR-compliant services for real-time or near-real-time EEG processing at scale for clinics and telehealth providers.
- Tools/workflows: MLOps (versioning, monitoring, drift detection); edge acceleration; audit trails.
- Assumptions/dependencies: Latency constraints; cybersecurity; cost models aligned with healthcare budgets.
- Reimbursement and regulatory pathways for EEG-based mental health tools
- Sectors: policy, payers, healthcare
- What: New CPT codes, coverage policies, and clinical guidelines recognizing validated EEG biomarkers in mental health care.
- Tools/workflows: Health economics studies; professional society endorsements; evidence synthesis.
- Assumptions/dependencies: Demonstrated clinical utility and cost-effectiveness; stakeholder consensus.
Notes on feasibility across applications:
- The reported accuracy (≈98%) was achieved on a relatively small, three-electrode dataset (MODMA, n=53 recordings) with specific preprocessing and epoching; broader generalization requires larger, diverse cohorts and multi-site validation.
- Device differences, recording conditions (resting vs. stimulation), and patient factors (comorbidities, medications) can materially affect performance; site-specific calibration and ongoing QA are critical.
- MRMR reduces feature dimensionality and can enable efficient inference, but feature selection stability across cohorts must be verified.
- All clinical uses require clear positioning as adjunctive tools until regulatory approvals are obtained; strong privacy, security, and fairness practices are essential for trust and adoption.
Glossary
- Actigraphy: A method of monitoring human rest/activity cycles using wearable motion sensors to infer sleep and activity patterns. "Actigraphy based methods have been explored as well."
- Area under the curve (AUC): A scalar summary of ROC performance measuring overall discriminative ability across all thresholds (1 is perfect). "The area under the curve (AUC) was close to 1"
- Backpropagation: The gradient-based procedure used to train neural networks by propagating errors backward to update weights. "Model parameters were optimized via backpropagation using a cross-entropy loss function."
- Bandpass filter: A signal processing filter that passes frequencies within a certain range and attenuates frequencies outside it. "A bandpass filter (4.5-45. Hz) was applied"
- Class imbalance: A condition where the number of samples in different classes is uneven, potentially biasing model training. "particularly in the presence of class imbalance."
- Confusion matrix: A table that summarizes a classifier’s performance by showing counts of true/false positives and negatives. "The confusion matrix (Figure 1) indicates that the model correctly classified"
- Convolutional neural network (CNN): A deep learning model using convolutional layers to capture spatial patterns (e.g., across EEG channels). "convolutional neural networks (CNNs) and gated recurrent units (GRUs)"
- Cross-entropy loss: A loss function commonly used in classification that measures the divergence between predicted and true distributions. "cross-entropy loss function."
- Cross-sectional: A study design that analyzes data from a population at a single point in time. "A cross sectional online survey analyzed clinical and demographic factors"
- DeprNet: A specific CNN-based architecture introduced for EEG-based depression classification. "a CNN based model known as DeprNet was introduced"
- Electroencephalography (EEG): A technique for recording electrical activity of the brain via scalp electrodes. "electroencephalography (EEG) has emerged as a promising objective tool"
- EEG band powers: Power (energy) measured within standard EEG frequency bands (e.g., alpha, beta, delta, theta) used as features. "alpha alpha one and alpha two, beta, delta, and theta band powers"
- Epoch (EEG): A fixed-length segment of an EEG recording used for analysis. "Each recording was segmented into 10 equal-length epochs"
- F1 score: The harmonic mean of precision and recall, balancing both metrics in a single measure. "F1 score"
- Feature fusion: The process of combining features from multiple sources or models into a joint representation. "in a feature fusion strategy"
- Functional connectivity: Statistical dependencies between neural signals reflecting interactions among brain regions. "functional connectivity"
- Fully connected layer: A neural network layer where each neuron connects to all neurons in the previous layer. "Fully Connected Layer"
- Gated recurrent unit (GRU): A type of recurrent neural network that models sequences using update/reset gates to control information flow. "gated recurrent units (GRUs)"
- Harmonic mean: A type of average especially suited for rates and ratios, used in the definition of F1 score. "The F1 score, defined as the harmonic mean of precision and recall"
- Hemispheric differences: Differences in neural measures between the left and right brain hemispheres. "significant hemispheric differences in theta power"
- Logistic regression: A statistical model for binary outcomes that estimates the probability of class membership. "Logistic regression revealed associations"
- Long Short-Term Memory (LSTM): A recurrent neural network architecture with gates designed to capture long-range temporal dependencies. "A generative long short term memory (LSTM) model was proposed"
- Minimum Redundancy Maximum Relevance (MRMR): A feature selection method that maximizes relevance to the target while minimizing redundancy among features. "The minimum redundancy maximum relevance (MRMR) algorithm is then applied"
- Multi Cluster Feature Selection (MCFS): An unsupervised feature selection technique leveraging spectral graph theory to preserve cluster structure. "Multi Cluster Feature Selection (MCFS) was applied"
- Multilayer perceptron (MLP): A feedforward neural network with one or more hidden layers of nonlinear units. "multilayer perceptron (MLP) networks"
- Noninvasive neuromodulation: Techniques that alter brain activity from outside the skull (e.g., tDCS, TMS) without surgery. "Noninvasive neuromodulation techniques, such as transcranial direct current stimulation (tDCS)"
- PHQ-9: A nine-item self-report questionnaire for screening and measuring the severity of depression. "using PHQ-9 scores as ground truth labels"
- Pooling layer: A CNN layer that reduces spatial dimensions (e.g., via max/average pooling) to summarize features. "Pooling Layer"
- Receiver operating characteristic (ROC): A curve plotting true positive rate against false positive rate across thresholds to evaluate classifiers. "The receiver operating characteristic (ROC) curves"
- Rectified Linear Unit (ReLU): A common activation function defined as f(x)=max(0,x) used in deep networks. "Relu Layer"
- Record-wise validation: Evaluation where data segments from the same subject can appear in both train and test sets at the record level. "record wise"
- ResNet-50: A 50-layer residual neural network architecture known for deep skip connections. "ResNet-50 + LSTM"
- Resting-state: A condition where participants are not performing an explicit task, used to assess baseline brain activity. "resting-state and stimulation conditions"
- Sequential Backward Feature Selection (SBFS): A wrapper feature selection method that iteratively removes features to optimize performance. "Sequential Backward Feature Selection (SBFS) was used"
- Softmax: A function that converts logits into a probability distribution over classes in the output layer. "a Softmax output layer for binary classification."
- Subject-wise validation: Evaluation where all data from a subject is confined to either training or testing to assess generalization across individuals. "subject wise"
- Support vector machine (SVM): A supervised learning algorithm that finds a maximum-margin hyperplane for classification. "support vector machine (SVM) for detecting persistent depressive disorder (PDD)"
- Theta asymmetry: A lateralization measure comparing theta-band power between hemispheres, often linked to affective processing. "theta asymmetry"
- Transcranial direct current stimulation (tDCS): A noninvasive brain stimulation method delivering weak direct currents via scalp electrodes. "transcranial direct current stimulation (tDCS)"
- Wavelet: A time-frequency analysis technique capturing localized spectral content at multiple scales. "wavelet based"
- Wearable three-electrode EEG device: A portable EEG system that records with three scalp electrodes, enabling ambulatory data collection. "a wearable three-electrode EEG device"
- XGBoost: An efficient gradient-boosted decision tree library widely used for structured data tasks. "with XGBoost achieving the best performance"
Collections
Sign up for free to add this paper to one or more collections.