This comprehensive review explores data fusion techniques for integrating electroencephalography (EEG) and functional near-infrared spectroscopy (fNIRS) signals to advance biomedical research and clinical applications.
This comprehensive review explores data fusion techniques for integrating electroencephalography (EEG) and functional near-infrared spectroscopy (fNIRS) signals to advance biomedical research and clinical applications. It covers the fundamental principles of these complementary neuroimaging modalities, detailed methodological approaches for multimodal fusion, practical strategies for optimizing data quality, and rigorous validation frameworks. Targeted at researchers, scientists, and drug development professionals, the article highlights how EEG-fNIRS fusion overcomes individual modality limitations by combining millisecond temporal resolution with improved spatial localization. Through clinical case studies and emerging research applications, we demonstrate how this integrated approach provides richer insights into brain function, enhances diagnostic capabilities, and accelerates therapeutic development for neurological and psychiatric disorders.
The human brain operates through two primary, interconnected physiological processes: rapid electrical activity from neuronal firing and slower hemodynamic responses that deliver energy resources. Electroencephalography (EEG) and functional near-infrared spectroscopy (fNIRS) have emerged as complementary neuroimaging techniques that capture these distinct phenomena directly related to neural function [1]. EEG measures the electrical potentials generated by synchronized pyramidal neuron activity, offering millisecond-level temporal resolution to track immediate brain dynamics [1]. In contrast, fNIRS detects hemodynamic changes by measuring concentration changes in oxygenated hemoglobin (HbO) and deoxygenated hemoglobin (HbR) in the blood, providing better spatial localization of active brain regions [2] [3]. Their integration creates a powerful multimodal approach for studying brain function, leveraging the neurovascular coupling mechanism that links neuronal electrical activity to subsequent vascular responses [1].
Table 1: Fundamental Comparison of EEG and fNIRS Neuroimaging Modalities
| Characteristic | EEG (Electrical Activity) | fNIRS (Hemodynamic Response) |
|---|---|---|
| Measured Parameter | Electrical potentials from synchronized post-synaptic activity | Concentration changes of HbO and HbR |
| Physiological Basis | Direct neuronal electrical activity | Neurovascular coupling |
| Temporal Resolution | Millisecond level (∼5 ms) [4] | Slower (∼0.1-1 Hz) [2] [1] |
| Spatial Resolution | Low (several centimeters) [5] [1] | Moderate (∼1-3 cm) [4] [5] |
| Penetration Depth | Superficial and deep sources (with volume conduction) | Cortical surface (1-3 cm) [2] |
| Primary Signal Origin | Pyramidal neurons in cortical layers [1] | Cerebral microvasculature (arterioles, capillaries, venules) |
| Noise Sensitivity | Sensitive to electromagnetic artifacts [6] | Sensitive to systemic physiological noise [7] |
Electroencephalography captures the macroscopic temporal dynamics of brain electrical activity through passive measurements of scalp voltages [2]. The recorded EEG signal primarily results from the summation of synchronized post-synaptic potentials in cortical pyramidal neurons [1]. For detectable EEG signals to occur, tens of thousands of pyramidal neurons within a cortical column must fire synchronously, with their dendritic trunks coherently oriented parallel to each other and perpendicular to the cortical surface [1]. This specific anatomical arrangement enables sufficient summation and propagation of electrical signals to the scalp surface where electrodes detect voltage fluctuations [1]. These neural oscillations are categorized into characteristic frequency bands—theta (4-7 Hz), alpha (8-14 Hz), beta (15-25 Hz), and gamma (>25 Hz)—each associated with different brain states and cognitive functions [1].
Functional near-infrared spectroscopy leverages the relative transparency of biological tissues to light in the near-infrared spectrum (650-950 nm) to measure hemodynamic changes associated with neural activity [2] [3]. Within this wavelength range, light penetrates the scalp and skull, reaching the cortical surface where it is predominantly absorbed by the chromophores oxyhemoglobin (HbO) and deoxyhemoglobin (HbR) [3]. The neurovascular coupling mechanism forms the foundation for fNIRS: when neurons become active, they trigger a complex cascade that increases local cerebral blood flow, delivering oxygen and nutrients to support metabolic demands [1] [3]. This hemodynamic response manifests as an initial, brief increase in deoxyhemoglobin concentration (the "initial dip") followed by a more pronounced increase in oxyhemoglobin and a decrease in deoxyhemoglobin concentration [3]. fNIRS systems typically utilize two or more wavelengths to distinguish between HbO and HbR based on their distinct absorption spectra [2].
Neurovascular coupling represents the critical physiological link between neuronal electrical activity and the hemodynamic responses measured by fNIRS [1]. When neurons fire, they consume energy, triggering a complex signaling cascade that involves astrocytes, neurons, and vascular cells [3]. This process begins with increased oxygen extraction from local blood vessels, creating a transient rise in deoxyhemoglobin [3]. Almost simultaneously, vasoactive signals cause arteriolar dilation, increasing local cerebral blood flow and delivering oxygenated blood that typically overshoots metabolic demands [3]. The resulting hemodynamic response peaks 4-6 seconds after neural activation, creating the characteristic fNIRS signal patterns [8]. This tightly regulated mechanism forms the theoretical basis for integrating EEG and fNIRS, as it connects the direct electrical measurements of EEG with the indirect metabolic-hemodynamic measurements of fNIRS [1].
Figure 1: Signaling Pathway Linking Neural Activity to Measurable Signals. The diagram illustrates how initial neural electrical activity triggers neurovascular coupling, leading to measurable EEG and fNIRS signals through different physiological pathways.
Successful simultaneous EEG-fNIRS recording requires careful hardware integration to ensure signal quality and synchronization. Two primary approaches exist for system integration: (1) separate systems synchronized via computer, and (2) unified systems with a single processor for both modalities [5]. While the first approach offers simplicity, the second provides more precise synchronization essential for capturing the temporal relationship between electrical and hemodynamic responses [5]. For helmet design, researchers typically use flexible EEG electrode caps as a foundation, creating punctures at specific locations to accommodate fNIRS probe fixtures [5]. Customized helmets using 3D printing or cryogenic thermoplastic sheets offer better fit and more consistent optode placement across subjects [5]. The integration must maintain proper source-detector distances (typically 2.5-3.5 cm for adults) for fNIRS while ensuring good electrode-scalp contact for EEG [2] [3]. Optodes and electrodes should be co-registered to anatomical landmarks (nasion, inion, preauricular points) using a 3D digitizer for accurate spatial localization [9].
Table 2: Research Reagent Solutions and Essential Materials
| Item Category | Specific Examples | Function/Purpose |
|---|---|---|
| fNIRS System Components | Continuous-wave NIRS system (e.g., Hitachi ETG-4100), optical fibers, laser diodes/LEDs (695 & 830 nm) [9] | Generate and detect NIR light to measure HbO/HbR concentration changes |
| EEG System Components | EEG amplifier system (e.g., BrainAMP), Ag/AgCl electrodes, electrolyte gel [5] | Measure electrical potentials from scalp surface |
| Integration Materials | Custom integration caps, 3D-printed helmet fixtures, thermoplastic sheets [5] | Co-register EEG and fNIRS components with consistent geometry |
| Auxiliary Equipment | 3D magnetic space digitizer (e.g., Polhemus Fastrak), response recording devices [9] | Record spatial coordinates of optodes/electrodes; capture behavioral data |
| Software Tools | MNE-Python, NIRS processing toolboxes, custom analysis scripts [8] | Preprocess, synchronize, and analyze multimodal data |
The motor execution, observation, and imagery paradigm provides an excellent experimental framework for studying the action observation network (AON) and comparing electrical and hemodynamic responses across different motor conditions [9]. This protocol employs a block design with three conditions: (1) Motor Execution (ME): participants physically perform a motor task (e.g., grasping and moving a cup) in response to an "your turn" auditory cue; (2) Motor Observation (MO): participants observe an experimenter performing the same motor task following a "my turn" cue; and (3) Motor Imagery (MI): participants mentally simulate the motor task without physical movement [9]. Each trial begins with a 2-second instruction period, followed by a 5-second task period, and a variable inter-trial interval (10-20 seconds) to allow hemodynamic responses to return to baseline [9]. For simultaneous EEG-fNIRS recording, participants should be seated comfortably facing an experimenter, with the fNIRS optodes placed over sensorimotor and parietal cortices and EEG electrodes arranged according to the international 10-20 system [9].
Figure 2: Experimental Protocol for Motor Paradigm. The workflow outlines the procedure for simultaneous EEG-fNIRS recording during motor execution, observation, and imagery tasks.
Preprocessing simultaneous EEG-fNIRS data requires modality-specific pipelines executed in parallel before multimodal fusion. For fNIRS data, begin by converting raw intensity signals to optical density, then to hemoglobin concentration changes using the modified Beer-Lambert law [8]. Apply bandpass filtering (0.01-0.5 Hz) to remove cardiac pulsation (∼1 Hz) and slow drifts while preserving the hemodynamic response [8]. For EEG data, implement high-pass filtering (0.5 Hz) to remove slow drifts, notch filtering (50/60 Hz) to remove line noise, and artifact removal techniques (e.g., ICA) to eliminate ocular and muscle artifacts [1]. Quality control metrics should include the scalp coupling index (SCI) for fNIRS (recommended threshold >0.5) and signal-to-noise ratio for EEG [8]. Epoch data according to experimental conditions, with appropriate baseline correction (typically -5 to 0 seconds before stimulus onset) and artifact rejection [8]. Finally, synchronize the preprocessed EEG and fNIRS data streams using recorded trigger pulses or shared clock signals [5].
Multimodal fusion of EEG and fNIRS data occurs at three primary levels: decision-level, feature-level, and data-level fusion [1]. Decision-level fusion involves processing and classifying each modality separately, then combining the results at the decision stage using methods like majority voting or meta-classifiers [4]. This approach demonstrated a 5-7% improvement in classification accuracy for motor imagery tasks compared to single-modality approaches [4]. Feature-level fusion combines extracted features from both modalities before classification, often employing techniques like canonical correlation analysis (CCA) or structured sparse multiset CCA (ssmCCA) to identify relationships between electrical and hemodynamic features [9] [4]. This method has shown particular promise in identifying shared neural regions associated with the action observation network during motor tasks [9]. Data-level fusion integrates raw or minimally processed data from both modalities, often using joint blind source separation or model-based approaches to identify common underlying components [7].
The complementary nature of EEG and fNIRS has enabled advanced applications across clinical and research domains. In clinical neuroscience, simultaneous EEG-fNIRS has been applied to study neurological and psychiatric disorders including epilepsy, Alzheimer's disease, stroke, attention-deficit hyperactivity disorder (ADHD), and amyotrophic lateral sclerosis (ALS) [4] [5]. For epilepsy monitoring, EEG provides precise temporal localization of epileptiform activity while fNIRS offers improved spatial localization of the epileptic focus [3]. In brain-computer interfaces (BCIs), hybrid EEG-fNIRS systems have significantly improved classification accuracy for motor imagery tasks compared to unimodal systems [4] [6]. Recent advances include deep learning approaches with cross-modal attention mechanisms (e.g., MBC-ATT framework) that dynamically weight the contribution of each modality based on task demands, further enhancing classification performance for cognitive state decoding [6].
The fundamental differences between electrical activity and hemodynamic responses, rather than presenting a challenge, create a powerful complementary relationship when captured through simultaneous EEG-fNIRS recording. EEG's millisecond temporal resolution provides an exquisite window into the direct electrical consequences of neural processing, while fNIRS's superior spatial localization captures the metabolically coupled hemodynamic responses that support such activity. The neurovascular coupling mechanism binds these phenomena together, offering researchers a more complete picture of brain function than either modality could provide alone. Through appropriate experimental protocols and advanced fusion methodologies, researchers can leverage these complementary strengths to advance our understanding of brain function in both healthy and pathological states, paving the way for more precise diagnostics and innovative therapeutic interventions in clinical neuroscience.
In non-invasive neuroimaging, the trade-off between temporal and spatial resolution is a fundamental concept. Electroencephalography (EEG) and functional near-infrared spectroscopy (fNIRS) represent two pillars of brain imaging, each with distinct resolution profiles. Temporal resolution refers to the precision with which a technique can measure when neural activity occurs, while spatial resolution describes its ability to pinpoint where in the brain this activity is generated [10] [11]. EEG is renowned for its millisecond-scale temporal resolution, enabling it to capture rapid neural dynamics. In contrast, fNIRS provides better spatial localization of cortical activity by measuring hemodynamic responses, though its temporal resolution is limited to seconds due to the slow nature of blood flow changes [10] [12]. This application note details the comparative strengths and limitations of these modalities and provides practical protocols for their integrated use in multimodal research, framed within the context of data fusion techniques for enhancing spatiotemporal imaging capabilities.
Table 1: Technical Specifications of EEG and fNIRS
| Feature | EEG (Electroencephalography) | fNIRS (Functional Near-Infrared Spectroscopy) |
|---|---|---|
| What It Measures | Electrical activity from postsynaptic potentials of cortical neurons [10] | Hemodynamic response (changes in oxygenated and deoxygenated hemoglobin) [10] |
| Temporal Resolution | High (millisecond scale) [10] [13] | Low (seconds scale) [10] [14] |
| Spatial Resolution | Low (centimeter-level) [10] [14] | Moderate (better than EEG, but limited to outer cortex) [10] [12] |
| Depth of Measurement | Cortical surface [10] | Outer cortex (~1–2.5 cm deep) [10] |
| Key Strength | Capturing rapid cognitive processes (e.g., sensory perception, ERPs) [10] | Localizing sustained cortical activity (e.g., workload, problem-solving) [10] |
| Primary Limitation | Spatial smearing due to volume conduction [13] | Indirect, slow hemodynamic response and superficial penetration [10] [12] |
The conventional depiction of temporal and spatial resolution as independent axes is an oversimplification. In practice, the factor that limits one resolution often degrades the other [13].
The complementary strengths of EEG and fNIRS make them ideal candidates for multimodal fusion, which aims to achieve a comprehensive view of brain activity with high spatiotemporal resolution [7] [14]. Fusion can be implemented at different stages of data processing.
Table 2: Data Fusion Approaches for EEG and fNIRS
| Fusion Stage | Description | Key Techniques | Benefits |
|---|---|---|---|
| Early-Stage Fusion | Raw or pre-processed data from both modalities are combined before feature extraction [15]. | Inputting concurrent EEG and fNIRS data into a Y-shaped neural network [15]. | Can capture underlying neurovascular coupling and subtle interactions; has shown higher classification performance in some BCI tasks [15]. |
| Data-Level / Symmetric Fusion | Data-driven methods that treat both modalities equally to find shared latent components [7]. | Joint Independent Component Analysis (jICA), Canonical Correlation Analysis (CCA) [7]. | Can reveal complex, shared neural processes without strong prior assumptions [7]. |
| Model-Based / Asymmetric Fusion | Using one modality to constrain or inform the analysis of the other [14]. | Using high-spatial-resolution fNIRS (or DOT) reconstruction as a spatial prior to constrain the EEG source localization inverse problem [14]. | Dramatically improves the spatial resolution of reconstructed neuronal sources; allows resolution of spatially close sources activated with small temporal separations (e.g., 50 ms) [14]. |
The rationale for fNIRS-EEG fusion is grounded in the principle of neurovascular coupling, where neuronal electrical activity triggers a subsequent hemodynamic response. The following diagram illustrates this relationship and a generalized experimental workflow.
Diagram Title: Neurovascular Coupling and Multimodal Fusion Workflow
This protocol ensures high-quality, synchronized data collection.
Equipment and Reagents:
Procedure:
This protocol uses fNIRS to spatially constrain EEG source analysis, enhancing spatial accuracy [14].
Equipment and Software:
Procedure:
This protocol uses a Y-shaped neural network to fuse EEG and fNIRS data at the raw data level for improved classification of mental states [15].
Equipment and Software:
Procedure:
Table 3: Key Materials for Multimodal EEG-fNIRS Research
| Item | Function/Description | Example Use Case |
|---|---|---|
| Integrated EEG-fNIRS Caps | Head caps with pre-defined placements for both electrodes and optodes, ensuring consistent co-registration across sessions [10]. | Essential for any multimodal study to ensure sensors are positioned correctly relative to each other and brain anatomy. |
| Conductive Electrolyte Gel | Reduces impedance between the scalp and EEG electrodes, facilitating the measurement of electrical potentials. | Standard requirement for obtaining high-quality EEG signals. |
| Short-Separation fNIRS Channels | fNIRS detectors placed very close (~8 mm) to a source, which are predominantly sensitive to systemic physiological noise in the scalp [7]. | Used as regressors in data processing to separate superficial (scalp) from deep (cortical) components of the fNIRS signal, improving signal quality [7]. |
| TTL Pulse Generator / Sync Box | A hardware device that sends precise digital timing pulses to multiple data acquisition systems. | Critical for synchronizing the clocks of separate EEG and fNIRS systems during simultaneous recording. |
| 3D Digitizer | A stylus-based system to record the precise 3D locations of EEG electrodes and fNIRS optodes relative to head landmarks. | Improves the accuracy of source reconstruction by providing exact sensor positions for the forward model. |
| Surface Laplacian (CSD) Toolbox | Software (e.g., in EEGLAB) that computes the Current Source Density (CSD) from scalp potentials. | Used to reduce the spatial blurring effect of volume conduction in EEG, thereby improving both spatial and temporal resolution of scalp-level data [13]. |
Neurovascular coupling (NVC) describes the fundamental biological mechanism that creates a tight temporal and regional linkage between neural activity and subsequent changes in cerebral blood flow (CBF). This physiological process ensures that active brain regions promptly receive an increased supply of oxygen and nutrients to meet heightened metabolic demands [17]. The adult human brain constitutes approximately 2% of total body weight yet consumes over 20% of the body's oxygen and glucose at rest, creating a critical dependence on precisely regulated blood flow delivery [17]. The neurovascular unit, composed of vascular smooth muscle cells, neurons, and astrocyte glial cells, forms the anatomical foundation for this sophisticated communication system [17] [18].
The significance of neurovascular coupling extends beyond basic physiology to underpin modern functional neuroimaging techniques. Both functional magnetic resonance imaging (fMRI) and functional near-infrared spectroscopy (fNIRS) rely on the hemodynamic changes triggered by neural activity to indirectly map brain function [19] [3]. Understanding NVC is therefore paramount for proper interpretation of neuroimaging data across both research and clinical settings. impairments in neurovascular coupling have been implicated in various pathological conditions including Alzheimer's disease, stroke, hypertension, and diabetes, highlighting its clinical relevance [18] [1].
The neurovascular unit operates through coordinated interactions between three primary cell types: neurons, astrocytes, and vascular smooth muscle cells. Upon neuronal activation, glutamate released from presynaptic terminals activates postsynaptic neurons and surrounding astrocytes [17]. Neurons respond by activating neuronal nitric oxide synthase (nNOS), producing nitric oxide (NO) that directly dilates parenchymal arterioles [17]. Concurrently, astrocytes respond to glutamate through metabotropic glutamate receptors, triggering calcium-dependent production of vasoactive compounds [17] [19].
The signaling cascade in astrocytes involves production of arachidonic acid, which is subsequently metabolized into several vasoactive messengers: prostaglandin E2 (PGE2) via cyclooxygenase-1 (COX-1), epoxyeicosatrienoic acid (EET) via cytochrome P450 (CYP) epoxygenase, and 20-Hydroxyeicosatetraenoic acid (20-HETE) via CYP4A [19]. These molecules collectively modulate vascular tone by acting on arterioles and capillaries. Additionally, GABA interneurons release various vasoactive substances including NO, acetylcholine, neuropeptide Y, and vasoactive intestinal peptide (VIP), which can produce both constrictive and dilatory effects on cerebral microvasculature [17].
Figure 1: Cellular Signaling Pathways in Neurovascular Coupling. This diagram illustrates the primary mechanisms through which neural activity triggers hemodynamic responses. Key pathways include neuronal nitric oxide production, astrocytic vasoactive compound synthesis, and interneuron-mediated regulation.
The canonical hemodynamic response to neural activity follows a characteristic pattern known as "functional hyperemia." During baseline conditions, cerebral blood flow is closely matched to metabolic demands. Upon neural activation, a biphasic response occurs: an initial brief increase in deoxygenated hemoglobin (the "initial dip") followed by a substantial increase in cerebral blood flow that delivers oxygenated hemoglobin beyond metabolic requirements [17] [3]. This overcompensation results in a 4-fold greater increase in CBF relative to the increase in ATP consumption, forming the physiological basis for blood-oxygen-level-dependent (BOLD) contrast used in fMRI [17].
The hemodynamic response profile varies between populations, with neonates and preterm infants demonstrating delayed and less pronounced responses compared to adults [18]. This developmental difference is attributed to ongoing maturation of neurovascular unit components including astrocytes and pericytes during early life [18]. The typical adult response to a stimulus produces a 10-20% increase in cerebral blood flow in the posterior cerebral artery and a 5-8% increase in the middle cerebral artery territory [18].
Computational models of neurovascular coupling integrate complementary data from animal and human studies to create quantitative frameworks for predicting hemodynamic responses. Sten et al. (2023) developed a comprehensive mathematical model that preserves mechanistic behaviors across species by translating cell-specific contributions identified in rodent optogenetics studies to human applications [19]. Their model identifies distinct neuronal subpopulations responsible for different temporal components of the vascular response: NO-producing interneurons mediate the initial rapid dilation, pyramidal neurons sustain the main dilation during prolonged stimuli, and NPY-interneurons contribute to the post-stimulus undershoot [19].
These models enable more accurate interpretation of neuroimaging data by accounting for the complex, non-linear relationships between neural activity and hemodynamic responses. For example, they incorporate the interplay between cerebral metabolic rate of oxygen (CMRO2), cerebral blood flow (CBF), and cerebral blood volume (CBV) that governs the BOLD response [19]. Such quantitative frameworks are particularly valuable for understanding pathophysiological conditions where neurovascular coupling is impaired, including stroke, neurodegenerative diseases, and developmental disorders [18] [19].
Advanced signal processing methods have been developed to characterize neurovascular coupling in both task-based and resting-state paradigms. For spatiotemporal studies involving controlled stimuli, the hemodynamic response function can be modeled using general linear models that account for the typical delay and dispersion of the blood flow response relative to neural activity [7] [18]. Resting-state analyses employ correlation techniques, coherence analysis, and graph theory to identify spontaneous coupling between electrical and hemodynamic fluctuations [18].
Recent methodological advances include the application of Dirichlet distribution parameter estimation to model uncertainty in multimodal data fusion and Dempster-Shafer theory for evidence combination from heterogeneous signal sources [20]. These approaches are particularly valuable for integrating EEG and fNIRS measurements, as they provide formal frameworks for handling the different temporal characteristics and physiological origins of electrical and hemodynamic signals [20] [1].
Table 1: Quantitative Parameters of Neurovascular Coupling
| Parameter | Typical Value/Range | Measurement Context | Technical Notes |
|---|---|---|---|
| CBF Response to CO₂ | 1-6% per mm Hg CO₂ change | Global cerebral hemodynamics | Particularly sensitive in hypercapnic range [17] |
| Temporal Delay | 1-5 seconds post-stimulus | Adult sensory stimulation | Longer delays observed in neonates [18] |
| Hemodynamic Response Duration | 10-20 seconds | Brief sensory stimulus (e.g., finger tapping) | Duration depends on stimulus length and intensity [18] |
| Spatial Extent of CBF Increase | 10-20% in PCA, 5-8% in MCA | Task-activated regions | Varies by vascular territory [18] |
| Initial Dip Onset | 1-2 seconds post-stimulus | BOLD fMRI/fNIRS | Not always reliably detected [3] |
The integration of electroencephalography (EEG) and functional near-infrared spectroscopy (fNIRS) provides a powerful multimodal approach for investigating neurovascular coupling by simultaneously capturing electrical neural activity and hemodynamic responses [1]. These modalities offer complementary technical properties: EEG delivers millisecond-level temporal resolution of neural oscillations but limited spatial resolution, while fNIRS provides better spatial localization of hemodynamic changes with reasonable temporal resolution (typically up to 10 Hz) [21] [1]. Furthermore, the signals originate from distinct yet coupled physiological processes—EEG primarily reflects postsynaptic potentials of cortical pyramidal neurons, while fNIRS measures hemodynamic changes associated with neurovascular coupling [1].
The practical advantages of combined EEG-fNIRS systems include portability, non-invasiveness, relative tolerance to movement artifacts, and compatibility with naturalistic environments [7] [1]. This makes them particularly suitable for studying brain function in special populations including children, patients with implants, and those requiring bedside monitoring [21] [1]. The built-in validation provided by measuring both electrical and hemodynamic responses to neural activation strengthens the interpretation of observed brain activity patterns [1].
Three primary methodological categories have emerged for analyzing concurrent EEG-fNIRS data:
EEG-informed fNIRS analysis: Using EEG features to guide the analysis of hemodynamic responses, particularly for event-related paradigms where precise timing from EEG can improve modeling of the delayed fNIRS response [1].
fNIRS-informed EEG analysis: Employing hemodynamic information to constrain source localization of electrical neural activity, addressing EEG's inherent spatial limitations [1].
Parallel EEG-fNIRS analysis: Processing both modalities independently followed by integrated interpretation, often using advanced fusion techniques such as classifier combination or evidence theory [20] [1].
Deep learning approaches have shown particular promise for multimodal fusion, with recent implementations using dual-scale temporal convolution and hybrid attention modules for EEG feature extraction combined with spatial convolution and gated recurrent units for fNIRS processing [20]. Decision-level fusion using Dempster-Shafer theory has achieved 83.26% accuracy in motor imagery classification, demonstrating the practical benefits of integrated analysis [20].
Figure 2: EEG-fNIRS Integration Workflow. This diagram outlines the major approaches for combining electrophysiological (EEG) and hemodynamic (fNIRS) data, including preprocessing steps, fusion strategies, and application domains.
Well-controlled stimulus paradigms are essential for reliable assessment of neurovascular coupling. In adult populations, standardized motor tasks such as finger tapping reliably activate contralateral motor cortex, producing measurable hemodynamic responses [18]. Visual stimuli using checkerboard patterns or flashing LEDs, and auditory stimuli using tone sequences or speech samples, effectively activate respective sensory cortices [18]. For neonatal and infant populations, flashing LEDs are particularly advantageous as they can be administered during sleep, minimizing movement artifacts [18].
Block-designed experiments, alternating between active and rest conditions, enhance the detection of stimulus-locked hemodynamic responses by improving signal-to-noise ratio through multiple trial averaging [3]. Event-related designs allow for more natural task structures and reduce habituation effects but typically require more sophisticated analysis approaches [3]. The selection of appropriate inter-stimulus intervals is critical, as the hemodynamic response requires 10-20 seconds to return to baseline following a brief stimulus [18].
To improve reliability and comparability across studies, standardized assessment protocols for neurovascular coupling should incorporate the following elements:
Physiological monitoring: Continuous measurement of end-tidal CO₂, blood pressure, and heart rate to account for systemic influences on cerebral hemodynamics [17].
Artifact handling: Implementation of robust artifact removal techniques, including short-separation channels for fNIRS to correct for superficial scalp blood flow contributions, and advanced algorithms for EEG artifact suppression [7] [21].
Stimulus characterization: Precise documentation of stimulus parameters (duration, intensity, modality) and timing synchronization between stimulus presentation and data acquisition [18].
Quality metrics: Establishment of predefined quality thresholds for signal-to-noise ratio, motion artifact contamination, and physiological noise levels for data inclusion [21] [3].
Recent methodological developments include automated software that coalesces repetitive trials into single contours and extracts multiple neurovascular coupling metrics including response latency, peak amplitude, and duration [17]. Such tools also facilitate normalization of responses to dynamic changes in arterial blood gases, which significantly influence the hyperemic response [17].
Table 2: Essential Research Materials for Neurovascular Coupling Studies
| Category | Specific Examples | Function/Application | Technical Notes |
|---|---|---|---|
| fNIRS Systems | Continuous-wave NIRS, Time-domain NIRS, Frequency-domain NIRS | Hemodynamic response measurement | CW-NIRS most common due to cost and simplicity [1] |
| EEG Systems | High-density EEG systems, Amplifiers, Active/passive electrodes | Neural electrical activity recording | Sampling rates typically 256-1024 Hz; higher than fNIRS [1] |
| Stimulation Equipment | LCD displays for visual patterns, Tactile stimulators, Auditory delivery systems | Controlled stimulus presentation | LED-based visual stimuli suitable for sleeping neonates [18] |
| Physiological Monitoring | Capnography, Pulse oximetry, Non-invasive BP monitoring | Control for systemic confounders | Essential for normalizing cerebral hemodynamic responses [17] |
| Analysis Software | Homer2, NIRS-SPM, EEGLAB, FieldTrip, Custom MATLAB scripts | Data processing and fusion | Increasing availability of open-source toolboxes [7] [1] |
| Computational Modeling Tools | MATLAB with custom scripts, Monte Carlo simulation packages | Quantitative modeling of NVC | Enables integration of animal and human data [19] |
Neurovascular coupling assessment has significant clinical utility across multiple neurological conditions. In critical care settings, fNIRS monitoring of cerebral oxygenation and autoregulation provides valuable information for managing patients with stroke, traumatic brain injury, and undergoing neurosurgical procedures [3]. The non-invasive nature and portability of fNIRS-EEG systems make them particularly suitable for prolonged monitoring in intensive care units [3].
In neurodegenerative disorders, impaired neurovascular coupling has been identified as a potential early biomarker for Alzheimer's disease and vascular dementia [18] [1]. The combination of fNIRS with EEG offers a practical method for repeated assessment of neurovascular function in these patient populations [1]. Additionally, epilepsy monitoring benefits from simultaneous electrical and hemodynamic recording, as the spread of seizure activity involves complex neurovascular interactions that can be localized through multimodal integration [3].
Future methodological developments are likely to focus on several key areas. First, the creation of high-density whole-head optode arrays with anatomical co-registration will improve spatial accuracy and enable more reliable individual subject analysis [3]. Second, advanced signal processing techniques including vector diagram analysis to detect the initial dip in the hemodynamic response may enhance temporal precision for event detection [3]. Third, the integration of additional physiological parameters such as cytochrome oxidase measurements will provide more comprehensive assessment of metabolic function alongside hemodynamic responses [21] [3].
The expanding application of machine learning and deep learning approaches to multimodal data fusion promises to advance both basic understanding of neurovascular coupling and its clinical applications. These techniques enable more sophisticated pattern recognition in complex datasets and improve real-time classification of brain states for brain-computer interfaces and clinical monitoring systems [20] [3]. As these methodologies mature, standardized protocols for assessing neurovascular coupling will become increasingly important for translating research findings into clinical practice.
{article}
Functional near-infrared spectroscopy (fNIRS) and electroencephalography (EEG) are revolutionizing cognitive neuroscience by enabling brain imaging in naturalistic, real-world settings. This paradigm shift addresses a critical limitation of traditional neuroimaging techniques like functional magnetic resonance imaging (fMRI) and positron emission tomography (PET), which require strict physical constraints and immobilization, thereby compromising ecological validity. This article details the technical advantages of portable fNIRS-EEG systems and provides structured Application Notes and Protocols for their use. Framed within a broader thesis on data fusion techniques, the content provides researchers and drug development professionals with practical methodologies for leveraging multimodal signal fusion to gain deeper, more authentic insights into brain function.
Traditional neuroimaging techniques, particularly fMRI and PET, have been foundational for human neuroscience. Functional MRI (fMRI) explores brain architecture and activity patterns by measuring the blood-oxygen-level-dependent (BOLD) signal, offering high spatial resolution for localizing neural activity across the entire brain [22]. PET imaging relies on the use of radiolabeled tracers to measure metabolic processes, such as glucose utilization. However, a key assumption of these methods is that findings from highly controlled laboratory settings can be generalized to mental processes in real-world scenarios [23]. This assumption is increasingly being challenged.
The pursuit of ecological validity—the extent to which experimental findings reflect real-world behavior and experience—is driving the adoption of naturalistic neuroimaging [23]. This approach utilizes portable technologies like fNIRS and EEG to study the brain in dynamic, interactive contexts. Research has revealed significant differences in brain activation between traditional paradigms and those that approximate real life. For example, the amygdala, often identified as a "fear center" in studies using static threatening images, is not engaged in the same way during dynamic and prolonged fear experiences, suggesting that different neural systems underpin acute versus sustained affective states [23]. Moving beyond the scanner is therefore not merely a technical convenience but a necessity for a complete understanding of brain function.
The following table summarizes the key characteristics of fMRI and PET compared to the portable modalities used for naturalistic imaging.
Table 1: Comparative Analysis of Neuroimaging Modalities for Real-World Settings
| Feature | fMRI | PET | fNIRS | EEG | fNIRS-EEG Fusion |
|---|---|---|---|---|---|
| Spatial Resolution | High (millimeters) [22] | High (millimeters) | Moderate (centimeters) [24] | Low (centimeters) [24] | Enhanced & Complementary [24] |
| Temporal Resolution | Slow (1-5 seconds) [22] | Very Slow (minutes) | Slow (1-5 seconds) | Very Fast (milliseconds) [24] | Enhanced & Complementary [24] |
| Portability | No | No | Yes [25] [24] | Yes [24] | Yes [26] |
| Tolerance to Motion | Low | Low | High [25] | Moderate | High [25] |
| Measurement Type | Hemodynamic (BOLD) [22] | Metabolic (Glucose) | Hemodynamic (HbO/HbR) [24] | Electrical Activity [24] | Neurovascular Coupling [7] |
| Key Real-World Advantage | N/A (Scanner-bound) | N/A (Scanner-bound) | Bedside/ecological use [25] | Real-time tracking | Comprehensive brain activity decoding [7] |
The fusion of fNIRS and EEG creates a synergistic system that overcomes the limitations of each individual modality. EEG provides excellent temporal resolution to track rapid neural events, while fNIRS offers better spatial resolution for localizing the hemodynamic response associated with neural activity [24]. This combination is particularly powerful for studying complex, real-world cognitive processes and for applications in clinical populations where traditional scanning is impractical [25].
Integrating signals from EEG and fNIRS requires sophisticated data fusion strategies. The choice of fusion stage presents a key research decision, with each level offering distinct advantages.
Table 2: Data Fusion Techniques for Multimodal EEG-fNIRS Data
| Fusion Stage | Description | Key Findings & Applications |
|---|---|---|
| Early-Stage (Data-Level) Fusion | Raw or pre-processed data from both modalities is combined into a single data structure for input into a model. | Shown to significantly improve classification of motor imagery tasks compared to middle- or late-stage fusion [15]. |
| Feature-Level Fusion | Features are extracted from each modality separately and then concatenated into a combined feature vector. | Direct concatenation has performed on par with the best decision-level techniques in emotion recognition tasks [27]. |
| Decision-Level (Late) Fusion | Each modality is processed through separate models, and the final decisions (e.g., classifications) are combined. | Average-based soft voting has shown strong performance in emotion recognition across valence, arousal, and dominance dimensions [27]. |
The following diagram illustrates the workflow of a Y-shaped neural network, a common architecture for investigating these fusion strategies:
Multimodal Fusion Network - This Y-shaped network processes EEG and fNIRS signals through separate encoders before fusing them for a unified output.
Background: Motor imagery (MI)-based neurofeedback is a promising tool for post-stroke motor rehabilitation, aiming to stimulate neuroplasticity in lesioned motor areas [26]. Unimodal neurofeedback has limitations, with approximately 30% of users unable to self-regulate brain activity effectively [26].
Objective: To assess the efficacy of multimodal EEG-fNIRS neurofeedback compared to unimodal (EEG-only or fNIRS-only) NF for upper-limb motor imagery.
Experimental Protocol:
Table 3: Research Reagent Solutions for Motor Imagery Protocol
| Item | Specification / Example | Function in Experiment |
|---|---|---|
| fNIRS System | NIRScout XP (NIRx) with 16 detectors, 16 LED sources | Measures hemodynamic response (HbO/HbR) in the cortex during motor imagery. |
| EEG System | ActiCHamp (Brain Products) 32-channel amplifier | Records electrical brain activity (ERD/ERS) with high temporal resolution. |
| Integrated Cap | EasyCap (CNX-128) with integrated EEG & fNIRS holders | Ensures stable and co-registered positioning of all sensors on the scalp. |
| Stimulus Presentation Software | MATLAB with Psychtoolbox or Presentation | Presents visual cues and the real-time neurofeedback metaphor to the participant. |
| Real-time Processing Platform | Custom software (e.g., via Lab Streaming Layer) | Synchronizes data streams, extracts features, computes NF score, and controls feedback. |
Background: Emotions are complex states that unfold dynamically in real-world contexts. Traditional experiments using static stimuli lack the ecological validity to capture these processes fully [23] [27].
Objective: To develop a personalized multi-modal fNIRS-EEG system for decoding the dynamic trajectories of emotional experiences in naturalistic settings.
Experimental Protocol:
The workflow for such a naturalistic emotion study is outlined below:
Emotion Decoding Workflow - This protocol uses naturalistic stimuli and fused neuroimaging to decode emotional states.
The integration of fNIRS and EEG represents a significant step toward truly naturalistic brain imaging, offering a compelling alternative to the scanner-bound environment of fMRI and PET. The synergistic combination of these modalities provides a more comprehensive picture of brain function by capturing complementary aspects of neural activity [24]. This is particularly valuable for tracking the dynamic, context-dependent, and personalized nature of cognitive and affective processes as they occur in real-world environments [23].
Future advancements in this field will likely focus on refining data-driven fusion techniques, such as source-decomposition methods, to better reveal latent neurovascular coupling processes [7]. Furthermore, the development of more miniaturized, robust, and user-friendly hardware will continue to push neuroimaging out of the laboratory and into clinics, homes, and real-world settings, opening unprecedented opportunities for diagnosing, monitoring, and treating neurological and psychiatric disorders [25] [28].
{/article}
The integration of electroencephalography (EEG) and functional near-infrared spectroscopy (fNIRS) represents a leading approach in non-invasive neuroimaging, capitalizing on the complementary strengths of each modality. This fusion provides a more comprehensive picture of brain activity by combining EEG's millisecond-level temporal resolution with fNIRS's superior spatial localization capabilities [7] [29] [6]. While EEG captures postsynaptic potentials from pyramidal neurons, reflecting electrical brain activity with high temporal precision, fNIRS measures hemodynamic responses associated with neural activity through near-infrared light, providing better spatial resolution and resistance to motion artifacts [29] [6]. This combination is particularly valuable for brain-computer interface (BCI) applications, clinical neurology, and neurorehabilitation research, where understanding both rapid neural dynamics and underlying metabolic processes is essential [7] [29]. The technical and physical foundations of these signal acquisition methods must be thoroughly understood to design effective experiments and implement appropriate data fusion strategies.
EEG measures electrical potentials generated by the synchronized postsynaptic activity of pyramidal neurons in the cerebral cortex. These electrical signals originate from ionic currents flowing during neuronal excitation and inhibition, creating dipole fields that can be detected on the scalp surface [6]. The conductive properties of biological tissues (scalp, skull, cerebrospinal fluid, and brain) volume-conduct these signals, which typically range from 10 to 100 microvolts in amplitude. EEG electrodes detect voltage differences between active sites and reference points, capturing oscillatory activity across multiple frequency bands (delta, theta, alpha, beta, gamma) that correlate with various brain states and cognitive processes.
Table 1: Key Characteristics of EEG Signal Acquisition
| Parameter | Specification | Physiological Basis | Technical Considerations |
|---|---|---|---|
| Physical Signal | Electrical potentials | Postsynaptic potentials of pyramidal neurons | Voltage differences measured at scalp surface |
| Temporal Resolution | Millisecond level (<100 ms) | Direct measurement of neural firing | Limited only by sampling rate (typically 250-2000 Hz) |
| Spatial Resolution | ~1-3 cm (limited) | Signal smearing by volume conduction | Improved with high-density arrays (64-256 channels) |
| Frequency Bands | Delta (0.5-4 Hz), Theta (4-8 Hz), Alpha (8-13 Hz), Beta (13-30 Hz), Gamma (30-100 Hz) | Different cognitive states and processes | Filter settings critical for isolating bands of interest |
| Artifact Sources | Ocular, muscle, cardiac, environmental noise | Non-neural bioelectric activity, external interference | Requires robust artifact detection/correction strategies |
fNIRS utilizes near-infrared light (typically 690-850 nm wavelengths) to measure hemodynamic changes in cortical tissue. The physical principle is based on the modified Beer-Lambert law, which describes light attenuation in scattering media like biological tissue [30]. At these wavelengths, light penetrates biological tissues effectively and is differentially absorbed by oxygenated (HbO) and deoxygenated hemoglobin (HbR), enabling quantification of changes in cerebral blood oxygenation associated with neural activity through neurovascular coupling.
The technique employs optical sources (emitting specific wavelengths) and detectors arranged on the scalp with specific separation distances (typically 3-4 cm), creating measurement channels. The differential pathlength factor (DPF) accounts for light scattering in tissue, while the partial volume factor (PVF) corrects for the fraction of the path that passes through brain tissue versus other tissues [30]. These are often combined into a partial pathlength factor (PPF) for practical application in concentration calculations.
Table 2: Key Characteristics of fNIRS Signal Acquisition
| Parameter | Specification | Physiological Basis | Technical Considerations |
|---|---|---|---|
| Physical Signal | Light attenuation | Hemoglobin absorption of NIR light | Measures HbO and HbR concentration changes |
| Temporal Resolution | ~0.1-1.0 seconds | Hemodynamic response delay (neurovascular coupling) | Limited by slow hemodynamic response (5-8 sec peak) |
| Spatial Resolution | 5-10 mm | Limited by source-detector separation (typically 3 cm) | High-density arrays enable tomographic reconstruction (HD-DOT) |
| Depth Sensitivity | Superficial cortex (1-3 cm) | Light scattering limits penetration | Short-separation channels help correct for superficial artifacts |
| Artifact Sources | Motion, scalp blood flow, systemic physiology | Non-cerebral hemodynamics, movement | Motion correction algorithms essential |
The synergy between EEG and fNIRS stems from their complementary measurement principles and spatiotemporal characteristics. EEG provides direct measurement of neural electrical activity with excellent temporal resolution, ideal for tracking rapid neural dynamics during cognitive tasks [6]. Conversely, fNIRS offers indirect measurement of neural activity via neurovascular coupling with better spatial specificity and resistance to movement artifacts, making it suitable for naturalistic environments [29] [6]. This complementary relationship enables more comprehensive brain monitoring, as each modality captures different aspects of brain function with different artifact profiles, providing built-in validation through neurovascular coupling principles [29].
Proper subject preparation is essential for high-quality simultaneous EEG-fNIRS data acquisition. Begin by measuring head circumference and selecting an appropriate integrated EEG-fNIRS cap size (e.g., Model M for 54-58 cm circumferences) [31]. Position the cap according to the international 10-20 or 10-5 system, ensuring comprehensive coverage of regions of interest (typically motor, prefrontal, and parietal cortices depending on the experimental paradigm) [31] [32].
For EEG setup, apply electrolytic gel to achieve electrode impedances below 10 kΩ, using the left or right mastoid as reference (M1/M2) [32] [33]. For fNIRS optode placement, ensure good skin contact without excessive pressure, verifying that source-detector pairs maintain approximately 3 cm separation to achieve optimal cortical sensitivity [31] [32]. Implement a synchronization trigger between EEG and fNIRS systems, typically using event markers transmitted from stimulus presentation software (e.g., E-Prime) to both acquisition systems simultaneously [31].
Table 3: Standardized Equipment Parameters for Simultaneous EEG-fNIRS
| Component | EEG Specifications | fNIRS Specifications | Integration Requirements |
|---|---|---|---|
| Acquisition System | g.HIamp amplifier (g.tec) or Neuroscan SynAmps2 | NirScan (Danyang Huichuang) or NIRScout (NIRx) | Synchronized triggering capability |
| Sampling Rate | 256-1000 Hz | 7.8125-11 Hz | Integer ratio recommended for data fusion |
| Channel Count | 32-64 electrodes | 24-90 measurement channels | Co-registered positioning via integrated cap |
| Reference Scheme | Mastoid (M1/M2) or average reference | Short-separation channels (<1 cm) for superficial artifact removal | Anatomical co-registration of all elements |
| Stimulus Synchronization | Event markers from E-Prime 3.0 or equivalent | Same event markers as EEG | Simultaneous trigger to both systems |
Motor imagery (MI) paradigms are widely used in BCI research and provide an excellent framework for demonstrating EEG-fNIRS integration. The following protocol is adapted from established methodologies used in recent multimodal datasets [31] [32]:
Pre-experiment Preparation: Conduct a grip strength calibration procedure using a dynamometer and stress ball to enhance motor imagery vividness. Have participants perform repeated 5 kg maximal force exertions followed by equivalent force applications using a stress ball, then practice grip training at one contraction per second [31].
Baseline Recording: Begin with 1-minute eyes-closed resting state followed by 1-minute eyes-open resting state, demarcated by an auditory cue (200 ms beep) [31].
Trial Structure: Implement the following sequence for each trial:
Session Structure: Include at least 2 sessions per subject, with each session containing 15-40 trials per condition. Implement 5-10 minute breaks between sessions to mitigate fatigue effects [31] [32].
Task Variants: For upper limb MI, include multiple joint movements: hand open/close, wrist flexion/extension, wrist abduction/adduction, elbow pronation/supination, elbow flexion/extension, shoulder pronation/supination, shoulder abduction/adduction, and shoulder flexion/extension [32].
For cognitive studies, the n-back task provides a well-established paradigm for investigating working memory load:
Instruction Display (2 s): Present task instructions indicating 0-back, 2-back, or 3-back condition [6].
Task Period (40 s): Display a random one-digit number every 2 s for 0.5 s, followed by a 1.5-s fixation cross [6]. In 0-back, participants press target button for a specific pre-defined number; in 2-back/3-back, participants indicate whether current number matches the number shown 2/3 trials earlier.
Rest Period (20 s): Participants focus on a fixation cross while minimizing movement [6].
Block Structure: Each participant should complete multiple blocks (e.g., 20 trials × 3 series × 3 sessions) to ensure adequate statistical power [6].
Diagram 1: EEG-fNIRS Preprocessing and Fusion Pathway. This workflow illustrates the parallel processing streams for EEG and fNIRS data before multimodal fusion, highlighting key steps including filtering, artifact correction, and feature extraction.
Preprocessing choices significantly impact downstream analysis and decoding performance. For EEG, optimal preprocessing typically includes high-pass filtering (≥1 Hz cutoff) to remove slow drifts, with evidence showing that higher high-pass filter cutoffs consistently increase decoding performance [34]. For fNIRS, motion correction is paramount, with algorithms like CBSI (correlation-based signal improvement), PCA, wavelet-based methods, and spline interpolation being commonly employed [30]. Physiological noise removal via bandpass filtering (0.01-0.5 Hz) effectively isolates hemodynamic responses from cardiac (~1-2 Hz), respiratory (~0.4 Hz), and Mayer wave (~0.1 Hz) interference [30].
Diagram 2: Multimodal Fusion Hierarchy. This diagram categorizes the primary fusion strategies for EEG-fNIRS integration, ranging from early data-level fusion to late decision-level fusion, with associated methodologies and applications.
Three primary fusion approaches dominate EEG-fNIRS integration:
Early Fusion: Concatenating raw data or low-level features before classification, potentially improving performance but requiring temporal alignment of fundamentally different signal dynamics [6].
Intermediate Fusion: Employing cross-modal attention mechanisms or joint feature learning to model dynamic dependencies between modalities, automatically focusing on relevant signals across time and space [6].
Late Fusion: Processing modalities independently through specialized architectures (e.g., dual-branch networks) then fusing decisions using methods like Dempster-Shafer Theory (DST) to model and combine uncertainty estimates from each modality [20] [6].
Deep learning approaches have shown particular promise, with architectures like MBC-ATT employing independent branches for each modality with cross-modal attention mechanisms, achieving significant improvements in classification accuracy for cognitive tasks [6]. Similarly, evidence theory-based fusion using Dirichlet distribution parameter estimation and DST has demonstrated 3-5% accuracy improvements in motor imagery classification compared to unimodal approaches [20].
Table 4: Essential Research Tools for EEG-fNIRS Acquisition and Analysis
| Tool Category | Specific Solutions | Function/Purpose | Implementation Examples |
|---|---|---|---|
| Acquisition Hardware | Integrated EEG-fNIRS caps (Model M) | Simultaneous signal acquisition with co-registered positioning | Custom-designed caps with 32 EEG electrodes, 32 optical sources, 30 photodetectors [31] |
| Amplification Systems | g.HIamp amplifier (g.tec), Neuroscan SynAmps2 | EEG signal amplification and digitization | 64-channel systems with 1000 Hz sampling rate, mastoid reference [32] |
| fNIRS Systems | NirScan (Danyang Huichuang), NIRScout (NIRx) | Optical signal transmission and detection | Continuous wave systems with 690/830 nm wavelengths, ~10 Hz sampling [31] [32] |
| Stimulus Presentation | E-Prime 3.0, PsychToolbox | Experimental paradigm implementation | Precise timing control with synchronized triggers to both acquisition systems [31] |
| Preprocessing Tools | MNE-Python, Homer3, EEGLAB | Signal preprocessing and artifact removal | Standardized pipelines for filtering, referencing, motion correction [34] [30] |
| Fusion Frameworks | MBC-ATT, EEG-fNIRS evidence theory fusion | Multimodal integration and classification | Deep learning with cross-modal attention; Dempster-Shafer theory for uncertainty modeling [20] [6] |
| Validation Datasets | HEFMI-ICH, TU-Berlin-A, multimodal n-back | Method benchmarking and comparison | Public datasets with simultaneous EEG-fNIRS recordings [20] [31] [6] |
The technical and physical basis of EEG and fNIRS signal acquisition provides the foundation for effective multimodal brain imaging research. EEG's millisecond temporal resolution for capturing electrical neural activity complements fNIRS's superior spatial localization of hemodynamic responses, creating a powerful synergistic combination. The experimental protocols and processing pathways detailed in this document establish standardized methodologies for acquiring high-quality simultaneous data. Current research indicates that sophisticated fusion strategies, particularly deep learning approaches with cross-modal attention and evidence theory-based decision fusion, consistently outperform unimodal classification, with improvements of 3-5% in accuracy for motor imagery and cognitive tasks [20] [6]. As the field advances, focus on standardized preprocessing, artifact handling, and open-access datasets will be crucial for accelerating developments in multimodal brain imaging and its applications in BCI, clinical neurology, and cognitive neuroscience.
Data-level fusion, often termed early fusion or feature-level fusion, is a data integration strategy wherein raw data or low-level features from multiple sources are combined before being input into a machine learning or statistical model [35] [36]. In the context of cognitive neuroscience and neuroimaging, this approach involves the direct combination of raw or minimally processed signals from different modalities, such as EEG and fNIRS, into a single, unified dataset for subsequent analysis. This method stands in contrast to late fusion (decision-level fusion), where each modality is processed independently and their outputs are combined at the decision stage [35].
The principal advantage of early fusion lies in its potential to model low-level, cross-modal interactions that might be lost when processing modalities separately [36]. For EEG and fNIRS, which capture complementary aspects of brain activity—electrical and hemodynamic, respectively—early fusion allows a model to learn the complex, non-linear relationships between immediate neuronal firing (EEG) and the delayed hemodynamic response (fNIRS) directly from the data. This is particularly valuable for researching brain-computer interfaces, cognitive workload assessment, and the neurovascular correlates of neurological disorders, offering a more holistic view of brain function. However, this approach presents significant challenges, including the need for precise temporal alignment of signals, handling the high-dimensional feature spaces that result from concatenation, and managing the different sampling rates inherent to each modality [35] [37].
A rigorous and standardized preprocessing pipeline for each modality is a critical prerequisite for successful data-level fusion. Inconsistent or inadequate preprocessing can introduce artifacts and noise that obscure genuine neural signals and lead to unreliable fusion outcomes.
The goal of fNIRS preprocessing is to isolate the task-related hemodynamic response function (HRF) by removing physiological, instrumental, and motion artifacts from the raw optical intensity signals [38]. The following protocol, adaptable using toolboxes like HOMER3 or MNE-Python, outlines the essential steps [39].
Protocol: fNIRS Data Preprocessing
EEG preprocessing aims to isolate neural activity from environmental and physiological artifacts to obtain clean event-related potentials (ERPs) or oscillatory features.
Protocol: EEG Data Preprocessing
cleanline function (in EEGLAB) to adaptively remove power line interference (e.g., 50/60 Hz and its harmonics) [40].Table 1: Summary of Key Preprocessing Steps for fNIRS and EEG
| Step | fNIRS | EEG |
|---|---|---|
| Primary Goal | Extract HRF | Extract ERP/Oscillations |
| Filtering | Bandpass (0.01-0.7 Hz) | High-pass (e.g., 1 Hz) & Line Noise Removal |
| Key Transformation | MBLL (to HbO/HbR) | Re-referencing, Spatial Filtering |
| Artifact Handling | Savitzky-Golay, SCI | Artifact Subspace Reconstruction (ASR) |
| Epoching | Around stimulus onset (e.g., -5 to 15 s) | Around stimulus onset (e.g., -0.2 to 1 s) |
Once the individual modalities are preprocessed, the core process of data-level fusion can begin. This involves bringing the data into a common representational space and combining them.
The first critical step is the precise temporal alignment of the fNIRS and EEG data streams.
After alignment, features are extracted from each modality and combined.
The following diagram illustrates the complete workflow from raw data to a fused feature vector.
Data-Level Fusion Workflow for EEG and fNIRS
This protocol provides a detailed methodology for a classic experiment that generates data suitable for EEG-fNIRS data-level fusion, based on a finger-tapping motor paradigm [39].
Protocol: Simultaneous EEG-fNIRS Recording During Motor Execution
[HbO_M1, Beta_ERD_C3].Table 2: Essential Equipment and Software for EEG-fNIRS Fusion Research
| Item | Function / Application |
|---|---|
| Continuous-Wave fNIRS System | Records changes in light intensity to calculate hemodynamic activity. Systems with multiple wavelengths (e.g., 760 nm, 850 nm) are required to resolve HbO and HbR [38]. |
| High-Density EEG System | Records electrical activity from the scalp. Systems with 64+ channels and capable of receiving external triggers are recommended for source localization and synchronization. |
| Trigger Interface Box | A critical hardware component for sending synchronized TTL pulse triggers from the stimulus computer to both EEG and fNIRS systems to ensure precise temporal alignment of data streams. |
| HOMER3 Software Package | A standard, open-source toolbox (running in MATLAB) specifically designed for fNIRS data preprocessing, visualization, and analysis [40] [39]. |
| EEGLAB Toolbox | A widely used, open-source MATLAB toolbox providing an interactive environment for processing and analyzing EEG data, including advanced techniques like ASR [40]. |
| MNE-Python Library | A powerful, open-source Python library for exploring, visualizing, and analyzing human neurophysiological data, including comprehensive support for both MEG, EEG, and fNIRS [39]. |
Within the broader scope of data fusion techniques for electroencephalography (EEG) and functional near-infrared spectroscopy (fNIRS) research, feature-level fusion represents a powerful strategy to overcome the inherent limitations of each modality when used in isolation. This approach involves the extraction and systematic integration of distinctive features from each signal type before they are fed into a classification or analysis model. The fundamental rationale is that EEG and fNIRS provide complementary insights into brain function: EEG captures millisecond-scale electrical activity with high temporal resolution but limited spatial precision, while fNIRS tracks slower hemodynamic changes (changes in oxyhemoglobin (HbO) and deoxyhemoglobin (HbR) concentrations) with better spatial localization on the cortical surface [41] [6] [9]. By fusing temporal and spectral features from EEG with hemodynamic patterns from fNIRS, researchers can achieve a more comprehensive and robust decoding of brain states, which is critical for applications in cognitive neuroscience, clinical diagnosis, and neurorehabilitation [7] [31].
The subsequent sections detail the core methodology, provide a step-by-step experimental protocol, present validation results, and list essential research tools for implementing this multimodal fusion approach.
The proposed feature-level fusion framework involves a structured pipeline from data acquisition to the integrated feature vector ready for classification. The core components are feature extraction from each modality and their subsequent fusion strategy.
EEG Feature Extraction: EEG signals are processed to extract features that characterize the brain's electrical activity in time, frequency, and space. Key steps include:
fNIRS Feature Extraction: fNIRS signals are processed to yield features related to blood oxygenation changes.
After extraction, features from both modalities are integrated into a single, comprehensive feature vector. The diagram below illustrates the complete workflow.
Beyond simple concatenation, advanced fusion mechanisms can be employed to optimize the interaction between features:
This protocol provides a detailed methodology for a motor imagery (MI) experiment, a common paradigm in brain-computer interface (BCI) research, using synchronous EEG-fNIRS recording and feature-level fusion.
Follow the workflow described in Section 2. The table below summarizes the quantitative performance of various classifiers when applied to fused EEG-fNIRS features, demonstrating the efficacy of this approach.
Table 1: Classifier Performance on Fused EEG-fNIRS Features for State Decoding
| Classifier Model | Reported Accuracy | Application Context |
|---|---|---|
| Support Vector Machine (SVM) | 82.10% | Etomidate use disorder identification [41] |
| Random Forest (RF) | 80.50% | Etomidate use disorder identification [41] |
| XGBoost | 78.40% | Etomidate use disorder identification [41] |
| Multimodal MBC-ATT (Deep Learning) | Outperformed conventional methods | Cognitive state (n-back, Word Generation) decoding [6] |
The validity of the feature-level fusion approach is confirmed by its ability to identify sensitive neural biomarkers and its superior performance compared to unimodal approaches.
The following table lists essential reagents, materials, and software required to establish a synchronous EEG-fNIRS recording and analysis system.
Table 2: Key Research Reagent Solutions for EEG-fNIRS Feature-Level Fusion
| Item Name | Function/Description | Example Specification/Provider |
|---|---|---|
| Hybrid EEG-fNIRS Cap | Integrated headgear with electrodes and optodes for simultaneous signal acquisition. | Custom design with 32 EEG electrodes, 32 fNIRS sources, 30 fNIRS detectors [31]. |
| EEG Amplifier | Records electrical brain activity with high temporal resolution. | g.HIamp (g.tec), NeuSen W (Neuracle) [41] [31]. |
| fNIRS System | Measures hemodynamic changes by emitting near-infrared light and detecting its attenuation. | NirSmart-6000A (Danyang Huichuang), NirScan (Danyang Huichuang), Hitachi ETG-4100 [41] [9] [31]. |
| Experiment Control Software | Presents stimuli and sends synchronization triggers to recording devices. | E-Prime (Psychology Software Tools) [31]. |
| Data Analysis Toolkit | Software for preprocessing, feature extraction, and fusion of multimodal data. | EEGLAB (for EEG), custom scripts in MATLAB/Python, Deep Learning frameworks (e.g., PyTorch) [41] [7]. |
| 3D Digitizer | Records the precise spatial coordinates of EEG electrodes and fNIRS optodes for accurate brain source localization. | Fastrak (Polhemus) [9]. |
Decision-level fusion represents a powerful strategy for combining the outputs of multiple classifiers, each potentially processing different data modalities or feature sets, to arrive at a final, more robust decision. Within the context of EEG and fNIRS signal analysis, this approach leverages the complementary strengths of these modalities—EEG provides fine temporal resolution (~5 msec) of electrophysiological activity, while fNIRS offers higher spatial resolution (~1 cm) of hemodynamic responses [4]. Unlike feature-level fusion which concatenates features before classification, decision-level fusion maintains separate processing pathways for each modality until the final decision stage, making it particularly valuable when dealing with inherently different signal characteristics and sampling rates [44]. This protocol document outlines standardized methodologies for implementing decision-level fusion techniques specifically tailored for hybrid EEG-fNIRS brain-computer interfaces and neuro-clinical applications.
Decision-level fusion operates on the principle that combining decisions from multiple classifiers can compensate for individual weaknesses and capitalize on their collective strengths. The mathematical foundation lies in probability theory and statistical inference, where the goal is to optimize the final classification accuracy beyond what any single classifier can achieve. Research has demonstrated that this approach can yield significant improvements in classification performance—one study on mental stress detection reported a +7.76% improvement compared to EEG alone and +10.57% compared to fNIRS alone [44]. Similarly, another implementation for COVID-19 patient health prediction achieved an accuracy of 97.24% by combining random forest, gradient boosting, and extreme gradient boosting classifiers [45].
Table 1: Quantitative Performance Improvements with Decision-Level Fusion
| Application Domain | Baseline Accuracy (Best Single Modality) | Decision-Level Fusion Accuracy | Performance Improvement | Citation |
|---|---|---|---|---|
| Mental Stress Detection (EEG vs fNIRS) | ~88% (EEG) | ~95.76% | +7.76% | [44] |
| COVID-19 Health Prediction | 94% (Single Model) | 97.24% | +3.24% | [45] |
| Motor Imagery Task Classification | 77.6% (Compact hBCI) | 96.74% | +19.14% | [44] |
| Mental Arithmetic Task Classification | 90.19% (Feature-level) | 98.42% | +8.23% | [44] |
The performance advantages stem from several factors: the ability to handle modality-specific characteristics optimally, reduced sensitivity to noise that may affect one modality but not the other, and the capacity to leverage temporal relationships between electrical and hemodynamic responses [4]. Furthermore, decision-level fusion provides a framework for integrating expert knowledge through the careful selection and weighting of individual classifiers, potentially incorporating domain-specific understanding of neurovascular coupling principles.
This protocol outlines the methodology for fusing decisions from multiple heterogeneous classifiers using soft voting, which aggregates probability estimates rather than hard decisions [45].
Step-by-Step Procedure:
Data Preparation and Preprocessing
Individual Classifier Training
Classifier Calibration
Decision Fusion Implementation
Performance Validation
This protocol adapts the Structural Causal Model (SCM) approach, which explicitly models causal relationships between decisions, for EEG-fNIRS fusion [46].
Step-by-Step Procedure:
Causal Graph Construction
Structural Equation Specification
Model Estimation and Training
Inference and Decision Making
Table 2: Research Reagent Solutions for EEG-fNIRS Decision-Level Fusion
| Reagent Category | Specific Tool/Algorithm | Function in Experimental Protocol |
|---|---|---|
| Base Classifiers | Random Forest, Gradient Boosting, XGBoost, SVM, k-NN | Generate preliminary decisions from features; provide diversity for ensemble [44] [47] [45] |
| Fusion Methods | Soft Voting, Structural Causal Models, Bayesian Networks, Neural Networks | Combine preliminary decisions into final output using specific mathematical frameworks [46] [45] |
| Feature Extraction | Fractal Dimension, Higher-Order Spectra, Entropy Measures, Band Power | Transform raw signals into informative features for classification [47] |
| Feature Selection | Genetic Algorithms, Mutual Information, Atomic Search Optimization | Identify most relevant features to reduce dimensionality and improve performance [4] [47] |
| Signal Processing | ICA, Bandpass Filters, Motion Artifact Removal Algorithms | Preprocess raw data to enhance signal quality and remove noise [4] [47] |
| Validation Metrics | Accuracy, F1-Score, Sensitivity, Specificity, Statistical T-Tests | Quantify and compare performance across different fusion strategies [44] [45] |
Successful implementation of decision-level fusion in EEG-fNIRS research requires careful attention to several practical considerations. First, temporal synchronization between EEG and fNIRS systems is crucial, as the hemodynamic response measured by fNIRS typically lags behind the electrical activity captured by EEG by several seconds [4]. Implementation should incorporate appropriate temporal alignment techniques, such as cross-correlation analysis or the use of shared trigger signals. Second, classifier diversity is essential for effective fusion—selecting classifiers with different inductive biases (e.g., tree-based, probabilistic, and margin-based models) creates complementarity that fusion can leverage [45]. Third, computational efficiency must be balanced with performance requirements, particularly for potential real-time applications like brain-computer interfaces where decision latency matters [44].
When deploying these methods in clinical or pharmaceutical research settings, particular attention should be paid to model interpretability. While complex ensemble methods may offer superior performance, the ability to understand and explain decisions is often critical for diagnostic applications. Techniques such as SHAP (SHapley Additive exPlanations) values or LIME (Local Interpretable Model-agnostic Explanations) can be incorporated to maintain this interpretability while benefiting from the performance advantages of decision-level fusion. Additionally, researchers should implement comprehensive validation protocols that include not only standard performance metrics but also assessments of robustness across patient populations, stability over multiple sessions, and generalizability to new data distributions.
The integration of electroencephalography (EEG) and functional near-infrared spectroscopy (fNIRS) represents a paradigm shift in non-invasive brain imaging, leveraging the complementary strengths of each modality to overcome their individual limitations. EEG provides millisecond-level temporal resolution crucial for capturing fast neural dynamics but suffers from poor spatial resolution and sensitivity to artifacts [6] [48]. In contrast, fNIRS measures hemodynamic responses with better spatial localization and robustness to motion artifacts but has limited temporal resolution [29]. This complementary relationship has motivated the development of advanced data fusion techniques, particularly cross-modal attention mechanisms and multi-branch neural networks, which enable more sophisticated integration of electrophysiological and hemodynamic brain activity patterns for enhanced brain-computer interface (BCI) performance and cognitive state decoding [6] [49] [48].
Multi-branch architectures have emerged as a dominant framework for multimodal brain signal processing, allowing specialized pathway design for modality-specific feature extraction.
MBC-ATT (Multi-Branch Convolutional Neural Network with Attention) employs independent branch structures to process EEG and fNIRS signals separately, thereby leveraging the unique advantages of each modality. The architecture implements a late fusion strategy with a cross-modal attention mechanism that selectively emphasizes relevant features across modalities, strengthening the model's ability to focus on task-relevant signals [6] [48]. Experimental results on n-back and word generation datasets demonstrate that this approach outperforms conventional methods in classification performance, validating its effectiveness for brain-computer interfaces [50] [48].
Multi-Branch GAT-GRU-Transformer utilizes three parallel branches for comprehensive feature extraction: a Graph Attention Network (GAT) for modeling spatial relationships among EEG channels, a hybrid Gated Recurrent Unit (GRU) and Transformer module for capturing temporal dependencies, and one-dimensional Convolutional Neural Networks (1D CNN) for extracting frequency-specific information [49]. This architecture achieved a classification accuracy of 55.76% on five-class motor imagery tasks, significantly outperforming single-dimensional approaches [49].
Cross-modal attention mechanisms represent a significant advancement beyond simple feature concatenation or decision-level fusion. These mechanisms dynamically model dependencies between modalities, allowing the model to automatically focus on the most relevant modalities and brain region signals according to specific task states [6] [48].
The attention mechanism in MBC-ATT implements a modality-guided attention strategy that selectively integrates information through joint modeling of cross-modal features. This approach addresses the limitations of static fusion strategies by dynamically adjusting the contribution of each modality, thereby enhancing the synergistic capability of fused signals and improving decoding accuracy and robustness [48]. This dynamic weighting is particularly valuable in complex cognitive tasks where the relative importance of EEG and fNIRS signals may vary throughout task execution.
Table 1: Performance Comparison of Multi-Branch Architectures
| Architecture | Modality | Task | Performance | Key Innovation |
|---|---|---|---|---|
| MBC-ATT [6] [48] | EEG-fNIRS | n-back & Word Generation | Outperformed conventional approaches | Cross-modal attention fusion |
| Multi-Branch GAT-GRU-Transformer [49] | EEG | Five-class Motor Imagery | 55.76% accuracy | Spatial, temporal, frequency feature integration |
| Mutual Information-Based Fusion [4] | EEG-fNIRS | Visuo-mental Task | Considerably improved hybrid classification | Mutual information feature selection |
Multiple standardized datasets have been developed to validate multimodal fusion architectures, each with specific experimental paradigms tailored to different research questions.
n-back Working Memory Task: Participants complete 0-back, 2-back, and 3-back conditions in randomized blocks. Each block consists of a 2-second instruction display, followed by a 40-second task period where a random one-digit number is displayed every 2 seconds for 0.5 seconds, followed by a 1.5-second fixation cross. After the task period, participants enter a 20-second rest period focusing on a fixed cross [6] [48]. Each participant typically performs 180 trials (20 trials × 3 series × 3 sessions) to ensure data adequacy and reliability.
Word Generation Task: This paradigm investigates neural activity characteristics of language-related brain regions. Each trial consists of a 2-second task prompt, a 10-second task execution period, and a 13-15 second rest period. During the task period, a random letter is displayed, and participants must quickly generate and silently list as many words as possible starting with that letter while avoiding repetition [48].
Motor Imagery Paradigms: Standardized motor imagery experiments typically involve visual cue presentation (2 seconds) followed by an execution phase (4-10 seconds) where participants perform kinesthetic motor imagery of specific movements, followed by rest periods (10-20 seconds) [31] [32]. The HEFMI-ICH dataset incorporates a unique grip strength calibration procedure using a dynamometer and stress ball to enhance motor imagery vividness and consistency [31].
Table 2: Standardized Data Acquisition Parameters
| Parameter | EEG Specifications | fNIRS Specifications |
|---|---|---|
| Sampling Rate | 256-1000 Hz [31] [32] | 7.8125-11 Hz [31] [32] |
| Electrode/Optode Configuration | 32-64 electrodes following international 10-20 system [31] [32] | 24-90 channels with 3cm source-detector separation [31] [32] |
| Key Metrics | Bandpower, amplitude, phase, frequency features [4] | Oxygenated (HbO) and deoxygenated hemoglobin (HbR) concentrations [4] |
| Reference Scheme | Left mastoid (M1) reference [32] | Short-separation channels for systemic artifact removal [29] |
| Filter Settings | 0.5-100 Hz bandpass, 50 Hz notch filter [32] | Bandpass filter to isolate hemodynamic responses (typically 0.01-0.2 Hz) [29] |
EEG preprocessing typically involves bandpass filtering (0.5-100 Hz), notch filtering at 50 Hz for power line interference removal, and artifact removal techniques including independent component analysis (ICA) for ocular and muscle artifact rejection [29] [32]. fNIRS processing includes converting raw light intensities to optical densities, motion artifact detection and correction, bandpass filtering to isolate hemodynamic responses, and conversion to hemoglobin concentration changes using the Modified Beer-Lambert Law [29]. Temporal synchronization between modalities is crucial and is typically achieved using event markers transmitted from experimental presentation software such as E-Prime 3.0 [31].
The following diagram illustrates the complete workflow for multimodal EEG-fNIRS processing using cross-modal attention mechanisms:
The cross-modal attention mechanism implements a sophisticated feature weighting system as shown in the following diagram:
Table 3: Essential Research Tools for EEG-fNIRS Fusion Studies
| Tool Category | Specific Examples | Function & Application |
|---|---|---|
| Acquisition Systems | g.HIamp amplifier (EEG) [31], NirScan fNIRS system [31], NIRScout system [32] | Synchronized multimodal data acquisition with precise temporal alignment |
| Experimental Paradigm Software | E-Prime 3.0 [31], PsychoPy, Presentation | Precise stimulus presentation and event marker generation for temporal synchronization |
| Signal Processing Tools | MATLAB with EEGLAB [29], NIRS-KIT [29], MNE-Python, Homer2 | Preprocessing, artifact removal, and feature extraction pipelines |
| Hybrid Caps | Custom EEG-fNIRS caps with 32 electrodes + 32 sources + 30 detectors [31] | Co-localized electrophysiological and hemodynamic measurement from same cortical regions |
| Deep Learning Frameworks | PyTorch, TensorFlow, Braindecode | Implementation of multi-branch architectures and attention mechanisms |
| Validation Metrics | Classification accuracy, mutual information [4], SHAP analysis [49] | Performance quantification and model interpretability |
The application of these advanced architectures spans multiple domains, demonstrating significant performance improvements over unimodal approaches. In cognitive state decoding, the MBC-ATT framework has shown superior performance in classifying working memory load during n-back tasks and language processing during word generation tasks [6] [48]. For motor imagery classification, multi-branch architectures have achieved accuracy improvements of 5-10% compared to unimodal systems, with particularly promising applications in neurorehabilitation for intracerebral hemorrhage patients [31].
Mutual information-based feature selection approaches have demonstrated considerable improvement in hybrid classification performance compared to individual modalities and conventional classification without feature selection [4]. This approach optimizes complementarity, redundancy, and relevance between multimodal features with respect to class labels, suggesting potential efficacy for wider neuro-clinical applications.
Validation methodologies typically include k-fold cross-validation, subject-independent testing to assess generalizability, and ablation studies to quantify the contribution of individual architectural components [49]. Additionally, interpretability techniques such as SHAP analysis and Phase Locking Value calculations help identify critical EEG channels and frequency bands, aligning model decisions with known neurophysiological principles [49].
Cross-modal attention mechanisms and multi-branch neural networks represent the cutting edge of EEG-fNIRS data fusion, offering sophisticated solutions that leverage the complementary nature of electrophysiological and hemodynamic brain signals. These architectures enable dynamic, task-adaptive integration of multimodal information, moving beyond the limitations of static fusion strategies. As these approaches continue to evolve, they hold significant promise for advancing BCI systems, cognitive state monitoring, and clinical applications in neurological rehabilitation and drug development research. The standardized protocols and architectural frameworks presented in this document provide a foundation for implementing these advanced data fusion techniques in both research and clinical settings.
The integration of electroencephalography (EEG) and functional near-infrared spectroscopy (fNIRS) represents a transformative approach in neuroimaging, leveraging the complementary strengths of each modality. EEG provides millisecond-level temporal resolution for capturing rapid neural electrical activity, while fNIRS offers superior spatial localization for tracking hemodynamic responses through cortical blood oxygen changes [6]. This synergistic combination enables more comprehensive brain decoding with high spatiotemporal resolution, particularly valuable for real-world clinical applications [7].
Data fusion techniques for EEG and fNIRS can be implemented at multiple levels, including early fusion (combining raw data or low-level features), feature-level fusion (integrating extracted features before classification), and late fusion (combining decisions or high-level features) [6]. Recent advances incorporate cross-modal attention mechanisms and deep learning architectures that dynamically weight the contribution of each modality based on task demands, significantly enhancing classification accuracy and robustness across various clinical applications [6].
Table 1: Comparative Analysis of Neuroimaging Modalities
| Modality | Temporal Resolution | Spatial Resolution | Artifact Resistance | Portability | Primary Applications |
|---|---|---|---|---|---|
| EEG | Millisecond-level | Limited (cm) | Low | High | Seizure detection, consciousness assessment, basic BCI |
| fNIRS | Seconds | Good (1-2 cm) | Moderate | High | Functional localization, cognitive monitoring |
| EEG-fNIRS Fusion | High (via EEG) | Good (via fNIRS) | Moderate-High | High | Advanced BCI, drug craving detection, disorder diagnosis |
| fMRI | Seconds | Excellent (mm) | High | Low | Gold standard for spatial localization |
Objective: To objectively classify individuals with substance addiction versus healthy controls using bimodal EEG-fNIRS data and deep learning.
Experimental Design: The protocol employs a visual trigger paradigm to elicit drug cravings in individuals with substance addiction, with concurrent EEG and fNIRS monitoring [51].
Participant Profile:
Data Acquisition Specifications:
Processing Workflow:
Performance Metrics: The protocol achieves 92.6% classification accuracy with an AUC of 0.903, significantly outperforming single-modal approaches [51].
Objective: To enhance motor recovery in stroke patients through BCI-based rehabilitation integrating motor imagery and motor attempts.
Experimental Design: Randomized double-blind controlled clinical trial with 48 ischemic stroke patients (25 BCI group, 23 control group) undergoing 20-minute upper and lower limb training sessions for two weeks [52].
Intervention Protocol:
Assessment Framework:
Key Findings: The BCI group demonstrated significantly greater improvement in upper extremity motor function (ΔFMA-UE: 4.0 vs. 2.0, p = 0.046) with enhanced functional connectivity in motor-related brain regions [52].
Objective: To identify patients with cognitive motor dissociation (CMD) among disorders of consciousness (DOC) populations using fNIRS with motor imagery tasks.
Experimental Design: fNIRS measurement of hemodynamic responses in 70 prolonged DOC patients during command-driven hand-open-close motor imagery tasks [53].
Participant Profile:
Task Paradigm:
Analysis Approach:
Key Findings: Identification of 7 CMD patients (4 VS/UWS, 3 MCS-) with significantly better outcomes (3/4 vs. 1/31, P = 0.014) compared to non-CMD patients [53].
The effectiveness of EEG-fNIRS integration relies fundamentally on the principle of neurovascular coupling - the tightly regulated relationship between neural activity and subsequent hemodynamic responses.
Temporal Relationship: The EEG signal detection precedes the fNIRS hemodynamic response by approximately 1-5 seconds, creating a complementary temporal signature that can be leveraged for more accurate brain state classification [7] [54].
Table 2: Essential Research Materials and Technologies for EEG-fNIRS Research
| Category | Specific Solutions | Technical Specifications | Primary Function | Example Applications |
|---|---|---|---|---|
| EEG Systems | 8-64 channel active electrode systems (e.g., g.tec LadyBird) | Sampling rate: 256-512 Hz, Impedance: <10 kΩ | Capture millisecond-level electrical brain activity | Motor intention decoding, cognitive state assessment [52] [55] |
| fNIRS Systems | Continuous-wave systems (e.g., NirScan-6000A) | Wavelengths: 703-850 nm, Sources: 24, Detectors: 24 | Measure hemodynamic responses via HbO/HbR concentration changes | Localization of neural activation, consciousness assessment [53] |
| Multimodal Caps | Integrated EEG-fNIRS headcaps | Predefined optode/electrode layouts per 10-20 system | Ensure proper spatial co-registration of modalities | Simultaneous bimodal data acquisition [51] |
| Stimulation Paradigms | Visual trigger systems, Auditory command interfaces | Precision timing, Counterbalanced design | Elicit specific neural responses for experimental control | Drug cue reactivity, motor imagery tasks [51] [53] |
| Analysis Platforms | MATLAB, Python with MNE, Homer2, BCILab | Custom scripts for fusion algorithms | Implement data fusion pipelines and classification | Feature extraction, cross-modal attention fusion [51] [6] |
| Validation Tools | Clinical assessment scales (FMA-UE, CRS-R) | Standardized scoring protocols | Ground truth measurement for algorithm validation | Outcome verification, clinical correlation [52] [53] |
Table 3: Quantitative Performance Comparison Across Applications
| Application Domain | Classification Accuracy | Superiority Over Single Modality | Key Performance Metrics | Clinical Validation |
|---|---|---|---|---|
| Drug Addiction Detection | 92.6% [51] | Significant improvement (p<0.05) | AUC: 0.903, F1-Score: 0.91 | 56 samples/participant, n=56 [51] |
| Stroke Rehabilitation | ΔFMA-UE: 4.0 vs. 2.0 (p=0.046) [52] | Clinical significance demonstrated | DAR decrease (p=0.031), DABR decrease (p<0.001) | RCT, n=48, double-blind [52] |
| Cognitive State Decoding | Superior to conventional approaches [6] | Enhanced cross-modal synergy | Improved robustness to artifacts | n-back and WG tasks, n=26 [6] |
| CMD Identification | Favorable outcome: 3/4 vs 1/31 (p=0.014) [53] | Identified 7 CMD from 70 DOC | 6-month GOSE follow-up | Diagnostic study, n=70 patients + 70 controls [53] |
Technical Integration Challenges:
Clinical Deployment Factors:
The integration of EEG and fNIRS through advanced data fusion techniques represents a paradigm shift in clinical neuroscience, enabling unprecedented insights into brain function across neurological and psychiatric conditions. These approaches show particular promise for conditions where traditional assessment methods face limitations, such as disorders of consciousness, addiction, and stroke recovery.
This application note details a pioneering study that successfully classified individuals with drug addiction from healthy controls with an accuracy of 92.6% [56] [57]. This work is situated within a broader research thesis focused on advancing data fusion techniques for Electroencephalogram (EEG) and functional Near-Infrared Spectroscopy (fNIRS) signals. The objective was to overcome the limitations of traditional, subjective assessment methods—such as psychological scales and self-reports—by developing an objective, physiological data-driven model [56]. The core innovation lies in the AR-TSNET deep learning algorithm, which performs feature-level fusion of EEG and fNIRS data, leveraging the complementary strengths of these two modalities to achieve superior classification performance [56] [57].
The experiment followed a structured protocol designed to elicit, capture, and analyze brain activity related to drug addiction.
The study compiled a dataset from 56 participants [56] [57]. Adherence to ethical guidelines was confirmed, including institutional ethical approval and obtaining informed consent from all participants [57].
Each participant contributed 56 samples, resulting in a total dataset corresponding to responses to drug-related image stimuli [56].
A visual trigger paradigm was employed to induce drug cravings in the addicted individuals [56] [57]. During this task, simultaneous EEG and fNIRS signals were recorded.
The acquired data were processed and classified using the novel AR-TSNET deep learning network. The following workflow outlines the key stages of the experimental protocol.
The AR-TSNET model was designed with specialized sub-networks for each modality, followed by fusion and classification.
Modality-Specific Feature Extraction:
Feature-Level Fusion: The features extracted by Tception and Sception were combined at a feature level, creating a unified, information-rich representation for the classifier [56].
Attention Mechanism and Residual Connections:
Model Training and Validation:
The model's performance was rigorously quantified using standard metrics. A comparison with single-modality approaches demonstrates the clear advantage of multimodal fusion.
The model achieved an average classification accuracy of 92.6% with a standard deviation of 5.56 across the six folds [56]. The area under the Receiver Operating Characteristic (ROC) curve (AUC) was 0.903, confirming the model's strong ability to discriminate between the two groups [56].
Table 1: Overall Performance Metrics of the AR-TSNET Model
| Metric | Value | Interpretation |
|---|---|---|
| Accuracy | 92.6% | Overall proportion of correct predictions [56] |
| AUC | 0.903 | Excellent model discriminative ability [56] |
| Standard Deviation | 5.56 | Variance in accuracy across cross-validation folds [56] |
A critical analysis was conducted to compare the performance of the fused model against models using only EEG or only fNIRS data. The results conclusively show that the bimodal approach provides superior performance across all evaluated metrics [56].
Table 2: Comparison of Single-Modal vs. Bimodal Fusion Performance
| Model Type | Precision | Recall | F1-Score |
|---|---|---|---|
| EEG Only | < 90% | < 90% | < 90% |
| fNIRS Only | < 90% | < 90% | < 90% |
| EEG-fNIRS Fusion (AR-TSNET) | > 90% | > 90% | > 90% |
Note: Exact values for single-modal results were not fully detailed in the source, but were reported to be below the performance of the fused model, which exceeded 90% on all metrics [56].
This section outlines the essential materials, software, and analytical tools used in this study, providing a resource for researchers seeking to replicate or build upon this work.
Table 3: Key Research Reagents and Tools for EEG-fNIRS Fusion Studies
| Item / Technique | Function / Description | Role in the Present Study |
|---|---|---|
| EEG System | Records electrical activity of the brain with high temporal resolution [6]. | Acquired millisecond-level neural potential fluctuations from 52 scalp electrodes [56]. |
| fNIRS System | Measures hemodynamic changes (oxy-/deoxy-hemoglobin) with good spatial resolution [6]. | Captured blood oxygen changes in 21 frontal cortex channels related to craving [56]. |
| Visual Trigger Paradigm | A set of standardized stimuli (e.g., images) designed to provoke a specific cognitive or emotional state. | Used drug-related images to reliably elicit craving responses in the addicted group [56] [57]. |
| AR-TSNET Algorithm | A custom deep learning network for feature-level EEG-fNIRS fusion. | Core classification model featuring Tception, Sception, attention, and residual connections [56] [57]. |
| Attention Mechanism | A deep learning component that dynamically weights the importance of input features. | Enhanced model focus on the most critical EEG and fNIRS features for classification [56]. |
| k-Fold Cross-Validation | A resampling procedure used to evaluate a model on limited data samples. | Employed sixfold cross-validation to ensure robust and generalizable performance metrics [56]. |
This case study demonstrates that deep learning-based fusion of EEG and fNIRS signals is a highly effective and promising objective method for assessing drug addiction severity [56] [57]. The achieved accuracy of 92.6% underscores the transformative potential of multimodal data fusion in neuroengineering and clinical diagnostics.
Within the broader thesis on data fusion techniques for EEG and fNIRS, this work provides a concrete implementation and validation of a feature-level fusion strategy. It highlights how overcoming the inherent limitations of each single modality—by combining the high temporal resolution of EEG with the better spatial resolution and artifact resistance of fNIRS—can yield a more comprehensive picture of brain activity and significantly enhance classification accuracy in real-world clinical applications, such as addiction medicine [7] [6] [58]. The success of the AR-TSNET model paves the way for its application in other neuropsychiatric disorders where objective biomarkers are urgently needed.
The fusion of Electroencephalography (EEG) and functional Near-Infrared Spectroscopy (fNIRS) provides a powerful neuroimaging approach with high spatiotemporal resolution, but its signal quality is fundamentally limited by various artifacts. Motion, physiological, and environmental noises can significantly obscure genuine neural signatures, leading to inaccurate data interpretation and conclusions. Effective artifact management is therefore not merely a preprocessing step but a critical foundation for reliable data fusion, particularly in studies involving drug development where subtle neurophysiological changes are of paramount interest. This document outlines standardized protocols for identifying and removing these contaminations to ensure data integrity in multimodal research.
Artifacts in combined EEG-fNIRS studies can be categorized based on their origin. The table below summarizes the primary types, their characteristics, and common removal approaches.
Table 1: Classification and Overview of Artifacts in EEG-fNIRS Signals
| Artifact Category | Specific Type | Typical Frequency Range | Main Causes | Common Removal Approaches |
|---|---|---|---|---|
| Motion Artifacts | Head Movement | Slow drifts to sharp spikes (< 0.1 Hz to >10 Hz) | Head displacement, optode movement [59] [60] | Accelerometer-based correction [59], Spline interpolation [59] [61], WPD-CCA [62] |
| Jaw/Facial Movement | Variable | Talking, chewing [59] | Computer vision tracking [60], Channel rejection | |
| Gait-Related | Periodic (~1-2 Hz) | Heel strike during walking [63] | Adaptive filtering, Deep Learning (e.g., Motion-Net) [63] | |
| Physiological Noise | Cardiac | ~1 Hz (0.8-1.5 Hz) | Heartbeat/pulse [64] [61] | Band-pass filtering, Adaptive filtering [61], WPD [62] |
| Respiratory | ~0.2-0.3 Hz (0.1-0.5 Hz) | Breathing [64] [61] | Band-pass filtering, MODWT [65] | |
| Mayer Waves | ~0.1 Hz | Very-low-frequency blood pressure oscillations [64] [61] | Band-pass filtering, MODWT-LSTM prediction [65] | |
| Systemic Hemodynamic | < 0.1 Hz | Global blood flow changes in scalp [66] [64] | Short-Separation Regression [66] [64], PCA/GLM [66] | |
| Environmental Noise | Powerline Interference | 50/60 Hz | AC power lines | Notch filtering |
| Technical Artifacts | Broad Spectrum | Equipment limitations, cable movement [63] | Proper grounding, Hardware inspection |
Objective: To acquire simultaneous EEG-fNIRS data with auxiliary information necessary for robust artifact identification and correction.
Materials:
Procedure:
Diagram 1: Artifact-Resilient Data Acquisition Workflow
Objective: To systematically remove motion-induced artifacts from fNIRS and EEG signals.
Background: Motion artifacts (MAs) are a dominant source of noise, characterized by sudden, high-amplitude spikes or shifts in the signal baseline [59] [60]. The following pipeline leverages a powerful two-stage algorithmic approach.
Procedure for fNIRS Signal Correction:
Procedure for EEG Signal Correction (Deep Learning Approach):
Diagram 2: Motion Artifact Removal Pipeline
Objective: To filter out systemic physiological noises that overlap with the task-evoked hemodynamic response.
Background: Physiological noises (cardiac, respiratory, Mayer waves) reside in frequency bands close to the hemodynamic response (~0.01-0.1 Hz). Simple band-pass filtering is often insufficient, especially for global systemic components captured in the superficial layers [64] [65]. This protocol uses a model-based approach.
Materials:
Procedure:
y(t) in a long-separation channel as a linear combination of components [61]:
y(t) = β * HRF(t) + α * SS(t) + Σ [b_m * sin(2πf_m t)] + ε(t)
where:
HRF(t) is the expected hemodynamic response function.SS(t) is the signal from a short-separation channel.Σ [b_m * sin(2πf_m t)] represents the sum of sinusoidal models for cardiac, respiratory, and Mayer wave frequencies.ε(t) is the residual error.Alternative/Complementary Method:
Table 2: Performance of Selected Artifact Removal Methods
| Method | Modality | Key Metric | Reported Performance | Primary Use Case |
|---|---|---|---|---|
| WPD-CCA [62] | fNIRS | ΔSNR (dB) | 16.55 dB | Motion Artifact Removal |
| Artifact Reduction (η) | 41.40% | |||
| WPD-CCA [62] | EEG | ΔSNR (dB) | 30.76 dB | Motion Artifact Removal |
| Artifact Reduction (η) | 59.51% | |||
| Motion-Net (CNN) [63] | EEG | ΔSNR (dB) | 20.00 dB | Motion Artifact Removal |
| Artifact Reduction (η) | 86.00% | |||
| RLSE with SS [61] | fNIRS | Contrast-to-Noise Ratio | Significant improvement vs. Kalman filter & ICA | Physiological & Superficial Noise |
| PCA/GLM with SS [66] | fNIRS | Contrast-to-Noise Ratio | Superior to minimal preprocessing & 3 other methods | Physiological Noise in whole-head montage |
Table 3: Key Materials and Tools for Artifact Management in EEG-fNIRS Research
| Item | Function/Description | Example Use Case |
|---|---|---|
| Short-Separation fNIRS Optodes | Source-detector pairs placed <1 cm apart to measure systemic hemodynamic noise from the scalp without cortical contribution [66] [64]. | Used as a regressor to remove global physiological noise from standard long-separation channels. |
| Tri-axial Accelerometer | A sensor attached to the participant's head to provide objective, ground-truth measurement of head movement dynamics [59]. | Used to identify motion artifact segments and validate motion correction algorithms. |
| Recursive Least-Squares Estimator (RLSE) | An adaptive filtering algorithm that efficiently estimates the parameters of a linear model, useful for separating signal from noise [61]. | Modeling and removing physiological noise and superficial signals in fNIRS data. |
| Wavelet Packet Decomposition (WPD) | A signal processing method that provides a more detailed time-frequency representation than standard wavelet transform [62]. | Decomposing signals to isolate motion artifact components for removal, especially in combination with CCA. |
| Mutual Information Feature Selection | A filter-based feature selection method that maximizes relevance and minimizes redundancy between features from different modalities [4]. | Optimizing feature sets for EEG-fNIRS data fusion by selecting the most complementary and informative features for classification. |
Effective artifact management is a critical, non-negotiable step in multimodal EEG-fNIRS research, directly impacting the validity of downstream data fusion and analysis. The protocols outlined herein provide a structured framework for tackling the primary challenges of motion, physiological, and environmental noise. By leveraging a combination of hardware solutions (e.g., short-separation channels, accelerometers) and advanced algorithmic approaches (e.g., WPD-CCA, RLSE, deep learning), researchers can significantly enhance data quality. This, in turn, ensures that subsequent fusion techniques uncover genuine neurovascular coupling dynamics, thereby increasing the reliability and interpretability of results in basic neuroscience and applied drug development studies.
The integration of electroencephalography (EEG) and functional near-infrared spectroscopy (fNIRS) provides a powerful multimodal approach for neuroscience research, combining EEG's millisecond-level temporal resolution with fNIRS's superior spatial localization of hemodynamic responses [67] [6]. However, this integration presents significant hardware challenges, particularly regarding accurate sensor placement, mechanical stability, and signal quality assurance. Traditional textile caps with manually prepared layouts often result in imprecise sensor positioning, leading to increased experimental variability [68]. This application note examines advanced hardware integration solutions, focusing on customizable 3D-printed headgear and integrated helmet systems that address these critical challenges for high-quality EEG-fNIRS data acquisition.
Recent advancements in neurotechnology have produced several innovative approaches to EEG-fNIRS integration. The table below summarizes the key performance characteristics of current hardware solutions.
Table 1: Performance Comparison of Integrated EEG-fNIRS Headgear Systems
| System Name | Customization Approach | Key Technical Features | Reported Placement Accuracy | Supported Modalities |
|---|---|---|---|---|
| ninjaCap [68] [69] | 3D-printed using atlas-based or subject-specific head models | Spring-relaxation algorithm for 3D-to-2D coordinate flattening; TPU flexible filament | 2.2 ± 1.5 mm | fNIRS, EEG, DOT, oMEG |
| Chameleon-1 [70] | Modular, adaptable headband design | Electrodes at F3/F4 positions (10-20 system); adjustable sizing | Not explicitly quantified | EEG |
| HEFMI-ICH Hybrid Cap [31] | Pre-configured hybrid design | 32 EEG electrodes, 32 fNIRS sources, 30 detectors; 90 fNIRS channels | Not explicitly quantified | EEG-fNIRS simultaneous recording |
The ninjaCap system represents a significant advancement through its fully customizable, 3D-printed design that enables accurate translation of 3D brain coordinates to 2D printable panels [68]. This approach addresses fundamental limitations of traditional EEG caps when used for fNIRS studies, particularly the suboptimal clustering of channel distances around typical fNIRS target separations [68]. The system utilizes thermoplastic polyurethane (TPU) filaments like ninjaFlex, offering the necessary flexibility, durability, and comfort for research applications.
The generation of a customized ninjaCap follows a structured workflow that ensures precise sensor placement tailored to individual research requirements and anatomical considerations.
Table 2: ninjaCap Generation Workflow Components
| Workflow Stage | Key Components | Output |
|---|---|---|
| Head Model & Probe Design | Colin 27 head model (default) or subject-specific MRI/photogrammetric models; Probe.SD file containing 3D node positions, anchor points, and grommet types | Registered probe layout on 3D head surface |
| 3D-to-2D Coordinate Flattening | Spring-relaxation algorithm; hex lattice structures for uniform deformation; 10-5 EEG coordinate reference | 2D panels representing flattened head surface sections |
| 3D Modeling & Printing | Blender3D software for extrusion and assembly; TPU flexible filament material; panel splitting for print bed compatibility | Set of 3-4 .STL files representing cap panels |
| Physical Assembly | Ultrasonic welding for panel connection; sensor and optode integration | Complete functional headgear |
The following diagram illustrates the complete ninjaCap generation workflow:
Successful implementation of integrated EEG-fNIRS systems requires specific materials and components. The table below details essential research reagents and their functions.
Table 3: Essential Research Reagents for Custom EEG-fNIRS Headgear
| Component Category | Specific Examples | Function & Application Notes |
|---|---|---|
| 3D Printing Materials | ninjaFlex TPU [68] | Flexible filament providing necessary stretchability, durability, and comfort for headgear |
| EEG Electrodes | Ag-AgCl (Silver/Silver Chloride) disks [71] | Standard wet electrodes for high-quality signal acquisition; require conductive gel |
| fNIRS Optodes | Sources & detectors (e.g., NIRScout system [32]) | Optical components for emitting and detecting near-infrared light through cerebral tissue |
| Hybrid Caps | Custom-designed EEG-fNIRS caps [31] | Integrated systems with pre-configured electrode and optode layouts for simultaneous recording |
| Conductivity Solutions | Electrolyte gels [70] | Ensure optimal electrical contact between EEG electrodes and scalp for signal quality |
| Adhesives & Stabilizers | Ultrasonic welding [68] | Panel connection method providing secure bonds without compromising material flexibility |
Head Model Selection: Choose between atlas-based models (e.g., Colin 27) or subject-specific models derived from structural MRI or photogrammetric scans [68]. For population studies, use the Colin 27 model with linear scaling based on head circumference (HC) measurements.
Probe Layout Design: Using AtlasViewer software, design the optode and electrode arrangement based on target brain regions. Specify source-detector distances according to fNIRS requirements (typically 20-40mm for adequate penetration depth). Save the configuration as a Probe.SD file containing 3D node positions, anchor points, and grommet type specifications [68].
Cap Generation Parameters: Input the Probe.SD file and head circumference measurement into the cloud-based pipeline at bfnirs.openfnirs.org. The system will output a set of 3-4 .STL files representing the cap panels optimized for standard 3D printer beds [68].
3D Printing Configuration: Use a fused deposition modeling (FDM) 3D printer with flexible TPU filament. Configure settings for optimal flexibility: 100% infill density, 0.15-0.20mm layer height, and slow print speed (20-30mm/s) to ensure quality [68].
Post-Processing: Remove support material carefully to avoid damaging the flexible structure. Inspect each panel for printing defects that might affect sensor placement accuracy.
Panel Assembly: Connect the printed panels using ultrasonic welding, ensuring precise alignment at joining points. Verify the complete assembly matches the intended head circumference before sensor integration [68].
Sensor Integration: Install fNIRS optodes and EEG electrodes into their designated grommets on the assembled cap. For EEG, apply conductive gel according to standard protocols to achieve impedance below 10 kΩ [32].
Subject Fitting: Position the cap on the participant's head using anatomical landmarks (nasion, inion, preauricular points) as reference. Ensure snug but comfortable fit to minimize motion artifacts.
Signal Quality Verification: Perform pre-experimental signal checks for both modalities. For fNIRS, verify signal-to-noise ratio and detector sensitivity. For EEG, confirm impedance values and check for abnormal noise patterns [31].
Proper hardware integration directly enables advanced data fusion techniques for EEG-fNIRS signals. The complementary nature of these modalities—with EEG capturing rapid neural electrical activity and fNIRS tracking slower hemodynamic responses—creates powerful opportunities for multimodal brain state decoding [6]. The precise colocation of sensors ensured by systems like the ninjaCap facilitates accurate cross-modal correlation analysis and neurovascular coupling investigation [68] [67].
Advanced fusion approaches include model-based architectures like the Multi-Branch Convolutional Neural Network with Attention (MBC-ATT), which processes EEG and fNIRS signals through separate branches before applying cross-modal attention mechanisms for feature discrimination [6]. Such methods typically achieve 5-10% improvement in classification accuracy compared to unimodal systems [31], but depend heavily on precise spatial alignment of acquisition hardware.
The hardware solutions described herein provide the foundation for exploiting these analytical advantages by ensuring consistent, accurate sensor placement across experimental sessions and between subjects—a critical requirement for reproducible multimodal research.
In concurrent electroencephalography (EEG) and functional near-infrared spectroscopy (fNIRS) studies, precise temporal synchronization of the two data streams is not merely beneficial—it is a fundamental prerequisite for valid data fusion and interpretation. These modalities capture distinct yet complementary physiological processes: EEG measures postsynaptic electrical potentials with millisecond temporal resolution, while fNIRS tracks hemodynamic responses related to blood oxygenation changes over seconds [72] [1]. This inherent difference in temporal dynamics means that without precise synchronization, any integrated analysis lacks temporal correspondence, potentially leading to misinterpretations of neurovascular coupling and brain function.
The core challenge lies in aligning these fundamentally different signals—one representing direct neuronal firing and the other reflecting metabolic consequences—to establish a meaningful relationship between electrical brain activity and hemodynamic responses. Furthermore, both data streams must be accurately synchronized with experimental events, ensuring that stimulus onsets and behavioral responses can be precisely mapped to both electrical and hemodynamic brain responses [73]. The selection of an appropriate synchronization strategy consequently directly impacts the validity and reliability of research findings in multimodal brain imaging.
Hardware synchronization establishes a direct physical connection between devices, creating a unified timekeeping system that is both robust and precise. This approach typically utilizes specialized devices that generate or relay trigger signals recognized simultaneously by both EEG and fNIRS acquisition systems.
Dedicated Synchronization Devices: Instruments like the PortaSync (a wireless, handheld device) offer a versatile hardware solution. They typically connect to the fNIRS software via Bluetooth and feature analog inputs and outputs. These can receive external trigger signals or generate their own, which are then recorded in the neuroimaging software while simultaneously sending the same signal to the EEG amplifier via its analog input [73]. Another example is the LabStreamer, which provides multiple high-speed analog inputs (up to 10kHz) suitable for capturing very fast-paced signals and connects to acquisition software via the Lab Streaming Layer protocol [73].
Parallel Port Triggers: A traditional yet effective method involves using the parallel port of a computer controlling the stimulus presentation software. The output signal can be split and sent to both the EEG amplifier and, through a device like the PortaSync, to the fNIRS system. This ensures both systems receive identical event markers [73].
Analog Output to Input Connection: In simpler setups, a direct cable can connect the analog output of one system (e.g., the stimulus computer or EEG amplifier) to the analog input of the other (the fNIRS system, or vice versa). The voltage change in the signal stream precisely marks the event onset in both recordings.
Table 1: Key Hardware Synchronization Equipment
| Equipment | Primary Function | Key Characteristics | Compatibility Notes |
|---|---|---|---|
| PortaSync | Generates/receives sync triggers | Wireless, Bluetooth, analog I/O, buttons for manual marking | Connects to OxySoft; analog out to EEG [73] |
| LabStreamer | High-speed data acquisition | Four analog inputs (up to 10kHz), LSL connection | Suitable for very fast EEG signals [73] |
| Parallel Port Sync Cable | Transmits TTL pulses | Direct computer connection, requires signal splitting | Classic, reliable method for sending triggers [73] |
Software synchronization relies on network communication protocols to achieve temporal alignment, eliminating the need for physical cables between systems while maintaining high precision when properly configured.
Lab Streaming Layer (LSL): This open-source ecosystem represents the current gold standard for software-based synchronization. LSL enables different software applications and acquisition devices on the same network to stream time-stamped data, including triggers and actual brain signals. The LSL system automatically handles the time-stamping, allowing the fNIRS software (e.g., OxySoft) to receive external event and data streams from EEG software or stimulus presentation packages [73] [74]. This framework is particularly advantageous for integrating multiple data streams beyond just triggers, such as simultaneous EEG and fNIRS signal acquisition into a single software.
Distributed Component Object Model (DCOM): This software interface allows different applications on the same computer to communicate. Stimulus presentation software can use DCOM to send triggers directly to the fNIRS software (e.g., OxySoft) without additional hardware [73]. This method is computer-dependent but can be highly effective for specific software combinations.
The following diagram illustrates the decision-making workflow for selecting the appropriate synchronization method based on equipment capabilities:
Each synchronization method offers distinct advantages and limitations, making them suitable for different experimental scenarios and equipment constraints. The choice between hardware and software approaches depends on multiple factors, including required precision, equipment compatibility, setup complexity, and experimental environment.
Table 2: Hardware vs. Software Synchronization Comparison
| Feature | Hardware Triggering | Software-Based Alignment |
|---|---|---|
| Primary Mechanism | Physical cables (analog, parallel port) | Network protocols (LSL, DCOM) |
| Typical Precision | Very high (sub-millisecond) | High (millisecond) |
| Setup Complexity | Moderate (requires cabling) | Low (requires network/software config.) |
| Key Advantage | Robustness, independence from OS delays | Flexibility, no additional cables |
| Key Limitation | Additional equipment cost, physical setup | Potential for software/network latency |
| Stimulus Locking | Excellent | Excellent |
| Best For | Environments with electrical noise; simple, robust setups | Flexible labs; streaming multimodal data beyond triggers |
This protocol provides a step-by-step methodology for achieving robust synchronization using the PortaSync device, suitable for experiments where software streaming is not feasible.
1. Equipment Setup and Connections:
2. Signal Configuration:
3. Experimental Recording and Validation:
This protocol outlines the procedure for implementing LSL-based synchronization, ideal for modern setups that benefit from cable-free operation and integrated data handling.
1. Network and Software Preparation:
2. Stream Configuration:
3. Recording and Synchronization Check:
Successful execution of synchronized EEG-fNIRS experiments requires specific hardware and software components. The following table details the essential "research reagents" and their critical functions in the experimental workflow.
Table 3: Essential Materials for Synchronized EEG-fNIRS Research
| Item | Category | Critical Function | Example Models/Protocols |
|---|---|---|---|
| Synchronization Hardware | Equipment | Generates or transmits precise physical trigger signals | PortaSync, LabStreamer [73] |
| Lab Streaming Layer (LSL) | Software | Enables software-based, time-stamped data streaming on a network | Open-source LSL library [73] [74] |
| Integrated EEG-fNIRS Cap | Consumable/Material | Holds both sensor types in a stable, co-registered spatial layout | actiCAP Xpress Twente Medical Systems, custom 3D-printed helmets [5] [74] |
| Stimulus Presentation Software | Software | Presforms experiments and generates event markers for synchronization | PsychoPy, Presentation, E-Prime |
| Analog Cables & Connectors | Accessory | Transmits analog trigger signals between devices | BNC, DIN, or other system-specific cables |
| Data Fusion Analysis Toolbox | Software | Performs joint analysis of synchronized multimodal data | MATLAB toolboxes, Python (MNE), Homer3, NIRS-KIT |
The strategic selection and proper implementation of synchronization techniques form the bedrock of any high-quality concurrent EEG-fNIRS study. Hardware triggering offers a physically robust solution, often preferred for its consistent precision and simplicity. In contrast, software-based alignment via frameworks like LSL provides immense flexibility and is well-suited to complex, multi-stream experimental designs. The optimal choice is not universal but must be determined by the specific research question, the available equipment's capabilities, and the constraints of the experimental environment. By adhering to the detailed protocols and utilizing the essential tools outlined in this document, researchers can ensure the temporal integrity of their multimodal data, thereby creating a reliable foundation for advanced data fusion and the generation of meaningful insights into brain function.
The accurate integration of electroencephalography (EEG) and functional near-infrared spectroscopy (fNIRS) data represents a powerful approach in neuroscience for investigating brain function across electrophysiological and hemodynamic domains. The cornerstone of effective multimodal data fusion is the precise co-registration of sensors to underlying brain anatomy. The International 10-20 system provides a standardized framework for electrode placement, but its effective implementation requires careful consideration of anatomical correlations, especially when integrating fNIRS optodes. This application note details standardized protocols for co-registering EEG and fNIRS sensors within the 10-20 framework, ensuring optimal data quality for subsequent fusion techniques in both clinical and research settings. The precise alignment of sensor positioning with cortical landmarks is fundamental to maximizing the spatial accuracy and interpretative power of multimodal brain imaging studies [75] [5].
The International 10-20 system standardizes sensor placement based on proportional distances between key cranial landmarks (nasion, inion, and preauricular points), ensuring consistent positioning across subjects despite variations in head size and shape. Understanding the relationship between these external scalp positions and underlying cerebral structures is critical for accurate source localization and interpretation of recorded signals.
Neuroimaging studies have validated the general consistency between 10-20 positions and their underlying cerebral structures. However, important developmental differences exist, particularly in infant populations. Research using MRI has demonstrated that while 10-20 electrode placements in infants generally project to the same cerebral structures as in adults, notable differences exist for occipital (O1, O2) and temporal (T5, T6) channels, which project over lower anatomical structures in infants [75]. This underscores the importance of age-specific anatomical atlases for developmental studies. Furthermore, anatomical asymmetries, such as the Yakovlean torque which displaces the right sylvian fissure dorsally and frontally, can result in sensors P3 and P4 being at different distances from this structure, highlighting the need for individualized co-registration where possible [75].
Table 1: Key Differences in 10-20 System Correlations Between Infants and Adults
| Parameter | Infant Population | Adult Population |
|---|---|---|
| O1/O2 Projection | Over lower anatomical structures | Standard projection patterns |
| T5/T6 Projection | Over lower anatomical structures | Standard projection patterns |
| Template Availability | Infant-specific probabilistic templates | Standard MNI adult templates |
| Sulcal Variability | Higher variability in folding patterns | More consistent sulcal patterns |
EEG and fNIRS offer complementary information about brain activity, making them ideal partners for multimodal fusion. EEG provides millisecond-level temporal resolution of electrical neural activity but suffers from relatively low spatial resolution and sensitivity to artifacts. Conversely, fNIRS measures hemodynamic responses with better spatial resolution and robustness to motion artifacts, but has significantly slower temporal resolution due to the inherent latency of neurovascular coupling [5] [6]. This technical complementarity enables more comprehensive brain monitoring when the signals are properly integrated, but necessitates precise sensor co-registration to ensure spatial correspondence between the different modalities.
Effective co-registration begins with physical sensor integration. Current approaches typically combine EEG electrodes and fNIRS optodes within a single acquisition helmet. Several design strategies exist:
The integration design must minimize crosstalk between modalities—ensuring that EEG electrodes do not interfere with optical paths and that fNIRS hardware does not introduce electrical noise into EEG recordings [76].
Precise alignment of sensor positions with individual anatomy significantly enhances the spatial accuracy of multimodal measurements. The following protocol outlines a comprehensive co-registration procedure:
Diagram 1: Anatomical Co-registration Workflow for Multimodal Setup
Table 2: Coregistration Methods and Their Applications
| Method | Procedure | Accuracy | Best Suited For |
|---|---|---|---|
| Landmark-based | Aligns sensors using nasion, inion, preauricular points, and Cz. | Moderate | Standardized studies without individual MRI |
| 3D Digitization | Records exact 3D coordinates of each sensor and landmark. | High | Studies requiring individual head models |
| Template Fitting | Warps standard sensor positions to fit individual head shape. | Moderate-High | Large cohort studies with structural MRI |
| Photogrammetry | Uses multiple 2D images to reconstruct 3D sensor positions. | High | Rapid acquisition without specialized digitizers |
Beyond standardized placement, optimizing electrode subsets for specific neural sources can significantly enhance signal quality while reducing system complexity. Genetic algorithms offer a powerful approach for this optimization:
Genetic Algorithm Optimization Protocol:
Research demonstrates that optimized low-density electrode subsets (e.g., 6-8 electrodes) can achieve accuracy comparable to full high-density arrays (200+ channels) for single-source localization tasks, highlighting the potential for more efficient experimental designs [77].
Rigorous validation ensures that co-registration accuracy translates to improved experimental outcomes across different application domains:
Table 3: Validation Protocols for Specific Application Domains
| Application Domain | Validation Task | Expected Outcome | Quality Metrics |
|---|---|---|---|
| Motor Imagery | Repetitive hand open/close, wrist flexion/extension motor imagery or execution [32]. | Contralateral primary motor cortex activation (EEG: µ/beta ERD; fNIRS: HbO increase). | Spatial consistency between EEG source localization and fNIRS activation focus. |
| Working Memory | n-back tasks with varying cognitive load (0-back, 2-back, 3-back) [6]. | Prefrontal cortex activation scaling with task difficulty. | Correlation between EEG frontal theta power and fNIRS HbO in DLPFC. |
| Language Processing | Word generation or semantic decision tasks [3]. | Left-lateralized frontal-temporal activation. | Laterality index consistency between modalities. |
| Epilepsy | Interictal spike or seizure recording [5]. | Concordant localization of epileptiform activity. | Spatial overlap between EEG spike source and fNIRS hemodynamic response. |
For all validation experiments, simultaneously record EEG and fNIRS data during task performance. Preprocess signals according to modality-specific pipelines (filtering, artifact removal), then analyze spatial concordance between EEG source reconstructions and fNIRS activation maps. Quantitative metrics should include distance between activation centroids, spatial correlation of activation maps, and temporal correlation between EEG features (e.g., band power) and fNIRS hemodynamic responses [5] [6].
Table 4: Essential Materials for EEG-fNIRS Co-registration Research
| Item | Specification/Example | Primary Function |
|---|---|---|
| EEG Acquisition System | 64+ channel systems (e.g., Neuroscan SynAmps2) [32]. | Records electrical brain activity with high temporal resolution. |
| fNIRS Acquisition System | Multichannel systems (e.g., NIRScout with 8+ sources/detectors) [32]. | Measures hemodynamic responses via light absorption. |
| Integrated Caps | Flexible caps with pre-configured EEG/fNIRS layouts or custom 3D-printed helmets [5]. | Ensures stable, consistent sensor placement. |
| 3D Digitizer | Electromagnetic (e.g., Polhemus) or optical (e.g., Structure Sensor) systems. | Records precise 3D coordinates of sensors and landmarks. |
| Head Model Software | EEGLAB, FieldTrip, SPM, NIRS Brain AnalyzINBOX [78]. | Creates forward models for source reconstruction. |
| Anatomical Templates | Age-appropriate templates (e.g., MNI for adults, infant templates) [75]. | Provides anatomical reference when individual MRI is unavailable. |
| Data Fusion Platforms | Custom MATLAB/Python scripts with toolboxes for multimodal analysis. | Implements feature-level or decision-level fusion algorithms. |
Diagram 2: System Components for Multimodal Co-registration
Precise co-registration of EEG and fNIRS sensors within the International 10-20 system framework is a critical prerequisite for meaningful multimodal data fusion. By implementing the anatomical coregistration protocols, optimization strategies, and validation frameworks outlined in this document, researchers can significantly enhance the spatial accuracy and interpretative power of their neuroimaging studies. Future directions in this field include the development of more sophisticated integrated helmet designs that ensure consistent scalp coupling, advances in real-time coregistration using computer vision, and the creation of age-specific and disease-specific head models to further improve source localization accuracy across diverse populations.
In multimodal electroencephalography (EEG) and functional near-infrared spectroscopy (fNIRS) research, data imbalance and poor model generalization present significant barriers to developing robust brain-computer interfaces (BCIs) and clinical diagnostic tools. The inherent differences in signal characteristics between electrophysiological (EEG) and hemodynamic (fNIRS) measurements create natural asymmetries that conventional machine learning approaches struggle to reconcile. EEG captures millisecond-level electrical neural activity with high temporal resolution but limited spatial precision, while fNIRS measures hemodynamic changes through oxygenation with superior spatial resolution but slower temporal dynamics [7] [79]. This fundamental discrepancy often leads to imbalanced feature learning during model training, where one modality dominates the optimization process, thereby limiting the synergistic potential of multimodal integration [79].
The problem extends beyond technical implementation to practical research constraints. Multimodal EEG-fNIRS datasets remain scarce, with limited public availability creating additional generalization challenges [7]. Furthermore, the temporal misalignment between rapid electrical signals and slower hemodynamic responses introduces representation learning complexities that unsupervised symmetric techniques attempt to resolve [7]. These challenges necessitate specialized frameworks that explicitly address imbalance while enhancing cross-subject and cross-task generalization, particularly in real-world applications such as Parkinson's disease detection [80], motor imagery classification [32], and semantic decoding [81].
A promising approach to addressing data imbalance involves feature decoupling through a dual-decoder architecture that separates modality-general and modality-specific features [79]. This framework employs dedicated global feature encoders for EEG and fNIRS domains, followed by two distinct decoders:
This architectural approach mitigates signal interference while enhancing extraction of task-relevant information. The separation allows models to leverage complementary strengths of both modalities - the temporal precision of EEG alongside the spatial specificity of fNIRS - without allowing either modality to dominate the learning process [79].
To explicitly counter learning imbalance, a Gradient Rebalancing (GradReb) strategy actively monitors and regulates discriminative differences between modalities during training [79]. This approach:
Implementation results demonstrate that GradReb significantly improves joint learning framework performance, particularly for classification tasks where baseline methods typically favor one modality [79].
For limited data scenarios, multimodal representation learning provides an alternative pathway to improved generalization. The EEG-fNIRS Representation Model (EFRM) employs a two-stage approach: [82]
This method enables effective adaptation to single-modality datasets when both modalities aren't available, enhancing practical flexibility. By leveraging the shared domain across EEG and fNIRS, this approach outperforms single-modality methods, especially for fNIRS classification where traditional transfer learning has shown limitations [82].
Table 1: Comparative Analysis of Technical Approaches for Addressing Data Imbalance
| Approach | Core Mechanism | Advantages | Implementation Considerations |
|---|---|---|---|
| Dual-Decoder Architecture [79] | Feature decoupling into modality-general and specific components | Reduces signal interference; Enhances task-relevant information extraction | Requires careful loss function balancing; Increases model complexity |
| Gradient Rebalancing (GradReb) [79] | Regulation of gradient flow during training | Prevents modality dominance; Ensures balanced feature learning | Needs monitoring of discriminative differences between modalities |
| Multimodal Representation Learning [82] | Two-stage pre-training and transfer learning | Effective for few-shot scenarios; Works with limited labeled data | Requires substantial unlabeled data for pre-training phase |
Proper data collection forms the foundation for addressing imbalance and generalization challenges. The following protocol ensures high-quality, temporally aligned multimodal data: [83] [84]
Robust evaluation methodologies are essential for properly assessing model generalization: [80]
To combat limited dataset sizes and class imbalances, employ multimodal data augmentation: [32]
In Parkinson's disease detection, fNIRS data from the frontopolar cortex (FPC) region can be leveraged with machine learning to achieve 85% accuracy using Support Vector Machines (SVM) [80]. The experimental protocol involved:
To address potential data imbalance between patient and control groups, the study employed SHapley Additive exPlanations (SHAP) to identify the most contributory channels (CH01, CH04, CH05, CH08), enabling feature selection that improved model generalization [80].
For motor imagery tasks involving multiple joint movements, a balanced multimodal approach achieves 65.49% classification accuracy between hand open/close and shoulder pronation/supination using EEG data with deep learning (ShallowConvNet) and data augmentation [32]. The protocol included:
Table 2: Performance Comparison Across Application Domains
| Application Domain | Modality | Key Methodology | Reported Performance | Generalization Approach |
|---|---|---|---|---|
| Parkinson's Detection [80] | fNIRS | SVM with channel selection | 85% accuracy, F1=0.85, AUC=0.95 | SHAP analysis for feature importance |
| Motor Imagery [32] | EEG-fNIRS | ShallowConvNet with augmentation | 65.49% accuracy (hand vs. shoulder tasks) | Large-scale dataset (5760 trials) |
| Semantic Decoding [81] | EEG-fNIRS | Multimodal fusion | Differentiating animals vs. tools | Sensory-based imagery tasks |
| Brain State Characterization [83] | EEG-fNIRS | Joint learning framework | Improved spatial-temporal resolution | Standardized acquisition protocol |
Table 3: Research Reagent Solutions for EEG-fNIRS Studies
| Item | Specification | Function/Application |
|---|---|---|
| EEG Acquisition System [32] | 64-channel Neuroscan SynAmps2 amplifier, 1000Hz sampling | Records electrical brain activity with high temporal resolution |
| fNIRS Acquisition System [32] | NIRScout system (NIRx), 8 sources, 8 detectors, 7.8125Hz sampling | Measures hemodynamic responses through light absorption |
| Multimodal Caps [32] | Integrated EEG electrodes + fNIRS optodes arranged in 10-5 system | Enables simultaneous co-located measurement |
| ETG-4000 System [80] | Near-infrared brain function imaging instrument, 695nm & 830nm wavelengths | Specialized for hemodynamic monitoring in clinical applications |
| HD-DOT System [7] | High-density diffuse optical tomography with 16 sources, 12 detectors | Provides high spatial resolution for cortical mapping |
| Data Processing Tools [83] | MATLAB/Python with MNE-Python, NIRSKit, HomER2 | Preprocessing, artifact removal, feature extraction |
| Synchronization Interface [85] | TTL trigger systems with LabView application | Ensures temporal alignment between modalities |
| Stimulus Presentation Software [86] | PsychToolbox, Presentation, OpenSesame | Controls experimental paradigms with precise timing |
The paradigm in neuroimaging is shifting from controlled laboratory settings toward wearable, ubiquitous monitoring in naturalistic environments [87] [88]. This transition to ecologically valid research is crucial for understanding brain function during real-world activities but introduces significant challenges from motion artifacts (MA). These artifacts, caused by subject movement during data acquisition with wearable devices, can severely corrupt electroencephalography (EEG) and functional near-infrared spectroscopy (fNIRS) signals, potentially rendering them unusable [87] [62]. For research relying on data fusion techniques from these complementary modalities—EEG providing millisecond temporal resolution and fNIRS offering better spatial localization and motion robustness [6] [29]—effective motion artifact mitigation becomes paramount. Without robust correction, subsequent fusion and interpretation risk significant error, undermining the validity of findings in cognitive neuroscience, clinical neurology, and drug development research [87] [29].
Motion artifacts manifest differently in EEG and fNIRS signals, presenting unique challenges for data fusion pipelines. In EEG, artifacts originate from muscle activity (EMG), electrode cable swings, and changes in electrode-skin impedance, leading to high-amplitude, high-frequency spikes and slow drifts that obscure neural signals [89] [29]. fNIRS artifacts arise from optode displacement, pressure changes, and altered scalp hemodynamics, causing baseline shifts, spikes, and increased variance in hemoglobin concentration measurements [90] [29].
The complementary nature of EEG and fNIRS artifacts complicates multimodal fusion. Artifacts are often temporally correlated with behavioral events but exhibit distinct morphological characteristics across modalities [29]. This differential contamination can desynchronize fused features, leading to misalignment in joint analysis and erroneous interpretation of neurovascular coupling relationships. Consequently, developing robust, integrated artifact mitigation protocols is essential for advancing data fusion techniques in naturalistic research [7] [29].
Traditional artifact correction methods include filtering, regression, and component analysis techniques. Recently, Wavelet Packet Decomposition (WPD) and its combination with Canonical Correlation Analysis (CCA) have demonstrated superior performance for single-channel denoising, a critical requirement for wearable systems with limited channels [87] [62].
Table 1: Performance of WPD and WPD-CCA for Motion Artifact Correction
| Method | Modality | Best Performing Wavelet | ΔSNR (dB) | Artifact Reduction (η%) |
|---|---|---|---|---|
| WPD (Single-Stage) | EEG | db2 | 29.44 | - |
| WPD (Single-Stage) | EEG | db1 | - | 53.48 |
| WPD-CCA (Two-Stage) | EEG | db1 | 30.76 | 59.51 |
| WPD (Single-Stage) | fNIRS | fk4 | 16.11 | 26.40 |
| WPD-CCA (Two-Stage) | fNIRS | db1 | 16.55 | - |
| WPD-CCA (Two-Stage) | fNIRS | fk8 | - | 41.40 |
The two-stage WPD-CCA method consistently outperforms single-stage WPD, with artifact reduction increasing by 11.28% for EEG and 56.82% for fNIRS when using the two-stage approach [62]. This demonstrates its particular effectiveness for fNIRS signals, making it highly suitable for multimodal studies where fNIRS signals are less inherently noisy but require cleaner baselines for accurate hemodynamic response interpretation [87].
Machine and deep learning methods represent the cutting edge of artifact mitigation research, showing promise for adaptive, data-driven correction.
Table 2: Learning-Based Approaches for fNIRS Motion Artifact Correction
| Method | Architecture | Key Mechanism | Reported Performance |
|---|---|---|---|
| Wavelet Regression ANN [90] | Artificial Neural Network | Unbalance index from entropy cross-correlation | Improved Contrast-to-Noise Ratio in gait tasks |
| MA Classifier [90] | LDA, SVM, KNN, GBT | Vigilance level detection during walking | Identified MA-induced accuracy reduction |
| U-Net HRF Reconstruction [90] | Convolutional Neural Network | Reconstruction of hemodynamic response | Lowest Mean Squared Error vs. wavelet/AR models |
| Denoising Auto-Encoder [90] | Auto-Encoder | Specialized loss function on synthetic data | Effective on open-access experimental datasets |
| sResFCNN + FIR Filter [90] | Fully Connected NN | Simplified residual architecture | Enhanced denoising capability |
These learning-based approaches are particularly valuable for data fusion applications as they can be designed to process both EEG and fNIRS signals within a unified architecture, potentially learning cross-modal artifact patterns [90] [58]. For instance, DeepSyncNet represents a novel deep learning framework that integrates EEG and fNIRS at the feature extraction stage using Attentional Fusion, enhancing classification for motor imagery and mental arithmetic tasks in BCIs [58].
Application: Correction of motion artifacts in single-channel EEG or fNIRS signals collected during naturalistic tasks.
Workflow:
Application: fNIRS data collection and analysis in real-world environments with unpredictable event timing.
Workflow:
This protocol was successfully applied in a naturalistic study investigating olfactory stimuli during laundry tasks, where fNIRS with AIDE identified differential prefrontal activation between fragranced and unfragranced conditions despite the movement-rich environment [88].
Application: Adaptive artifact correction for multimodal EEG-fNIRS data using deep learning architectures.
Workflow:
Table 3: Key Research Materials for Movement Artifact Mitigation Studies
| Item | Function/Application | Example Specifications |
|---|---|---|
| Wearable EEG System | Mobile neural electrical activity recording | Science-grade, 8-64 channels, wireless capability [89] |
| Portable fNIRS Device | Hemodynamic response monitoring in natural settings | Continuous wave system, multiple wavelengths, wearable design [88] [29] |
| Motion Tracking System | Quantification of subject movement for artifact reference | IMU sensors, accelerometers, gyroscopes [62] |
| Wavelet Toolbox | Signal decomposition for WPD-based artifact removal | MATLAB, Python PyWavelets, with multiple wavelet families [87] |
| Deep Learning Framework | Implementation of neural networks for artifact correction | TensorFlow, PyTorch with custom architectures [90] [58] |
| Benchmark Datasets | Method validation and performance comparison | Publicly available EEG-fNIRS datasets with motion artifacts [62] [58] |
| AIDE Software Platform | Automatic identification of functional events in naturalistic fNIRS data | fNIRS analysis toolbox for naturalistic paradigms [88] |
| Homer2/3 Toolkit | Standard fNIRS processing including motion correction | MATLAB-based environment for fNIRS analysis [90] |
Effective motion artifact mitigation is foundational for successful EEG-fNIRS data fusion in naturalistic research environments. The quantitative evidence demonstrates that two-stage approaches like WPD-CCA significantly outperform single-stage methods, particularly for fNIRS signals. Emerging deep learning architectures show promise for adaptive, multimodal artifact correction through attention mechanisms and specialized loss functions. The experimental protocols presented provide actionable methodologies for implementing these techniques across various research scenarios, from controlled behavioral tasks to fully naturalistic settings. As wearable neuroimaging continues to evolve toward real-world applications, robust artifact mitigation will remain crucial for validating data fusion techniques and extracting meaningful insights from brain dynamics in ecologically valid contexts.
In the realm of data fusion techniques for electroencephalography (EEG) and functional near-infrared spectroscopy (fNIRS) research, selecting appropriate performance metrics is paramount for accurately evaluating classification models. EEG captures electrical brain activity with high temporal resolution but limited spatial resolution, while fNIRS measures hemodynamic responses with better spatial resolution but lower temporal resolution [5]. Their integration in a hybrid Brain-Computer Interface (BCI) aims to leverage these complementary characteristics to improve overall system performance [92] [93]. However, this multimodal approach introduces unique challenges for performance assessment, particularly given the typically imbalanced nature of brain signal datasets where target events (e.g., specific mental states) are often rare compared to baseline activity. This application note provides detailed protocols for using accuracy, AUC-ROC, and F1-score to evaluate classification performance within EEG-fNIRS fusion research, enabling researchers to select metrics that align with their specific experimental objectives and data characteristics.
Binary classification metrics are derived from four fundamental outcomes organized in a confusion matrix [94] [95]:
Table 1: Fundamental Classification Metrics
| Metric | Formula | Interpretation | Perfect Value |
|---|---|---|---|
| Accuracy | (TP + TN) / (TP + TN + FP + FN) | Overall correctness across both classes | 1.0 (100%) |
| Precision | TP / (TP + FP) | Accuracy when predicting positive class | 1.0 |
| Recall (Sensitivity) | TP / (TP + FN) | Ability to identify all positive instances | 1.0 |
| F1-score | 2 × (Precision × Recall) / (Precision + Recall) | Harmonic mean of precision and recall | 1.0 |
| False Positive Rate (FPR) | FP / (FP + TN) | Proportion of negatives incorrectly classified as positive | 0.0 |
Accuracy provides an intuitive measure of overall correctness by measuring the proportion of all correct predictions among the total predictions made [94]. It is mathematically defined as:
[ \text{Accuracy} = \frac{TP + TN}{TP + TN + FP + FN} ]
In Python, accuracy can be calculated using scikit-learn:
Application Context: Accuracy works well for balanced datasets where both classes are equally important and represented [96] [95]. However, in EEG-fNIRS research, accuracy becomes problematic for imbalanced datasets, which are common when detecting rare cognitive events or specific mental states. For instance, in a BCI where 95% of samples represent baseline activity and only 5% represent target mental commands, a model that always predicts the majority class would achieve 95% accuracy while being practically useless for detecting the target commands [94].
The Receiver Operating Characteristic (ROC) curve visualizes the trade-off between the True Positive Rate (TPR) and False Positive Rate (FPR) across all possible classification thresholds [96]. The Area Under the ROC Curve (AUC-ROC) provides a single number summarizing the model's ability to distinguish between classes, independent of any specific threshold.
Interpretation: AUC-ROC represents the probability that a randomly chosen positive instance is ranked higher than a randomly chosen negative instance [96]. A perfect model has an AUC of 1.0, while a random classifier has an AUC of 0.5.
Implementation:
Application Context: AUC-ROC is particularly valuable in EEG-fNIRS research when you care equally about both positive and negative classes and ultimately care about ranking predictions rather than outputting well-calibrated probabilities [96]. However, for heavily imbalanced datasets where the negative class dominates, the ROC curve can be overly optimistic because the false positive rate is pulled down due to the large number of true negatives [96].
The F1-score is the harmonic mean of precision and recall, providing a balanced metric that considers both false positives and false negatives [96] [94]. It is mathematically defined as:
[ \text{F1} = 2 \times \frac{\text{precision} \times \text{recall}}{\text{precision} + \recall} = \frac{2TP}{2TP + FP + FN} ]
Implementation:
Application Context: F1-score is the recommended metric for most EEG-fNIRS classification problems where you care more about the positive class [96]. It is especially valuable for imbalanced datasets common in BCI applications, such as detecting specific cognitive states or motor imagery tasks against background brain activity. The F1-score gives equal weight to precision and recall, making it suitable when both false positives and false negatives have consequences [94].
Table 2: Metric Selection Guide for EEG-fNIRS Research
| Research Scenario | Recommended Metric | Rationale | EEG-fNIRS Example |
|---|---|---|---|
| Balanced dataset, equal importance of classes | Accuracy | Provides good overall assessment when classes are balanced | Subject-independent classification with equal trial distribution |
| Imbalanced dataset, focus on positive class | F1-score | Robust to class imbalance; balances precision and recall | Detection of rare epileptic spikes or specific cognitive states |
| Ranking predictions, overall separability | AUC-ROC | Measures ranking quality independent of threshold | Model selection during initial development phase |
| High cost of false positives | Precision | Emphasizes correctness of positive predictions | BCI communication systems where false commands are critical |
| High cost of false negatives | Recall | Emphasizes identifying all positive instances | Disease diagnosis or detection of critical neurological events |
Objective: To classify left-hand vs. right-hand motor imagery tasks using fused EEG-fNIRS features.
Dataset: Public dataset from Shin et al. (2017) containing 29 participants performing kinesthetic motor imagery (grasping a ball) with 30 trials per task [15].
Preprocessing Pipeline:
Feature Extraction:
Classification & Evaluation:
Expected Performance: State-of-the-art methods achieve approximately 76% accuracy for bimodal classification, outperforming unimodal approaches (EEG-only: ~65%, fNIRS-only: ~57%) [15].
Objective: To distinguish between mental arithmetic (MA) and resting state using multimodal EEG-fNIRS.
Experimental Design:
Data Acquisition Specifications:
Multimodal Fusion Strategies:
Evaluation Framework:
Table 3: Essential Research Materials for EEG-fNIRS Classification Experiments
| Category | Item | Specification | Function/Purpose |
|---|---|---|---|
| Hardware Equipment | EEG Amplifier | BrainAmp, g.tec, or similar; ≥32 channels, ≥200Hz sampling | Records electrical brain activity with high temporal resolution |
| fNIRS System | NIRScout, NIRx, or similar; ≥16 sources, ≥16 detectors | Measures hemodynamic responses via near-infrared light | |
| Integrated Cap | EEG electrodes + fNIRS optodes in unified assembly | Ensures precise co-registration and synchronization of modalities | |
| Software Tools | Preprocessing | EEGLAB, NIRS-KIT, MNE-Python, Homer2 | Signal filtering, artifact removal, and basic feature extraction |
| Classification | scikit-learn, TensorFlow, PyTorch | Implementation of machine learning models and metrics | |
| Custom Analysis | MATLAB, Python with custom scripts | Specialized fusion algorithms and statistical analysis | |
| Experimental Materials | Task Presentation | PsychoPy, Presentation, E-Prime | Controlled delivery of cognitive paradigms |
| Data Synchronization | Lab Streaming Layer (LSL), trigger interface | Precise temporal alignment of multimodal data streams | |
| Validation Resources | Public Datasets | Shin et al. (2017) motor imagery/mental arithmetic | Benchmarking and method comparison |
| Performance Metrics | Accuracy, F1-score, AUC-ROC implementations | Comprehensive evaluation of classification performance |
In practical BCI applications, severe class imbalance is common, particularly in asynchronous BCIs where control states are rare compared to non-control states. Under these conditions:
Clinical Diagnostic Applications (e.g., epilepsy detection, ADHD assessment):
BCI Communication Applications:
Neuromarketing/Cognitive Research:
Comprehensive reporting should include:
The selection of appropriate performance metrics is crucial for advancing EEG-fNIRS multimodal fusion research. While accuracy provides an intuitive overall measure, F1-score offers robustness against class imbalance common in BCI applications, and AUC-ROC assesses the overall ranking capability of classifiers. The experimental protocols and metric selection framework presented in this application note provide researchers with standardized methodologies for comprehensive evaluation of classification performance in multimodal brain signal research. By aligning metric selection with specific research objectives and data characteristics, the field can move toward more reproducible and comparable results across studies, ultimately accelerating progress in hybrid BCI development and clinical applications.
The integration of electroencephalography (EEG) and functional near-infrared spectroscopy (fNIRS) represents a paradigm shift in neuroimaging research, offering a powerful solution to overcome the inherent limitations of single-modality approaches. Multimodal fusion leverages the complementary strengths of these two modalities: EEG provides millisecond-level temporal resolution of electrophysiological brain activity, while fNIRS offers superior spatial resolution for mapping hemodynamic responses [4] [6]. This synergy is particularly valuable for brain-computer interfaces (BCIs), cognitive state monitoring, and clinical diagnostics, where comprehensive neural decoding is essential.
The fundamental premise of EEG-fNIRS fusion stems from their complementary nature. EEG captures macroscopic cortical dynamics with fine temporal resolution (~5 msec) but suffers from low spatial resolution and sensitivity to artifacts [4]. Conversely, fNIRS measures hemodynamic properties with higher spatial resolution (~1 cm) and better resistance to motion artifacts, though it has significantly poorer temporal resolution [97] [98]. Through neurovascular coupling, these signals provide interrelated yet distinct perspectives on neural activity, creating a robust foundation for multimodal integration [99].
This Application Note provides a comprehensive quantitative comparison between single-modality and multimodal fusion approaches, detailing performance metrics, experimental protocols, and methodological considerations for researchers and drug development professionals working with neural signal analysis.
Table 1: Quantitative performance comparison of single-modality versus multimodal fusion approaches
| Modality | Fusion Method | Task Domain | Classification Accuracy | Performance Improvement | Citation |
|---|---|---|---|---|---|
| EEG-only | ShallowConvNet | Motor Imagery (Multiple Joints) | 65.49% | Baseline | [32] |
| fNIRS-only | Feature Selection | Visuo-Mental Task | ~57% | Baseline | [15] |
| EEG-fNIRS | Mutual Information Feature Selection | Visuo-Mental Task | Significant improvement | ~5% over single modalities | [4] |
| EEG-fNIRS | Early-Stage Fusion (Y-shaped Network) | Left/Right Motor Imagery | 76.21% | ~11% over EEG-only; ~19% over fNIRS-only | [15] |
| EEG-fNIRS | EFDA-CDG Framework | Mental Arithmetic | 91.93% | Substantial improvement over single modalities | [98] |
| EEG-fNIRS | EFDA-CDG Framework | n-back Task | 90.54% | Substantial improvement over single modalities | [98] |
| EEG-fNIRS | Multimodal MBC-ATT | Cognitive Tasks (n-back, WG) | Outperformed conventional approaches | Statistically significant | [6] |
| EEG-fNIRS | EFRM (Few-shot Learning) | Multiple Tasks | Competitive with state-of-the-art | High performance with minimal labeled data | [99] |
Table 2: Advantages and limitations of fusion approaches
| Fusion Category | Examples | Advantages | Limitations |
|---|---|---|---|
| Early Fusion | Feature concatenation, Y-shaped network [15] | Maximizes information sharing; Higher performance in motor imagery [15] | Susceptible to overfitting; Requires temporal alignment |
| Late Fusion | Classifier output combination, Fuzzy fusion [4] | Leverages modality-specific processing; More robust to noise | Potentially misses cross-modality interactions |
| Hybrid Methods | Mutual information feature selection [4], Cross-modal attention [6] | Optimizes complementarity; Reduces redundancy | Computationally intensive; Complex implementation |
Multimodal EEG-fNIRS fusion consistently demonstrates statistically significant improvements in classification accuracy across diverse task domains. The performance gains range from approximately 5% to over 19% compared to single-modality approaches, with the most substantial improvements observed in complex cognitive tasks [4] [15]. Early fusion strategies have shown particular promise for motor imagery tasks, achieving accuracy levels of 76.21% in left/right hand discrimination [15].
Advanced deep learning frameworks incorporating data augmentation and attention mechanisms have pushed performance boundaries further, achieving exceptional accuracy rates exceeding 90% for mental arithmetic and n-back tasks [98]. These approaches effectively address the challenge of limited training data through sophisticated augmentation techniques while leveraging the complementary information from both modalities.
Objective: To classify fused EEG-fNIRS data by optimizing complementarity, redundancy, and relevance between multimodal features using mutual information metrics [4].
Equipment Setup:
Experimental Procedure:
Data Acquisition (60-90 minutes)
Signal Processing Pipeline
Mutual Information Feature Selection
Classification & Validation
Objective: To investigate fusion stages and determine optimal integration point for EEG-fNIRS signals in motor imagery classification [15].
Dataset:
Processing Workflow:
Diagram 1: Early fusion experimental workflow
EEG Processing Branch:
fNIRS Processing Branch:
Fusion & Classification:
Objective: To implement cross-modal attention mechanism for selectively integrating EEG and fNIRS signals in cognitive state decoding [6].
Dataset Preparation:
Network Architecture:
Cross-Modal Attention Module
Fusion & Classification
Validation Method:
The effectiveness of EEG-fNIRS multimodal fusion is grounded in the fundamental principles of neurovascular coupling, the biological process that links neural activity to subsequent hemodynamic responses [99].
Diagram 2: Neurovascular coupling in multimodal fusion
Temporal Relationships:
Neurophysiological Basis:
This biological relationship creates the foundation for effective multimodal fusion, where electrophysiological and hemodynamic measures provide complementary views of the same underlying neural processes.
Table 3: Essential equipment and computational tools for EEG-fNIRS multimodal research
| Category | Specific Tool/Solution | Function/Purpose | Example Specifications |
|---|---|---|---|
| Acquisition Hardware | EEG Cap & Amplifier | Records electrical brain activity | 64-channel setup; 1000 Hz sampling; <10 kΩ impedance [32] |
| Acquisition Hardware | fNIRS System with Optodes | Measures hemodynamic responses | 8 sources, 8 detectors; 7.8125 Hz sampling; 24 channels [32] |
| Software & Analysis | Mutual Information Toolkits | Feature selection and optimization | Maximizes complementarity, minimizes redundancy [4] |
| Software & Analysis | Deep Learning Frameworks | Implements fusion architectures | PyTorch/TensorFlow with custom Y-shaped networks [15] |
| Software & Analysis | Data Augmentation Tools | Expands limited training datasets | Denoising Diffusion Probabilistic Models (DDPM) [98] |
| Experimental Materials | Multimodal Recording Caps | Integrated EEG electrodes & fNIRS optodes | Custom designs with 10-20 system compatibility [32] |
| Experimental Materials | Signal Quality Verification Kits | Ensures data integrity during acquisition | Impedance checkers, light intensity monitors [32] |
Multimodal EEG-fNIRS fusion represents a significant advancement over single-modality approaches, consistently demonstrating superior performance across diverse applications from motor imagery to cognitive state assessment. The quantitative evidence confirms that properly implemented fusion strategies can achieve performance improvements of 5-19% over individual modalities, with advanced frameworks reaching exceptional accuracy levels exceeding 90% for specific tasks [4] [98] [15].
The optimal fusion strategy depends on the specific application requirements: early fusion approaches excel in motor imagery tasks, while attention-based mechanisms show promise for complex cognitive state decoding. Future directions include standardized preprocessing pipelines, transfer learning approaches for limited data scenarios, and real-time implementation for clinical and BCI applications [98] [99]. As multimodal methodologies continue to mature, they hold tremendous potential to transform neuroimaging research and clinical brain monitoring applications.
The integration of electroencephalography (EEG) and functional near-infrared spectroscopy (fNIRS) represents a promising direction for brain activity decoding with high spatiotemporal resolution in naturalistic scenarios [7]. However, the advancement of data fusion techniques for these multimodal signals faces significant challenges due to the lack of standardized public datasets and robust benchmarking frameworks. This creates substantial variability in how machine learning is applied to fNIRS data and how methodology and results are reported, making it difficult to evaluate whether some approaches are genuinely superior to others [100]. This application note provides a comprehensive resource for researchers seeking to navigate the current landscape of public datasets, benchmarking standards, and experimental protocols for EEG-fNIRS fusion research, with particular emphasis on their application within therapeutic development contexts.
The emergence of publicly available, multimodal datasets is crucial for developing and validating data fusion algorithms. These datasets enable direct comparison of methods and facilitate reproducible research. Several recent datasets have been curated to support specific research applications and are summarized in Table 1.
Table 1: Publicly Available Multimodal EEG-fNIRS Datasets
| Dataset Name | Primary Research Focus | Subjects (Healthy/Patients) | Recorded Tasks | Key Modalities | Data Format & Access |
|---|---|---|---|---|---|
| HEFMI-ICH [31] | Intracerebral Hemorrhage Rehabilitation | 17 Normal / 20 ICH Patients | Left/Right Hand Motor Imagery | 32-ch EEG, 90-ch fNIRS (HbO, HbR) | Raw & pre-processed trial data; Clinical scores (FMA-UE, MBI) |
| Multimodal Upper-Limb Joint MI [32] | Fine Motor Imagery Decoding | 18 Healthy | 8 MI tasks for Hand, Wrist, Elbow, Shoulder | 64-ch EEG, 24-ch fNIRS | Raw recordings; 5760 total trials |
| REFED [101] | Affective Computing (aBCI) | 32 Healthy | Emotion evocation via video | EEG, fNIRS (HbO, HbR, HbT) | Raw data; Real-time valence/arousal annotations |
| Shin et al. Dataset A [15] | Basic Motor Imagery Benchmarking | 29 Healthy | Left/Right Hand Motor Imagery | EEG, fNIRS (HbO) | Pre-processed; Standard left/right MI |
| Semantic Imagery Dataset [81] | Semantic Neural Decoding | 12 Healthy (+7 for EEG-only) | Silent Naming, Visual, Auditory, Tactile Imagery of Animals/Tools | Simultaneous EEG-fNIRS | Raw data; Semantic category labels |
These datasets support various research applications, from motor rehabilitation to affective computing. The HEFMI-ICH dataset is particularly notable for including patient data, addressing a critical gap in clinical translation [31]. For drug development professionals, such patient-derived datasets provide invaluable resources for evaluating neurotherapeutics targeting motor recovery or cognitive function.
To address the lack of community standards for applying machine learning to fNIRS data, the BenchNIRS framework has been developed as an open-source benchmarking tool [100]. This framework establishes a best-practice methodology using nested cross-validation to optimize models and evaluate them without bias, providing standardized metrics and figures for performance comparison.
Performance benchmarking across multiple datasets reveals that actual classification accuracy is often lower than scores frequently reported in literature, with minimal differences between model types when evaluated rigorously [100]. This underscores the importance of standardized benchmarking to prevent overestimation of model capabilities, particularly when transitioning from research to clinical applications.
A common experimental paradigm for EEG-fNIRS research involves motor imagery tasks with standardized timing. A typical protocol follows this structure [31]:
This protocol is implemented in the HEFMI-ICH dataset and can be adapted for various patient populations [31]. For pharmacological studies, the baseline and post-intervention assessments can be incorporated within this framework to evaluate drug effects on motor imagery-related brain activity.
Simultaneous EEG-fNIRS acquisition requires careful hardware integration. A standard setup includes [31]:
A standardized preprocessing pipeline is essential for reproducible results. The workflow below outlines the critical steps for preparing raw EEG and fNIRS signals for subsequent analysis or fusion experiments.
Multiple fusion strategies have been developed to integrate EEG and fNIRS signals, each with distinct advantages. Research comparing these approaches has demonstrated that early-stage fusion significantly outperforms middle-stage and late-stage fusion for classifying motor imagery tasks [15]. The diagram below illustrates the architectural differences and information flow between these three primary fusion strategies.
Successful execution of EEG-fNIRS experiments requires specific hardware, software, and analytical tools. Table 2 catalogues the essential "research reagents" for multimodal brain signal research.
Table 2: Essential Research Reagents for EEG-fNIRS Experiments
| Category | Item | Specification/Example | Primary Function |
|---|---|---|---|
| Hardware | EEG Amplifier | g.HIamp (g.tec), Neuroscan SynAmps2 | High-fidelity electrical signal acquisition (≥256 Hz) |
| fNIRS System | NirScan (Danyang Huichuang), NIRScout (NIRx) | Hemodynamic response measurement via near-infrared light | |
| Hybrid Cap | Custom designs with integrated EEG electrodes & fNIRS optodes | Simultaneous signal acquisition with precise spatial co-registration | |
| Software | Stimulus Presentation | E-Prime, PsychoPy, Presentation | Precise experimental paradigm control and event marker generation |
| Data Analysis | EEGLAB, FieldTrip, MNE-Python, HomER2, NIRS-KIT | Signal processing, visualization, and statistical analysis | |
| Machine Learning | BenchNIRS [100], Scikit-learn, PyTorch, TensorFlow | Model development, training, and benchmarking | |
| Analytical Components | Clinical Scales | Fugl-Meyer Assessment (FMA-UE), Modified Barthel Index (MBI) | Quantitative assessment of motor function and independence |
| Self-Assessment Scales | SAM (Valence/Arousal), PANAS | Subjective emotion and affective state measurement |
The developing ecosystem of public datasets and benchmarking standards for EEG-fNIRS research provides critical infrastructure for advancing data fusion techniques. The available resources support diverse applications from motor rehabilitation to affective computing, while emerging frameworks like BenchNIRS address the crucial need for standardized evaluation. For researchers in academia and drug development, these tools enable more reproducible, comparable, and clinically relevant studies of brain function and therapeutic interventions. Future work should focus on expanding patient-specific datasets, validating fusion algorithms in clinical populations, and establishing consensus standards for reporting results across the research community.
The integration of electroencephalography (EEG) and functional near-infrared spectroscopy (fNIRS) represents a transformative approach in clinical neuroscience, enabling a more comprehensive investigation of brain function by capturing complementary neural signatures. EEG provides millisecond-level temporal resolution of electrophysiological activity, while fNIRS tracks hemodynamic responses with superior spatial localization [5]. This multimodal data fusion framework offers a powerful tool for identifying robust biomarkers and elucidating the underlying neuropathophysiology of neurological and psychiatric disorders, ultimately advancing objective diagnostic and therapeutic monitoring capabilities [4] [5].
This application note details experimental protocols and presents clinical validation case studies for three distinct conditions: Parkinson's disease, Attention-Deficit/Hyperactivity Disorder, and Major Depressive Disorder. The synergistic combination of EEG and fNIRS leverages complementary information; EEG reflects neuronal electrical discharges, whereas fNIRS measures hemodynamic changes associated with regional neural metabolism [5]. This dual-modality approach is particularly valuable for investigating complex phenomena such as cognitive-motor interference and neurovascular coupling, which are often impaired in various brain disorders [102].
Successful execution of simultaneous EEG-fNIRS studies requires specific hardware, software, and methodological components. The table below catalogues the essential solutions for integrated system configuration, signal acquisition, and data processing.
Table 1: Key Research Reagent Solutions for Integrated EEG-fNIRS Studies
| Item Name | Function/Application | Technical Notes |
|---|---|---|
| Integrated EEG-fNIRS Helmet | Co-registration of electrophysiological and hemodynamic signals from homologous brain regions. | Custom designs via 3D printing or cryogenic thermoplastic sheets improve scalp coupling and placement consistency [5]. |
| Synchronization Module | Temporal alignment of EEG and fNIRS data streams with high precision. | A unified processor for simultaneous acquisition is preferred over post-hoc software synchronization [5]. |
| fNIRS Probe Configurations | Measures concentration changes in oxygenated (HbO) and deoxygenated (HbR) hemoglobin. | Source-detector distances of 3 cm are typical to ensure adequate cortical penetration [5]. |
| EEG Recording System | Records electrical potentials from the scalp with high temporal resolution. | A 32-channel system following the 10-20 international standard is common for full-brain coverage [103]. |
| Task-Related Component Analysis (TRCA) | A computational algorithm to extract task-reproducible components from bimodal signals, enhancing signal-to-noise ratio. | Effectively improves the reproducibility and discriminability of neural patterns for subsequent analysis [102]. |
| Mutual Information-Based Feature Selection | A filter method for selecting an optimized subset of features from the fused EEG-fNIRS dataset by maximizing relevance and minimizing redundancy. | Helps improve hybrid classification performance by exploiting feature complementarity with respect to class labels [4]. |
| Correlation-Based Signal Improvement (CBSI) | A combined hemoglobin measure used in fNIRS analysis to mitigate motion artifacts and improve signal quality. | Validated as a robust measure of executive demand during walking tasks in Parkinson's disease [104]. |
Background & Validation Objective: Parkinson's disease (PD) is characterized by motor symptoms and cognitive decline. Cognitive-motor interference (CMI) occurs when simultaneous performance of a motor and cognitive task leads to deterioration in one or both, a common challenge in PD that increases fall risk [102] [104]. This study validated fNIRS measurement of executive demand during walking with and without a dual-task in younger adults, older adults, and people with PD [104].
Experimental Protocol:
Key Findings & Clinical Relevance: The fNIRS CBSI measure was validated as an indicator of executive demand during walking. A positive relationship was found between dlPFC activity during dual-task walking and dual-task cost in younger adults. In the PD group, dlPFC activity during single-task walking was positively related to step time variability and negatively related to walking speed [104]. This protocol provides an objective method to quantify the cognitive burden of walking in PD, which can be used to assess disease progression and the efficacy of rehabilitative interventions.
Background & Validation Objective: Major Depressive Disorder diagnosis often relies on subjective clinical interviews and rating scales, leading to a need for objective biomarkers. This study aimed to establish an automated diagnostic assessment for MDD by combining EEG and fNIRS to improve classification accuracy [103].
Experimental Protocol:
Key Findings & Clinical Relevance: The hybrid EEG-fNIRS model achieved a classification accuracy of 92.7%, significantly outperforming the model using EEG features alone (81.8%) [103]. Key biomarkers included brain network local efficiency in the delta band, hemispheric asymmetry in the theta band, and brain oxygen sample entropy. This demonstrates the substantial potential of multimodal data fusion in providing an objective, physiological-based diagnostic tool for depression.
Background & Validation Objective: ADHD is a heterogeneous neurodevelopmental disorder. Machine learning (ML) applied to multimodal data can help unravel its complex neural mechanisms beyond what is possible with single-modality studies [105].
Experimental Protocol:
Key Findings & Clinical Relevance: Machine learning models that integrate multimodal data have successfully identified that core characteristics of ADHD extend beyond traditional symptom descriptions. For instance, models have highlighted the discriminative power of teachers' ratings on oppositional questions and the significance of factors like family history and academic difficulties [105]. The application of ML to EEG and fNIRS data holds promise for discovering objective biomarkers that can aid in subtyping ADHD and predicting treatment response, moving towards precision medicine.
The following table synthesizes key performance metrics and findings from the featured case studies and related research, highlighting the efficacy of the EEG-fNIRS fusion approach.
Table 2: Summary of Quantitative Results from EEG-fNIRS Clinical Validation Studies
| Clinical Condition | Primary Metric | EEG Alone | fNIRS Alone | Fused EEG-fNIRS | Key Biomarkers Identified |
|---|---|---|---|---|---|
| Major Depressive Disorder (MDD) [103] | Classification Accuracy | 81.8% | Information Missing | 92.7% | Delta band local efficiency, Theta band asymmetry, Brain oxygen sample entropy |
| Psychological Stress [4] [103] | Classification Accuracy | ~3.4% improvement vs. single modality | ~11% improvement vs. single modality | ~5-11% increase vs. best single modality | Mutual information-based selected features |
| Mental Workload [103] | Classification Accuracy | 85.9% | 74.8% | 90.9% | Hybrid temporal-spectral-hemodynamic features |
| Amyotrophic Lateral Sclerosis (ALS) [4] | Hybrid Classification Performance | Information Missing | Information Missing | Significantly improved vs. single modality and conventional classification | Optimized feature subset from mutual information selection |
The diagram below illustrates the end-to-end pipeline for acquiring, processing, and analyzing simultaneous EEG-fNIRS data in clinical validation studies.
This diagram conceptualizes the neural processes and their measurement during a dual-task paradigm, relevant to Parkinson's disease and other conditions affecting cognitive-motor function.
Motor Imagery (MI) decoding is a cornerstone of modern Brain-Computer Interface (BCI) systems, enabling direct communication between the brain and external devices by interpreting the neural activity associated with imagined movements without any physical execution. This technology has transformative potential, particularly in neurorehabilitation, assisting patients with motor disabilities, stroke, ALS, or spinal cord injuries in regaining communication and control capabilities [106] [107].
The core challenge in MI-BCI systems lies in accurately classifying intention from noisy, non-stationary Electroencephalogram (EEG) signals. Recent breakthroughs have significantly enhanced performance through two key paradigms: the integration of source localization techniques with deep learning, and the application of sophisticated domain adaptation frameworks to overcome the variability in brain signals across different individuals [108] [109].
Table 1: Quantitative Performance of Advanced MI Decoding Models
| Model / Approach | Core Methodology | Dataset(s) | Performance (Accuracy) | Key Advantage |
|---|---|---|---|---|
| Beamforming + ResNet CNN [108] | Source localization (Beamforming) transformed to cortical maps, analyzed by a custom CNN. | 4-class MI (Left hand, Right hand, Both feet, Tongue) | 99.15% (within-subject) | Superior spatial resolution; outperforms sensor-domain approaches. |
| TFANet [110] | Temporal Fusion Attention Network with Multi-Scale Temporal Self-Attention (MSTSA). | BCIC-IV-2a, BCIC-IV-2b | 84.92% (2a), 88.41% (2b) (within-subject) | Effectively captures complex temporal dependencies in EEG signals. |
| MSDCDA [109] | Multi-source Dynamic Conditional Domain Adaptation Network with dynamic residual modules. | BCIC-IV-2a, BCIC-IV-2b | 78.55% (2a), 85.08% (2b) (cross-subject) | Mitigates inter-subject variability; enables effective cross-user decoding. |
The performance gains from these models are a substantial improvement over traditional machine learning methods, which often rely on hand-crafted features and struggle with signal noise and subject-specific variations [110]. Furthermore, the fusion of EEG with other modalities like functional near-infrared spectroscopy (fNIRS) represents a promising direction. This multimodal approach compensates for the inherent limitations of each technique—combining the excellent temporal resolution of EEG with the superior spatial resolution and metabolic information of fNIRS—paving the way for more robust and naturalistic brain imaging systems [7].
This protocol details the methodology for achieving state-of-the-art within-subject classification accuracy [108].
A. Data Acquisition & Preprocessing
B. Source Localization via Beamforming
C. Deep Learning Model Training & Classification
This protocol addresses the challenge of applying a model trained on one subject to a new, unseen subject, a critical requirement for real-world BCI deployment [109].
A. Multi-Source Data Preparation
B. Dynamic Domain Adaptation
C. Model Evaluation
Table 2: Essential Materials and Computational Tools for MI-Decoding Research
| Item | Function / Application in MI Research |
|---|---|
| High-Density EEG System | Records electrical brain activity from the scalp with high temporal resolution. Essential for capturing event-related desynchronization (ERD) during motor imagery [109]. |
| fNIRS/HD-DOT System | Measures hemodynamic changes (blood oxygenation) in the cortex. Used in multimodal fusion with EEG to provide complementary spatial information and validate neuronal activity [7]. |
| Public BCI Datasets (e.g., BCIC-IV-2a/2b) | Provide standardized, annotated EEG data for benchmarking and training new decoding algorithms, ensuring reproducibility and comparability of results [110]. |
| Source Localization Algorithms (Beamforming, MNE) | Computational methods to solve the EEG inverse problem, projecting sensor-level signals to their origins on the cortical surface, thereby enhancing spatial information [108]. |
| Domain Adaptation Frameworks (e.g., MSDCDA) | Software and model architectures designed to mitigate the performance drop caused by inter-subject variability, a major hurdle for practical BCI systems [109]. |
| Synthetic fNIRS-EEG Datasets | Computer-generated datasets that simulate realistic, co-registered multimodal signals. They are crucial for developing and testing fusion algorithms in the absence of large, real-world multimodal datasets [7]. |
The integration of electroencephalography (EEG) and functional near-infrared spectroscopy (fNIRS) represents a paradigm shift in neuroimaging, enabling unprecedented insights into brain function through complementary data streams. This multimodal approach leverages EEG's excellent temporal resolution with fNIRS's superior spatial specificity to overcome the limitations of individual modalities [5]. The convergence of these technologies creates a powerful framework for cross-disciplinary validation, bridging the gap between basic neuroscience discoveries and clinical applications in psychiatry and neurology [111]. This protocol details standardized methodologies for fNIRS-EEG data fusion, with particular emphasis on translational applications in therapeutic development and clinical diagnostics.
The fundamental rationale for EEG-fNIRS integration stems from their complementary nature: EEG captures millisecond-scale neuronal electrical activity, while fNIRS measures hemodynamic responses with spatial precision superior to EEG [5]. This combination provides a more complete picture of brain activity by simultaneously tracking electrophysiological and neurovascular components [7]. Furthermore, both modalities share practical advantages including non-invasiveness, portability, and relatively low cost compared to fMRI and PET, making them ideal for naturalistic experimental settings and diverse patient populations [5].
Table 1: Technical Specifications and Comparative Analysis of Neuroimaging Modalities
| Technique | Temporal Resolution | Spatial Resolution | Invasiveness | Key Measured Parameters | Primary Applications |
|---|---|---|---|---|---|
| EEG | Millisecond-scale (excellent) | ~1-2 cm (low) | Non-invasive | Electrical potentials from synchronized neuronal firing | Epilepsy monitoring, sleep studies, brain-computer interfaces |
| fNIRS | Seconds (moderate) | ~1-2 cm (moderate) | Non-invasive | Concentration changes in oxygenated (HbO) and deoxygenated hemoglobin (HbR) | Functional brain mapping, cognitive neuroscience, clinical monitoring |
| fMRI | 1-2 seconds (moderate) | 1-3 mm (high) | Non-invasive | Blood oxygenation level dependent (BOLD) signal | Pre-surgical mapping, cognitive task localization, connectomics |
| PET | Minutes (poor) | 3-5 mm (high) | Injective (radioactive tracers) | Metabolic activity, receptor binding, neurotransmitter dynamics | Oncology, neuropharmacology, metabolic studies |
| MEG | Millisecond-scale (excellent) | 3-5 mm (high) | Non-invasive | Magnetic fields generated by neuronal activity | Source localization, epilepsy focus identification |
The synergistic potential of EEG and fNIRS emerges directly from their complementary technical profiles. EEG's exceptional temporal resolution (milliseconds) captures neural dynamics with precision unmatched by hemodynamic methods, while fNIRS provides spatial localization superior to EEG alone [5]. Crucially, both modalities operate without electromagnetic interference, enabling truly simultaneous data acquisition – a significant advantage over combinations involving fMRI or MEG [5].
Two primary approaches exist for hardware integration of EEG-fNIRS systems:
Synchronized Separate Systems: EEG and fNIRS data are acquired using independent equipment (e.g., NIRScout and BrainAMP systems) with post-hoc synchronization during analysis [5]. While simpler to implement, this approach may suffer from precision limitations for microsecond-scale EEG analysis.
Unified Processor Systems: A single integrated unit processes and acquires both EEG and fNIRS signals simultaneously [5]. This architecture provides superior synchronization precision and streamlines analytical workflows, making it the preferred approach despite greater system complexity.
Helmet design represents a critical integration challenge. Current solutions include:
Table 2: Data Fusion Techniques for EEG-fNIRS Integration
| Fusion Level | Description | Common Algorithms/Methods | Advantages | Limitations |
|---|---|---|---|---|
| Early Fusion (Data-level) | Raw or preprocessed data combined before feature extraction | Y-shaped neural networks [15] | Preserves temporal relationship; enables shared representation learning | Susceptible to noise propagation; requires temporal alignment |
| Middle Fusion (Feature-level) | Features extracted separately then combined | Common Spatial Pattern (CSP) + hemoglobin mean/slope [15], Canonical Correlation Analysis [15] | Leverages modality-specific processing; more robust to noise | May miss cross-modal interactions; requires careful feature selection |
| Late Fusion (Decision-level) | Separate classification with combined decisions | Weighted voting, Bayesian fusion | Flexible to modality-specific classifiers; robust to missing data | Loses low-level interactions; may yield suboptimal performance |
Evidence suggests that early-stage fusion of EEG and fNIRS significantly outperforms middle and late fusion approaches in classification tasks. In motor imagery experiments, Y-shaped neural networks implementing early fusion demonstrated statistically higher performance (N = 57, P < 0.05), achieving average accuracy of 76.21% in left/right hand discrimination tasks [15].
Protocol 1: Simultaneous EEG-fNIRS Acquisition for Motor Imagery Classification
This protocol details the experimental setup based on Dataset A from Shin et al. (2017), which has been widely used in hybrid BCI research [15].
I. Equipment Setup
II. Participant Preparation
III. Experimental Procedure
IV. Data Acquisition Parameters
V. Preprocessing Pipeline
VI. Feature Extraction
This protocol has demonstrated classification accuracy of approximately 76% using early fusion approaches, significantly outperforming unimodal classification (EEG-only: ~65%, fNIRS-only: ~57%) [15].
Figure 1: Comprehensive Workflow for EEG-fNIRS Multimodal Experiment
Table 3: Clinical Applications of EEG-fNIRS Fusion Technology
| Clinical Domain | Specific Application | Fusion Benefits | Reported Performance |
|---|---|---|---|
| Psychiatry | Schizophrenia diagnosis [111] | Combined neurovascular coupling and electrical activity reveals default mode network abnormalities | Identification of inhibitory PFC→insular connectivity [111] |
| Epilepsy | Seizure focus localization [5] | Improved spatial precision for EEG spikes with hemodynamic correlation | Enhanced surgical planning accuracy |
| Neurodegenerative Disorders | Parkinson's disease monitoring [7] | Complementary motor cortex signatures during movement tasks | Successful distinction from healthy controls [7] |
| Anesthesiology | Depth of anesthesia monitoring [5] | Multimodal confirmation of anesthetic effects on brain function | Reduced incidence of intraoperative awareness |
| Pediatric Neurology | Infantile spasm evaluation [5] | Non-invasive monitoring suitable for developing brains | Improved seizure characterization |
| Stroke Rehabilitation | Motor recovery assessment | Tracking both electrical reorganization and hemodynamic adaptation | More comprehensive recovery biomarkers |
The translational potential of EEG-fNIRS fusion is particularly promising in psychiatry, where it enables a novel cross-validation framework between clinical phenomenology and neurobiological mechanisms [111]. This approach involves simultaneous administration of psychological assessment tools during neuroimaging acquisition, creating direct "translational bridges" between subjective experience and objective brain measures [112].
For instance, in schizophrenia research, concurrent fMRI and paranoid-depressive scale administration revealed robust activations in default mode network structures (precuneus, posterior cingulate cortex, angular gyrus) during paranoid ideation, alongside aberrant inhibitory connectivity from prefrontal cortex to anterior insula [111]. This methodology can be adapted for EEG-fNIRS studies to establish clinically practical biomarkers for diagnostic assessment and treatment monitoring.
Protocol 2: Simultaneous Neuroimaging and Clinical Assessment for Psychiatric Disorders
This protocol adapts the translational cross-validation model proposed by Stoyanov et al. for EEG-fNIRS applications in psychiatric disorders [111] [112].
I. Clinical Assessment Integration
II. Experimental Design Considerations
III. Data Integration and Analysis
IV. Validation Framework
This protocol enables the development of clinically practical biomarkers where validated psychological assessments can eventually serve as proxies for more costly neuroimaging procedures in treatment monitoring and medication selection [112].
Table 4: Key Research Reagent Solutions for EEG-fNIRS Studies
| Category | Specific Tool/Resource | Function | Application Notes |
|---|---|---|---|
| Hardware Platforms | NIRScout System (fNIRS) | fNIRS signal acquisition with multiple wavelength capability | Compatible with BrainAMP EEG for synchronized separate systems |
| BrainAMP System (EEG) | High-quality EEG acquisition with low noise characteristics | ||
| Integrated EEG-fNIRS Caps | Customizable headgear for simultaneous acquisition | 3D-printed or thermoplastic options improve fit and stability | |
| Software Tools | Homer2 / NIRS Toolbox | fNIRS data processing and hemoglobin calculation | Standardized pipeline for optical data preprocessing |
| EEGLAB / FieldTrip | EEG signal processing and artifact removal | Comprehensive toolbox for electrophysiological analysis | |
| Custom Y-shaped Networks | Early fusion deep learning architecture | Implemented in Python (TensorFlow/PyTorch) or MATLAB | |
| Data Resources | Public Dataset A (Shin et al.) | Left/right hand motor imagery benchmark | 29 subjects, 30 trials/condition [15] |
| Public Dataset B (Shin et al.) | Mental arithmetic vs. relax imagery | Useful for cognitive state classification | |
| Analytical Modules | Common Spatial Patterns (CSP) | EEG feature extraction for motor imagery | Maximizes variance between conditions |
| Canonical Correlation Analysis | Multimodal feature fusion | Identifies correlated components across modalities | |
| Short-Time Fourier Transform | Time-frequency analysis for EEG | Enables spectrogram imaging for deep learning [113] |
Rigorous statistical validation is essential given the high-dimensional nature of multimodal neuroimaging data. Recent research highlights critical considerations for cross-validation in neuroimaging-based classification [114]:
Robust artifact removal remains challenging in multimodal studies:
Future methodological developments should focus on standardized artifact handling, improved motion correction, and open benchmarking datasets to enhance reproducibility across research sites [7] [5].
The integration of EEG and fNIRS through advanced data fusion techniques represents a transformative approach in neuroscience and clinical medicine. The protocols and methodologies outlined herein provide a framework for rigorous cross-disciplinary validation, enabling researchers to leverage the complementary strengths of electrophysiological and hemodynamic measures. As hardware integration becomes more sophisticated and analytical methods more powerful, this multimodal approach promises to accelerate the translation of basic neuroscience discoveries into clinically actionable tools for diagnosis, therapeutic monitoring, and personalized treatment in neurology and psychiatry.
EEG-fNIRS data fusion represents a transformative approach in neuroimaging that successfully bridges the temporal-spatial resolution divide, enabling more comprehensive brain activity monitoring in naturalistic settings. The integration of these complementary modalities has demonstrated significant improvements in classification accuracy across numerous applications, from brain-computer interfaces to clinical diagnostics, with some studies achieving over 92% classification performance. Future directions should focus on developing more sophisticated unsupervised fusion algorithms, creating larger open-access multimodal datasets, improving hardware portability and integration, and establishing standardized validation frameworks. As these technologies become more accessible and analytical methods more refined, EEG-fNIRS fusion holds tremendous potential to advance personalized medicine, accelerate therapeutic development, and provide deeper insights into brain function across diverse populations and real-world environments. The continued collaboration between engineers, neuroscientists, and clinical researchers will be essential to fully realize the promise of multimodal neuroimaging in understanding and treating neurological and psychiatric disorders.