高级检索

    多层次特征建模与时空相关性挖掘的自监督脑电图分类

    Self-supervised EEG Classification with Multi-level Feature Modeling and Spatial-temporal Correlation Mining

    • 摘要: 多通道脑电图(electroencephalography,EEG)作为一种非侵入性技术,通过在头皮上布置多个电极记录大脑电活动,帮助理解个体的心理状态和辅助诊断多种疾病.鉴于标记大量EEG数据的高昂成本和技术挑战,自监督学习(self-supervised learning,SSL)作为一种无需标签的学习范式,通过挖掘数据内在结构进行学习,可有效提升模型的泛化性能,已在EEG领域获得广泛关注.尽管当前研究已取得了显著进展,但仍面临以下挑战:首先,多通道EEG数据蕴含复杂的时空关联,然而众多当前方法仅局限于时间或空间单一维度的建模,未能全面融合两者,限制了对EEG信号内在复杂特征的深入理解;其次,许多现有方法未能有效结合利用片段级与示例级信息,前者有助于提升模型的泛化能力,而后者则能帮助模型更好地适应下游(分类)任务.针对上述挑战,本文提出并实现一个结合对比和重建的自监督预训练框架.具体而言,在时间维度掩码重建的基础上引入通道掩码策略,有效捕捉EEG数据的时空关系,并通过细粒度地捕捉片段间的关系,提升了模型的性能和泛化能力.同时,结合示例级对比学习与掩码重建任务,帮助模型学习到具有示例区分性和局部感知能力的表示,从而更好地适应下游任务.此外,引入自步学习机制进一步增强了模型训练的稳定性.最后,多个EEG任务上的实验结果验证了所提方法的有效性.

       

      Abstract: Multi-channel electroencephalography (EEG), as a non-invasive technique, records brain electrical activity through multiple electrodes placed on the scalp, aiding in understanding individual psychological states and assisting in the diagnosis of various diseases. Given the high cost and technical challenges of annotating large-scale EEG data, self-supervised learning (SSL), as a label-free learning paradigm, has garnered widespread attention in the EEG domain. SSL leverages the intrinsic structure of data for learning, effectively enhancing model generalization ability. Despite significant advancements in current research, it encounters the following challenges: firstly, multi-channel EEG data contains complex spatiotemporal correlations, yet many existing methods are confined to modeling either temporal or spatial dimensions independently, rather than integrating both comprehensively and thus limiting a profound understanding of the intrinsic complexities of EEG signals; secondly, many current approaches have failed to integrate segment-level and instance-level information effectively. The former contributes to improving the model's generalization capability, while the latter aids the model in better adapting to downstream (classification) tasks. To address the aforementioned challenges, this study proposes and implements a self-supervised pre-training framework that combines contrastive and reconstruction strategies. Specifically, by introducing a channel masking strategy on the basis of temporal dimension mask reconstruction, the spatiotemporal relationships of EEG data are effectively captured. Through fine-grained capturing of inter-segment relationships, the performance and generalization ability of the model are improved. Meanwhile, combining instance-level contrastive learning with masked reconstruction tasks helps the model learn representations with instance discriminability and local perception ability, thus better adapting to downstream tasks. Furthermore, the introduction of a self-paced learning mechanism further enhances the stability of model training. Finally, experimental results across multiple EEG tasks validate the effectiveness of the proposed approach.

       

    /

    返回文章
    返回