高级检索

    多层次特征建模与时空依赖挖掘的自监督脑电分类

    Self-Supervised EEG Classification with Multi-Level Feature Modeling and Spatiotemporal Dependence Mining

    • 摘要: 多通道脑电图(electroencephalography,EEG)作为一种非侵入性技术,通过在头皮上布置多个电极记录大脑电活动,帮助理解个体的心理状态和辅助诊断多种疾病. 鉴于标记大量EEG数据的高昂成本和技术挑战,自监督学习(self-supervised learning,SSL)作为一种无需标签的学习范式,通过挖掘数据内在结构进行学习,可有效提升模型的泛化性能,已在EEG领域获得广泛关注. 尽管当前研究已取得了显著进展,但仍面临以下挑战:首先,多通道EEG数据蕴含复杂的时空关联,然而众多当前方法仅局限于时间或空间单一维度的建模,未能全面融合两者,限制了对EEG信号内在复杂特征的深入理解;其次,许多现有方法未能有效结合利用片段级与示例级信息,前者有助于提升模型的泛化能力,而后者则能帮助模型更好地适应下游(分类)任务. 针对上述挑战,提出并实现一个结合对比和重建的自监督预训练框架. 具体而言,在时间维度掩码重建的基础上引入通道掩码策略,有效捕捉EEG数据的时空关系,并通过细粒度地捕捉片段间的关系,提升了模型的性能和泛化能力. 同时,结合示例级对比学习与掩码重建任务,帮助模型学习到具有示例区分性和局部感知能力的表示,从而更好地适应下游任务. 此外,引入自步学习机制进一步增强了模型的泛化能力. 最后,多个EEG任务上的实验结果验证了所提方法的有效性.

       

      Abstract: Multi-channel electroencephalography (EEG), as a non-invasive technique, records brain electrical activity through multiple electrodes placed on the scalp, aiding in understanding individual psychological states and assisting in the diagnosis of various diseases. Given the high cost and technical challenges of annotating large-scale EEG data, self-supervised learning (SSL), as a label-free learning paradigm, has garnered widespread attention in the EEG domain. SSL leverages the intrinsic structure of data to learn representations, thereby effectively enhancing the model's generalization ability. Despite significant advancements in current research, the field still faces the following challenges: first, multi-channel EEG data contain complex spatiotemporal correlations, yet many existing methods are confined to modeling either temporal or spatial dimensions in isolation, rather than integrating both comprehensively, thus limiting the ability to profoundly understand the intrinsic complexities of EEG signals; second, many current approaches have not effectively integrated segment-level and instance-level information. The former contributes to improving the model's generalization ability, while the latter aids the model in better adapting to downstream (classification) tasks. To address these challenges, we propose and implement a self-supervised pre-training framework that combines contrastive and reconstruction strategies. Specifically, by introducing a channel masking strategy based on temporal mask reconstruction, our framework effectively captures the spatiotemporal relationships of EEG data. By capturing inter-segment relationships in a fine-grained manner, the framework enhances both the model's performance and its generalization ability. Meanwhile, integrating instance-level contrastive learning with masked reconstruction tasks helps the model learn representations with instance discriminability and local perceptual ability, thus facilitating better adaptation to downstream tasks. Additionally, the introduction of the self-paced learning mechanism further enhances the model's generalization ability. Finally, experimental results across multiple EEG tasks demonstrate the effectiveness of the proposed approach.

       

    /

    返回文章
    返回