Skip to yearly menu bar Skip to main content


Poster

A Vector Quantization Pretraining Method for EEG Time Series with Random Projection and Phase Alignment

Haokun Gui · Xiucheng Li · Xinyang Chen


Abstract:

In this paper, we propose a BERT-style self-supervised learning model, VQ-MTM (Vector Quantization Masked Time-Series Modeling), for the EEG time series data analysis. At its core, VQ-MTM comprises a theoretically grounded random-projection quantization module and a phase-aligning module guided by the Time-Phase-Shift Equivariance of DFT, the two modules can generate well-defined semantic units for the corrupted and periodic time series, thus offering robust and consistent learning signals for the EEG self-supervised learning. VQ-MTM also owns low model complexity and can easily adapt to large-scale datasets. We conduct experiments on five real-world datasets including two large-scale datasets to verify the efficacy of our proposed model, the experiment results show that VQ-MTM is able to consistently surpass the existing methods by large margins on both seizure detection and classification tasks. Our code is available at https://anonymous.4open.science/r/Time-Series-Pretrain-24F6.

Live content is unavailable. Log in and register to view live content