IEEE Transactions on Neural Systems and Rehabilitation Engineering (Jan 2024)

Simplifying Multimodal With Single EOG Modality for Automatic Sleep Staging

  • Yangxuan Zhou,
  • Sha Zhao,
  • Jiquan Wang,
  • Haiteng Jiang,
  • Zhenghe Yu,
  • Shijian Li,
  • Tao Li,
  • Gang Pan

DOI
https://doi.org/10.1109/TNSRE.2024.3389077
Journal volume & issue
Vol. 32
pp. 1668 – 1678

Abstract

Read online

Polysomnography (PSG) recordings have been widely used for sleep staging in clinics, containing multiple modality signals (i.e., EEG and EOG). Recently, many studies have combined EEG and EOG modalities for sleep staging, since they are the most and the second most powerful modality for sleep staging among PSG recordings, respectively. However, EEG is complex to collect and sensitive to environment noise or other body activities, imbedding its use in clinical practice. Comparatively, EOG is much more easily to be obtained. In order to make full use of the powerful ability of EEG and the easy collection of EOG, we propose a novel framework to simplify multimodal sleep staging with a single EOG modality. It still performs well with only EOG modality in the absence of the EEG. Specifically, we first model the correlation between EEG and EOG, and then based on the correlation we generate multimodal features with time and frequency guided generators by adopting the idea of generative adversarial learning. We collected a real-world sleep dataset containing 67 recordings and used other four public datasets for evaluation. Compared with other existing sleep staging methods, our framework performs the best when solely using the EOG modality. Moreover, under our framework, EOG provides a comparable performance to EEG.

Keywords