This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
1. Introduction
Sleep plays an important role in human health [1]. The sleep monitoring of human has significant implications for medical research and practice [2]. Sleep specialists usually evaluate the quality of sleep by analyzing signals from sensors connected to different parts of body in accordance with the Rechtschaffen and Kales Rules [3] or the American Academy of Sleep Medicine (AASM) sleep score manual [4]. In particular, polysomnography (PSG), which records EEGs, electrooculograms (EOG), electrocardiograph (ECG), electromyography (EMG), respiratory effort, leg movement, and blood oxygen saturation over several nights in a sleep laboratory, is considered as a gold standard for evaluating sleep status of subjects [5]. In order to improve the efficiency of sleep monitoring, several effective sleep staging methods based on EEG, ECG, and EMG signals have been proposed in recent years [6]. However, wearing too many sensors during sleep is obtrusive and uncomfortable, the silver/silver chloride electrodes with certain adhesive or conductive paste the signal acquisition are adopted mostly, and the placement of them is demanded carefully in hairy regions of scalp to minimize movement-related noise, which affects the natural sleep of subjects and is not suitable for long-term sleep monitoring in home environment.
In recent years, noninvasive [7, 8] or noncontact [9, 10] measurement of cardiac, respiratory, and body movement signals which offers the potential of low cost and easy operation for long term dynamic sleep monitoring has gradually gained the favour of researchers. But its performances badly depend on the quality of signal acquisition and complex signal processing. EEG recordings play a crucial role in the classification of sleep stages. In order to solve the problem of wearers’ comfort, classification of sleep stages based on single-channel EEG signals has been extensively investigated [11, 12], because compared to complex PSG devices, the corresponding dual-electrode device has the advantages of wearable and less interference.
Compared with frontal electrodes, EEG recordings from central, occipital, and parietal electrodes could better detect sleep spindles, vertex waveforms, and rhythms [13]. However, channel selection is important for measurement convenience. Fp1, Fp2, and Fpz electrodes below hairline are suitable for wearing and data collection, so do F3 and F4 electrodes. Researchers have tried to address this problem by extracting the same relevant information from frontal electrodes [14]. But the information which can support for W, N1, N2, and N3 stages is different from that extracted from the parietal lobe, so it can affect the credibility of staging results.
The performance of most sleep stage classification methods relies on the selection of representative features for different sleep stages. Frequency domain [15], time-domain, and time-frequency domain [16] decompositions are the common steps for processing time signals and extracting features directly, and various mathematical models have been established in the process of discovering hidden features. After feature extraction, various machine learning algorithms are used for classification [17, 18], such as linear discriminant analysis, nearest neighbour classifier, decision tree, support vector machine, random forest, and ensemble learning. It also shows the good results with combinatorial machine learning models [19]. In recent years, deep learning methods such as convolutional neural network (CNN) [20], recurrent neural network, and other deep neural networks have become common in pattern recognition in biomedical signal processing. In [21], long short-term memory (LSTM) model which takes advantage of sequential data learning to optimize classification performance was proposed for automated sleep stage. Since feature-based approaches may not be suitable for a comprehensive description of subject heterogeneity, CNNs were also applied to learn multiple filters to extract time-invariant features from raw EEG channel [22].
To solve both the subject heterogeneity and temporal pattern recognition problems, the combination of CNN and LSTM has shown a good performance for the usage of precomputed spectrograms in [23]. However, most representative deep learning models rely heavily on hyperparameter tuning, which is challenging to extend. Although some studies such as [24] consider the temporal context, training must be divided into pretraining and fine-tuning. In addition, since the learning rate is set to a very low value during fine-tuning, it takes more time to reach the optimum or may not even be optimal. Therefore, the computation cannot be performed in parallel, which also prolongs the training process.
Increasing the network layer number of CNNs can improve the extraction effect of signal features. Multiscale CNNs were also proposed to perform multiscale feature extraction and classification simultaneously [25]. However, gradient dispersion or gradient explosion is likely to occur when the depth of CNNs increases. The residual network (ResNet) proposed by He et al. [26] was aimed to solve the degeneration problem of network. The method was applied to the machine fault detection and achieved good results [27]. The sleep stage classification based on residual-based attention model was also adopted in [28], but only
Considering that multiscale convolutional neural network can capture the detailed signal features required for pattern classification, the idea was adopted to realize a wearable smart eye-mask in our prior study [29]. This method uses single-channel original EEG signals and omits the process of special feature extraction. It has good performance and application potential and can provide support for clinical applications such as screening and diagnosis of sleep disorders. The main contributions of this work are as follows:
(1) A deep learning architecture of ResNet with different filter sizes was developed. Time-invariant features from original single-channel EEG signals can be extracted by training learning filters, so it can save the time of features computation, while residual can be trained to encode temporal information such as sleep phase transition rules into the model
(2) A training algorithm that can effectively train the model end-to-end was developed. Single-channel EEG from forehead was adapted to reduce the patients impact and increase usability
(3) The proposed approach is evaluated on two publicly available datasets: sleep-EDF-expanded [30] and CinC2018 [31] through subject cross-validation experiment with a strong robustness performance. The results are compared with state-of-the-art results in the field of sleep stage classification to demonstrate the superiority of our method, and it solves the problem that too little data may lead to poor generalization effect of sleep stage model
2. Methods
Sleep staging is a problem of recognition or regression of time series signals. According to the sleep manual, experts divided the EEG, EOG, EMG, and other signals obtained by sensors at intervals of 30 s into W, REM, N1, N2, and N3. In this paper, the multiscale feature learning was integrated into a residual network (ResNet) to automatically learn sleep features of original physiological signals at different time scales, and classification in a parallel way from complex original physiological signals of sleep was achieved.
2.1. Residual Block
In general, the more layers in convolutional neural network, the more diverse features can be extracted. But previous experiments show that there exists a degradation problem in deep networks: when the network depth of network increases, accuracy saturates and then degrades, and the addition of more layers can lead to an even higher training error, which is not caused by overfitting.
Although the CNN of dozens in layers can be trained by normalized initialization and batch normalization (BN), it is prone to degenerate as the number of layers increases. Theoretically, if the additional layer of a certain layer in the deep network does not learn anything, but just copies the features of last layer, it is called identity mapping, and the training error should not increase. A deep residual learning framework was proposed to solve the degradation phenomenon [26]: if an identity mapping was optimal, it would be easier to push the residual to zero than to fit an identity mapping by a stack of nonlinear layers.
For the network with an input
[figure omitted; refer to PDF]
The output
2.2. MRCNN
In order to extract features from different receiving domains, we used three ResNet pathways constructed by multiple ResNet units as shown in Figure 2, and time series signal fragments were directly used as inputs of the network. Each path contained four ResNet units with two convolutional layers and a shortcut. Each convolution layer was followed by the BN and the activation function ReLU. The solid line shortcut means that it can be added data directly, dashed line shortcuts indicated that they need to be added the
[figure omitted; refer to PDF]
According to the above description, the expression of ResNet unit constructed can be expressed by the following equations.
where
CNN is a unique artificial neural network inspired by the cerebral cortex, which uses convolution method to extract signal features and compress signal size. In order to retain valid information, it reduces the amount of input data by several orders of magnitude, which can form an optimal network and reduce the risk of network overfitting. CNN contains two core layers: convolution and pooling. The purpose of convolution is to extract features at different levels from original data, and a certain number of filters are used to extract feature maps of input data. The pooling layer is periodically inserted between CNN’s convolutional layers, and a subsampling operation is used to reduce the number of parameters. At the end of the algorithm, the
In a CNN, receptor field is defined as the size of the region mapped on the input image by pixel points on the output feature map in each layer of CNN. In layman’s terms, this is the area where the input feature is “seen” by the output feature point. The receptive field is calculated by the following equation.
The EEG signals input to the network are processed by the convolution layer with a convolution kernel length of 15 and transmitted to the BN and ReLU activation functions for maximum pooling. Then, the output data was sent to three channels of different sizes of the convolution kernel for calculation. Finally, the characteristics of three channels were combined and connected to the full connection layer with 1536 neurons, and the network classification results were obtained by the softmax function.
The receptive field of each convolution layer can be calculated according to equation (1). The receptive field of the output characteristic graph of the last convolution layer in the network was 563 in the input data. EEG signals had a sampling frequency of 100 Hz, and the effective frequency resolution of 563 data points was 0.35 Hz, which can meet the frequency resolution requirements of all rhythmic waves.
2.3. Network Training
In order to realize the loss calculation of multiclassification of sleep stage, cross entropy (CE) was used as the loss function, and its definition was shown as the following equation.
For the loss function, the adaptive moment estimation solver (Adam) was used for optimization. Learning rate was set to
2.4. Performance Evaluation
Recall
3. Experiments
3.1. Datasets
In order to ensure the robustness and reproducibility of the results, two public datasets were conducted experiments. The first dataset used in the experiment was sleep-EDF-expanded and contains two different groups of subjects, named as Sleep Cassette (SC) group and Sleep Telemetry (ST) group, respectively. The annotation files included sleep stages W, REM, Stage 1 (S1), Stage 2 (S2), Stage 3 (S3), Stage 4 (S4), movement time (M), and UNKNOWN, and it consisted of a manual score by a skilled technician. Stage M and UNKNOWN were deleted for their extremely small percentage. At the same time, according to the latest AASM sleep scoring manual, S1 and S2 were corresponded to N1 and N2, respectively, and S3 and S4 were combined into N3. In this study, we cropped the SC
The second dataset used in the experiment was provided by the CinC2018, which contains 1,985 samples. The sleep stages of each sample were labeled by Massachusetts General Hospital clinical staff and divided into six stages: W, N1, N2, N3, REM, and Undefined. For research and application consideration, data from F4-M1 channels was selected here. The data was divided into a training set (
Table 1
The number and proportion of epochs of various sleep stages.
Dataset | W | N1 | N2 | N3 | REM | Total |
Sleep-EDF-expanded (197 samples) | 71197 (29.8%) | 25169 (10.5%) | 88975 (37.2%) | 19454 (8.1%) | 34181 (14.3%) | 238976 (100%) |
CinC2018 (500 samples) | 82368 (17.95%) | 66243 (14.4%) | 197338 (43.0%) | 54318 (11.835%) | 58672 (12.78%) | 458939 (100%) |
3.2. Preprocessing
In order to adapt to different PSG devices and individual differences of subjects, EEG signals were normalized by using the 5-th and 95-th quantiles [32], as shown in the following equation.
In order to expand the dataset and improve the network generalization ability, each input data should be randomly clipped (the 3000 data points in each epoch were randomly clipped to 2700), the flip probability was 50%, and 0.01 times random noise was added. Finally, the preprocessed data were integrated, and then added batch size and channel number, and it was converted to tensor data type, as shown in the following equation.
3.3. Performance Evaluation
The cross-validation used in this paper included
In order to evaluate the performance of proposed method, 5-fold cross-validation and subject cross-validation were completed on sleep-EDFx dataset, and data of 197 subjects from FPz-Cz channels was used. Subject cross-validation divided the dataset into training set and verification set with a partition ratio of 8 : 2, and statistical information is shown in Tables 2 and 3. In order to evaluate the effect in practical application, subject cross-validation was also completed on CinC2018 dataset, and the data of 500 subjects from F4-M1 channel was used.
Table 2
Subject cross-validation statistics in sleep-EDF-expanded dataset.
Samples | W (%) | N1 (%) | N2 (%) | N3 (%) | REM (%) | |
Total | 197 | 71199 (29.8%) | 25174 (10.5%) | 88983 (37.2%) | 19454 (8.1%) | 34184 (14.3%) |
Train | 157 | 53014 (27.7%) | 20236 (10.6%) | 73213 (38.3%) | 16235 (8.5%) | 28497 (14.9%) |
Eval | 40 | 18185 (38.0%) | 4938 (10.3%) | 15770 (33.0%) | 3219 (6.7%) | 5687 (11.9%) |
Table 3
Subject cross-validation statistics in CinC2018 dataset.
Samples | W (%) | N1 (%) | N2 (%) | N3 (%) | REM (%) | |
Total | 500 | 82368 (18.0%) | 66243 (14.4%) | 197338 (43.0%) | 54318 (11.8%) | 58672 (12.8%) |
Train | 400 | 64477 (17.8%) | 48227 (13.3%) | 155927 (43.1%) | 45090 (12.5%) | 47942 (13.3%) |
Eval | 100 | 17891 (18.4%) | 18016 (18.5%) | 41411 (42.6%) | 9228 (9.5%) | 10730 (11.0%) |
4. Results and Discussion
4.1. Cross-Validation
The results of 5-fold cross-validation are shown in Table 4. The table contains the classification performance index of each sleep stage and the overall performance index of the original confusion matrix obtained after cross-validation. The overall recall rate, accuracy rate, specificity, and kappa coefficient are 78.94%, 92.06%, 95.13%, and 0.7360, respectively. It can be seen that the network proposed in this paper can provide good classification performance with high resolution for the period W, but poor resolution for N1. In particular, it is found by the confusion matrix that N1 is easily misjudged as N2 and REM, which is consistent with the results in [33]. This may be due to the relatively small proportion of sleep time in N1, resulting in less training data. In the meantime, N1 is a transitional period between waking state and sleep state and contains both
Table 4
5-fold cross-validation confusion matrix and performance in sleep-EDF-expanded.
W | N1 | N2 | N3 | REM | Kappa | ||||
W | 63232 | 6813 | 254 | 58 | 840 | 88.81 | 95.93 | 98.38 | \ |
N1 | 1871 | 16228 | 3869 | 141 | 3060 | 64.48 | 86.91 | 89.55 | \ |
N2 | 345 | 10401 | 68533 | 5401 | 4295 | 77.02 | 87.87 | 94.30 | \ |
N3 | 30 | 93 | 2566 | 16738 | 27 | 86.04 | 96.51 | 97.44 | \ |
REM | 470 | 5036 | 1864 | 25 | 26786 | 78.36 | 93.46 | 95.99 | \ |
Overall | 78.94 | 92.06 | 95.13 | 0.7360 |
The results of subject cross-validation on the dataset are shown in Table 5. The overall recall rate, accuracy rate, specificity, and kappa coefficient of classification are 75.81%, 91.13%, 94.65%, and 0.7001, respectively. Results compared with the existed sleep staging methods are shown in Table 6. It can maintain better system performance for the data of invisible subjects in the case of large amount of data and reach the similar effect of CNN-LSTM [21]. This cross-validation method can effectively prove the generalization ability of the system on unknown subjects and has high practical application value. In terms of N1 resolution, the recall rate of our proposed network for N1 is 60%, which is much more than other deep learning models.
Table 5
Subject cross-validation confusion matrix and performance in sleep-EDF-expanded.
W | N1 | N2 | N3 | REM | Kappa | ||||
W | 16097 | 1483 | 63 | 17 | 517 | 88.56 | 94.13 | 97.55 | \ |
N1 | 532 | 2971 | 668 | 14 | 753 | 60.17 | 85.78 | 88.73 | \ |
N2 | 78 | 2494 | 11064 | 1006 | 1141 | 70.10 | 86.83 | 95.07 | \ |
N3 | 6 | 78 | 450 | 2670 | 16 | 82.92 | 96.67 | 97.67 | \ |
REM | 110 | 778 | 397 | 6 | 4399 | 77.31 | 92.22 | 94.24 | \ |
Overall | 75.81 | 91.13 | 94.65 | 0.7001 |
Table 6
Performance comparison of different sleep staging models based on subject cross-validation.
Method | Kappa | |||||||
W | N1 | N2 | N3 | REM | ||||
LSTM | 83 | 39 | 80 | 74 | 68 | 80 | 68.8 | 0.64 |
CNN | 80 | 40 | 81 | 69 | 64 | 78 | 66.8 | 0.58 |
CNN-LSTM | 84 | 45 | 82 | 83 | 82 | 86 | 75.2 | 0.71 |
ResNet18 | 89 | 49 | 71 | 81 | 82 | 91 | 74.4 | 0.70 |
Our MRCNN | 89 | 60 | 70 | 83 | 77 | 91 | 75.8 | 0.70 |
In [29], several methods based on residual network were compared by using sleep-EDF-expanded dataset, and it can be seen that our proposed MRCNN performs better than other residual networks also in terms of N1 resolution. In this paper, ResNet18 [26] and MRCNN were compared in same circumstances. In order to adapt to the input of one-dimensional data, all two-dimensional layers in the network structure of ResNet18 were modified to one-dimensional layers. It can also be seen that our proposed MRCNN performs better than ResNet18 in terms of N1 resolution. From the data shown in Table 6, it can be also seen that the proposed network can provide poor resolution for N2. In particular in Table 5, it was found by the confusion matrix that N2 was easily misjudged as N1 or N3. This may be due to N2 is a transitional period between N1 and N3 and contains both sleep spindles and
4.2. Performance Evaluation in Wearable Application
Especially, the channel used in the experiments is F4-M1, which is suitable for wearable application. Training set of 400 samples and test set of 100 samples from CinC2018 were used for subject cross-validation. The results of performance are shown in Table 7. Even though the performance of the system decreased a little when it applied to large amount of data, it still demonstrated the generalization ability of the system in the presence of unknown subjects.
Table 7
Subject cross-validation confusion matrix and performance in CINC2018.
W | N1 | N2 | N3 | REM | Kappa | ||||
W | 14518 | 2442 | 427 | 62 | 440 | 81.15 | 92.49 | 95.05 | \ |
N1 | 3131 | 8588 | 2670 | 33 | 3590 | 47.67 | 83.57 | 91.00 | \ |
N2 | 582 | 4416 | 29880 | 4077 | 2454 | 72.16 | 83.75 | 91.62 | \ |
N3 | 22 | 9 | 1469 | 7663 | 65 | 83.04 | 94.33 | 95.47 | \ |
REM | 193 | 973 | 537 | 6 | 9918 | 84.04 | 92.04 | 92.96 | \ |
Overall | 73.61 | 89.24 | 93.22 | 0.6224 |
4.3. Comparison of Automatic and Artificial Sleep Staging
Automated sleep staging was performed by using the trained model and compared with manual scoring results of expert, as shown in Figure 3, the automatic staging results are close to the manual staging results.
[figures omitted; refer to PDF]
4.4. Automatic Extraction of EEG Features by MRCNN
The results of MRCNN’s extraction of effective EEG features are shown here. EEG data in W stage and N3 stage from CinC2018 dataset were fed into the trained network model, respectively, and then the output of the first convolution layer of the model was visualized. It contains 64 convolution kernels, corresponding to the output of 64 channels, and the final results are shown in Figure 4. It can be seen from the figure that for the same input signal, different convolution kernels have different outputs. In addition, by comparing the convolution layer output results of W stage and N3 stage, it can be found that there are significant differences in the output of the same convolution kernel of different input signals, which fully demonstrates that the convolutional neural network can automatically extract the features of EEG signals.
[figures omitted; refer to PDF]
5. Conclusions
In this paper, a new sleep staging method based on multiscale residual network was proposed. It can automatically extract useful information from original single-channel EEG signals and classify sleep stages. By the cross-validation of datasets, the system performance can be maintained for the data of invisible subjects in the case of large data volume. Compared with other deep learning methods, our method only uses a single-channel EEG, and it does not require complex data preprocessing and specialized feature extraction processes to achieve better system performance, which provides the possibility for the clinical application of automated sleep staging. In addition, the multiscale residual network could be further deepened when the computing capacity was enough, and then, a larger amount of data could be used for training, so that the model with better robustness and system performance can be obtained theoretically.
Acknowledgments
The research was supported by the Science and Technology Program of Guangzhou (2019050001), Program for Chang Jiang Scholars and Innovative Research Teams in Universities (no. IRT_17R40), Guangdong Provincial Key Laboratory of Optical Information Materials and Technology (2017B030301007), Guangzhou Key Laboratory of Electronic Paper Displays Materials and Devices (201705030007), and MOE International Laboratory for Optical Information Technologies and the 111 Project. “A multiscale residual convolutional neural network for sleep staging based on single channel electroencephalography signals” as the preliminary study has been presented as a preprint according to the following link: https://www.researchsquare.com/article/rs-554671/v1.
[1] L. Xie, H. Kang, Q. Xu, M. J. Chen, Y. Liao, M. Thiyagarajan, J. O’Donnell, D. J. Christensen, C. Nicholson, J. J. Iliff, T. Takano, R. Deane, M. Nedergaard, "Sleep drives metabolite clearance from the adult brain," Science, vol. 342 no. 6156, pp. 373-377, DOI: 10.1126/science.1241224, 2013.
[2] A. M. Bianchi, M. O. Martin, S. Cerutti, "Processing of signals recorded through smart devices: sleep-quality assessment," IEEE Transactions on Information Technology in Biomedicine, vol. 14 no. 3, pp. 741-747, DOI: 10.1109/TITB.2010.2049025, 2010.
[3] E. A. Wolpert, "A manual of standardized terminology, techniques and scoring system for sleep stages of human subjects," Electroencephalography & Clinical Neurophysiology, vol. 26 no. 2, pp. 644-644, 1969.
[4] R. B. Berry, R. Budhiraja, D. J. Gottlieb, D. Gozal, C. Iber, V. K. Kapur, C. L. Marcus, R. Mehra, S. Parthasarathy, S. F. Quan, S. Redline, K. P. Strohl, S. L. Davidson Ward, M. M. Tangredi, American Academy of Sleep Medicine, "Rules for scoring respiratory events in sleep: update of the 2007 AASM manual for the scoring of sleep and associated events," Journal of Clinical Sleep Medicine, vol. 8 no. 5, pp. 597-619, DOI: 10.5664/jcsm.2172, 2012.
[5] H. T. Wu, R. Talmon, Y. L. Lo, "Assess sleep stage by modern signal processing techniques," IEEE Transactions on Biomedical Engineering, vol. 62 no. 4, pp. 1159-1168, DOI: 10.1109/TBME.2014.2375292, 2015.
[6] S. Khalighi, T. Sousa, G. Pires, U. Nunes, "Automatic sleep staging: a computer assisted approach for optimal combination of features and polysomnographic channels," Expert Systems with Applications, vol. 40 no. 17, pp. 7046-7059, DOI: 10.1016/j.eswa.2013.06.023, 2013.
[7] M. Takano, A. Ueno, "Noncontact in-bed measurements of physiological and behavioral signals using an integrated fabric-sheet sensing scheme," IEEE Journal of Biomedical and Health Informatics, vol. 23 no. 2, pp. 618-630, DOI: 10.1109/JBHI.2018.2825020, 2019.
[8] J. Lomaliza, H. Park, M. Billinghurst, "Combining photoplethysmography and ballistocardiography to address voluntary head movements in heart rate monitoring," IEEE Access, vol. 8, pp. 226224-226239, DOI: 10.1109/ACCESS.2020.3045387, 2020.
[9] G. Scebba, G. Poian, W. Karlen, "Multispectral video fusion for non-contact monitoring of respiratory rate and apnea," IEEE Transactions on Biomedical Engineering, vol. 68 no. 1, pp. 350-359, DOI: 10.1109/TBME.2020.2993649, 2021.
[10] J. Liu, Y. Chen, Y. Wang, X. Chen, J. Cheng, J. Yang, "Monitoring vital signs and postures during sleep using WiFi signals," IEEE Internet of Things Journal, vol. 5 no. 3, pp. 2071-2084, DOI: 10.1109/JIOT.2018.2822818, 2018.
[11] S. Liang, C. Kuo, Y. Hu, Y. H. Pan, Y. H. Wang, "Automatic stage scoring of single-channel sleep EEG by using multiscale entropy and autoregressive models," IEEE Transactions on Instrumentation and Measurement, vol. 61 no. 6, pp. 1649-1657, DOI: 10.1109/TIM.2012.2187242, 2012.
[12] D. Jiang, Y. Lu, Y. Ma, Y. Wang, "Robust sleep stage classification with single-channel EEG signals using multimodal decomposition and HMM-based refinement," Expert Systems with Applications, vol. 121, pp. 188-203, DOI: 10.1016/j.eswa.2018.12.023, 2019.
[13] G. Zhu, Y. Li, P. Wen, "Analysis and classification of sleep stages based on difference visibility graphs from a single-channel EEG signal," IEEE journal of biomedical and health informatics, vol. 18 no. 6, pp. 1813-1821, DOI: 10.1109/JBHI.2014.2303991, 2014.
[14] C. Huang, C. Lin, L. Ko, S. Y. Liu, T. P. Su, T. P. Lin, "Knowledge-based identification of sleep stages based on two forehead electroencephalogram channels," Frontiers in Neuroscience, vol. 8,DOI: 10.3389/fnins.2014.00263, 2014.
[15] Y. Hsu, Y. Yang, J. Wang, C. Hsu, "Automatic sleep stage recurrent neural classifier using energy features of EEG signals," Neurocomputing, vol. 104 no. 1, pp. 105-114, DOI: 10.1016/j.neucom.2012.11.003, 2013.
[16] O. Tsinalis, P. M. Matthews, Y. Guo, "Automatic sleep stage scoring using time-frequency analysis and stacked sparse autoencoders," Annals of Biomedical Engineering, vol. 44 no. 5, pp. 1587-1597, 2016.
[17] R. Boostani, F. Karimzadeh, M. Nami, "A comparative review on sleep stage classification methods in patients and healthy individuals," Computer Methods and Programs in Biomedicine, vol. 140, pp. 77-91, DOI: 10.1016/j.cmpb.2016.12.004, 2017.
[18] T. Lajnef, S. Chaibi, P. Ruby, P. E. Aguera, J. B. Eichenlaub, M. Samet, A. Kachouri, K. Jerbi, "Learning machines and sleeping brains: automatic sleep stage classification using decision-tree multi-class support vector machines," Journal of Neuroscience Methods, vol. 250, pp. 94-105, DOI: 10.1016/j.jneumeth.2015.01.022, 2015.
[19] I. Fernandez-Varela, E. Hernandez-Pereira, D. Alvarez-Estevez, V. Moret-Bonillo, "Combining machine learning models for the automatic detection of EEG arousals," Neurocomputing, vol. 268 no. 11, pp. 100-108, DOI: 10.1016/j.neucom.2016.11.086, 2017.
[20] A. Sors, S. Bonnet, S. Mirekc, L. Vercueil, J. Payen, "A convolutional neural network for sleep stage scoring from raw single-channel EEG," Biomedical Signal Processing and Control, vol. 42, pp. 107-114, DOI: 10.1016/j.bspc.2017.12.001, 2018.
[21] N. Michielli, U. Acharya, U. Rajendra, F. Molinari, "Cascaded LSTM recurrent neural network for automated sleep stage classification using single-channel EEG signals," Computers in Biology and Medicine, vol. 106, pp. 71-81, DOI: 10.1016/j.compbiomed.2019.01.013, 2019.
[22] M. Sokolovsky, F. Guerrero, S. Paisarnsrisomsuk, C. Ruiz, S. A. Alvarez, "Deep learning for automated feature discovery and classification of sleep stages," IEEE/ACM transactions on computational biology and bioinformatics, vol. 17 no. 6, pp. 1835-1845, DOI: 10.1109/TCBB.2019.2912955, 2020.
[23] A. Supratak, H. Dong, C. Wu, Y. Guo, "DeepSleepNet: a model for automatic sleep stage scoring based on raw single-channel EEG," IEEE Transactions on Neural Systems and Rehabilitation Engineering, vol. 25 no. 11, pp. 1998-2008, DOI: 10.1109/TNSRE.2017.2721116, 2017.
[24] H. Dong, A. Supratak, W. Pan, C. Wu, P. M. Matthews, Y. Guo, "Mixed neural network approach for temporal sleep stage classification," IEEE Transactions on Neural Systems and Rehabilitation Engineering, vol. 26 no. 2, pp. 324-333, DOI: 10.1109/TNSRE.2017.2733220, 2018.
[25] G. Jiang, H. He, J. Yan, P. Xie, "Multiscale convolutional neural networks for fault diagnosis of wind turbine gearbox," IEEE Transactions on Industrial Electronics, vol. 66 no. 4, pp. 3196-3207, DOI: 10.1109/TIE.2018.2844805, 2019.
[26] K. He, X. Zhang, S. Ren, J. Sun, Deep Residual Learning for Image Recognition, 2016.
[27] R. Liu, F. Wang, B. Yang, S. Qin, "Multiscale kernel based residual convolutional neural network for motor fault diagnosis under nonstationary conditions," IEEE Transactions on Industrial Informatics, vol. 16 no. 6, pp. 3797-3806, DOI: 10.1109/TII.2019.2941868, 2020.
[28] W. Qu, Z. Wang, H. Hong, Z. Chi, D. D. Feng, R. Grunstein, C. Gordon, "A residual based attention model for EEG based sleep staging," IEEE Journal of Biomedical and Health Informatics, vol. 24 no. 10, pp. 2833-2843, DOI: 10.1109/JBHI.2020.2978004, 2020.
[29] Q. Zhong, H. Lei, Q. Chen, G. Zhou, "A multiscale residual convolutional neural network for sleep staging based on single channel electroencephalography signals," ,DOI: 10.21203/rs.3.rs-554671/v1, .
[30] "The sleep-EDF database [Expanded]," . http://www.physionet.org/physiobank/database/sleep-edfx/
[31] The PhysioNet Computing in Cardiology Challenge, 2018. https://www.physionet.org/content/challenge-2018/1.0.0/
[32] A. N. Olesen, P. Jennum, P. Peppard, E. Mignot, H. B. D. Sorensen, "Deep residual networks for automatic sleep stage classification of raw polysomnographic waveforms," 2018 IEEE 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), .
[33] P. Memar, F. Faradji, "A novel multi-class EEG-based sleep stage classification system," IEEE Transactions on Neural Systems and Rehabilitation Engineering, vol. 26 no. 1, pp. 84-95, DOI: 10.1109/TNSRE.2017.2776149, 2018.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Copyright © 2021 Qinghua Zhong et al. This is an open access article distributed under the Creative Commons Attribution License (the “License”), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License. https://creativecommons.org/licenses/by/4.0/
Abstract
Sleep disorder is a serious public health problem. Unobtrusive home sleep quality monitoring system can better open the way of sleep disorder-related diseases screening and health monitoring. In this work, a sleep stage classification algorithm based on multiscale residual convolutional neural network (MRCNN) was proposed to detect the characteristics of electroencephalogram (EEG) signals detected by wearable systems and classify sleep stages. EEG signals were analyzed in each epoch of every 30 seconds, and then 5-class sleep stage classification, wake (W), rapid eye movement sleep (REM), and nonrapid eye movement sleep (NREM) including N1, N2, and N3 stages was outputted. Good results (accuracy rate of 92.06% and 91.13%, Cohen’s kappa of 0.7360 and 0.7001) were achieved with 5-fold cross-validation and independent subject cross-validation, respectively, which performed on European Data Format (EDF) dataset containing 197 whole-night polysomnographic sleep recordings. Compared with several representative deep learning methods, this method can easily obtain sleep stage information from single-channel EEG signals without specialized feature extraction, which is closer to clinical application. Experiments based on CinC2018 dataset also proved that the method has a good performance on large dataset and can provide support for sleep disorder-related diseases screening and health surveillance based on automatic sleep staging.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Details




1 Guangdong Provincial Key Laboratory of Optical Information Materials and Technology & Institute of Electronic Paper Displays, South China Academy of Advanced Optoelectronics, South China Normal University, Guangzhou 510006, China; School of Physics and Telecommunication Engineering, South China Normal University, Guangzhou 510006, China
2 School of Physics and Telecommunication Engineering, South China Normal University, Guangzhou 510006, China
3 Guangdong Provincial Key Laboratory of Optical Information Materials and Technology & Institute of Electronic Paper Displays, South China Academy of Advanced Optoelectronics, South China Normal University, Guangzhou 510006, China; Shenzhen Guohua Optoelectronics Tech. Co. Ltd., Shenzhen 518110, China; Academy of Shenzhen Guohua Optoelectronics, Shenzhen 518110, China