Skip to main content

Detection of sleep disordered breathing severity using acoustic biomarker and machine learning techniques



Breathing sounds during sleep are altered and characterized by various acoustic specificities in patients with sleep disordered breathing (SDB). This study aimed to identify acoustic biomarkers indicative of the severity of SDB by analyzing the breathing sounds collected from a large number of subjects during entire overnight sleep.


The participants were patients who presented at a sleep center with snoring or cessation of breathing during sleep. They were subjected to full-night polysomnography (PSG) during which the breathing sound was recorded using a microphone. Then, audio features were extracted and a group of features differing significantly between different SDB severity groups was selected as a potential acoustic biomarker. To assess the validity of the acoustic biomarker, classification tasks were performed using several machine learning techniques. Based on the apnea–hypopnea index of the subjects, four-group classification and binary classification were performed.


Using tenfold cross validation, we achieved an accuracy of 88.3% in the four-group classification and an accuracy of 92.5% in the binary classification. Experimental evaluation demonstrated that the models trained on the proposed acoustic biomarkers can be used to estimate the severity of SDB.


Acoustic biomarkers may be useful to accurately predict the severity of SDB based on the patient’s breathing sounds during sleep, without conducting attended full-night PSG. This study implies that any device with a microphone, such as a smartphone, could be potentially utilized outside specialized facilities as a screening tool for detecting SDB.


Sleep disordered breathing (SDB) is characterized by repetitive episodes of partial or complete obstruction of the upper airway during sleep. This disease is frequently observed in patients with hypertension, cardiac arrhythmia, coronary artery disease, stroke, and type 2 diabetes. It has also been revealed as an independent risk factor for these disorders. Even though the prevalence of SDB is as high as 24% in men and 9% in women, based on the apnea–hypopnea index (AHI) criterion of ≥ 5/h, SDB remains under-recognized and underdiagnosed [1]. Although polysomnography (PSG) is a standard examination for diagnosing SDB, it is subjected to several limitations. These include high cost, insufficient capacity for examination compared to the large number of patients, and the first night effect leading to reduced sleep efficiency and quality [2,3,4,5]. Because of these limitations, assessing the treatment outcome is difficult. Home screening can be useful to address these problems, and ambulatory sleep monitoring devices are used for this purpose. However, all the devices used in practice are placed on the patient’s body and may lead to patient discomfort due to the recording equipment.

Previous studies have analyzed snoring sounds occurring during sleep and identified their relationship with AHI, as snoring is a common symptom of SDB [6,7,8,9,10,11,12]. In addition to snoring sounds, various abnormalities of the sound per se, or the sound pattern, are observed in patients with SDB [13]. Respiratory sounds during inhalation and exhalation may be noisy when the upper airway space is compromised. Breathing intervals and patterns during sleep are also disturbed. Irregular breathing sounds will be detected when partial or complete breathing cessations occur due to hypopnea or apnea. Breath holding, breathing interruptions, gasping, and choking may also be detected [14].

Considering the acoustic properties of SDB, many studies have investigated the relationships between audio features and pathological symptoms or AHI scores. Common approaches for the analysis of sleep sound are physical measurements of the sound strength such as frequency spectrum, weighted sound intensity, and root mean square value of a signal [15]. Alternatively, different approaches such as the analysis of inter-SDB event silence and identification of snoring irregularity have also been attempted. Ben-Israel et al. found that the inter-event silence, which is associated with the acoustic energy patterns of SDB events, was significantly correlated with AHI [16]. They also estimated AHI using a multivariate linear regression model and found that the estimated AHI correlated with AHI measured on PSG. Another approach is to tailor the feature extraction process using signal patterns on the temporal or spectral domain. Mesquita et al. proposed a method identifying two distinct types of snoring: non-regular and regular snoring [17]. They extracted snoring sounds from respiratory sound signals and analyzed time intervals between regular snoring sounds in short segments of overnight recordings. They found that subjects with severe SDB had a shorter time interval between regular snoring and less dispersion on the time interval features. However, many recent studies have adopted the use of event detectors for estimation of snoring events [8, 18,19,20,21,22,23]; such methods may have disadvantages when using noisy data, leading to low detection performance and a decrease in system robustness.

In our study, we aimed to extract an acoustic biomarker from breathing sounds recorded during the entire overnight sleep that would allow us to classify the severity of SDB in patients using deep neural network. The acoustic biomarker consists of several audio features that describe the acoustic characteristics of the patient’s breathing sound during sleep. We hypothesized that breathing sounds may carry important information, and thus, by analyzing such audio signals obtained from a large number of subjects, it may be possible to identify acoustic biomarkers that indicate the severity of SDB. Thus, we collected breathing sound data from full-night PSG and extracted several standard audio features, as well as audio features used in sleep sound analysis. We then used statistical analysis to identify acoustic biomarkers and verified their effectiveness using a train-test classification framework based on the SDB severity.



The patients who visited a Seoul National University Bundang Hospital (SNUBH) sleep center between October 2013 and March 2014, because of snoring or cessation of breathing during sleep, were recruited as subjects. The subjects underwent a full-night PSG (Embla® N7000, Natus neurology). We classified the subjects into four SDB severity groups according to their AHI: normal (AHI < 5), mild (5 ≤ AHI < 15), moderate (15 ≤ AHI < 30), and severe (AHI ≥ 30). Each group consisted of 30 patients, with a total of 120 patients included the study. Among the 120 subjects, there were 3 children (age < 8) and 4 adolescents (8 ≤ age < 18).

Acquisition of sleep sounds

All the subjects underwent an attended full-night PSG in the sleep laboratory of the SNUBH, South Korea. All procedures performed on human participants were in accordance with the ethical standards of the Institutional Review Boards at the SNUBH (IRB-B-1404/248-109). The breathing sound during sleep was recorded as a part of the PSG using a PSG-embedded microphone (SUPR-102, ShenZhen YIANDA Electronics Co. Ltd., Shenzhen, China) placed on the ceiling above the patient’s bed, at a distance of 1.7 m as shown in Fig. 1. Since the microphone was used for recording environmental sounds during PSG, we used breathing sounds contained within the environmental sounds as an input signal. The sampling frequency of the recordings was 8 kHz. No additional analog filters were applied for signal correction, and raw recordings were used for the study. The mean recording time was 7 h 10 m 30 s.

Fig. 1
figure 1

Acquisition of sleep sounds and PSG reports in sleep laboratory. Audio data and PSG reports were recorded from the PSG system. After acquisition, two filtering stages were adopted to eliminate unwanted noises for 120 patients


The recorded audio contained sounds from various sources of noise (e.g. PSG machine noise, duvet noise, conversation between a clinician and a patient) in addition to the target sleep sound. Our study aimed to extract the acoustic marker representing the severity of SDB; the noises not related to breathing mentioned above can create problems during the analysis of features extracted from recorded sleep sounds. Overcoming this problem required the use of several sound filtering processes to distinguish the meaningful sleep sounds from unwanted noise.

To this end, we performed a two-stage filtering process to remove various unwanted noises and purify the sleep breathing sounds. First, we filtered breathing sounds using spectral subtraction filtering method [24] given that the spectra of noises does not change the target signal and that subtractive filtering method is computationally efficient [25]. We also applied sleep stage filtering to eliminate the noises originating from conversations and the sound of duvet. During stages 2 and 3 non-REM (rapid eye movement) sleeps, that comprise most of the sleep [26], the respiration is quite stable and regular compared to stage 1 and REM sleeps. Also, muscle activity during the REM and stage 1 sleeps results into unintended noises such as the duvet noise [9]. Therefore, we focused on sleep breathing sounds occurring during stages 2 and 3 sleeps. The mean length of filtered sounds was 4 h 1 m 55 s, which is 56.19% of the mean recording time.

Extraction of audio features

Several audio features were extracted from the preprocessed signals to estimate the severity of SDB. These audio features included the mel frequency cepstral coefficients (MFCCs), spectral flux, and zero crossing rate, and represented a variety of temporal and spectral characteristics of an audio signal. Figure 2 outlines the overall feature extraction process used in this study.

Fig. 2
figure 2

Audio feature extraction framework. Audio features were extracted in every 5 s windows. Then statistical values (means and standard deviations) of features were calculated over whole sleep period

The whole-night sleep breathing sound was divided into multiple windowed signals. A window size of 5 s was adopted, and the above-mentioned audio features were then extracted from each windowed signal. This gave a sequence of 5760 values per each audio feature, assuming an 8-h-long sleep period. We calculated the mean and standard deviation of the values for all features as their representative values. Table 1 shows the list of extracted features and their descriptions according to the feature extraction framework. All features were extracted with jAudio (java-based audio feature extraction software), and statistical analyses were performed using MATLAB R2016a (MathWorks, Inc., MA, USA).

Table 1 List of extracted audio features

In addition, other audio features that were demonstrated as effective for sleep sound analysis in previous studies [15, 27, 28] were adopted for analysis. Some of these features were added to the extraction framework as detailed below.

First, the formants of sleep breathing sounds were extracted. We extracted the amplitudes of the initial three peaks in the frequency domain (F1, F2, and F3 formants) from overnight sleep breathing sounds, and calculated the maximum, minimum, mean, and standard deviation of the formants of each patient across the windows. Second, we computed a sub-band energy distribution of the sleep breathing sound, which is the energy distribution of a given signal among partitioned frequency bands. The formants and sub-band energy distribution are both useful to distinguish snoring events from non-snoring events. A Gammatone filter bank, a widely used method for auditory signal processing, was applied for sub-band analysis, as it reflects the frequency analysis of the human ear. Likewise, simple statistics of the sub-band energy distribution, such as maximum, minimum, mean, and standard deviation, were calculated and regarded as candidate features. The weighted sound intensities, which describe the loudness of a sound as perceived by a human, were computed to function as another extra audio feature. This computation used various sound pressure levels (SPLs), such as the A-weighted sound pressure level (dBA), the C-weighted sound pressure level (dBC), and the linear-weighted (unweighted) sound pressure level (dB). The A-weighting scale was modeled to reflect the frequency sensitivity of the human ear and the C-weighting scale mimicked the frequency sensitivity of the human ear within a loud noise environment.

SDB severity group discriminators and the acoustic biomarker

To identify an acoustic biomarker, we first defined a concept of SDB severity group discriminators. A one-way ANOVA test was performed on the extracted audio features to evaluate significant differences across the SDB severity groups. A Tukey honest significant difference (HSD) test was used to evaluate significant differences between each pair of SDB severity groups, and to identify group discriminators. We assumed that features with p value < 0.05 can be used to distinguish among the SDB severity groups. The SDB severity group discriminators were the features that could statistically differentiate between a specific SDB severity group and the other groups. For example, features showing a statistical significance between the normal and other groups (mild, moderate, and severe) were defined as normal group discriminators. Other severity group discriminators were also defined in the same way. Consequently, the collection of discriminators for the four groups (normal, mild, moderate, and severe SDB severity) was defined as the acoustic biomarker. That is, the acoustic marker is a set of statistically significant features that can discriminate among the SDB severity groups. Figure 3 shows relations between audio features and the acoustic biomarker. We also applied the feature selection algorithm based on SVM to acoustic biomarker and selected.

Fig. 3
figure 3

Selection of acoustic biomarker. Among statistical values of audio features, SDB-severity group discriminators were determined with Tukey-HSD tests. The union of all SDB-severity group discriminators is defined as the acoustic biomarker

Extraction of quantized transition matrix

We adopted first derivatives of audio features to reflect temporal changes of sleep sound. However, these may not directly indicate the changes in the magnitude of sound. To overcome this issue, we imported quantized transition matrix (qTM) showing a simplified distribution of transition patterns of the signal magnitude. Figure 4 shows the overall process and an example of qTM extraction. First, for simplification, the absolute magnitudes of signals are quantized into several levels. In this paper, the magnitudes were quantized into three levels, i.e., silence, low-level signal, and high-level signal.

Fig. 4
figure 4

Process of qTM extraction. First, absoluted magnitude values are quantized into three levels (silence, low-level signal, high-level signal) for simplification. Among silence periods, apnea candidate periods were determined under standards of AASM and finally signals were quantized into four levels. Temporal transitions of quantized magnitudes were derived and transition probabilities were calculated over whole sleep

Zero-magnitude (silence) signals can be caused by either apnea or quiet breathing. Discriminating the cause of zero-magnitude signal is important since apnea and quiet breathing differentially affect the severity of sleep breathing disorder. According to the American Association of Sleep Medicine (AASM), episodes of breathing cessation lasting longer than 20 s are considered as obstructive apnea. If a zero-magnitude signal continues for 20–60 s, it is classified as apnea-candidate period and classified as level 4. Otherwise, the signal remains classified as level 0 (silence).

Last, we defined the transition matrix Q representing the probability distribution with the size of m × n, which contains the component q mn representing the probability of transition from the magnitude level m to the level n at the subsequent time point. By using the components of transition matrix Q, temporal changes in the signals can be easily observed, and Q is denominated as qTM.

Evaluation of the acoustic biomarker

The effectiveness of the defined acoustic biomarker was verified using a classification scenario involving the deep neural network learning technique. A tenfold cross-validation method was adopted for the classification experiments. For every iteration, the patients were disjointly assigned to a training or test group in the ratio of 9:1. The acoustic biomarkers were extracted from every patient in the training group and used to train the model according to the subjects’ SDB severity groupings using three machine learning techniques: simple logistics, support vector machine (SVM), and deep neural network. The SDB severity group of each subject in the test group was then predicted based on their acoustic biomarkers.

In case of the SVM, polykernel was adopted with an exponent of 1.0 and a c of 1.0. The structure of the deep neural network is described in Fig. 5. The network contained two hidden layers with 50 and 25 nodes respectively, two dropout layers, and an output layer with 4 nodes for 4-class classification. Rectified linear unit (ReLU) was adopted as an activation function. All output values of hidden layers were calculated with the ReLU and used as input values of the next layer. Between hidden layers and the output layer, the dropout technique was adopted to avoid the so-called overfitting problem. Randomly selected 20% of the nodes were eliminated through dropout processes.

Fig. 5
figure 5

Structure of the deep neural network. The network contains two hidden layers with 50 and 25 nodes respectively, two dropout layers, and an output layer with 4 nodes for 4-class classification

Four-group classification tasks were conducted. The classification aimed to predict the AHI severity group of each subject in the test group. The patients were assigned the most suitable severity group based on the extracted acoustic biomarker.


Polysomnographic findings

The subjects were 80 men and 40 women, with a mean age of 50.7 (SD 15.7) years. The mean AHI was 22.4 (SD 23.8)/h, and the mean body mass index was 25.4 (SD 4.0) kg/m2. General characteristics of each group are shown in Table 2. As shown in Table 2, we conducted ANOVA for each category and found significant differences in the body mass index and the AHI between the groups (p < 0.05). As the age difference of the subject did not show a significant difference according to AHI at the significance level of 0.05, four severity levels were classified based on AHI than age, and in reality, it is general to divide into four severity levels according to AHI.

Table 2 Clinical statistics of the population for each group (N = 120)

Statistical evaluation of audio features

A total of 20 audio features frequently used in the MIR field were extracted (Table 1). Most of the features were composed of a single variable. However, four audio features had multiple variables: the MFCC, linear predictive coding, area method of moments, and area method of moments of MFCC. Forty-eight average values and 110 standard deviation values of audio features were statistically significant. In particular, the p values of 120 statistically significant values were found to be less than 0.001, which is much smaller than the reference value of 0.05, indicating that many audio features differed significantly among the AHI severity groups. However, in the case of the first derivative of the audio features, only 41 values were determined to be significant. Therefore, a small number of values among the significant features was indicative of the change with time.

The F1, F2, and F3 formants were extracted from each analysis window, and several statistical measures were calculated across the windows: the maximum value, minimum value, mean, and standard deviation. Among statistical values, only standard deviation of F1 was determined as significant; however, according to the results of Tukey HSD test, the standard deviation of F1 was not significant for every pair of severity groups.

For sub-band energy distribution, we subdivided the frequency band ranging from 0 to 4000 Hz into eight sub-bands of 500 Hz bandwidth (named sb1, sb1,…,sb8). We derived a Gammatonegram with Gammatone filter bank and calculated simple energy distribution statistics for each sub-band: maximum value, minimum value, mean, and standard deviation, as in the formant analysis. Although many features were revealed as significant by the ANOVA test, most of these were not significant according to Tukey HSD test. Only a few features (standard deviations of sb1, sb2, sb3, sb4, sb5, sb6, sb7, sb8, and minimum values of sb6, sb8) in sub-band analysis could be used as normal discriminators.

For weighted sound intensity, we extracted dBA, dBC, dB and their peak pressure values for every window. The statistical features presented in the two examples above were calculated. We found that the results were similar to those of sub-band analysis. Only standard deviations could be used as discriminators since only these features were revealed as significant by the Tukey-HSD test.

SDB severity group discriminators

A total of 132 features were identified as group discriminators (62 normal discriminators, 7 moderate discriminators, 63 severe discriminators), as listed in Table 3. No feature was found to discriminate the mild group from the other groups. Final discriminators comprised of general audio features and weighted sound intensity features. In addition, many first derivatives were also shown to be effective, implying the importance of temporal changes in breathing sounds. A set of combined SDB severity group discriminators was regarded as the final acoustic biomarker, which contained 98 audio features collected by taking the ensemble of the discriminators from each group.

Table 3 List of SDB severity group discriminators

To visually inspect the effectiveness of the discriminators, we applied t-SNE, a widely-used algorithm for dimension reduction and visualization [29], and projected the subjects onto a two-dimensional plane. Figure 6 illustrates the distribution of 120 subjects when using whole audio features (Fig. 6a) or only the discriminators (Fig. 6b). Dashed circles represent the confidence ellipse, the region that contains 76% of the samples that can be drawn from the underlying Gaussian distribution. As clearly shown in the figure, the overlaps between the groups become much smaller when using the discriminators as compared to whole features. This group separability has a direct effect on the classification performance, as described in the following section.

Fig. 6
figure 6

Distribution of subject groups. Using the t-SNE algorithm, distributions of 120 subjects using a whole audio features and b discriminators (acoustic biomarker) respectively. When using the acoustic biomarker, group separability was increased and it has a direct effect on the classification performance

Classification experiments using the acoustic biomarker

In order to investigate the effect of each technology element presented in this paper, the baseline was obtained by inputting the statistical values of all extracted audio features and performing four-group classification using the simple logistics classifier. After that, we investigated the change in performance by adding the selected acoustic biomarker and extracted qTM, and compared the results obtained through various classifiers. Finally, the performance of binary classification was compared under various thresholds, and the applicability to screening test and mobile devices examined.

Classification results using all statistical values of audio features

Figure 7 shows classification performance when the statistical values of all audio features extracted from the sleep sound are used. Based on the four-group classification, the classification accuracy was of about 65.8%, and the classification accuracy of the normal and severe groups was particularly high. In the case of Mild and Moderate group, we observed that classification accuracy was relatively low. Confusion matrix showed that classification error between the two groups was large.

Fig. 7
figure 7

Performance of classification using all audio features (baseline). Specificity, sensitivity and area under ROC curve are depicted when all audio features are adopted as input features. A confusion matrix of the 4-group classification is also presented

Effect of the window size

We adopted four window sizes (2.5, 5, 7.5 and 10 s) when extracting audio features from sleep sounds. Figure 8 shows classification performances using all audio features extracted from segmented sleep sounds according to the window sizes. The graph is concave and the window size of 5 s shows the highest performance of all. Since statistics of audio features are used for learning and the periods of sleep events are multiple seconds in general, fine segmentation is needless and relatively large window size is adequate in our study.

Fig. 8
figure 8

Comparison of performance of using audio features extracted under various window sizes

Effect of the qTM

Classification performance shown in Fig. 9 was obtained when the features representing the temporal change in the sleep sound were added by learning using both the statistical values of all audio features and the qTM extracted from the sleep sound. In general, the accuracy of classification improved in all severity groups. Especially, classification performance of mild group was increased to the level of normal and severe groups. Notably, the performance of the ROC area rose to over 90%.

Fig. 9
figure 9

Performance of classification when components of qTM are added as input features

Effect of the acoustic biomarker

The performance shown in Fig. 10 was obtained when only the feature that distinguishes between the severity groups was selected by acoustic biomarkers and used for learning. In addition, when acoustic biomarkers and qTMs were used for learning, the resulting performance was as the one shown in Fig. 11. The acoustic biomarker alone showed a 66.7% increase in performance, even though the size of input features was reduced to one-third. However, the classification performance of mild and moderate groups was still low, but the addition of the qTM increased the classification performance to 88.3% due to synergy effect. Finally, the use of both acoustic biomarkers and qTMs shows the best performance, and the number of features employed is significantly reduced.

Fig. 10
figure 10

Performance of classification when the acoustic biomarkers are adopted instead of all audio features

Fig. 11
figure 11

Performance of classification when both the acoustic biomarker and the qTM are adopted

Comparison of the results obtained using various classifiers

In this section, we compare the results of learning using the three classifiers mentioned in the previous chapter to establish which classifier is suitable. Learning was performed using simple logistics, SVM, and deep neural networks with two hidden layers. It was assumed that both the acoustic biomarker and the qTM were used as input features, and the results of the four-group classification using the above-mentioned three classifiers were compared. The results of the comparison are shown in Fig. 12. Simple logistics showed the best performance and the deep neural network showed low performance due to the small dataset and input features.

Fig. 12
figure 12

Comparison of performance of using various classifiers

Effect of SVM-based feature selection on acoustic biomarker

We additionally adopted SVM-based feature selection and chose top-50, 75 ranked features. Figure 13 shows performances when all final features, top-75 features and top-50 features are used for classification respectively. We used SVMAttributeEval function in the Weka software [29] for the feature selection. We can find out that there is no difference between performances. However, when we adopt chosen features as input, it may be preferable in terms of the reduction of computational power and time required for the algorithm.

Fig. 13
figure 13

Comparison of performance of using different feature sets chosen with SVM-based feature selection algorithm

Binary classification using multiple thresholds

Unlike the four-group classification, the binary classification divides the patients into two groups according to whether the patient’s AHI value is lower than a certain threshold or not. The results of binary classifications are shown in Fig. 14. The thresholds were set at 5, 15, and 30, which are the boundaries of the AHI range in each severity group. The acoustic biomarker and qTM were used as input features, and the experiments were performed using simple logistics, which was shown to be the best classifier. The accuracy of 92.5% was achieved under the thresholds 15 and 30. In addition, the sensitivity, specificity, and area under ROC were near 90% under all conditions.

Fig. 14
figure 14

Performance of binary classifications under various thresholds


The extracted audio features used in this study successfully represented signal patterns in the temporal and spectral domains. Respiratory events during sleep influence the power of the signal of sleep breathing sounds, and many of the extracted audio features illustrate this phenomenon. Certain features also contain several statistical measures of audio signal samples from both domains. Therefore, the audio features can easily represent the great majority of useful traits in the original audio signals, without requiring the use of the whole signal. Moreover, the first derivatives of each feature were also calculated to observe temporal changes in the audio signal. Along with other features, means and standard deviations of the derivatives were calculated, and used as candidate features for the acoustic biomarker. However, not all the extracted features were useful for generating acoustic biomarkers for each subject, as some of extracted features showed no significant differences among the different patient severity groups. It was therefore necessary to verify whether each feature was useful for classifying the severity of SDB.

As shown by the classification task result, use of only the selected acoustic biomarker appeared more advantageous in terms of performance and computation speed than the use of statistical values of all audio features. In addition, when using both the acoustic biomarker and the qTM, the performance was greatly improved because, as the first derivative of the audio feature was generally not statistically significant, the acoustic biomarker did not contain enough elements to represent the temporal change in the sleep sound. Regarding the qTM, this feature indicates the temporal change of the signal, and a significant increase in performance demonstrates that the temporal variation of the signal plays an important role in the prediction of SDB severity.

Although the ANOVA test indicated a significance of various features, it remained difficult to make conclusions regarding the discriminatory power of candidate features. The final determination of acoustic biomarker components was therefore performed based on the results of Tukey HSD test. After determining the acoustic biomarker, a simple classification task was conducted to validate its effectiveness. The means of the features constitute a relatively small portion of all discriminators, suggesting that standard deviations of the features may, in general, be more effective as discriminators. Moreover, since breathing sounds during sleep have time-sequential patterns, many means and standard deviations of the first derivatives could also be used as discriminators.

In our study, both specificity and sensitivity were as high as the accuracy, demonstrating the reliability of the test. While many previous studies conducted binary classifications, it is unusual to evaluate performances using four-group classifications. Another reason for performing four-group classification was to precisely examine the performance. By using both the acoustic biomarker and qTM, it was possible to predict the severity of SDB in subjects with a rather high degree of sensitivity and specificity. This suggests that our model may be adequate for obstructive sleep apnea screening.

Nevertheless, it appears that the deep neural network has a slightly lower classification performance than other classifiers. This is because its learning structure does not satisfy the number of patients or the number of features that can sufficiently learn it, even if it is a simple form with a small number of layers and nodes. In particular, unlike in other studies, this study did not consider snoring events as units of learning. Instead, the features were extracted from the entire sleep surface of the patient and learnt. However, the number of samples was not large enough which is somewhat unsuitable for deep neural network learning. However, if enough sleep sounds are used for learning, the deep neural network will perform better than the results described here.

It should be noted that we conducted a patient-wise classification, while previous studies [4, 30,31,32,33] focused on event-wise classifications. Generally, event-wise classification tasks use more training data than patient-wise classification tasks, and thus, derive better results. However, our patient-wise classification still had a higher performance, which proves that the acoustic biomarker effectively represents the characteristics of sleep breathing sounds in patients with SDB.


We demonstrated that generation of an acoustic biomarker representing the SDB severity score may be used to accurately predict the SDB severity of a patient by analyzing the overnight recording of sleep breathing sound, without conducting the attended full-night PSG.

Simple classification tasks, using extracted features compiled into an acoustic biomarker, were highly performant. This method may be useful for the actual diagnosis because of its high prediction performance, i.e., specificity and accuracy. Therefore, in future, subjects might be able to perform a screening test by themselves with simple devices such as smartphones and smartwatches, using the acoustic biomarker and several machine learning algorithms. Regarding the detection of sleep stages, widely used actigraphy-based sleep stage detection embedded in smartwatches can be adopted in home environment.

Furthermore, our framework can be employed using any recording system since we analyzed sleep sounds recorded by a monitoring video with a low-quality microphone. In addition, the research framework considered in this study may be easily applied to other analogous studies. Improving the quality of recorded sleep breathing sounds may increase the classification performance using the acoustic biomarker.

A more precise prediction model may be derived with more extensive PSG data, since deep neural network is more suitable for learning large scale-data. Thus, additional PSG data should be acquired in the future to increase the diagnostic accuracy of the model.

Additional standard or preprocessed audio features may represent the sleep sound signals, and it would be beneficial to add them as components of the acoustic biomarker. Although temporal change patterns in the audio features within each window are also important in diagnosing SDB, they were not analyzed in this study. Representing these patterns would require identification of additional features. Future work should therefore focus on discovering new features and adding them to the acoustic biomarker to improve the accuracy of classification.



sleep disordered breathing




apnea–hypopnea index


Seoul National University Bundang Hospital


rapid eye movement


mel frequency cepstrum coefficient


spectral centroid


spectral rolloff point


spectral flux


spectral variability


root mean square


fraction of low energy windows


zero crossings


strongest beat


beat sum


strength of strongest beat




linear predictive coding


method of moments


relative difference function


area methods of moments


log of CQ transform


sound pressure level


A-weighted sound pressure level


C-weighted sound pressure level


linear weighted sound pressure level


honest significant difference


quantized transition matrix


American Association of Sleep Medicine


support vector machine


rectified linear unit


standard deviation


  1. Young T, Peppard PE, Gottlieb DJ. Epidemiology of obstructive sleep apnea: a population health perspective. Am J Respir Crit Care Med. 2002;165(9):1217–39.

    Article  Google Scholar 

  2. Behar J, Roebuck A, Shahid M, Daly J, Hallack A, Palmius N, et al. SleepAp: an automated obstructive sleep apnoea screening application for smartphones. In: Proc. computing in cardiology conference. 2013. p. 257–60.

  3. Roebuck A, Monasterio V, Gederi E, Osipov M, Behar J, Malhotra A, et al. A review of signals used in sleep analysis. Physiol Meas. 2013;35(1):R1–57.

    Article  Google Scholar 

  4. Abeyratne U, Wakwella A, Hukins C. Pitch jump probability measures for the analysis of snoring sounds in apnea. Physiol Meas. 2005;26(5):779–98.

    Article  Google Scholar 

  5. McCombe A, Kwok V, Hawke W. An acoustic screening test for obstructive sleep apnoea. Clin Otolaryngol. 1995;20(4):348–51.

    Article  Google Scholar 

  6. Emoto T, Abeyratne U, Akutagawa M, Nagashino H, Kinouchi Y. Feature extraction for snore sound via neural network processing. In: Proc. annual international conference of the IEEE engineering in Medicine and Biology Society. 2007. p. 5477–80.

  7. Fiz J, Jané R, Solà-Soler J, Abad J, García M, Morera J. Continuous analysis and monitoring of snores and their relationship to the apnea–hypopnea index. Laryngoscope. 2010;120(4):854–62.

    Article  Google Scholar 

  8. Nakano H, Hirayama K, Sadamitsu Y, Toshimitsu A, Fujita H, Shin S, et al. Monitoring sound to quantify snoring and sleep apnea severity using a smartphone: proof of concept. J Clin Sleep Med. 2014;10(1):73–8.

    Google Scholar 

  9. Fiz J, Abad J, Jané R, Riera M, Mañanas M, Caminal P, et al. Acoustic analysis of snoring sound in patients with simple snoring and obstructive sleep apnoea. Eur Respir J. 1996;9(11):2365–70.

    Article  Google Scholar 

  10. Michael H, Andreas S, Thomas B, Beatrice H, Werner H, Holger K. Analysed snoring sounds correlate to obstructive sleep disordered breathing. Eur Arch Otorhinolaryngol. 2007;265(1):105–13.

    Article  Google Scholar 

  11. Snider B, Kain A. Automatic classification of breathing sounds during sleep. In: 2013 IEEE international conference on proc. acoustics, speech and signal processing. 2013.

  12. Dafna E, Tarasiuk A, Zigel Y. OSA severity assessment based on sleep breathing analysis using ambient microphone. In: Proc. annual international conference of the IEEE engineering in Medicine and Biology Society. 2013. p. 2044–7.

  13. Azarbarzin A, Moussavi Z. Snoring sounds variability as a signature of obstructive sleep apnea. Med Eng Phys. 2013;35(4):479–85.

    Article  Google Scholar 

  14. Epstein L, Kristo D, Strollo P, Friedman N, Malhotra A, Patil S, et al. Clinical guideline for the evaluation, management and long-term care of obstructive sleep apnea in adults. J Clin Sleep Med. 2009;5(3):263–76.

    Google Scholar 

  15. Pevernagie D, Aarts R, De Meyer M. The acoustics of snoring. Sleep Med Rev. 2010;14(2):131–44.

    Article  Google Scholar 

  16. Ben-Israel N, Tarasiuk A, Zigel Y. Obstructive apnea hypopnea index estimation by analysis of nocturnal snoring signals in adults. Sleep. 2012;35(9):1299–305.

    Article  Google Scholar 

  17. Mesquita J, Sola-Soler J, Fiz J, Morera J, Jane R. All night analysis of time interval between snores in subjects with sleep apnea hypopnea syndrome. Med Biol Eng Comput. 2012;50(4):373–81.

    Article  Google Scholar 

  18. Ankishan H, Ari F. Snore-related sound classification based on time-domain features by using ANFIS model. In: Proc. innovations in intelligent systems and applications, international symposium on. 2011. p. 441–4.

  19. Ankışhan H, Yılmaz D. Comparison of SVM and ANFIS for snore related sounds classification by using the largest Lyapunov exponent and entropy. Comput Math Methods Med. 2013;2013(3):1–13.

    MathSciNet  MATH  Google Scholar 

  20. Azarbarzin A, Moussavi Z. Automatic and unsupervised snore sound extraction from respiratory sound signals. IEEE Trans Biomed Eng. 2010;58(5):1156–62.

    Article  Google Scholar 

  21. Duckitt W, Tuomi S, Niesler T. Automatic detection, segmentation and assessment of snoring from ambient acoustic data. Physiol Meas. 2006;27(10):1047–56.

    Article  Google Scholar 

  22. Sola-Soler J, Jané R, Fiz J, Morera J. Spectral envelope analysis in snoring signals from simple snorers and patients with obstructive sleep apnea. In: Proc. annual international conference of the IEEE engineering in Medicine and Biology Society. 2003. p. 2527–30.

  23. Yadollahi A, Moussavi Z. Acoustic obstructive sleep apnea detection. In: Proc. annual international conference of the IEEE engineering in Medicine and Biology Society. 2009. p. 7110–3.

  24. Boll SF. Suppresion of acoustic noise in speech using spectral subtraction. IEEE Trans Acoust Speech Signal Process. 1979;27(2):113–20.

    Article  Google Scholar 

  25. Kim J, Kim T, Lee D, Kim JH, Lee K. Exploiting temporal and nonstationary features in breathing sound analysis for multiple obstructive sleep apnea severity classification. Biomed Eng Online. 2017;16:6.

    Article  Google Scholar 

  26. Stevens D. Sleep medicine secrets. London: Elsevier Health Sciences; 2004.

    Google Scholar 

  27. Yadollahi A, Moussavi Z. Formant analysis of breath and snore sounds. In: Proc. annual international conference of the IEEE engineering in Medicine and Biology Society. 2009. p. 563–66.

  28. Cavusoglu M, Kamasak M, Erogul O, Ciloglu T, Serinagaoglu Y, Akcam T. An efficient method for snore/nonsnore classification of sleep sounds. Physiol Meas. 2007;28(8):841–53.

    Article  Google Scholar 

  29. Hall M, Frank E, Holmes G, Pfahringer B, Reutemann P, Witten IH. The WEKA data mining software: an update. ACM SIGKDD Explor Newsl. 2009;11(1):10–8.

    Article  Google Scholar 

  30. Ng AK, Koh TS, Baey E, Puvanendran K. Diagnosis of obstructive sleep apnea using formant features of snore signals, vol. 14, no. 2. Berlin: Springer; 2007. p. 967–70.

  31. Ng A, Koh T, Abeyratne U, Puvanendran K. Investigation of obstructive sleep apnea using nonlinear mode interactions in nonstationary snore signals. Ann Biomed Eng. 2009;37:1796–806.

    Article  Google Scholar 

  32. Nakano H, Hayashi M, Ohshima E, Nishikata N, Shinohara T. Validation of a new system of tracheal sound analysis for the diagnosis of sleep apnea–hypopnea syndrome. Sleep. 2004;27(5):951–7.

    Article  Google Scholar 

  33. Willemen T, Van Deun D, Verhaert V, Vandekerckhove M, Exadaktylos V, Verbraecken J, Van Huffel S, Haex B, Vander Sloten J. An evaluation of cardio-respiratory and movement features with respect to sleep stage classification. IEEE J Biomed Health Inform. 2013;18(2):661–9.

    Article  Google Scholar 

Download references

Authors’ contributions

TK, JK and KL participated in the design and coordination of the study. TK conceived the study, implemented the algorithm, and drafted the manuscript. TK, JK and KL critically reviewed the manuscript. All authors read and approved the final manuscript.

Competing interests

The authors declare that they have no competing interests.

Availability of data and materials

Please contact author for data requests.

Consent for publication

Informed consent was obtained from all individual participants included in the study. In case of children participants, the written informed consent was obtained from the parents or guardians on behalf of them.

Ethics approval and consent to participate

The protocol number (IRB-B-1404/248-109) was approved by the institutional review boards (IRBs) at SNUBH (Seoul National University Bundang Hospital).


The work was partly supported by the SNUBH Grant #06-2014-157 and the Bio and Medical Technology Development Program of the National Research Foundation (NRF) funded by the Korean government, Ministry of Science, ICT & Future Planning (MSIP) (NRF-2015M3A9D7066972, NRF-2015M3A9D7066980).

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Author information

Authors and Affiliations


Corresponding authors

Correspondence to Jeong-Whun Kim or Kyogu Lee.

Rights and permissions

Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (, which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver ( applies to the data made available in this article, unless otherwise stated.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Kim, T., Kim, JW. & Lee, K. Detection of sleep disordered breathing severity using acoustic biomarker and machine learning techniques. BioMed Eng OnLine 17, 16 (2018).

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: