Estimating Respiratory Rate From Breath Audio Obtained Through Wearable Microphones

07/28/2021 ∙ by Agni Kumar, et al. ∙ Apple Inc. 0

Respiratory rate (RR) is a clinical metric used to assess overall health and physical fitness. An individual's RR can change from their baseline due to chronic illness symptoms (e.g., asthma, congestive heart failure), acute illness (e.g., breathlessness due to infection), and over the course of the day due to physical exhaustion during heightened exertion. Remote estimation of RR can offer a cost-effective method to track disease progression and cardio-respiratory fitness over time. This work investigates a model-driven approach to estimate RR from short audio segments obtained after physical exertion in healthy adults. Data was collected from 21 individuals using microphone-enabled, near-field headphones before, during, and after strenuous exercise. RR was manually annotated by counting perceived inhalations and exhalations. A multi-task Long-Short Term Memory (LSTM) network with convolutional layers was implemented to process mel-filterbank energies, estimate RR in varying background noise conditions, and predict heavy breathing, indicated by an RR of more than 25 breaths per minute. The multi-task model performs both classification and regression tasks and leverages a mixture of loss functions. It was observed that RR can be estimated with a concordance correlation coefficient (CCC) of 0.76 and a mean squared error (MSE) of 0.2, demonstrating that audio can be a viable signal for approximating RR.



There are no comments yet.


page 1

page 5

This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.

I Introduction

Breathlessness, or dyspnea, is a common symptom in many acute and chronic clinical conditions. Acute breathlessness often occurs during an asthmatic episode [1] or heart attack [18], while chronic breathlessness is frequently a symptom of low cardiovascular fitness and obesity [2, 10], chronic obstructive pulmonary disease (COPD) [6], and congestive heart failure (CHF) [8, 23]. Breathlessness on exertion is also a strong independent predictor of mortality [22] and is a commonly-used clinical metric for assessing and monitoring disease progression. The primary classifications of heart failure (NYHA Class I-IV) are defined in terms of breathlessness, either occurring at rest or during normal levels of physical activity [21].

Breathlessness scores as quantified by the Borg Dyspnea Scale, used to assess clinical severity of diseases such as peripheral artery disease (PAD) and other respiratory disorders, are subjective patient-reported measures

[4]. Individuals with such conditions are typically required to interface with a healthcare provider in order for their symptoms to be recognized. Creating an objective method for breathlessness detection would lower the burden in identifying this symptom and may even alert healthcare providers to patients’ underlying medical conditions long before disease progression would have been observed in a clinical setting.

In this paper, we take the first step towards developing a breathlessness measurement tool by estimating respiratory rate (RR) on exertion in a healthy population using audio from wearable headphones. Given this focus, such a capability also offers a cost-effective method to track cardio-respiratory fitness over time. While sensors such as thermistors, respiratory gauge transducers, and acoustic sensors provide the most accurate estimation of a person’s breathing patterns, they are intrusive and may not be comfortable for everyday use. In contrast, wearable headphones are relatively economical, accessible, comfortable, and aesthetically acceptable.

Previous work on breath sounds detection from audio has focused on the detection and categorization of particular breath sounds using breath characteristics to distinguish between healthy and abnormal breath sounds [14, 5], as well as RR estimation from both contact-based sensors to obtain tracheal sounds [26, 25] and non-contact-based sensors, like smartphones, to acquire nasal breath recordings [24]. The goal of this study is to validate whether versatile non-contact sensors, such as wearable near-field microphones, can provide incoming audio data sufficient to distinguish between normal and heavy breathing and also estimate RR by the sensing of breath sound patterns. Though we focus on discerning RR in the context of fitness activity and not clinical breathlessness detection specifically, findings could in turn be used for medical applications.

Ii Data

We chose to focus on the exercise context to obtain breath samples of varying intensities. Data was collected from 21 healthy individuals from both indoor and outdoor environments. Participants spanned the ages of 22 to 60 and were split fairly evenly between genders. Several participants provided six pulse rate measurements per audio sample submitted, each of which spanned a six-minute active period.

All data was recorded using microphone-enabled, near-range headphones, specifically Apple’s AirPods. These particular wearables were selected because they are owned by millions and utilized in a wide array of contexts, from speaking on the phone to listening to music during exercise. Unlike prior studies in which the acoustic environment and background noise were artificially altered [24], our study did not place artificial acoustic constraints on the data collection setup.

Each trial of the data collection asked participants to record four one-minute audio clips before, during, immediately after, and while cooling down following completion of a nine-minute workout session, in which six minutes involved physical exercise. Workout types were selected to induce heavy breathing, with the goal of doubling participants’ resting heart rates at the peak of physical exertion.

  • Before workout (minute 1)

  • During the workout, towards the end (minute 5)

  • Immediately after the workout (minute 6)

  • While cooling down (minute 8)

Participants were also asked to note pulse rates in beats per minute (BPM) as obtained from an Apple Watch at six points in the exercise session: before recording an audio clip of type , before recording an audio clip of type (minute 5), and before and after recording an audio clip of each of type (minutes 6 and 7) and type (minutes 8 and 9). The data collection protocol is summarized in Figure 1.

Fig. 1: One trial of data collection included four recording types corresponding to distinct workout stages.

All audio sessions were segmented into randomly-selected lengths between 4-7 seconds, to increase the probability that a segment would contain at least one breath cycle. 3003 segments were manually annotated for RR. The annotation process involved counting the number of inhale-exhale cycles that could be heard in each audio sample and dividing the breath count by the clip duration in minutes to achieve respiratory rate measures in breaths per minute.

299 segments of type , 457 of type , 762 of type , and 285 of type , along with 1200 randomly chosen segments, were annotated and served as ground truth values. The perceptual annotation was motivated by the fact that perceived audio respiration events should be represented well by the acoustic features, which in turn would facilitate the model training. In cases of high background noise, audio samples could be rendered imperceptible to the human ear and result in distorted acoustic features.

In the spectrogram in Figure 2(a), indicative of normal breathing, two prominent inhalation signals are observed with no harmonic structure. The intense exercise sample in Figure 2(b), associated with heavy breathing, shows more frequent energy bursts but also lacks harmonic structure, indicating both a higher RR and greater background noise. These observations support the use of temporal spectral representations in distinguishing between normal and heavy breathing, which prompted using temporal convolution and recurrent layers in the model described in Section IV.

(a) normal breathing
(b) heavy breathing
Fig. 2: Breath audio spectrograms

In total, over 3000 annotated audio signals were analyzed. We are not aware of any audio data resources that have perceptually graded RR labels that are bigger in volume than that collected for this work. Analyses by breathing intensity and gender are thus provided in Section V.

Iii Analysis

We observe several interesting relations in the data. In Figure 3, we note a right-tailed distribution of RR, likely due to some participants having not exerted themselves to the level requested, perhaps because it was not achievable during a six-minute period with particular workout types. In a future data collection, efforts will be made to either upsample data yielding higher respiratory rates or extend the workout duration to increase the chances of observing heavy breathing.

Fig. 3: Distribution of annotated respiratory rates
Fig. 4: Respiratory and pulse rates over workout stages

Participants’ exercises varied in intensity and included running, biking, HIIT (high-intensity interval training), calisthenics, fast walking, and walking up an incline. 60% of audio samples were recorded indoors while 40% were taken outdoors. It was observed that most of the data that was not entirely noisy, i.e., breath-only or a mix of breath and noise, comprised samples from during and immediately after workouts, when breathing was heaviest. Moreover, participants’ noise contributions differed significantly, likely due to differing workout environments and home exercise equipment. Data from users who completed repeated exercises in close time proximity tended to exhibit higher-than-normal resting RR and pulse rates, since breaks taken in between were sometimes not long enough to reestablish a completely non-exerted state.

In Figure 4, we visualize how participants’ respiratory and pulse rates vary over time across workout stages, represented by individual blue boxes. As expected, the median pulse rate is lower for stages and than for and , and that the pulse rate associated with is higher than that associated with . RR varies in a similar manner, and the high RR values achieved align with those collected in similar studies [19]. Each overlaid white circle is associated with a single audio segment. The numbers of white circles differed between minutes because participants could choose to opt-in or opt-out of submitting audio samples at any point in time, as the study was designed to ensure high privacy.

Iv System Description

We investigated mel-filterbank energy (MFB) acoustic features to parameterize the audio streams [9] and trained an LSTM network and a time-convolutional LSTM (TC-LSTM) network using multi-task learning (MTL) [11]. Our study presents, for the first time, the use of filterbank energies processed by time-convolutional layers. We hypothesize that such a setup is effective as it captures both signal envelope-level information as well as longer time contexts to better assess respiration information. The learning network, an end-to-end model, is not standard in that it simultaneously encompasses both regression and classification tasks. Its architecture was motivated by the nature of the specific data collected, described in Section II.

Iv-a Acoustic Features

We examined standard 40-dimensional MFB features with upper and lower cut-off frequencies of 0 and 7500 Hz. We investigated MFBs with both more than and less than 40 filteranks, and observed that 40 filterbanks offered the best performance on the held-out validation set.

This frequency range was chosen as exhalation signals contain wideband information, with [20] noting that frequencies between 20 to 6000 Hz can encompass useful information. From our analysis, we observed that nasal and oral exhalation had very different spectral characteristics, with the former having low frequency band-pass characteristics and the latter having more energy in high-pass regions. (Similarly, when comparing nasal and oral snores, [17] found nasal snores to exhibit frequencies around 500 Hz, while oral snores were associated with wide-band spectra ranging from 500 Hz to 1000 Hz.) Given that this study considers data collected from subjects after intensive workouts that may result in more oral breathing due to exhaustion, we hypothesized that energies in high frequency spectrums could be quite useful. Moreover, as one of the multi-task learned model’s objective functions relates to noise detection, the selected MFB feature frequency range is fitting.

Fig. 5: Multi-task learning architecture with LSTM

Iv-B Acoustic Modeling

We investigated single layer LSTMs with 16, 32, and 64 neurons in the recurrent and embedding layers. The network had three outputs:

  • A 2-dimensional output corresponding to RR and respiration count (RC) estimates

  • A 3-dimensional output reflecting the classes of no breathing, normal breathing (RR between 5 to 30 breaths per minute), and heavy breathing (RR greater than 30 breaths per minute)

  • A 2-dimensional output corresponding to noise or noiseless cases

The model, depicted in Figure 5, was trained with multiple objective functions as a multi-task learning (MTL) network, where the tasks were RR estimation, heavy breathing detection, and noise detection, represented by the three outputs mentioned above. As the audio segments used to train the model had varying lengths, meaning that two segments of differing durations could have the same RR but different RCs, RC was included as a target in the multi-task objective function to ensure the model’s ability to generalize across temporal durations and also learn correlations between RC and RR.

The individual losses from each task are given below, where concordance correlation coefficient (CCC) loss is used on the RR and RC outputs and weighted cross-entropy (CE) loss is used on the breath and noise classification tasks. The interpretation of CCC

used here is similar to that of Pearson’s product moment correlation (

PPMC), where CCC ranges between to and a higher value indicates a better fit [3]. CCC was chosen as a performance metric based on findings from the literature showing that it captures both aspects of the correlation and the error between target and estimated scores [13, 12].

Additionally, a focal loss term [15, 7] was used for the breath detection task, and a convex mixture of all the losses after dynamic weight averaging [16], with weighting factor , was used as the MTL loss to train the network shown in Figure 5.

The CCC for each of the RR and RC outputs is below, where and are the means, and

are the corresponding variances for the estimated and ground truth variables, and

describes the correlation between these two variables.

The model was trained with a mini-batch size of 64 using an Adam optimizer, with a learning rate of 0.01 and a momentum of 0.9. For all the model training steps, early stopping was allowed based on cross-validation error. The embedding layer between the breath classification and RR and RC estimation tasks was shared. The noise classification task had a separate embedding layer, which was a fourth of the size of the breath embedding layer.

We also investigated a TC-LSTM network, where a convolution operation with a filter size of 3 was performed across time. The number of convolution filters was same as the number of input feature dimensions. The feature maps from the convolutional layers were fed as inputs to the succeeding LSTM network.

During training, the CCC

for RR estimation on a held-out validation set was used to select the best epoch, and the model from that epoch was used to obtain the performance on the held-out evaluation set. Results on the evaluation set are provided in terms of both

CCC and mean squared error (MSE).

V Results

We investigated model size by exploring LSTM models with 16, 32, and 64 neurons in the recurrent and embedding layers. The evaluation metrics used were

scores for the classification task and CCC for the regression task.

Table I shows the variation in performance by model size, where the model with 32 neurons in the LSTM layer, 32 neurons in the breath embedding layer, and 8 neurons in the noise embedding layer was found to perform the best on both the validation and evaluation sets. The model with 64 neurons shows some degree of overfitting, where the performance gap between the validation and evaluation set was larger than that of the model with 32 neurons. This could be a consequence of the data volume limitation, as larger datasets may enable using models with more parameters.

0.79 0.85 57.55 88.82
0.88 0.88 64.76 89.61
0.85 0.85 56.53 93.65
0.59 0.65 49.20 66.91
0.73 0.73 66.33 76.81
0.62 0.62 52.82 65.04
TABLE I: CCC, PPMCC () for RR estimations, and scores for breath and noise classification tasks for the validation and evaluation sets, from the MTL-LSTM network

Next, we explored a time-convolution layer in the acoustic model and modulation features as an alternative feature representation, and whether such selections could facilitate better detection of respiration from audio. Table II presents the results from these experiments, for which there were 32 neurons in the LSTM and breath embedding layers. RR was estimated with a CCC as high as 0.76 and a detection accuracy of breathing at 66%.

LSTM 0.73 0.73 0.32 66.33
TC-LSTM 0.76 0.78 0.31 63.00
TABLE II: CCC, PPMCC (), and MSE scores for RR estimation, and scores for breath classification tasks for the evaluation set
Fig. 6: Performance (MSE) comparisons for various RR ranges for female and male participants in the test set

We then investigated how RR estimation varied for different RR ranges and how data augmentation could help improve the performance of the RR estimation model. Recall that the dataset used in this study included audio in both indoor and outdoor conditions. The outdoor data already contained natural ambient noise such as wind and traffic sounds, so the indoor data was further augmented by pseudo-stationary noise reflective of appliance sounds at various signal-to-noise ratios (SNRs) between 20 to 40 dBs. For each indoor data file, noise was added at three different SNR levels, each of which was selected from a uniform distribution between 10 to 20 dB, 20 to 30dB and 30 to 40dB, respectively. Data augmentation was applied only on the training partition.

Model below 15 15 to 25 above 25
LSTM 0.33 0.43 0.29
TC-LSTM 0.32 0.42 0.22
LSTM (augumented data) 0.28 0.24 0.29
TC-LSTM (augmented data) 0.21 0.20 0.21
TABLE III: MSE values across the LSTM and TC-LSTM models before and after data augmentation across RR ranges

Table III shows how the performance of RR estimation in terms of MSE varied at low (less than 15 breaths per minute), medium (between 15 and 25 breaths per minute), and high (greater than 25 breaths per minute) RR rates. It is evident that data augmentation helped reduce the MSE for almost all the RR ranges. Moreover, as displayed in Figure 6 above, the TC-LSTM model performed better than the LSTM model both with and without data augmentation across female and male participants in the test set. Interestingly, estimates for female participants were almost always better than those for the male participants. It is possible that differing anatomical breathing apertures, age distributions, or ratios of indoor to outdoor samples collected between the two groups could have contributed to this observation.

Vi Conclusions

In this study, we investigated the feasibility of estimating respiratory rate from audio captured using wearable, near-field microphones. We found a time convolution LSTM network to be effective at generating respiratory rate estimates and robust against data sparsity.

The work is unique in three main ways, in that it estimates respiratory rate from a wearable microphone under natural ambient conditions both indoors and outdoors, uses a model-driven approach to estimate respiratory rate directly from filterbank energies, and introduces situational awareness through multi-task learning so that the model could discern high SNR conditions from low ones. To the best of our knowledge, no prior study has investigated data collected from natural conditions from both indoor and outdoor background conditions, used perceptually graded data, and attempted to build an end-to-end system that can consume filterbank energies to directly predict respiratory rates and make heavy breathing classifications. Hence, we hope our proposed system will be used as a baseline for future studies on end-to-end respiratory rate estimation models.

Results presented validate that RR can be estimated from audio captured using wearable microphones, enabling the detection of heavy breathing conditions and the monitoring of RR changes, a measure of cardio-respiratory fitness, over time. Data augmentation with simple acoustic distortion was demonstrated to be an effective tool to reduce error rates. The findings show promise for further development of a respiratory health tool with a larger study cohort.


We would like to thank Siddharth Khullar, Jamie Cheng, Suki Lee, Christopher Webb, Kate Neihaus, and many other colleagues for their comments, suggestions, and help at various stages of this effort.


  • [1] M. Barbaro, D. Lacedonia, G. Palladino, L. Bergantino, C. Ruggeri, D. Martinelli, and G. Carpagnano (2011) Dyspnea perception in asthma: role of airways inflammation, age and emotional status. Respiratory medicine 105 (2), pp. 195–203. Cited by: §I.
  • [2] V. Bernhardt and T. Babb (2014) Respiratory symptom perception differs in obese women with strong or mild breathlessness during constant-load exercise. Chest 145 (2), pp. 361–369. Cited by: §I.
  • [3] J. M. Bland and D. Altman (1986) Statistical methods for assessing agreement between two methods of clinical measurement. The lancet 327 (8476), pp. 307–310. Cited by: §IV-B.
  • [4] G. Borg (1982) Psychophysical bases of perceived exertion.. Medicine & Science in Sports & Exercise. Cited by: §I.
  • [5] J. Castro and P. Marti-Puig (2014)

    Real-time identification of respiratory movements through a microphone. adcaij: advances in distributed computing and artificial intelligence journal (issn: 2255-2863)

    Salamanca 3 (3). Cited by: §I.
  • [6] B. Celli, C. Cote, S. Lareau, and P. Meek (2008) Predictors of survival in copd: more than just the fev1. Respiratory medicine 102, pp. S27–S35. Cited by: §I.
  • [7] H. Gil, S. Park, Y. Park, W. Han, J. Hong, and J. Jung (2020) Balance-oriented focal loss with linear scheduling for anchor free object detection. arXiv preprint arXiv:2012.13763. Cited by: §IV-B.
  • [8] N. Hendren, M. Drazner, A. Pandey, W. Tang, and J. Grodin (2019) Implications of perceived dyspnea and global well-being measured by visual assessment scales during treatment for acute decompensated heart failure. The American journal of cardiology 124 (3), pp. 402–408. Cited by: §I.
  • [9] H. Heo, J. Jung, I. Yang, S. Yoon, H. Shim, and H. Yu (2019)

    End-to-end losses based on speaker basis vectors and all-speaker hard negative mining for speaker verification

    arXiv preprint arXiv:1902.02455. Cited by: §IV.
  • [10] D. Jensen, D. Ofir, and D. O’Donnell (2009) Effects of pregnancy, obesity and aging on the intensity of perceived breathlessness during exercise in healthy humans. Respiratory physiology & neurobiology 167 (1), pp. 87–100. Cited by: §I.
  • [11] A. Kendall, Y. Gal, and R. Cipolla (2018) Multi-task learning using uncertainty to weigh losses for scene geometry and semantics. In

    Proceedings of the IEEE conference on computer vision and pattern recognition

    pp. 7482–7491. Cited by: §IV.
  • [12] V. Kowtha, V. Mitra, C. Bartels, E. Marchi, S. Booker, W. Caruso, S. Kajarekar, and D. Naik (2020) Detecting emotion primitives from speech and their use in discerning categorical emotions. arXiv preprint arXiv:2002.01323. Cited by: §IV-B.
  • [13] I. Lawrence and K. Lin (1989) A concordance correlation coefficient to evaluate reproducibility. Biometrics, pp. 255–268. Cited by: §IV-B.
  • [14] S. Li, B. Lin, C. Tsai, C. Yang, and B. Lin (2017) Design of wearable breathing sound monitoring system for real-time wheeze detection. Sensors 17 (1), pp. 171. Cited by: §I.
  • [15] T. Lin, P. Goyal, R. Girshick, K. He, and P. Dollár (2017) Focal loss for dense object detection. Proceedings of the IEEE international conference on computer vision, pp. 2980–2988. Cited by: §IV-B.
  • [16] S. Liu, E. Johns, and A. Davison (2019) End-to-end multi-task learning with attention. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1871–1880. Cited by: §IV-B.
  • [17] T. MIKAMI, Y. KOJIMA, K. YONEZAWA, M. YAMAMOTO, and M. FURUKAWA (2012) Classification of oral/nasal simulated snores based on the acoustic properties. Journal of Biomechanical Science and Engineering 7 (4), pp. 433–448. Cited by: §IV-A.
  • [18] R. Nakanishi, J. Rana, A. Rozanski, V. Cheng, H. Gransar, L. Thomson, R. Miranda-Peats, S. Hayes, J. Friedman, and D. Berman (2013) Relationship of dyspnea vs. typical angina to coronary artery disease severity, burden, composition and location on coronary ct angiography. Atherosclerosis 230 (1), pp. 61–66. Cited by: §I.
  • [19] A. Nicolò, C. Massaroni, and L. Passfield (2017) Respiratory frequency during exercise: the neglected physiological measure. Frontiers in physiology 8, pp. 922. Cited by: §III.
  • [20] J. Niu, M. Cai, Y. Shi, S. Ren, W. Xu, W. Gao, Z. Luo, and J. M. Reinhardt (2019) A novel method for automatic identification of breathing state. Scientific reports 9 (1), pp. 1–13. Cited by: §IV-A.
  • [21] C. C. of the New York Heart Association et al. (1994) Nomenclature and criteria for diagnosis of diseases of the heart and great vessels. Vol. 253, Boston: Little, Brown & Co. Cited by: §I.
  • [22] J. Pan, P. Adab, C. Jiang, W. Zhang, F. Zhu, Y. Jin, G. Thomas, and T. Lam (2019) All-cause and cause-specific mortality from restrictive and obstructive spirometric patterns in chinese adults with and without dyspnea: guangzhou biobank cohort study. Respiratory medicine 151, pp. 66–80. Cited by: §I.
  • [23] P. Pang, K. Lane, M. Tavares, A. Storrow, C. Shen, W. Peacock, R. Nowak, A. Mebazaa, S. Laribi, and J. Hollander (2017) Is there a clinically meaningful difference in patient reported dyspnea in acute heart failure? an analysis from urgent dyspnea. Heart & Lung 46 (4), pp. 300–307. Cited by: §I.
  • [24] Y. Ren, C. Wang, J. Yang, and Y. Chen (2015) Fine-grained sleep monitoring: hearing your breathing with smartphones. In 2015 IEEE Conference on Computer Communications (INFOCOM), pp. 1194–1202. Cited by: §I, §II.
  • [25] G. Sierra, V. Telfort, B. Popov, L. Durand, R. Agarwal, and V. Lanzo (2004) Monitoring respiratory rate based on tracheal sounds. first experiences. In The 26th Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Vol. 1, pp. 317–320. Cited by: §I.
  • [26] G. Sierra, V. Telfort, B. Popov, M. Pelletier, P. Despault, V. Lanzo, and R. Agarwal (2006) Comparison of respiratory rate estimation based on tracheal sounds versus a capnograph. In 2005 IEEE Engineering in Medicine and Biology 27th Annual Conference, pp. 6145–6148. Cited by: §I.