-
J Clin Monit Comput · Oct 2023
Deep learning classification of capnography waveforms: secondary analysis of the PRODIGY study.
- Aaron Conway, Mohammad Goudarzi Rad, Wentao Zhou, Matteo Parotto, and Carla Jungquist.
- Peter Munk Cardiac Centre, University Health Network, Toronto, Canada. aaron.conway@utoronto.ca.
- J Clin Monit Comput. 2023 Oct 1; 37 (5): 132713391327-1339.
AbstractCapnography monitors trigger high priority 'no breath' alarms when CO2 measurements do not exceed a given threshold over a specified time-period. False alarms occur when the underlying breathing pattern is stable, but the alarm is triggered when the CO2 value reduces even slightly below the threshold. True 'no breath' events can be falsely classified as breathing if waveform artifact causes an aberrant spike in CO2 values above the threshold. The aim of this study was to determine the accuracy of a deep learning approach to classifying segments of capnography waveforms as either 'breath' or 'no breath'. A post hoc secondary analysis of data from 9 North American sites included in the PRediction of Opioid-induced Respiratory Depression In Patients Monitored by capnoGraphY (PRODIGY) study was conducted. We used a convolutional neural network to classify 15 s capnography waveform segments drawn from a random sample of 400 participants. Loss was calculated over batches of 32 using the binary cross-entropy loss function with weights updated using the Adam optimizer. Internal-external validation was performed by iteratively fitting the model using data from all but one hospital and then assessing its performance in the remaining hospital. The labelled dataset consisted of 10,391 capnography waveform segments. The neural network's accuracy was 0.97, precision was 0.97 and recall was 0.96. Performance was consistent across hospitals in internal-external validation. The neural network could reduce false capnography alarms. Further research is needed to compare the frequency of alarms derived from the neural network with the standard approach.© 2023. The Author(s), under exclusive licence to Springer Nature B.V.
Notes
Knowledge, pearl, summary or comment to share?You can also include formatting, links, images and footnotes in your notes
- Simple formatting can be added to notes, such as
*italics*
,_underline_
or**bold**
. - Superscript can be denoted by
<sup>text</sup>
and subscript<sub>text</sub>
. - Numbered or bulleted lists can be created using either numbered lines
1. 2. 3.
, hyphens-
or asterisks*
. - Links can be included with:
[my link to pubmed](http://pubmed.com)
- Images can be included with:
![alt text](https://bestmedicaljournal.com/study_graph.jpg "Image Title Text")
- For footnotes use
[^1](This is a footnote.)
inline. - Or use an inline reference
[^1]
to refer to a longer footnote elseweher in the document[^1]: This is a long footnote.
.