Vis enkel innførsel

dc.contributor.advisorEngan, Kjersti
dc.contributor.advisorEftestøl, Trygve
dc.contributor.advisorAustvoll, Ivar
dc.contributor.authorMeinich-Bache, Øyvind
dc.date.accessioned2020-10-22T11:23:56Z
dc.date.available2020-10-22T11:23:56Z
dc.date.issued2020-01
dc.identifier.citationAutomatic Video Analysis in Resuscitation by Øyvind Meinich-Bache, Stavanger : University of Stavanger, 2020 (PhD thesis UiS, no. 499)en_US
dc.identifier.isbn978-82-7644-900-6
dc.identifier.issn1890-1387
dc.identifier.urihttps://hdl.handle.net/11250/2684476
dc.descriptionPhD thesis in Information technologyen_US
dc.description.abstractThis thesis investigates possibilities for applying automatic video analysis in the medical context of resuscitation of a patient. Two situations are investigated: 1) Out-of-hospital cardiac arrest (OHCA) where there is a need for cardiopulmonary resuscitation (CPR) and 2) newborn resuscitation where the newborn is in need of various resuscitation activities, such as stimulation and ventilation support. Both situations suffer from high mortality rates and measurement of resuscitation parameters and activities to evaluate if the performed resuscitation complies with the recommended guidelines, could contribute to ensure provision of quality treatment. Currently there are no clinical solutions utilizing automatic video analysis to improve the quality of the resuscitation in the two situations approached in this thesis. In this work, conventional image processing methods, such as segmentation and frequency analysis approaches have been used to perform measurementof the CPR quality during simulated OHCA situations. The methods for measurement of chest compression rate and CPR summary parameters are implemented in a smartphone app which performs real-time measurements and communicate the information to a webserver that could be monitored by the emergency unit. The system performance is satisfactory with accurate measurements and could add valuable information to the communication between the caller and the emergency unit in OHCA situations. Deep learning and convolutional neural network (CNN) approaches have been used for activity recognition from newborn resuscitation videos. The proposed system, ORAA-net, is a two-step approach consisting of 1) Object detection and Region proposal using a 2D CNN and post-processing, and 2) Activity recognition and generation of Activity timelines using 3D CNNs. The system provides promising results on a dataset of noisy low quality newborn resuscitation videos. By detecting and quantifying the amount of the relevant activities for each episode, a better understanding of the effect of the different resuscitation activities can be achieved, and potentially contribute to optimize patient treatment in newborn resuscitations situations.en_US
dc.language.isoengen_US
dc.publisherUniversity of Stavanger, Norwayen_US
dc.relation.ispartofseriesPhD thesis UiS;
dc.relation.ispartofseries;499
dc.rightsCopyright the author, all right reserved
dc.subjectautomatisk videoanalyseen_US
dc.subjectgjenopplivingen_US
dc.subjecthjerteinfarkten_US
dc.subjectbildebehandlingen_US
dc.subjectinformasjonsteknologien_US
dc.subjectmedisinsk teknologien_US
dc.titleAutomatic Video Analysis in Resuscitationen_US
dc.typeDoctoral thesisen_US
dc.rights.holder© Øyvind Meinich-Bache, 2020en_US
dc.subject.nsiVDP::Mathematics and natural science: 400::Information and communication science: 420en_US
dc.subject.nsiVDP::Teknologi: 500::Medisinsk teknologi: 620en_US


Tilhørende fil(er)

Thumbnail

Denne innførselen finnes i følgende samling(er)

Vis enkel innførsel