Please use this identifier to cite or link to this item: https://hdl.handle.net/20.500.13091/2419
Title: Medical thermograms' classification using deep transfer learning models and methods
Authors: Örnek, Ahmet Haydar
Ceylan, Murat
Keywords: Classification
Convolutional neural networks
Deep learning
Medicine
Neonate
Thermography
Transfer learning
Svm
Publisher: Springer
Abstract: Infrared thermal imaging and deep learning provide intelligent monitoring systems that detect diseases in early phases. However, deep learning models require thousands of labeled images to be effectively trained from scratch. Since such a dataset cannot be collected from a neonatal intensive care unit (NICU), deep transfer learning models and methods were used for the first time in this study to classify neonates in the NICU as healthy and unhealthy. When nine different pre-trained models (VGG16, VGG19, Xception, ResNet101, ResNet50, Inceptionv3, InceptionResNetv2, MobileNet and DenseNet201) and two different classification methods (Multilayer Perceptrons and Support Vector Machines (SVMs)) were compared, best results were obtained as 100.00% specificity, sensitivity and accuracy with VGG19, and SVMs. This study proposes highest classification performance when comparing other studies that detect health status of neonates.
URI: https://doi.org/10.1007/s11042-021-11852-6
https://hdl.handle.net/20.500.13091/2419
ISSN: 1380-7501
1573-7721
Appears in Collections:Mühendislik ve Doğa Bilimleri Fakültesi Koleksiyonu
Scopus İndeksli Yayınlar Koleksiyonu / Scopus Indexed Publications Collections
WoS İndeksli Yayınlar Koleksiyonu / WoS Indexed Publications Collections

Files in This Item:
File SizeFormat 
s11042-021-11852-6.pdf
  Until 2030-01-01
2.52 MBAdobe PDFView/Open    Request a copy
Show full item record



CORE Recommender

WEB OF SCIENCETM
Citations

2
checked on Apr 20, 2024

Page view(s)

64
checked on Apr 22, 2024

Download(s)

6
checked on Apr 22, 2024

Google ScholarTM

Check




Altmetric


Items in GCRIS Repository are protected by copyright, with all rights reserved, unless otherwise indicated.