Effect of image resolution on automated classification of chest X-rays.
Autor: | Haque MIU; University of Tennessee, The Bredesen Center, Knoxville, Tennessee, United States., Dubey AK; Oak Ridge National Laboratory, Computational Sciences and Engineering Division, Oak Ridge, Tennessee, United States., Danciu I; Oak Ridge National Laboratory, Computational Sciences and Engineering Division, Oak Ridge, Tennessee, United States., Justice AC; VA Connecticut Healthcare, West Haven, Connecticut, United States.; VA Connecticut Healthcare System, Pain Research, Informatics, Multimorbidities, Education (PRIME) Center, West Haven, Connecticut, United States.; Yale School of Medicine, Department of Medicine, New Haven, Connecticut, United States.; Yale University, School of Public Health, New Haven, Connecticut, United States., Ovchinnikova OS; University of Tennessee, The Bredesen Center, Knoxville, Tennessee, United States.; Oak Ridge National Laboratory, Computational Sciences and Engineering Division, Oak Ridge, Tennessee, United States.; University of Tennessee, Materials Science and Engineering, Knoxville, Tennessee, United States., Hinkle JD; Oak Ridge National Laboratory, Computational Sciences and Engineering Division, Oak Ridge, Tennessee, United States. |
---|---|
Jazyk: | angličtina |
Zdroj: | Journal of medical imaging (Bellingham, Wash.) [J Med Imaging (Bellingham)] 2023 Jul; Vol. 10 (4), pp. 044503. Date of Electronic Publication: 2023 Aug 04. |
DOI: | 10.1117/1.JMI.10.4.044503 |
Abstrakt: | Purpose: Deep learning (DL) models have received much attention lately for their ability to achieve expert-level performance on the accurate automated analysis of chest X-rays (CXRs). Recently available public CXR datasets include high resolution images, but state-of-the-art models are trained on reduced size images due to limitations on graphics processing unit memory and training time. As computing hardware continues to advance, it has become feasible to train deep convolutional neural networks on high-resolution images without sacrificing detail by downscaling. This study examines the effect of increased resolution on CXR classification performance. Approach: We used the publicly available MIMIC-CXR-JPG dataset, comprising 377,110 high resolution CXR images for this study. We applied image downscaling from native resolution to 2048 × 2048 pixels , 1024 × 1024 pixels , 512 × 512 pixels , and 256 × 256 pixels and then we used the DenseNet121 and EfficientNet-B4 DL models to evaluate clinical task performance using these four downscaled image resolutions. Results: We find that while some clinical findings are more reliably labeled using high resolutions, many other findings are actually labeled better using downscaled inputs. We qualitatively verify that tasks requiring a large receptive field are better suited to downscaled low resolution input images, by inspecting effective receptive fields and class activation maps of trained models. Finally, we show that stacking an ensemble across resolutions outperforms each individual learner at all input resolutions while providing interpretable scale weights, indicating that diverse information is extracted across resolutions. Conclusions: This study suggests that instead of focusing solely on the finest image resolutions, multi-scale features should be emphasized for information extraction from high-resolution CXRs. (© 2023 Society of Photo-Optical Instrumentation Engineers (SPIE).) |
Databáze: | MEDLINE |
Externí odkaz: |