Please use this identifier to cite or link to this item:
http://dspace.iitrpr.ac.in:8080/xmlui/handle/123456789/4436
Full metadata record
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Sambyal, A S | - |
dc.contributor.author | Niyaz, U | - |
dc.contributor.author | Krishnan, N C. | - |
dc.contributor.author | Bathula, D R. | - |
dc.date.accessioned | 2024-05-08T12:54:19Z | - |
dc.date.available | 2024-05-08T12:54:19Z | - |
dc.date.issued | 2024-05-08 | - |
dc.identifier.uri | http://dspace.iitrpr.ac.in:8080/xmlui/handle/123456789/4436 | - |
dc.description.abstract | Abstract: Background and Objective – In the field of medical image analysis, achieving high accuracy is not enough; ensuring well-calibrated predictions is also crucial. Confidence scores of a deep neural network play a pivotal role in explainability by providing insights into the model's certainty, identifying cases that require attention, and establishing trust in its predictions. Consequently, the significance of a well-calibrated model becomes paramount in the medical imaging domain, where accurate and reliable predictions are of utmost importance. While there has been a significant effort towards training modern deep neural networks to achieve high accuracy on medical imaging tasks, model calibration and factors that affect it remain under-explored. Methods – To address this, we conducted a comprehensive empirical study that explores model performance and calibration under different training regimes. We considered fully supervised training, which is the prevailing approach in the community, as well as rotation-based self-supervised method with and without transfer learning, across various datasets and architecture sizes. Multiple calibration metrics were employed to gain a holistic understanding of model calibration. Results – Our study reveals that factors such as weight distributions and the similarity of learned representations correlate with the calibration trends observed in the models. Notably, models trained using rotation-based self-supervised pretrained regime exhibit significantly better calibration while achieving comparable or even superior performance compared to fully supervised models across different medical imaging datasets. Conclusion – These findings shed light on the importance of model calibration in medical image analysis and highlight the benefits of incorporating self-supervised learning approach to improve both performance and calibration. | en_US |
dc.language.iso | en_US | en_US |
dc.subject | Calibration | en_US |
dc.subject | Fully-supervised | en_US |
dc.subject | Deep neural network | en_US |
dc.subject | Self-supervised | en_US |
dc.subject | Transfer learning | en_US |
dc.subject | Medical imaging | en_US |
dc.title | Understanding calibration of deep neural networks for medical image classification | en_US |
dc.type | Article | en_US |
Appears in Collections: | Year-2023 |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
Full Text.pdf | 3.76 MB | Adobe PDF | View/Open Request a copy |
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.