TY - JOUR
T1 - DistilIQA
T2 - Distilling Vision Transformers for no-reference perceptual CT image quality assessment
AU - Baldeon-Calisto, Maria
AU - Rivera-Velastegui, Francisco
AU - Lai-Yuen, Susana K.
AU - Riofrío, Daniel
AU - Pérez-Pérez, Noel
AU - Benítez, Diego
AU - Flores-Moyano, Ricardo
N1 - Publisher Copyright:
© 2024 Elsevier Ltd
PY - 2024/7
Y1 - 2024/7
N2 - No-reference image quality assessment (IQA) is a critical step in medical image analysis, with the objective of predicting perceptual image quality without the need for a pristine reference image. The application of no-reference IQA to CT scans is valuable in providing an automated and objective approach to assessing scan quality, optimizing radiation dose, and improving overall healthcare efficiency. In this paper, we introduce DistilIQA, a novel distilled Vision Transformer network designed for no-reference CT image quality assessment. DistilIQA integrates convolutional operations and multi-head self-attention mechanisms by incorporating a powerful convolutional stem at the beginning of the traditional ViT network. Additionally, we present a two-step distillation methodology aimed at improving network performance and efficiency. In the initial step, a ”teacher ensemble network” is constructed by training five vision Transformer networks using a five-fold division schema. In the second step, a ”student network”, comprising of a single Vision Transformer, is trained using the original labeled dataset and the predictions generated by the teacher network as new labels. DistilIQA is evaluated in the task of quality score prediction from low-dose chest CT scans obtained from the LDCT and Projection data of the Cancer Imaging Archive, along with low-dose abdominal CT images from the LDCTIQAC2023 Grand Challenge. Our results demonstrate DistilIQA‘s remarkable performance in both benchmarks, surpassing the capabilities of various CNNs and Transformer architectures. Moreover, our comprehensive experimental analysis demonstrates the effectiveness of incorporating convolutional operations within the ViT architecture and highlights the advantages of our distillation methodology.
AB - No-reference image quality assessment (IQA) is a critical step in medical image analysis, with the objective of predicting perceptual image quality without the need for a pristine reference image. The application of no-reference IQA to CT scans is valuable in providing an automated and objective approach to assessing scan quality, optimizing radiation dose, and improving overall healthcare efficiency. In this paper, we introduce DistilIQA, a novel distilled Vision Transformer network designed for no-reference CT image quality assessment. DistilIQA integrates convolutional operations and multi-head self-attention mechanisms by incorporating a powerful convolutional stem at the beginning of the traditional ViT network. Additionally, we present a two-step distillation methodology aimed at improving network performance and efficiency. In the initial step, a ”teacher ensemble network” is constructed by training five vision Transformer networks using a five-fold division schema. In the second step, a ”student network”, comprising of a single Vision Transformer, is trained using the original labeled dataset and the predictions generated by the teacher network as new labels. DistilIQA is evaluated in the task of quality score prediction from low-dose chest CT scans obtained from the LDCT and Projection data of the Cancer Imaging Archive, along with low-dose abdominal CT images from the LDCTIQAC2023 Grand Challenge. Our results demonstrate DistilIQA‘s remarkable performance in both benchmarks, surpassing the capabilities of various CNNs and Transformer architectures. Moreover, our comprehensive experimental analysis demonstrates the effectiveness of incorporating convolutional operations within the ViT architecture and highlights the advantages of our distillation methodology.
KW - Convolutional operations
KW - Distillation
KW - Image quality assessment
KW - Low-dose computed tomography
KW - No-reference quality assessment
KW - Self-attention mechanism
KW - Vision transformers
UR - http://www.scopus.com/inward/record.url?scp=85195044272&partnerID=8YFLogxK
U2 - 10.1016/j.compbiomed.2024.108670
DO - 10.1016/j.compbiomed.2024.108670
M3 - Artículo
AN - SCOPUS:85195044272
SN - 0010-4825
VL - 177
JO - Computers in Biology and Medicine
JF - Computers in Biology and Medicine
M1 - 108670
ER -