News Release

AI can help optimize CT scan X-ray radiation dose

Machine-learning models can accurately mimic how radiologists assess the image quality of computed tomography reconstructions, to achieve the highest quality image with the lowest radiation dose possible

Peer-Reviewed Publication

SPIE--International Society for Optics and Photonics

Researchers from Italy developed and trained neural-network-based models to analyze CT scan images. These models accurately matched the assessment of a human professional and can help optimize radiation dose without compromising a patient’s diagnosis.

image: Researchers from Italy developed and trained neural-network-based models to analyze CT scan images. These models accurately matched the assessment of a human professional and can help optimize radiation dose without compromising a patient’s diagnosis. view more 

Credit: Valeri et al., doi 10.1117/1.JMI.10.S1.S11904.

Computed tomography (CT) is one of the most powerful and well-established diagnostic tools available to modern medicine. An increasing number of people have been opting for CT scans, raising concerns about the amount of X-ray radiation that patients are exposed to. Ideally, a patient is exposed to minimum radiation levels during treatments or diagnostic procedures, while still receiving the expected benefit.

In practice, this is known as the ALARA principle, which stands for “As Low As Reasonably Achievable.” However, this principle results in a trade-off because CT image quality decreases with a decrease in radiation power. Thus, medical staff usually aim to strike a balance between a patient’s exposure to X-rays and obtaining good quality CT images to avoid misdiagnosis.

This balance can be achieved through an optimization strategy, in which healthcare professionals, primarily radiologists, observe real images generated by the tomographer and try to identify features, such as tumors or abnormal tissue. Following this, a specialist employs statistical methods to calculate the optimal radiation dose and configuration of the tomographer. This procedure can be generalized by employing reference CT images obtained by scanning specifically designed phantoms containing inserts of different sizes and contrasts, which represent standardized abnormalities. Nevertheless, such manual image analyses are very time-consuming.

To address this issue, a team of researchers from Italy led by Dr. Sandra Doria and members of the Physics Department at the University of Florence, in collaboration with radiologists and medical physicists from Florence Hospital, explored the possibility of automating this process using artificial intelligence (AI).

As reported in Journal of Medical Imaging (JMI), the team created and trained an algorithm—a “model observer”—based on convolutional neural networks (CNNs), which could analyze the standardized abnormalities in CT images just as well as a professional.

To do so, the team had to generate enough training and testing data for the model. Thirty healthcare professionals visually examined 1000 CT images, each consisting in a phantom that mimics human tissue. Aptly termed “phantom,” this material contained cylindrical inserts of different diameters and contrasts. The observers were asked to identify if and where the inserted object appeared in each of the images and state how confident they were in their assessment. This resulted in a dataset of 30,000 labeled CT images taken using different tomographic reconstruction configurations, accurately reflecting human interpretation.

Next, the team implemented two AI models based on different architectures—UNet and MobileNetV2. They modified the base design of these architectures to enable them to perform both classification (“Is there an unusual object in the CT image?”) and localization (“Where is the unusual object?”). Then, they trained and tested the models using images from the dataset.

Through statistical analyses, the research team evaluated various performance metrics to verify that the model observers could accurately emulate how a human would assess the CT images of the phantom. “Our results were very promising, as both trained models performed remarkably well and achieved an absolute percentage error of less than 5 percent. This indicated that the models could identify the object inserted in the phantom with similar accuracy and confidence as a human professional, for almost all reconstruction configurations and abnormalities sizes and contrasts,” remarked Doria, while discussing their findings.

Doria and her team believe that with additional efforts, their model could become a viable strategy to automatically assess CT image quality. She further adds, “Our CNN-based model observers could greatly simplify the process of optimizing the radiation dose used in CT protocols, thereby minimizing health risks to the patient, and help avoid the time-consuming limitations of medical evaluations.”

Doria expressed confidence that the team will succeed in applying their AI model observers on a larger scale, making CT evaluations faster and safer than ever before.

Read the Open Access article by Valeri et al., “UNet and MobileNet CNN-based model observers for CT protocol optimization: comparative performance evaluation by means of phantom CT images,” J. Med. Imaging 10(S1), S11904 (2023), doi 10.1117/1.JMI.10.S1.S11904.

The article is part of the JMI Special Issue on Medical Image Perception and Observer Performance (currently in progress), guest edited by Elizabeth A. Krupinski, Asli Kumcu, and Karla Evans.

Disclaimer: AAAS and EurekAlert! are not responsible for the accuracy of news releases posted to EurekAlert! by contributing institutions or for the use of any information through the EurekAlert system.