CT image reconstruction: integrating iterative methods with Ml-EM algorithm and deep learning models

Computed Tomography (CT) imaging faces limitations, including low spatial resolution and noise, particularly in low-radiation-dose imaging. To address these challenges, researchers are exploring CT image reconstruction from sinogram data. Sinograms represent X-ray absorption throughout the body, and...

Celý popis

Uloženo v:
Podrobná bibliografie
Vydáno v:Cybernetics and physics Ročník 13; číslo 2; s. 130 - 141
Hlavní autoři: Pham, Cong Thang, Tran, Thi Thu Thao, Huynh, Duc Anh Bao, Nguyen, Quoc Cuong, Nguyen, Tien Hung
Médium: Journal Article
Jazyk:angličtina
Vydáno: 30.09.2024
ISSN:2226-4116, 2226-4116
On-line přístup:Získat plný text
Tagy: Přidat tag
Žádné tagy, Buďte první, kdo vytvoří štítek k tomuto záznamu!
Popis
Shrnutí:Computed Tomography (CT) imaging faces limitations, including low spatial resolution and noise, particularly in low-radiation-dose imaging. To address these challenges, researchers are exploring CT image reconstruction from sinogram data. Sinograms represent X-ray absorption throughout the body, and sophisticated image reconstruction methods, including machine learning algorithms and generative adversarial networks (GANs), can improve precision and resolution without increasing patient radiation exposure. This study proposes an iterative reconstruction approach that combines filters from deep learning models (Convolutional Neural Networks and UNet) with the Maximum Likelihood Expectation Maximization (ML-EM) algorithm and the Enhanced Super-Resolution Generative Adversarial Networks (ESRGAN) model. Our method aims to enhance image quality and reconstruction speed. Experimental results show significant improvements in image quality and resolution, with the proposed method (DL-MLEM-IR-UNET-ESRGAN) achieving an average SSIM of 0.9980 and PSNR of 53.2119, outperforming other methods. Additionally, our method reduces reconstruction time, with an average runtime of 131 seconds.
ISSN:2226-4116
2226-4116
DOI:10.35470/2226-4116-2024-13-2-130-141