ISSN 1004-4140
CN 11-3017/P

基于改进型CycleGAN的CBCT合成CT图像质量改进研究

Quality Improvement of CBCT-Synthesized CT Images Based on Improved CycleGAN

  • 摘要: 目的:本研究提出一种基于改进型CycleGAN的无监督学习ViTD-CycleGAN,以从锥形束计算机断层扫描(CBCT)合成计算机断层扫描(CT)图像,旨在提高合成CT(sCT)图像质量和真实性。方法:ViTD-CycleGAN在生成器中引入了基于视觉转换器(ViT)的U-Net框架和深度卷积(DW),结合Transformer的自注意力机制,以提取并保留重要特征和细节信息。同时,引入梯度惩罚(GP)和像素级损失函数(PL),以增强模型训练的稳定性和图像一致性。结果:在头颈部和胸部数据集上的定量评价指标(MAE、PSNR、SSIM)均优于现有无监督学习方法。消融实验显示,DW对模型性能提升最为显著。视觉可视化分析进一步证实ViTD-CycleGAN生成的sCT图像具有更高的图像质量和真实性。结论:本研究提出的方法与其他无监督学习方法相比,可提高CBCT合成CT图像质量,具有一定的临床应用价值。

     

    Abstract: Objective: This study proposes an unsupervised learning model, ViTD-CycleGAN, based on an improved CycleGAN to synthesize computed tomography (CT) images from cone-beam computed tomography (CBCT) images. Our aim is to enhance the quality and realism of synthetic CT (sCT) images. Methods: ViTD-CycleGAN incorporates a U-Net framework based on a vision Transformer (ViT) and depth-wise convolution (DW) into its generator, where the self-attention mechanism of the Transformer is leveraged to extract and preserve crucial features and detailed information. Additionally, a gradient penalty and pixel-wise loss function are introduced to enhance the stability of the model training and image consistency. Results: Quantitative evaluation metrics (MAE, PSNR, and SSIM) for head and neck as well as chest datasets indicate the superior performance of the proposed model compared with existing unsupervised learning methods. Ablation experiments show that the DW significantly improved the model performance. Visual-display analysis confirms that the sCT images generated using the ViTD-CycleGAN exhibit higher image quality and realism. Conclusion: Compared with other unsupervised learning methods, the proposed method can improve the quality of CBCT-synthesized CT images and thus offer potential clinical application value.

     

/

返回文章
返回