Abstract

Diffusion models have demonstrated significant potential in producing high-quality images in medical image translation to aid disease diagnosis, localization, and treatment. Nevertheless, current diffusion models often fall short when it comes to faithfully translating medical images. They struggle to accurately preserve anatomical structures, especially when working with unpaired datasets. In this study, we introduce the frequency decoupled diffusion model (FDDM) for magnetic resonance (MR)-to-computed tomography (CT) conversion. The differences between MR and CT images lie in both anatomical structures (e.g. the outlines of organs or bones) and the data distribution (e.g. intensity values and contrast within). Therefore, FDDM first converts anatomical information using an initial conversion module. Then, the converted anatomical information guides a subsequent diffusion model to generate high-quality CT images. Our diffusion model uses a dual-path reverse diffusion process for low-frequency and high-frequency information, achieving a better balance between image quality and anatomical accuracy. We extensively evaluated FDDM using two public datasets for brain MR-to-CT and pelvis MR-to-CT translations. The results show that FDDM outperforms generative adversarial network (GAN)-based, variational autoencoder (VAE)-based, and diffusion-based models. The evaluation metrics included Fréchet inception distance (FID), mean absolute error, mean squared error, structural similarity index measure, and Dice similarity coefficient (DICE). FDDM achieved the best scores on all metrics for both datasets, particularly excelling in FID, with scores of 25.9 for brain data and 29.2 for pelvis data, significantly outperforming the other methods. These results demonstrate that FDDM can generate high-quality target domain images while maintaining the accuracy of translated anatomical structures, thereby facilitating more precise/accurate downstream tasks including anatomy segmentation and radiotherapy planning.

Details

Title
FDDM: unsupervised medical image translation with a frequency-decoupled diffusion model
Author
Li, Yunxiang 1   VIAFID ORCID Logo  ; Hua-Chieh Shao  VIAFID ORCID Logo  ; Qian, Xiaoxue  VIAFID ORCID Logo  ; Zhang, You 1   VIAFID ORCID Logo 

 Department of Radiation Oncology, UT Southwestern Medical Center , Dallas, TX 75390, United States of America 
First page
025007
Publication year
2025
Publication date
Jun 2025
Publisher
IOP Publishing
e-ISSN
26322153
Source type
Scholarly Journal
Language of publication
English
ProQuest document ID
3187548804
Copyright
© 2025 The Author(s). Published by IOP Publishing Ltd. This work is published under https://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.