TY - GEN
T1 - DIReCT
T2 - 29th International Conference on Information Processing in Medical Imaging, IPMI 2025
AU - Liu, Tuo
AU - Wang, Haifeng
AU - Chang, Heng
AU - Wang, Fan
AU - Lian, Chunfeng
AU - Ma, Jianhua
N1 - Publisher Copyright:
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2026.
PY - 2026
Y1 - 2026
N2 - Recent advancements in generative learning have enabled PET image synthesis from relatively more accessible MRI scans, offering a safer, cost-effective, and scalable alternative to traditional PET imaging, e.g., for Alzheimer’s disease (AD) diagnosis. However, current MRI-to-PET translation methods face limitations in controllability and fidelity, often failing to capture personalized metabolic activations and fine-grained structural details in critical regions. To address these challenges, we propose a novel controllable MRI-to-PET translation framework, termed DIReCT, which leverages rectified flow to generate high-fidelity PET images tailored to downstream diagnostic and analytical needs. By injecting cross-modal guidance from a pretrained vision-language model (BiomedCLIP), DIReCT incorporates both common imaging knowledge and individualized clinical information to enhance the personalization of PET synthesis. Extensive experiments on the ADNI dataset demonstrate that DIReCT significantly outperforms existing methods across various image quality metrics. Notably, the synthesized FDG-PET images by DIReCT achieve analytical performance comparable to real FDG-PET scans, excelling in capturing AD-related pathological features for reliable group comparisons and personalized diagnosis.
AB - Recent advancements in generative learning have enabled PET image synthesis from relatively more accessible MRI scans, offering a safer, cost-effective, and scalable alternative to traditional PET imaging, e.g., for Alzheimer’s disease (AD) diagnosis. However, current MRI-to-PET translation methods face limitations in controllability and fidelity, often failing to capture personalized metabolic activations and fine-grained structural details in critical regions. To address these challenges, we propose a novel controllable MRI-to-PET translation framework, termed DIReCT, which leverages rectified flow to generate high-fidelity PET images tailored to downstream diagnostic and analytical needs. By injecting cross-modal guidance from a pretrained vision-language model (BiomedCLIP), DIReCT incorporates both common imaging knowledge and individualized clinical information to enhance the personalization of PET synthesis. Extensive experiments on the ADNI dataset demonstrate that DIReCT significantly outperforms existing methods across various image quality metrics. Notably, the synthesized FDG-PET images by DIReCT achieve analytical performance comparable to real FDG-PET scans, excelling in capturing AD-related pathological features for reliable group comparisons and personalized diagnosis.
KW - Controllable Cross-Modal Synthesis
KW - Domain-Knowledge Encoding
KW - PET Imaging
UR - https://www.scopus.com/pages/publications/105014494659
U2 - 10.1007/978-3-031-96628-6_15
DO - 10.1007/978-3-031-96628-6_15
M3 - 会议稿件
AN - SCOPUS:105014494659
SN - 9783031966279
T3 - Lecture Notes in Computer Science
SP - 218
EP - 231
BT - Information Processing in Medical Imaging - 29th International Conference, IPMI 2025, Proceedings
A2 - Oguz, Ipek
A2 - Zhang, Shaoting
A2 - Metaxas, Dimitris N.
PB - Springer Science and Business Media Deutschland GmbH
Y2 - 25 May 2025 through 30 May 2025
ER -