Shaoyan Pan, Elham Abouei, Jacob Wynne, Tonghe Wang, Richard L. J. Qiu, Yuheng Li, Chih-Wei Chang, Junbo Peng, Justin Roper, Pretesh Patel, David S. Yu, Hui Mao, Xiaofeng Yang
Magnetic resonance imaging (MRI)-based synthetic computed tomography (sCT) simplifies radiation therapy treatment planning by eliminating the need for CT simulation and error-prone image registration, ultimately reducing patient radiation dose and setup uncertainty. We propose an MRI-to-CT transformer-based denoising diffusion probabilistic model (MC-DDPM) to transform MRI into high-quality sCT to facilitate radiation treatment planning. MC-DDPM implements diffusion processes with a shifted-window transformer network to generate sCT from MRI. The proposed model consists of two processes: a forward process which adds Gaussian noise to real CT scans, and a reverse process in which a shifted-window transformer V-net (Swin-Vnet) denoises the noisy CT scans conditioned on the MRI from the same patient to produce noise-free CT scans. With an optimally trained Swin-Vnet, the reverse diffusion process was used to generate sCT scans matching MRI anatomy. We evaluated the proposed method by generating sCT from MRI on a brain dataset and a prostate dataset. Qualitative evaluation was performed using the mean absolute error (MAE) of Hounsfield unit (HU), peak signal to noise ratio (PSNR), multi-scale Structure Similarity index (MS-SSIM) and normalized cross correlation (NCC) indexes between ground truth CTs and sCTs. MC-DDPM generated brain sCTs with state-of-the-art quantitative results with MAE 43.317 HU, PSNR 27.046 dB, SSIM 0.965, and NCC 0.983. For the prostate dataset, MC-DDPM achieved MAE 59.953 HU, PSNR 26.920 dB, SSIM 0.849, and NCC 0.948. In conclusion, we have developed and validated a novel approach for generating CT images from routine MRIs using a transformer-based DDPM. This model effectively captures the complex relationship between CT and MRI images, allowing for robust and high-quality synthetic CT (sCT) images to be generated in minutes.
Junbo Peng, Chih-Wei Chang, Huiqiao Xie, Richard L. J. Qiu, Justin Roper, Tonghe Wang, Beth Bradshaw, Xiangyang Tang, Xiaofeng Yang
Background: Dual-energy CT (DECT) and material decomposition play vital roles in quantitative medical imaging. However, the decomposition process may suffer from significant noise amplification, leading to severely degraded image signal-to-noise ratios (SNRs). While existing iterative algorithms perform noise suppression using different image priors, these heuristic image priors cannot accurately represent the features of the target image manifold. Although deep learning-based decomposition methods have been reported, these methods are in the supervised-learning framework requiring paired data for training, which is not readily available in clinical settings. Purpose: This work aims to develop an unsupervised-learning framework with data-measurement consistency for image-domain material decomposition in DECT.
Junbo Peng, Yuan Gao, Chih-Wei Chang, Richard Qiu, Tonghe Wang, Aparna Kesarwala, Kailin Yang, Jacob Scott, David Yu, Xiaofeng Yang
Background: Cone-beam computed tomography (CBCT) scans, performed fractionally (e.g., daily or weekly), are widely utilized for patient alignment in the image-guided radiotherapy (IGRT) process, thereby making it a potential imaging modality for the implementation of adaptive radiotherapy (ART) protocols. Nonetheless, significant artifacts and incorrect Hounsfield unit (HU) values hinder their application in quantitative tasks such as target and organ segmentations and dose calculation. Therefore, acquiring CT-quality images from the CBCT scans is essential to implement online ART in clinical settings. Purpose: This work aims to develop an unsupervised learning method using the patient-specific diffusion model for CBCT-based synthetic CT (sCT) generation to improve the image quality of CBCT. Methods: The proposed method is in an unsupervised framework that utilizes a patient-specific score-based model as the image prior alongside a customized total variation (TV) regularization to enforce coherence across different transverse slices. The score-based model is unconditionally trained using the same patient's planning CT (pCT) images to characterize the manifold of CT-quality images and capture the unique anatomical information of the specific patient. The efficacy of the proposed method was assessed on images from anatomical sites including head and neck (H&N) cancer, pancreatic cancer, and lung cancer. The performance of the proposed CBCT correction method was evaluated using quantitative metrics including mean absolute error (MAE), peak signal-to-noise ratio (PSNR), and normalized cross-correlation (NCC). Additionally, the proposed algorithm was benchmarked against two other unsupervised diffusion model-based CBCT correction algorithms.
Junbo Peng, Tonghe Wang, Huiqiao Xie, Richard L. J. Qiu, Chih-Wei Chang, Justin Roper, David S. Yu, Xiangyang Tang, Xiaofeng Yang
Background: Limited-angle (LA) dual-energy (DE) cone-beam CT (CBCT) is considered as a potential solution to achieve fast and low-dose DE imaging on current CBCT scanners without hardware modification. However, its clinical implementations are hindered by the challenging image reconstruction from LA projections. While optimization-based and deep learning-based methods have been proposed for image reconstruction, their utilization is limited by the requirement for X-ray spectra measurement or paired datasets for model training. Purpose: This work aims to facilitate the clinical applications of fast and low-dose DECBCT by developing a practical solution for image reconstruction in LA-DECBCT. Methods: An inter-spectral structural similarity-based regularization was integrated into the iterative image reconstruction in LA-DECBCT. By enforcing the similarity between the DE images, LA artifacts were efficiently reduced in the reconstructed DECBCT images. The proposed method was evaluated using four physical phantoms and three digital phantoms, demonstrating its efficacy in quantitative DECBCT imaging. Conclusions: The proposed method achieves accurate image reconstruction without the need for X-ray spectra measurement for optimization or paired datasets for model training, showing great practical value in clinical implementations of LA-DECBCT.
Junbo Peng, Chih-Wei Chang, Richard L. J. Qiu, Tonghe Wang, Justin Roper, Beth Ghavidel, Xiangyang Tang, Xiaofeng Yang
Background: Dual-energy imaging on cone-beam CT (CBCT) scanners has great potential in different clinical applications, including image-guided surgery and adaptive proton therapy. However, the clinical practice of dual-energy CBCT (DE-CBCT) has been hindered by the requirement of sophisticated hardware components. Purpose: In this work, we aim to propose a practical solution for single-scan dual-energy imaging on current CBCT scanners without hardware modifications, using two complementary limited-angle scans with a projection-consistent diffusion model. Methods: Our approach has two major components: data acquisition using two complementary limited-angle scans, and dual-energy projections restoration with subsequent FDK reconstruction. Two complementary scans at different kVps are performed in a single rotation by switching the tube voltage at the middle of the source trajectory, acquiring the mixed-spectra projection in a single CBCT scan. Full-sampled dual-energy projections are then restored by a projection-consistent diffusion model in a slice-by-slice manner, followed by the DE-CBCT reconstruction using the FDK algorithm. Results: The proposed method was evaluated in a simulation study of digital abdomen phantoms and a study of real rat data. In the simulation study, the proposed method produced DE-CBCT images at a mean absolute error (MAE) of 20 HU. In the small-animal study, reconstructed DE-CBCT images using the proposed method gave an MAE of 25 HU. Conclusion: This study demonstrates the feasibility of DE-CBCT imaging using two complementary limited-angle scans with a projection-consistent diffusion model in both half-fan and short scans. The proposed method may allow quantitative applications of DE-CBCT and enable DE-CBCT-based adaptive proton therapy.
Shaoyan Pan, Elham Abouei, Junbo Peng, Joshua Qian, Jacob F Wynne, Tonghe Wang, Chih-Wei Chang, Justin Roper, Jonathon A Nye, Hui Mao, Xiaofeng Yang
Objective: Positron Emission Tomography (PET) has been a commonly used imaging modality in broad clinical applications. One of the most important tradeoffs in PET imaging is between image quality and radiation dose: high image quality comes with high radiation exposure. Improving image quality is desirable for all clinical applications while minimizing radiation exposure is needed to reduce risk to patients. Approach: We introduce PET Consistency Model (PET-CM), an efficient diffusion-based method for generating high-quality full-dose PET images from low-dose PET images. It employs a two-step process, adding Gaussian noise to full-dose PET images in the forward diffusion, and then denoising them using a PET Shifted-window Vision Transformer (PET-VIT) network in the reverse diffusion. The PET-VIT network learns a consistency function that enables direct denoising of Gaussian noise into clean full-dose PET images. PET-CM achieves state-of-the-art image quality while requiring significantly less computation time than other methods. Results: In experiments comparing eighth-dose to full-dose images, PET-CM demonstrated impressive performance with NMAE of 1.278+/-0.122%, PSNR of 33.783+/-0.824dB, SSIM of 0.964+/-0.009, NCC of 0.968+/-0.011, HRS of 4.543, and SUV Error of 0.255+/-0.318%, with an average generation time of 62 seconds per patient. This is a significant improvement compared to the state-of-the-art diffusion-based model with PET-CM reaching this result 12x faster. Similarly, in the quarter-dose to full-dose image experiments, PET-CM delivered competitive outcomes, achieving an NMAE of 0.973+/-0.066%, PSNR of 36.172+/-0.801dB, SSIM of 0.984+/-0.004, NCC of 0.990+/-0.005, HRS of 4.428, and SUV Error of 0.151+/-0.192% using the same generation process, which underlining its high quantitative and clinical precision in both denoising scenario.
Shaoyan Pan, Chih-Wei Chang, Junbo Peng, Jiahan Zhang, Richard L. J. Qiu, Tonghe Wang, Justin Roper, Tian Liu, Hui Mao, Xiaofeng Yang
This study aims to develop a novel Cycle-guided Denoising Diffusion Probability Model (CG-DDPM) for cross-modality MRI synthesis. The CG-DDPM deploys two DDPMs that condition each other to generate synthetic images from two different MRI pulse sequences. The two DDPMs exchange random latent noise in the reverse processes, which helps to regularize both DDPMs and generate matching images in two modalities. This improves image-to-image translation ac-curacy. We evaluated the CG-DDPM quantitatively using mean absolute error (MAE), multi-scale structural similarity index measure (MSSIM), and peak sig-nal-to-noise ratio (PSNR), as well as the network synthesis consistency, on the BraTS2020 dataset. Our proposed method showed high accuracy and reliable consistency for MRI synthesis. In addition, we compared the CG-DDPM with several other state-of-the-art networks and demonstrated statistically significant improvements in the image quality of synthetic MRIs. The proposed method enhances the capability of current multimodal MRI synthesis approaches, which could contribute to more accurate diagnosis and better treatment planning for patients by synthesizing additional MRI modalities.
Junbo Peng, Ailin Wu, Lei Zhu
Beam angle optimization (BAO) largely determines the performance of fixed-field intensity modulated radiation therapy (IMRT), and it is usually considered as non-convex optimization and an NP hard problem. In this work, BAO is reformulated into a highly efficient framework of standard quadratic optimization. The maximum of beamlet intensities for each incident field as the surrogate variable indicating whether one radiation field has been selected. By converting the function of maximum value in the objective into a set of linear constraints, the problem is solved as standard quadratic optimization via reweighting l1-norm scheme. The performance of the proposed BAO has been verified on a digital phantom and two patients. And the conclusion is drawn: the proposed convex optimization framework is able to find an optimal set of beam angles, leading to improved dose sparing on OARs in fixed-field IMRT.
Junbo Peng, Richard L. J. Qiu, Jacob F Wynne, Chih-Wei Chang, Shaoyan Pan, Tonghe Wang, Justin Roper, Tian Liu, Pretesh R. Patel, David S. Yu, Xiaofeng Yang
Background: Daily or weekly cone-beam computed tomography (CBCT) scans are commonly used for accurate patient positioning during the image-guided radiotherapy (IGRT) process, making it an ideal option for adaptive radiotherapy (ART) replanning. However, the presence of severe artifacts and inaccurate Hounsfield unit (HU) values prevent its use for quantitative applications such as organ segmentation and dose calculation. To enable the clinical practice of online ART, it is crucial to obtain CBCT scans with a quality comparable to that of a CT scan. Purpose: This work aims to develop a conditional diffusion model to perform image translation from the CBCT to the CT domain for the image quality improvement of CBCT. Methods: The proposed method is a conditional denoising diffusion probabilistic model (DDPM) that utilizes a time-embedded U-net architecture with residual and attention blocks to gradually transform standard Gaussian noise to the target CT distribution conditioned on the CBCT. The model was trained on deformed planning CT (dpCT) and CBCT image pairs, and its feasibility was verified in brain patient study and head-and-neck (H&N) patient study. The performance of the proposed algorithm was evaluated using mean absolute error (MAE), peak signal-to-noise ratio (PSNR) and normalized cross-correlation (NCC) metrics on generated synthetic CT (sCT) samples. The proposed method was also compared to four other diffusion model-based sCT generation methods. Conclusions: The proposed conditional DDPM method can generate sCT from CBCT with accurate HU numbers and reduced artifacts, enabling accurate CBCT-based organ segmentation and dose calculation for online ART.
Junbo Peng
Cone-beam CT (CBCT) is installed in the treatment room to facilitate online clinical applications, including image guidance in radiation and surgery. Half-fan and short-can are the commonly used modes in clinical applications to expand the imaging volume and reduce radiation exposure. Anatomical structures are faithfully reconstructed using CBCT while deep information including material composition is not feasible in its current form. Dual-energy computed tomography (DECT) is an advanced medical imaging technology with superior capability of material differentiation by scanning the patient using two different energy spectra. Conventional DECT systems require expensive and sophisticated hardware components, significantly limiting DECT applications to cone-beam CT (CBCT) machines. In this work, we propose a dual-energy cone-beam CT (DECBCT) system by placing a dedicatedly designed beam modulator in front of the detector surface to acquire dual-energy projection data in a single scan. A deep learning-based data restoration model is introduced to generate complete dual-energy data from the detector-modulated data for DECT image reconstruction and material decomposition. The performance of the proposed method is evaluated using an electron density phantom and clinical CT. The mean error for electron density calculation is lower than 1.16% in the electron density phantom study. The mean errors for reconstructed linear attenuation coefficient (LAC) are lower than 1.41% and 0.89% in the head and pelvis studies, respectively. The results indicate that the proposed method is a promising solution for single-scan DECBCT in clinical practice.
Shaoyan Pan, Vanessa Su, Junbo Peng, Junyuan Li, Yuan Gao, Chih-Wei Chang, Tonghe Wang, Zhen Tian, Xiaofeng Yang
We present a new imaging system to support real-time tumor tracking for surface-guided radiotherapy (SGRT). SGRT uses optical surface imaging (OSI) to acquire real-time surface topography images of the patient on the treatment couch. However, OSI cannot visualize internal anatomy. This study proposes an Advanced Surface Imaging (A-SI) framework to address this issue. In the proposed A-SI framework, a high-speed surface imaging camera consistently captures surface images during radiation delivery, and a CBCT imager captures single-angle X-ray projections at low frequency. The A-SI then utilizes a generative model to generate real-time volumetric images with full anatomy, referred to as Optical Surface-Derived cone beam computed tomography (OSD-CBCT), based on the real-time high-frequent surface images and the low-frequency collected single-angle X-ray projections. The generated OSD-CBCT can provide accurate tumor motion for precise radiation delivery. The A-SI framework uses a patient-specific generative model: physics-integrated consistency-refinement denoising diffusion probabilistic model (PC-DDPM). This model leverages patient-specific anatomical structures and respiratory motion patterns derived from four-dimensional CT (4DCT) during treatment planning. It then employs a geometric transformation module (GTM) to extract volumetric anatomy information from the single-angle X-ray projection. A simulation study with 22 lung cancer patients evaluated the A-SI framework supported by PC-DDPM. The results showed that the framework produced real-time OSD-CBCT with high reconstruction fidelity and precise tumor localization. This study demonstrates the potential of A-SI to enable real-time tumor tracking with minimal imaging dose, advancing SGRT for motion-associated cancers and interventional procedures.
Junbo Peng, Tonghe Wang, Shaoyan Pan, Xiaofeng Yang
This study proposes a novel imaging and reconstruction framework for dual-energy cone-beam CT (DECBCT) using only two orthogonal X-ray projections at different energy levels (2V-DECBCT). The goal is to enable fast and low-dose DE volumetric imaging with high spectral fidelity and structural accuracy, suitable for DECBCT-guided radiation therapy. We introduce a framework for 2V-DECBCT based on physics-informed dual-domain diffusion models. A cycle-domain training strategy is employed to enforce consistency between projection and volume reconstructions through a differentiable physics-informed module. Furthermore, a spectral-consistency loss is introduced to preserve inter-energy contrast during the generative process. The model is trained and evaluated using 4D XCAT phantom data under realistic anatomical motion. The method produces high-fidelity DECBCT volumes from only two views, accurately preserving anatomical boundaries and suppressing artifacts. Subtraction maps computed from the reconstructed energy volumes show strong visual and numerical agreement with ground truth. This work presents the first diffusion model-based framework for 2V-DECBCT reconstruction, demonstrating accurate structural and spectral recovery from extremely sparse inputs.
Junbo Peng, Huiqiao Xie, Tonghe Wang, Xiangyang Tang, Xiaofeng Yang
Daily or weekly cone-beam computed tomography (CBCT) is employed in image-guided radiotherapy (IGRT) for precise patient alignment. However, its clinical utility in quantitative tasks is hindered by severe artifacts and inaccurate Hounsfeld unit (HU). It is essential to enhance CBCT image quality to a level comparable with that of conventional CT scans. This study proposed a conditional flow matching model that gradually transforms a sample from normal distribution to the corresponding CT sample conditioned on the input CBCT image. The proposed model was trained using CBCT and deformed planning CT (dpCT) image pairs in a supervised learning scheme. The feasibility of the conditional flow matching model was verified using studies of brain, head-and-neck (HN), and lung patients. The quantitative performance was evaluated using three metrics, including mean absolute error (MAE), peak signal-to-noise ratio (PSNR), and normalized cross-correlation (NCC). The proposed flow matching model was also compared to other flow matching and diffusion-based generative models for sCT generation. The proposed flow matching model effectively reduced multiple types of artifacts on CBCT images in all the studies. In the study of brain patient, the MAE, PSNR, and NCC of the sCT were improved to 26.02 HU, 32.35 dB, and 0.99, respectively, from 40.63 HU, 27.87 dB, and 0.98 on the CBCT images. In the study of HN patient, the metrics were improved to 33.17 HU, 28.68 dB, 0.98 from 38.99 HU, 27.00 dB, 0.98. In the lung patient study, the metrics were 25.09 HU, 32.81 dB, 0.99 and 32.90 HU, 30.48 dB, 0.98 for sCT and CBCT, respectively. The proposed conditional flow matching model effectively synthesizes high-quality CT-like images from CBCT, achieving accurate HU representation and artifact reduction. This enables more reliable organ segmentation and dose calculation in CBCT-guided online ART workflows.