DiffusionBlend: Learning 3D Image Prior through Position-aware Diffusion Score Blending for 3D Computed Tomography Reconstruction
This repository contains the official implementation of "DiffusionBlend: Learning 3D Image Prior through Position-aware Diffusion Score Blending for 3D Computed Tomography Reconstruction", published at NeurIPS 2024.
DiffusionBlend introduces a novel method for 3D computed tomography (CT) reconstruction using position-aware diffusion score blending. By leveraging position-specific priors, the framework achieves enhanced reconstruction accuracy while maintaining computational efficiency.
- Position-aware Diffusion Blending: Incorporates spatial information to refine 3D reconstruction quality.
- Triplane-based 3D Representation: Utilizes a position encoding to model 3D patch priors efficiently.
- Scalable and Generalizable: Designed for both synthetic and real-world CT reconstruction tasks.
The code is implemented in Python and requires the following dependencies:
torch
(>=1.9.0)torchvision
numpy
You can install the dependencies via:
pip install torch torchvision numpy
To train the model on synthetic volume CT data, use the following script:
bash train_SVCT_3D_triplane.sh
To perform inference and evaluate 3D reconstruction using diffusion score blending, use:
bash eval_3D_blend_cond.sh



If you find this work useful in your research, please cite:
@inproceedings{diffusionblend2024,
title={DiffusionBlend: Learning 3D Image Prior through Position-aware Diffusion Score Blending for 3D Computed Tomography Reconstruction},
author={Song, Bowen and Hu, Jason and Luo, Zhaoxu and Fessler, Jeffrey A and Shen, Liyue},
booktitle={Advances in Neural Information Processing Systems (NeurIPS)},
year={2024}
}
We thank the contributors and the NeurIPS community for their valuable feedback and discussions.
This project is licensed under the MIT License. See the LICENSE file for details.