Skip to content

Commit 11d0585

Browse files
adding support for mixed precision training in attention block
1 parent 952e5d1 commit 11d0585

File tree

2 files changed

+120
-91
lines changed

2 files changed

+120
-91
lines changed

README.md

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
# diffusion.cu
22

3-
This project is a from-scratch implementation of diffusion model training in C++/CUDA. Inspired by Andrej Karpathy's [llm.c](https://github.com/karpathy/llm.c) and Chen Lu's [unet.cu](https://github.com/clu0/unet.cu). The implementation is based on the U-Net architecture in the paper [Diffusion Models Beat GANs on Image Synthesis](https://arxiv.org/abs/2105.05233).
3+
This project is a from-scratch implementation of diffusion model training in C++/CUDA. This project is currently in-progress. Inspired by Andrej Karpathy's [llm.c](https://github.com/karpathy/llm.c) and Chen Lu's [unet.cu](https://github.com/clu0/unet.cu). The implementation is based on the architecture in the paper [Diffusion Models Beat GANs on Image Synthesis](https://arxiv.org/abs/2105.05233).
44

55
## Training
66

@@ -26,6 +26,7 @@ This currently supports unconditional diffusion model training, and the end-to-e
2626
In Progress:
2727
- support for distributed training via MPI
2828
- support for mixed precision training
29+
- support for DiT as another architecture we can add, because we can re-use the same components of llm.c
2930

3031
### **My Motivation:**
3132

0 commit comments

Comments
 (0)