You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+8-6Lines changed: 8 additions & 6 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,14 +1,15 @@
1
1
# Latent Diffusion MNIST Experiment
2
2
3
-
Latent Diffusion MNIST Experiment is a Python project exploring the implementation of a latent diffusion model using a variational autoencoder (VAE) and a conditional U-Net. It currently targets the MNIST dataset, but the architecture and training pipeline are intentionally dataset-agnostic and can scale to larger and more complex image domains.
3
+
Latent Diffusion MNIST Experiment is a Python project exploring the implementation of a latent diffusion model using a variational autoencoder (VAE) and a conditional U-Net. It currently targets the MNIST dataset, but the architecture and training pipeline are intentionally dataset-agnostic and can scale to larger and more complex image domains with minimal changes.
4
4
5
5
## Features
6
6
-**Modular Architecture** – separate training routines for VAE and U-Net components built on top of Hugging Face Diffusers.
7
-
-**Dataset Agnostic Pipeline** – although the repository demonstrates MNIST, the data loaders and model design can be extended to other datasets with minimal changes, retaining most of the training logic from this experiment.
7
+
-**Dataset Agnostic Pipeline** – although the repository demonstrates MNIST, the data loaders and model design can be easily extended to other datasets, retaining most of the training logic from this experiment.
8
8
-**Configurable Training** – hyperparameters (batch size, learning rates, epochs, etc.) are managed via `config.yaml` for reproducible experiments.
9
9
-**Efficient & Stable Training** – leverages `Accelerate` for device management and distributed training, cosine learning rate schedules, and EMA tracking. Added gradient clipping.
10
10
-**Visualization Utilities** – automatic saving of reconstruction and generation plots for monitoring model performance.
11
-
-**Gradio App** – ready-to-deploy web app for interactive predictions. Hosted on [Huggingface Spaces](https://huggingface.co/spaces/codinglabsong/aging-gan).
11
+
-**Gradio App** – *Work in Progress*
12
+
<!-- - **Gradio App** – ready-to-deploy web app for interactive predictions. Hosted on [Huggingface Spaces](https://huggingface.co/spaces/codinglabsong/aging-gan). -->
12
13
-**Developer Tools & CI** – Linting with ruff and black, unit tests with pytest, end‐to‐end smoke tests in GitHub Actions.
13
14
14
15
## Installation
@@ -61,7 +62,7 @@ The script will load the VAE and EMA-smoothed U-Net weights and produce images s
61
62
## Results
62
63
### Example Outputs
63
64
64
-
*Placeholder for generated image examples.*
65
+
*Work in Progress*
65
66
66
67
### Considerations for Improvements
67
68
@@ -70,8 +71,9 @@ The script will load the VAE and EMA-smoothed U-Net weights and produce images s
70
71
- Integrate more advanced schedulers or guidance techniques.
71
72
72
73
## Running the Gradio Inference App
74
+
*Work in Progress*
73
75
74
-
This project includes an interactive Gradio app for making predictions with the trained model.
76
+
<!-- This project includes an interactive Gradio app for making predictions with the trained model.
75
77
76
78
1. **Obtain the Trained Model:**
77
79
- Ensure that a trained model directory (`models/vae.pth` and `models/ema-unet.pth`) is available in the project root.
@@ -84,7 +86,7 @@ This project includes an interactive Gradio app for making predictions with the
84
86
```
85
87
- Visit the printed URL (e.g., `http://127.0.0.1:7860`) to interact with the model.
86
88
87
-
> You can also access the hosted demo on [Huggingface Spaces](https://huggingface.co/spaces/codinglabsong/aging-gan)
89
+
> You can also access the hosted demo on [Huggingface Spaces](https://huggingface.co/spaces/codinglabsong/aging-gan) -->
0 commit comments