# Diffusion-Models-pytorch **Repository Path**: Heconnor/Diffusion-Models-pytorch ## Basic Information - **Project Name**: Diffusion-Models-pytorch - **Description**: No description available - **Primary Language**: Unknown - **License**: Apache-2.0 - **Default Branch**: main - **Homepage**: None - **GVP Project**: No ## Statistics - **Stars**: 0 - **Forks**: 0 - **Created**: 2024-04-29 - **Last Updated**: 2024-04-30 ## Categories & Tags **Categories**: Uncategorized **Tags**: DiffusionModel ## README # Diffusion Models This is an easy-to-understand implementation of diffusion models within 100 lines of code. Different from other implementations, this code doesn't use the lower-bound formulation for sampling and strictly follows Algorithm 1 from the [DDPM](https://arxiv.org/pdf/2006.11239.pdf) paper, which makes it extremely short and easy to follow. There are two implementations: `conditional` and `unconditional`. Furthermore, the conditional code also implements Classifier-Free-Guidance (CFG) and Exponential-Moving-Average (EMA). Below you can find two explanation videos for the theory behind diffusion models and the implementation. Qries Qries
## Train a Diffusion Model on your own data: ### Unconditional Training 1. (optional) Configure Hyperparameters in ```ddpm.py``` 2. Set path to dataset in ```ddpm.py``` 3. ```python ddpm.py``` ### Conditional Training 1. (optional) Configure Hyperparameters in ```ddpm_conditional.py``` 2. Set path to dataset in ```ddpm_conditional.py``` 3. ```python ddpm_conditional.py``` ## Sampling The following examples show how to sample images using the models trained in the video on the [Landscape Dataset](https://www.kaggle.com/datasets/arnaud58/landscape-pictures). You can download the checkpoints for the models [here](https://drive.google.com/drive/folders/1beUSI-edO98i6J9pDR67BKGCfkzUL5DX?usp=sharing). ### Unconditional Model ```python device = "cuda" model = UNet().to(device) ckpt = torch.load("unconditional_ckpt.pt") model.load_state_dict(ckpt) diffusion = Diffusion(img_size=64, device=device) x = diffusion.sample(model, n=16) plot_images(x) ``` ### Conditional Model This model was trained on [CIFAR-10 64x64](https://www.kaggle.com/datasets/joaopauloschuler/cifar10-64x64-resized-via-cai-super-resolution) with 10 classes ```airplane:0, auto:1, bird:2, cat:3, deer:4, dog:5, frog:6, horse:7, ship:8, truck:9``` ```python n = 10 device = "cuda" model = UNet_conditional(num_classes=10).to(device) ckpt = torch.load("conditional_ema_ckpt.pt") model.load_state_dict(ckpt) diffusion = Diffusion(img_size=64, device=device) y = torch.Tensor([6] * n).long().to(device) x = diffusion.sample(model, n, y, cfg_scale=3) plot_images(x) ```
A more advanced version of this code can be found [here](https://github.com/tcapelle/Diffusion-Models-pytorch) by [@tcapelle](https://github.com/tcapelle). It introduces better logging, faster & more efficient training and other nice features and is also being followed by a nice [write-up](https://wandb.ai/capecape/train_sd/reports/Training-a-Conditional-Diffusion-model-from-scratch--VmlldzoyODMxNjE3).