diff --git a/README.md b/README.md index 09956b1..561a06b 100644 --- a/README.md +++ b/README.md @@ -1,6 +1,6 @@ # ddpo-pytorch -This is an implementation of [Denoising Diffusion Policy Optimization (DDPO)](https://rl-diffusion.github.io/) in PyTorch with support for [low-rank adaptation (LoRA)](https://huggingface.co/docs/diffusers/training/lora). Unlike our original research code (which you can find [here](https://github.com/jannerm/ddpo)), this implementation runs on GPUs, and if LoRA is enabled, requires less than 10GB of GPU memory to finetune a Stable Diffusion-sized model! +This is an implementation of [Denoising Diffusion Policy Optimization (DDPO)](https://rl-diffusion.github.io/) in PyTorch with support for [low-rank adaptation (LoRA)](https://huggingface.co/docs/diffusers/training/lora). Unlike our original research code (which you can find [here](https://github.com/jannerm/ddpo)), this implementation runs on GPUs, and if LoRA is enabled, requires less than 10GB of GPU memory to finetune Stable Diffusion! ![DDPO](teaser.jpg) @@ -42,4 +42,4 @@ The image at the top of this README was generated using LoRA! However, I used a You can find the exact configs I used for the 4 experiments in `config/dgx.py`. For example, to run the aesthetic quality experiment: ```bash accelerate launch scripts/train.py --config config/dgx.py:aesthetic -``` \ No newline at end of file +```