Commit
·
96fe19f
1
Parent(s):
72b584c
Update README.md
Browse files
README.md
CHANGED
|
@@ -189,51 +189,6 @@ Based on that information, we estimate the following CO2 emissions using the [Ma
|
|
| 189 |
- **Carbon Emitted (Power consumption x Time x Carbon produced based on location of power grid):** 11250 kg CO2 eq.
|
| 190 |
|
| 191 |
|
| 192 |
-
## Citation
|
| 193 |
-
|
| 194 |
-
```bibtex
|
| 195 |
-
@InProceedings{Rombach_2022_CVPR,
|
| 196 |
-
author = {Rombach, Robin and Blattmann, Andreas and Lorenz, Dominik and Esser, Patrick and Ommer, Bj\"orn},
|
| 197 |
-
title = {High-Resolution Image Synthesis With Latent Diffusion Models},
|
| 198 |
-
booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
|
| 199 |
-
month = {June},
|
| 200 |
-
year = {2022},
|
| 201 |
-
pages = {10684-10695}
|
| 202 |
-
}
|
| 203 |
-
```
|
| 204 |
-
|
| 205 |
-
*This model card was written by: Robin Rombach and Patrick Esser and is based on the [DALL-E Mini model card](https://huggingface.co/dalle-mini/dalle-mini).*
|
| 206 |
-
olution `512x512` on "laion-improved-aesthetics" (a subset of laion2B-en,
|
| 207 |
-
filtered to images with an original size `>= 512x512`, estimated aesthetics score `> 5.0`, and an estimated watermark probability `< 0.5`. The watermark estimate is from the LAION-5B metadata, the aesthetics score is estimated using an [improved aesthetics estimator](https://github.com/christophschuhmann/improved-aesthetic-predictor)).
|
| 208 |
-
- `sd-v1-3.ckpt`: Resumed from `sd-v1-2.ckpt`. 195k steps at resolution `512x512` on "laion-improved-aesthetics" and 10\% dropping of the text-conditioning to improve [classifier-free guidance sampling](https://arxiv.org/abs/2207.12598).
|
| 209 |
-
|
| 210 |
-
|
| 211 |
-
- **Hardware:** 32 x 8 x A100 GPUs
|
| 212 |
-
- **Optimizer:** AdamW
|
| 213 |
-
- **Gradient Accumulations**: 2
|
| 214 |
-
- **Batch:** 32 x 8 x 2 x 4 = 2048
|
| 215 |
-
- **Learning rate:** warmup to 0.0001 for 10,000 steps and then kept constant
|
| 216 |
-
|
| 217 |
-
## Evaluation Results
|
| 218 |
-
Evaluations with different classifier-free guidance scales (1.5, 2.0, 3.0, 4.0,
|
| 219 |
-
5.0, 6.0, 7.0, 8.0) and 50 PLMS sampling
|
| 220 |
-
steps show the relative improvements of the checkpoints:
|
| 221 |
-
|
| 222 |
-

|
| 223 |
-
|
| 224 |
-
Evaluated using 50 PLMS steps and 10000 random prompts from the COCO2017 validation set, evaluated at 512x512 resolution. Not optimized for FID scores.
|
| 225 |
-
|
| 226 |
-
## Environmental Impact
|
| 227 |
-
|
| 228 |
-
**Stable Diffusion v1** **Estimated Emissions**
|
| 229 |
-
Based on that information, we estimate the following CO2 emissions using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). The hardware, runtime, cloud provider, and compute region were utilized to estimate the carbon impact.
|
| 230 |
-
|
| 231 |
-
- **Hardware Type:** A100 PCIe 40GB
|
| 232 |
-
- **Hours used:** 150000
|
| 233 |
-
- **Cloud Provider:** AWS
|
| 234 |
-
- **Compute Region:** US-east
|
| 235 |
-
- **Carbon Emitted (Power consumption x Time x Carbon produced based on location of power grid):** 11250 kg CO2 eq.
|
| 236 |
-
|
| 237 |
## Citation
|
| 238 |
|
| 239 |
```bibtex
|
|
|
|
| 189 |
- **Carbon Emitted (Power consumption x Time x Carbon produced based on location of power grid):** 11250 kg CO2 eq.
|
| 190 |
|
| 191 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 192 |
## Citation
|
| 193 |
|
| 194 |
```bibtex
|