We’re happy to bring you the latest release of Stable Diffusion, Version 2.1. We promised faster releases after releasing Version 2,0, and we’re delivering only a few weeks later. The Version 2 model line is trained using a brand new text encoder (OpenCLIP), developed by LAION, that gives us a deeper range of expression than Version 1. Within a few days of releasing SD v2, people started getting fantastic results as they learned some new ways to prompt, and you’ll be happy to discover that 2.1 supports the new prompting style and brings back many of the old prompts too! The differences are more data, training, and less restrictive dataset filtering.
Notes: 312000000000000 FLOP / GPU / sec [A100 reported, bf16 assumed] * 200000 GPU-hours * 3600 sec / hour * 0.3 [assumed utilization] = 6.7392e+22 FLOP