Open-source text-to-image AI model
Stable Diffusion is an open-source deep learning text-to-image model developed by Stability AI that generates detailed and photorealistic images from textual descriptions using a latent diffusion model architecture. It consists of a variational autoencoder, U-Net, and text encoder, enabling efficient operation on consumer-grade GPUs. The model supports tasks like inpainting, outpainting, and image-to-image translation, with recent versions like Stable Diffusion 3.5 offering improved prompt adherence, typography, and diverse style generation.
API usage is based on credits (1 credit = $0.01 USD). No explicit price for general 'Stable Diffusion'; SDXL 1.0 from 0.9 credits/image, Stable Diffusion 3.5 Large from 6.5 credits/image. Model is open-source with free self-hosting for < $1M revenue. See official page for details.
No official exact price per image for 'Stable Diffusion'; values above reflect credit-based system and variant caveats from Stability AI's pricing page.