🤗 Diffusers: State-of-the-art diffusion models for image and audio generation in PyTorch
Awesome pre-trained models toolkit based on PaddlePaddle. (400+ models including Image, Text, Audio, Video and Cross-Modal with Easy Inference & Serving)
OpenMMLab Multimodal Advanced, Generative, and Intelligent Creation Toolbox. Unlock the magic 🪄: Generative-AI (AIGC), easy-to-use APIs, awsome model zoo, diffusion models, for text-to-image generation, image/video restoration/enhancement, etc.
Just playing with getting VQGAN+CLIP running locally, rather than having to use colab.
Kandinsky 2 — multilingual text2image latent diffusion model
Stable Diffusion in pure C/C++
Beautiful Stable Diffusion API and UI with support for AITemplate acceleration
Personalization for Stable Diffusion via Aesthetic Gradients 🎨
Templating language written for Stable Diffusion workflows. Available as an extension for the Automatic1111 WebUI.
基于Stable Diffusion优化的AI绘画模型。支持输入中英文文本,可生成多种现代艺术风格的高质量图像。| An optimized text-to-image model based on Stable Diffusion. Both Chinese and English text inputs are available to generate images. The model can generate high-quality images in several modern art styles.
[SIGGRAPH Asia 2022] Text2Light: Zero-Shot Text-Driven HDR Panorama Generation
Text, video or image to video, image and audio in Blender Video Sequence Editor using Zeroscope (SD, XL, upscale to XL), Animov, Potat1, Stable Diffusion(1.5, 2.0, XL), Deep Floyd IF, AudioLDM2 and Bark.
Just playing with getting CLIP Guided Diffusion running locally, rather than having to use colab.
Get updates on the fastest growing repos and cool stats about GitHub right in your inbox
Once per month. No spam.