Search Results for "stable diffusion" - Page 2

Showing 82 open source projects for "stable diffusion"

View related business solutions
  • $300 in Free Credit Towards Top Cloud Services Icon
    $300 in Free Credit Towards Top Cloud Services

    Build VMs, containers, AI, databases, storage—all in one place.

    Start your project in minutes. After credits run out, 20+ products include free monthly usage. Only pay when you're ready to scale.
    Get Started
  • Earn up to 16% annual interest with Nexo. Icon
    Earn up to 16% annual interest with Nexo.

    Let your crypto work for you

    Put idle assets to work with competitive interest rates, borrow without selling, and trade with precision. All in one platform. Geographic restrictions, eligibility, and terms apply.
    Get started with Nexo.
  • 1
    tinygrad

    tinygrad

    Deep learning framework

    This may not be the best deep learning framework, but it is a deep learning framework. Due to its extreme simplicity, it aims to be the easiest framework to add new accelerators to, with support for both inference and training. If XLA is CISC, tinygrad is RISC.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 2
    Sygil WebUI

    Sygil WebUI

    Stable Diffusion web UI

    Sygil WebUI is a browser-based interface for running Stable Diffusion image generation locally or on a server, wrapping common text-to-image and image-to-image workflows into a practical UI. It provides multiple UI modes (including a legacy Gradio interface) and focuses on making iterative prompting, parameter tuning, and post-processing accessible without writing code. The UI exposes core generation controls like resolution, CFG guidance, sampling steps, samplers, seeds, and batch generation so users can reproduce results and refine outputs systematically. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 3
    InvokeAI

    InvokeAI

    InvokeAI is a leading creative engine for Stable Diffusion models

    InvokeAI is an implementation of Stable Diffusion, the open source text-to-image and image-to-image generator. It provides a streamlined process with various new features and options to aid the image generation process. It runs on Windows, Mac and Linux machines, and runs on GPU cards with as little as 4 GB or RAM. InvokeAI is a leading creative engine built to empower professionals and enthusiasts alike.
    Downloads: 20 This Week
    Last Update:
    See Project
  • 4
    PersonaLive

    PersonaLive

    Expressive Portrait Image Animation for Live Streaming

    ...PersonaLive’s architecture balances visual quality and efficiency by combining motion encoding, temporal modules, and hybrid implicit control signals to preserve identity and stable expression through long sequences.
    Downloads: 0 This Week
    Last Update:
    See Project
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • 5
    Text-to-image Playground

    Text-to-image Playground

    A playground to generate images from any text prompt using SD

    dalle-playground is an open-source web application that allows users to generate images from natural language text prompts using modern text-to-image generative models. Originally built around DALL-E Mini, the project later transitioned to using Stable Diffusion, enabling more detailed and higher-quality image synthesis. The system combines a backend machine learning service with a browser-based frontend interface that lets users experiment interactively with prompt engineering and generative AI. Developers can run the application locally or deploy it using cloud infrastructure, making it accessible both for experimentation and educational use. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    Auto-Photoshop-StableDiffusion-Plugin

    Auto-Photoshop-StableDiffusion-Plugin

    Plug-in that makes it easy to generate stable diffusion images

    Auto-Photoshop-StableDiffusion-Plugin is an open-source extension that seamlessly integrates Stable Diffusion image generation directly into Adobe Photoshop, allowing artists and designers to generate, edit, and refine AI-generated imagery without leaving the Photoshop environment. It bridges Photoshop with popular diffusion backends like AUTOMATIC1111 or ComfyUI, effectively embedding powerful generative tools into a familiar creative workflow so users can apply AI creation to layers, selections, and masks while retaining Photoshop’s full editing capabilities. ...
    Downloads: 20 This Week
    Last Update:
    See Project
  • 7
    StabilityMatrix

    StabilityMatrix

    Multi-Platform Package Manager for Stable Diffusion

    StabilityMatrix is a project that helps organize, evaluate, and compare generative AI models and their behavior across prompts, datasets, or configuration settings. It provides a framework to run experiments systematically—capturing inputs, model configurations, outputs, and metrics—so researchers and practitioners can reason about differences in quality, robustness, and failure modes. The repository often bundles tooling for automated prompt sweeping, scoring heuristics (such as diversity,...
    Downloads: 198 This Week
    Last Update:
    See Project
  • 8
    IOPaint

    IOPaint

    Image inpainting tool powered by SOTA AI Model

    IOPaint is a powerful open-source image editing tool focused on inpainting, outpainting, object removal, and general image manipulation driven by state-of-the-art AI models, delivering these capabilities through both local and hosted workflows. Designed to be fully self-hosted and flexible, IOPaint supports a variety of underlying generators and inpaint models — from LaMa erase networks to Stable Diffusion-based replace/object generation — giving users multiple ways to refine or reconstruct images by removing unwanted elements or expanding artwork beyond its original boundaries. Its feature set includes erasing people, watermarks, or defects, adding or replacing objects, applying text-aware edits, and extending images outward (outpainting) to fill contours or expand compositions.
    Downloads: 18 This Week
    Last Update:
    See Project
  • 9
    SwarmUI

    SwarmUI

    Modular AI image and video generation web UI with extensible tools

    SwarmUI is a modular web-based user interface designed for AI-driven image generation, with a strong focus on usability, performance, and extensibility. It serves as a unified environment for working with multiple AI models, including Stable Diffusion and newer image and video generation systems, allowing users to create and manage outputs through a browser interface. SwarmUI is built to accommodate both beginners and advanced users by offering a simple “Generate” interface alongside more advanced workflow tools that expose deeper configuration options. It integrates with underlying systems like node-based workflows, enabling flexible and customizable pipelines for complex generation tasks. ...
    Downloads: 12 This Week
    Last Update:
    See Project
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 10
    SD.Next

    SD.Next

    All-in-one WebUI for AI generative image and video creation

    SD.Next is an all-in-one web user interface for generative image creation that expands beyond basic Stable Diffusion workflows to cover broader image and video generation, captioning, and processing tasks. It is designed as a power-user environment where model management, generation features, and workflow controls are centralized in a single UI rather than spread across separate scripts and utilities. The project emphasizes broad model support and includes mechanisms for discovering, downloading, and configuring models through integrated tooling, lowering the setup burden for experimentation. ...
    Downloads: 10 This Week
    Last Update:
    See Project
  • 11
    Matcha-TTS

    Matcha-TTS

    A fast TTS architecture with conditional flow matching

    ...It models speech as an ODE-based generative process, and conditional flow matching lets it reach high-quality audio in only a few synthesis steps, which greatly reduces latency compared to score-matching diffusion approaches. The model is fully probabilistic, so it can generate diverse realizations of the same text while still sounding stable and intelligible. The repository provides an end-to-end TTS pipeline: a PyTorch/Lightning training stack, configuration files, pre-trained checkpoints, a command-line interface, and a Gradio app for interactive testing. ...
    Downloads: 16 This Week
    Last Update:
    See Project
  • 12
    Pixelization

    Pixelization

    Stable-diffusion-webui-pixelization

    This is a specialized extension for the popular Stable Diffusion Web UI (AUTOMATIC1111) that focuses on converting or “pixelizing” images into a pixel-art aesthetic. It's designed as a plugin you install into the Web UI so that in the “Extras” or “Pixelization” tab you can drag in an input image and produce a stylized, block-based version with control over cell size, color depth, and segmentation.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    Luna AI

    Luna AI

    Virtual AI anchor that combines state-of-the-art technology

    Luna AI is a virtual AI streamer framework designed to power an interactive VTuber that can go live on major platforms and chat with viewers in real time. It is built around a core assistant persona called “Luna AI,” which can be driven by a wide range of large language models and platforms, including GPT-style APIs, Claude, LangChain-based backends, ChatGLM, Kimi, Ollama, and many others. The project supports multiple rendering backends for the avatar, such as Live2D, Unreal Engine (UE),...
    Downloads: 18 This Week
    Last Update:
    See Project
  • 14
    DeepSpeed MII

    DeepSpeed MII

    MII makes low-latency and high-throughput inference possible

    ...The Deep Learning (DL) open-source community has seen tremendous growth in the last few months. Incredibly powerful text generation models such as the Bloom 176B, or image generation model such as Stable Diffusion are now available to anyone with access to a handful or even a single GPU through platforms such as Hugging Face. While open-sourcing has democratized access to AI capabilities, their application is still restricted by two critical factors: inference latency and cost. DeepSpeed-MII is a new open-source python library from DeepSpeed, aimed towards making low-latency, low-cost inference of powerful models not only feasible but also easily accessible. ...
    Downloads: 7 This Week
    Last Update:
    See Project
  • 15
    HunyuanWorld 1.0

    HunyuanWorld 1.0

    Generating Immersive, Explorable, and Interactive 3D Worlds

    HunyuanWorld-1.0 is an open-source, simulation-capable 3D world generation model developed by Tencent Hunyuan that creates immersive, explorable, and interactive 3D environments from text or image inputs. It combines the strengths of video-based diversity and 3D-based geometric consistency through a novel framework using panoramic world proxies and semantically layered 3D mesh representations. This approach enables 360° immersive experiences, seamless mesh export for graphics pipelines, and...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 16
    Grounded-Segment-Anything

    Grounded-Segment-Anything

    Marrying Grounding DINO with Segment Anything & Stable Diffusion

    Grounded-Segment-Anything is a research-oriented project that combines powerful open-set object detection with pixel-level segmentation and subsequent creative workflows, effectively enabling detection, segmentation, and high-level vision tasks guided by free-form text prompts. The core idea behind the project is to pair Grounding DINO — a zero-shot object detector that can locate objects described by natural language — with Segment Anything Model (SAM), which can produce detailed masks for...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    WhisperSpeech

    WhisperSpeech

    An Open Source text-to-speech system built by inverting Whisper

    WhisperSpeech is an open-source text-to-speech system created by “inverting” OpenAI’s Whisper, reusing its strengths as a semantic audio model to generate speech instead of only transcribing it. The project aims to be for speech what Stable Diffusion is for images: powerful, hackable, and safe for commercial use, with code under Apache-2.0/MIT and models trained only on properly licensed data. Its architecture follows a token-based, multi-stage pipeline inspired by AudioLM and SPEAR-TTS: Whisper is used to produce semantic tokens, EnCodec compresses the waveform into acoustic tokens, and Vocos reconstructs high-fidelity audio from those tokens. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 18
    Flow Matching

    Flow Matching

    A PyTorch library for implementing flow matching algorithms

    flow_matching is a PyTorch library implementing flow matching algorithms in both continuous and discrete settings, enabling generative modeling via matching vector fields rather than diffusion. The underlying idea is to parameterize a flow (a time-dependent vector field) that transports samples from a simple base distribution to a target distribution, and train via matching of flows without requiring score estimation or noisy corruption—this can lead to more efficient or stable generative training. The library supports both continuous-time flows (via differential equations) and discrete-time analogues, giving flexibility in design and tradeoffs. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    AnimateDiff

    AnimateDiff

    Plug-n-play module turning text-to-image models into animation

    AnimateDiff is an open-source project designed to enhance text-to-image diffusion models by adding animation capabilities. It allows users to turn static images generated by popular text-to-image models into animated sequences without requiring additional model training. This plug-and-play tool is compatible with a wide range of community models and facilitates the generation of animation directly from pre-existing text-to-image models. It supports various configurations to create animations...
    Leader badge
    Downloads: 33 This Week
    Last Update:
    See Project
  • 20
    OnnxStream

    OnnxStream

    Lightweight inference library for ONNX files, written in C++

    The challenge is to run Stable Diffusion 1.5, which includes a large transformer model with almost 1 billion parameters, on a Raspberry Pi Zero 2, which is a microcomputer with 512MB of RAM, without adding more swap space and without offloading intermediate results on disk. The recommended minimum RAM/VRAM for Stable Diffusion 1.5 is typically 8GB. Generally, major machine learning frameworks and libraries are focused on minimizing inference latency and/or maximizing throughput, all of which at the cost of RAM usage. ...
    Downloads: 8 This Week
    Last Update:
    See Project
  • 21
    PromptSniffer

    PromptSniffer

    View Extract & Remove AI generation metadata with right click

    A powerful tool for reading, extracting, and removing AI generation metadata from image files. Specifically designed to handle metadata from AI image generation tools like ComfyUI, Stable Diffusion, SwarmUI, InvokeAI, and more. Core Functionality Read EXIF/Metadata: Extract and display comprehensive metadata from images AI Metadata Detection: Automatically identify and highlight AI generation metadata Metadata Removal: Strip AI generation metadata while preserving image quality Batch Processing: Handle multiple files with wildcard patterns Cross-Platform: Works on Windows, macOS, and Linux AI Tool Support ComfyUI: Detects and extracts workflow JSON data Stable Diffusion: Identifies prompts, parameters, and generation settings SwarmUI/StableSwarmUI: Handles JSON-formatted metadata Midjourney, DALL-E, NovelAI: Recognizes generation signatures Automatic1111, InvokeAI: Extracts generation parameters
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    1654 — encryption system

    1654 — encryption system

    1024-bit encryption system (functional safe or disk) | С++20

    This release marks the first production-ready publication of 1654, a dependency-free cryptographic vault engine built around a strict vault abstraction and a wide-state cryptographic design. 1654 is not a wrapper around existing crypto libraries. It is a self-contained system whose cryptographic behavior is fully defined within the repository. Key Characteristics Vault-oriented model Encrypted data is treated as a sealed container-safe with explicit operations, not as a...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 23
    KoboldCpp

    KoboldCpp

    Run GGUF models easily with a UI or API. One File. Zero Install.

    KoboldCpp is an easy-to-use AI text-generation software for GGML and GGUF models, inspired by the original KoboldAI. It's a single self-contained distributable that builds off llama.cpp and adds many additional powerful features.
    Leader badge
    Downloads: 423 This Week
    Last Update:
    See Project
  • 24
    Extended Dreambooth How-To Guides

    Extended Dreambooth How-To Guides

    Implementation of Dreambooth

    Extended Dreambooth How-To Guides is an implementation and extended toolkit for fine-tuning Stable Diffusion models using the DreamBooth technique, enabling users to train AI image generators to reproduce specific subjects, styles, or identities from a small set of reference images. The project adapts and expands upon earlier DreamBooth research by providing practical scripts, notebooks, and workflows that allow users to train personalized models on local machines, cloud environments, or platforms such as Google Colab. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    Prompt-to-Prompt

    Prompt-to-Prompt

    Latent Diffusion and Stable Diffusion Implementation

    Prompt-to-Prompt is a research codebase that demonstrates how to edit images generated by diffusion models using only changes to the text prompt. Instead of retraining or heavy fine-tuning, it manipulates the model’s cross-attention maps so the structure of the original image is largely preserved while semantics shift according to the revised prompt. The method supports gentle edits (e.g., style, color, lighting) as well as stronger semantic substitutions, and it can localize edits to...
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB