Showing 90 open source projects for "flow"

View related business solutions
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 1
    Claude-Flow

    Claude-Flow

    The leading agent orchestration platform for Claude

    Claude-Flow v2 Alpha is an advanced AI orchestration and automation framework designed for enterprise-grade, large-scale AI-driven development. It enables developers to coordinate multiple specialized AI agents in real time through a hive-mind architecture, combining swarm intelligence, neural reasoning, and a powerful set of 87 Modular Control Protocol (MCP) tools.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 2
    AI-Flow

    AI-Flow

    UI application to connect multiple AI models together

    Open-source tool to seamlessly connect multiple AI models in repeatable flow. While a live demo is available for convenience, for the best experience we recommend running the application directly on your local machine. AI Flow is an open source, user-friendly UI application that empowers you to seamlessly connect multiple AI models together, specifically leveraging the capabilities of ChatGPT. This unique tool paves the way to creating interactive networks of different AI models, fostering a multi-perspective response to various prompts.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 3
    Prompt flow

    Prompt flow

    Build high-quality LLM apps

    Prompt flow is a suite of development tools designed to streamline the end-to-end development cycle of LLM-based AI applications, from ideation, prototyping, testing, and evaluation to production deployment and monitoring. It makes prompt engineering much easier and enables you to build LLM apps with production quality.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 4
    Flow Matching

    Flow Matching

    A PyTorch library for implementing flow matching algorithms

    ...The library supports both continuous-time flows (via differential equations) and discrete-time analogues, giving flexibility in design and tradeoffs. It provides examples across modalities (images, toy 2D distributions) to help users understand how to apply flow matching in practice. The codebase includes notebooks illustrating 2D flow matching, discrete flows, and Riemannian flow matching on curved manifolds (e.g. flat torus) for non-Euclidean support.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Custom VMs From 1 to 96 vCPUs With 99.95% Uptime Icon
    Custom VMs From 1 to 96 vCPUs With 99.95% Uptime

    General-purpose, compute-optimized, or GPU/TPU-accelerated. Built to your exact specs.

    Live migration and automatic failover keep workloads online through maintenance. One free e2-micro VM every month.
    Try Free
  • 5
    Matcha-TTS

    Matcha-TTS

    A fast TTS architecture with conditional flow matching

    Matcha-TTS is a non-autoregressive neural text-to-speech architecture that uses conditional flow matching to generate speech quickly while maintaining natural quality. It models speech as an ODE-based generative process, and conditional flow matching lets it reach high-quality audio in only a few synthesis steps, which greatly reduces latency compared to score-matching diffusion approaches. The model is fully probabilistic, so it can generate diverse realizations of the same text while still sounding stable and intelligible. ...
    Downloads: 16 This Week
    Last Update:
    See Project
  • 6
    Flowise

    Flowise

    Drag & drop UI to build your customized LLM flow

    Open source UI visual tool to build your customized LLM flow using LangchainJS, written in Node Typescript/Javascript. Conversational agent for a chat model which utilizes chat-specific prompts and buffer memory. Open source is the core of Flowise, and it will always be free for commercial and personal usage. Flowise support different environment variables to configure your instance. You can specify the following variables in the .env file inside the packages/server folder.
    Downloads: 20 This Week
    Last Update:
    See Project
  • 7
    DeepSeek-OCR 2

    DeepSeek-OCR 2

    Visual Causal Flow

    DeepSeek-OCR-2 is the second-generation optical character recognition system developed to improve document understanding by introducing a “visual causal flow” mechanism, enabling the encoder to reorder visual tokens in a way that better reflects semantic structure rather than strict raster scan order. It is designed to handle complex layouts and noisy documents by giving the model causal reasoning capabilities that mimic human visual scanning behavior, enhancing OCR performance on documents with rich spatial structure. ...
    Downloads: 7 This Week
    Last Update:
    See Project
  • 8
    HunyuanVideo-I2V

    HunyuanVideo-I2V

    A Customizable Image-to-Video Model based on HunyuanVideo

    ...The repository includes pretrained weights, inference and sampling scripts, training code for LoRA effects, and support for parallel inference via xDiT. Resolution, video length, stability mode, flow shift, seed, CPU offload etc. Parallel inference support using xDiT for multi-GPU speedups. LoRA training / fine-tuning support to add special effects or customize generation.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 9
    Turing.jl

    Turing.jl

    Bayesian inference with probabilistic programming

    Bayesian inference with probabilistic programming.
    Downloads: 6 This Week
    Last Update:
    See Project
  • Full-stack observability with actually useful AI | Grafana Cloud Icon
    Full-stack observability with actually useful AI | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • 10
    GLM-TTS

    GLM-TTS

    Controllable & emotion-expressive zero-shot TTS

    GLM-TTS is an advanced text-to-speech synthesis system built on large language model technologies that focuses on producing high-quality, expressive, and controllable spoken output, including features like emotion modulation and zero-shot voice cloning. It uses a two-stage architecture where a generative LLM first converts text into intermediate speech token sequences and then a Flow-based neural model converts those tokens into natural audio waveforms, enabling rich prosody and voice character even for unseen speakers. The system introduces a multi-reward reinforcement learning framework that jointly optimizes for voice similarity, emotional expressiveness, pronunciation, and intelligibility, yielding output that can rival commercial options in naturalness and expressiveness. ...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 11
    MCP Server Chart

    MCP Server Chart

    A visualization mcp contains 25+ visual charts

    ...Out of the box it exposes more than 20–25 chart generators—covering staples like bar, line, area, histogram and pie, plus advanced visuals such as dual-axes, heatmaps, radar, flow and fishbone diagrams—so an AI client can request a chart and receive an image URL in return. The server can run over stdio for desktop IDEs or via SSE/“streamable” HTTP transport, making it easy to plug into MCP-capable clients and platforms (including Dify) without custom glue code. A simple CLI and environment variables control behavior, including disabling specific tools, selecting a visualization request service, or tagging a service instance for multi-tenant setups. ...
    Downloads: 5 This Week
    Last Update:
    See Project
  • 12
    WanGP

    WanGP

    AI video generator optimized for low VRAM and older GPUs use

    ...Wan2GP provides a full web-based interface that simplifies interaction with complex generative pipelines, making it easier to configure prompts, models, and rendering settings. It also integrates a wide range of utilities such as prompt enhancement, mask editing, motion design, and extraction tools for pose, depth, and flow data to support advanced video workflows.
    Downloads: 39 This Week
    Last Update:
    See Project
  • 13
    Zygote

    Zygote

    21st century AD

    Zygote provides source-to-source automatic differentiation (AD) in Julia, and is the next-gen AD system for the Flux differentiable programming framework. For more details and benchmarks of Zygote's technique, see our paper. You may want to check out Flux for more interesting examples of Zygote usage; the documentation here focuses on internals and advanced AD usage.
    Downloads: 7 This Week
    Last Update:
    See Project
  • 14
    tt-metal

    tt-metal

    TT-NN operator library, and TT-Metalium low level kernel programming

    ...Instead of following a traditional GPU model centered on massive thread parallelism, the platform is built around a grid of specialized compute nodes called Tensix cores, each with local SRAM, dedicated compute units, and multiple RISC-V control processors. The SDK provides the abstractions and APIs needed to manage data movement, compute kernels, memory coordination, and execution flow across this architecture.
    Downloads: 31 This Week
    Last Update:
    See Project
  • 15
    DeerFlow

    DeerFlow

    Deep Research framework, combining language models with tools

    ...“planner,” “searcher,” “coder,” “report generator”) that collaborate in a structured workflow, allowing tasks like literature reviews, data gathering, data analysis, code execution, and final report generation to be largely automated. It supports asynchronous task coordination, modular tool integration, and orchestrates the data flow between agents — making it suitable for large-scale or multi-stage research pipelines. Users can deploy it locally or on server infrastructure, integrate custom tools, and benefit from its flexible configuration.
    Downloads: 72 This Week
    Last Update:
    See Project
  • 16
    SGLang

    SGLang

    SGLang is a fast serving framework for large language models

    SGLang is a fast serving framework for large language models and vision language models. It makes your interaction with models faster and more controllable by co-designing the backend runtime and frontend language.
    Downloads: 7 This Week
    Last Update:
    See Project
  • 17
    Agentic Commerce Protocol (ACP)

    Agentic Commerce Protocol (ACP)

    Interaction model for connecting buyers to complete purchases

    ...The repository organizes the spec as human-readable RFCs plus machine-readable OpenAPI and JSON Schema definitions, along with worked examples and a changelog so integrators can track breaking changes. Practically, ACP defines three main pieces; a Product Feed for discovery, an Agentic Checkout API for stateful, in-conversation checkout, and a Delegated Payment flow so a merchant’s existing PSP securely handles payment credentials and settlement. Merchants remain the merchant-of-record—orders, fraud controls, payment authorization/capture, refunds, and post-purchase communication all stay on their systems while the agent surfaces status to the buyer.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 18
    IQuest-Coder-V1 Model Family

    IQuest-Coder-V1 Model Family

    New family of code large language models (LLMs)

    IQuest-Coder-V1 is a cutting-edge family of open-source large language models specifically engineered for code generation, deep code understanding, and autonomous software engineering tasks. These models range from tens of billions to smaller footprints and are trained on a novel code-flow multi-stage paradigm that captures how real software evolves over time — not just static code snapshots — giving them a deeper semantic understanding of programming logic. They support native long contexts up to 128K tokens, enabling them to reason across large codebases and multi-file interactions without context fragmentation, and include “Thinking” variants optimized for complex reasoning and “Loop” variants with recurrent mechanisms to improve inference efficiency. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    FLUX.2

    FLUX.2

    Official inference repo for FLUX.2 models

    ...It supports high-resolution output (up to ~4 megapixels), which allows for photography-quality images, detailed product shots, infographics or UI mockups rather than just low-resolution drafts. FLUX.2 is built with a modern architecture (a flow-matching transformer + a revamped VAE + a strong vision-language encoder), enabling strong prompt adherence, correct rendering of text/typography in images, reliable lighting, layout, and physical realism, and consistent style/character/product identity across multiple generations or edits.
    Downloads: 50 This Week
    Last Update:
    See Project
  • 20
    Lemonade

    Lemonade

    Lemonade helps users run local LLMs with the highest performance

    Lemonade is a local LLM runtime that aims to deliver the highest possible performance on your own hardware by auto-configuring state-of-the-art inference engines for both NPUs and GPUs. The project positions itself as a “local LLM server” you can run on laptops and workstations, abstracting away backend differences while giving you a single place to serve and manage models. Its README emphasizes real-world adoption across startups, research groups, and large companies, signaling a focus on...
    Downloads: 14 This Week
    Last Update:
    See Project
  • 21
    Mistral Vibe CLI

    Mistral Vibe CLI

    Minimal CLI coding agent by Mistral

    Mistral Vibe is an AI-powered “vibe-coding” command-line interface (CLI) and coding-assistant framework built by Mistral AI to let developers write, refactor, search, and manage code through natural language and context-aware automation, rather than manual typing only. It aims to take developers out of repetitive boilerplate and let them stay “in the flow”: you can ask the tool to generate functions, refactor code, search across the codebase, manipulate files, commit changes via Git, or run commands — all from a unified CLI interface. Behind the scenes, it leverages Mistral’s coding-optimized LLM stack (including models tuned for code understanding and generation), with project-wide context awareness: it scans your file structure, Git status, and recent history to inform suggestions so that generated code aligns with existing context.
    Downloads: 39 This Week
    Last Update:
    See Project
  • 22
    Firecrawl MCP Server

    Firecrawl MCP Server

    Adds powerful web scraping and search to Cursor and Claude

    ...The server is designed to run with Firecrawl’s hosted API or self-hosted deployments, making it flexible for enterprise data-governance requirements. Built-in behaviors include JavaScript rendering, automatic retries, and streamable HTTP so long pages and large crawls can flow incrementally into agents. Because it’s an MCP server, clients get typed inputs/outputs and standardized error handling rather than ad-hoc prompt instructions. The repository is active, widely starred, and includes quick starts that make it easy to add web research to an agent stack.
    Downloads: 11 This Week
    Last Update:
    See Project
  • 23
    LangGraph

    LangGraph

    Build resilient language agents as graphs

    ...LangGraph allows you to define flows that involve cycles, essential for most agentic architectures, differentiating it from DAG-based solutions. As a very low-level framework, it provides fine-grained control over both the flow and state of your application, crucial for creating reliable agents. Additionally, LangGraph includes built-in persistence, enabling advanced human-in-the-loop and memory features.
    Downloads: 11 This Week
    Last Update:
    See Project
  • 24
    OpenClaw Opik Observability Plugin

    OpenClaw Opik Observability Plugin

    Official plugin for OpenClaw that exports agent traces to Opik

    ...Each time an AI agent performs an action—such as calling a large language model, invoking a tool, accessing memory, or delegating to a sub-agent—the plugin records the full interaction and sends it to Opik for analysis and visualization. This allows developers to inspect inputs, outputs, token usage, latency, and execution flow across complex multi-step agent workflows. The goal of the project is to provide transparency into the internal reasoning and operational pipeline of agent systems so developers can diagnose failures, control costs, and improve reliability.
    Downloads: 9 This Week
    Last Update:
    See Project
  • 25
    WebLLM

    WebLLM

    Bringing large-language models and chat to web browsers

    WebLLM is a modular, customizable javascript package that directly brings language model chats directly onto web browsers with hardware acceleration. Everything runs inside the browser with no server support and is accelerated with WebGPU. We can bring a lot of fun opportunities to build AI assistants for everyone and enable privacy while enjoying GPU acceleration. WebLLM offers a minimalist and modular interface to access the chatbot in the browser. The WebLLM package itself does not come...
    Downloads: 4 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • 3
  • 4
  • Next
MongoDB Logo MongoDB