• MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • $300 in Free Credit Towards Top Cloud Services Icon
    $300 in Free Credit Towards Top Cloud Services

    Build VMs, containers, AI, databases, storage—all in one place.

    Start your project in minutes. After credits run out, 20+ products include free monthly usage. Only pay when you're ready to scale.
    Get Started
  • 1
    MLX Engine

    MLX Engine

    LM Studio Apple MLX engine

    MLX Engine is the Apple MLX-based inference backend used by LM Studio to run large language models efficiently on Apple Silicon hardware. Built on top of the mlx-lm and mlx-vlm ecosystems, the engine provides a unified architecture capable of supporting both text-only and multimodal models. Its design focuses on high-performance on-device inference, leveraging Apple’s MLX stack to accelerate computation on M-series chips.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 2
    FlowGram

    FlowGram

    Extensible workflow development framework

    FlowGram is an open-source, node-based workflow development framework and toolkit aimed at helping developers build custom AI-workflow platforms or automation systems through a visual, drag-and-drop interface. Instead of shipping as a ready-made product, it provides the building blocks — a canvas for wiring together nodes, a form engine for configuring node parameters, a variable-scope and type-inference engine, and a set of “materials” (pre-built node types such as code execution, conditional logic, LLM calls, etc.) that can be composed into larger workflows. This makes FlowGram highly flexible: you can prototype data-processing pipelines, AI-agent flows, automation scripts, or even business process automation without writing all the plumbing yourself. ...
    Downloads: 6 This Week
    Last Update:
    See Project
  • 3
    Perplexica

    Perplexica

    Perplexica is an AI-powered answering engine.

    Perplexica is a privacy-focused AI answering engine like Perplexity that you can self-host on your own hardware for private, source-cited web research. It combines live internet search results with AI models, letting you use local LLMs via Ollama or connect to providers like OpenAI, Claude, Gemini, and Groq. Powered by SearxNG, it aggregates results from multiple search engines while keeping your identity and queries private.
    Downloads: 28 This Week
    Last Update:
    See Project
  • 4
    Search with Lepton

    Search with Lepton

    Lightweight demo to build a conversational AI search engine quickly

    ...It includes both a backend service written in Python and a web interface that allows users to interact with the search engine in a conversational format. Developers can configure different search providers and language models through environment variables, making it flexible for experimentation and prototyping.
    Downloads: 3 This Week
    Last Update:
    See Project
  • Earn up to 16% annual interest with Nexo. Icon
    Earn up to 16% annual interest with Nexo.

    More flexibility. More control.

    Generate interest, access liquidity without selling, and execute trades seamlessly. All in one platform. Geographic restrictions, eligibility, and terms apply.
    Get started with Nexo.
  • 5
    QMD

    QMD

    mini cli search engine for your docs, knowledge bases, etc.

    QMD is a powerful and lightweight command-line tool that acts as an on-device search engine for your personal knowledge base, allowing you to index and search files like Markdown notes, meeting transcripts, technical documentation, and other text collections without depending on cloud services. Designed to keep all search activity local, it combines classic full-text search techniques with modern semantic features such as vector similarity and hybrid ranking so that queries return not just literal matches but conceptually relevant results. ...
    Downloads: 10 This Week
    Last Update:
    See Project
  • 6
    NextChat

    NextChat

    Light and Fast AI Assistant

    ...It supports context-aware chatting using page content, quick queries, and code assistance without opening a new tab. Developers appreciate its customization options—like hotkeys, themes, and engine selection—while everyday users enjoy handy tools for writing, summarizing, or brainstorming on the fly.
    Downloads: 36 This Week
    Last Update:
    See Project
  • 7
    OpenWork AI

    OpenWork AI

    An open-source alternative to Claude Cowork, powered by opencode

    ...This project also includes moderation and reputation layers so that contributor trust and quality can be assessed and integrated into future task assignments. Rather than a single monolithic workflow engine, it emphasizes openness — providing APIs and interfaces so communities can build custom dashboards, integrate specialized agents, or add bespoke evaluation criteria.
    Downloads: 56 This Week
    Last Update:
    See Project
  • 8
    LaVague

    LaVague

    Framework for building AI agents that automate complex web tasks

    ...These agents can navigate web pages, retrieve information, fill out forms, and execute multi-step workflows automatically. LaVague is centered around a World Model that analyzes the current webpage state and determines the next set of instructions, combined with an Action Engine that converts those instructions into executable automation code. It can use browser automation tools such as Selenium or Playwright to interact with websites programmatically. Developers can integrate various language models and configure the agent’s reasoning and execution behavior to suit different automation scenarios.
    Downloads: 7 This Week
    Last Update:
    See Project
  • 9
    AstronRPA

    AstronRPA

    Agent-ready RPA suite with visual workflow automation tools engine

    Astron RPA is an enterprise-grade robotic process automation platform designed to help organizations and developers build automated workflows for desktop and web applications. It provides a visual workflow designer that supports low-code and no-code development, allowing users to create automation processes through a drag-and-drop interface instead of writing extensive code. It enables automation of common desktop software and browser-based tasks, making it suitable for repetitive business...
    Downloads: 5 This Week
    Last Update:
    See Project
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 10
    OpenMemory

    OpenMemory

    Local long-term memory engine for AI apps with persistent storage

    OpenMemory is a self-hosted memory engine designed to provide long-term, persistent storage for AI and LLM-powered applications. It enables developers to give otherwise stateless models a structured memory layer that can store, retrieve, and manage contextual information over time. OpenMemory is built around a hierarchical memory architecture that organizes data into semantic sectors and connects them through a graph-based structure for efficient retrieval.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 11
    Mastra

    Mastra

    The TypeScript AI agent framework

    ...Model routing lets you connect to dozens of providers (OpenAI, Anthropic, Gemini, and others) through a single standardized interface, while agents orchestrate LLM calls and tools to solve open-ended tasks with internal reasoning loops. When explicit control is needed, Mastra’s workflow engine uses a graph-style API (.then(), .branch(), .parallel()) to orchestrate multi-step processes.
    Downloads: 7 This Week
    Last Update:
    See Project
  • 12
    Satori

    Satori

    Enlightened library to convert HTML and CSS to SVG

    ...In general, only these static and visible elements and properties that are implemented. Also, Satori does not guarantee that the SVG will 100% match the browser-rendered HTML output since Satori implements its own layout engine based on the SVG 1.1 spec.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 13
    Dagger

    Dagger

    Containerized automation engine for programmable CI/CD workflows

    Dagger is an open source automation engine designed to build, test, and deliver software in a consistent and programmable way. It enables developers to define software delivery workflows using code instead of complex shell scripts or configuration files. Dagger executes tasks inside containers, ensuring that automation runs in identical environments across local machines, CI servers, or cloud infrastructure.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    Farfalle

    Farfalle

    AI search engine - self-host with local or cloud LLMs

    Farfalle is an open-source AI-powered search engine designed to provide an answer-centric search experience similar to modern conversational search systems. The project integrates large language models with multiple search APIs so that the system can gather information from external sources and synthesize responses into concise answers. It can run either with local language models or with cloud-based providers, allowing developers to deploy it privately or integrate with hosted AI services. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    node-llama-cpp

    node-llama-cpp

    Run AI models locally on your machine with node.js bindings for llama

    node-llama-cpp is a JavaScript and Node.js binding that allows developers to run large language models locally using the high-performance inference engine provided by llama.cpp. The library enables applications built with Node.js to interact directly with local LLM models without requiring a remote API or external service. By using native bindings and optimized model execution, the framework allows developers to integrate advanced language model capabilities into desktop applications, server software, and command-line tools. ...
    Downloads: 18 This Week
    Last Update:
    See Project
  • 16
    WrenAI

    WrenAI

    Open-source SQL AI Agent for Text-to-SQL. Make Text2SQL Easy

    Wren AI is a SQL AI Agent for data teams to get results and insights faster by asking business questions without writing SQL, and it's open-source. Wren AI has implemented a semantic engine architecture to provide the LLM context of your business; you can easily establish a logical presentation layer on your data schema that helps LLM learn more about your business context. With Wren AI, you can process metadata, schema, terminology, data relationships, and the logic behind calculations and aggregations with “Modeling Definition Language”, to generate accurate SQL queries with semantic context. ...
    Downloads: 21 This Week
    Last Update:
    See Project
  • 17
    Nestia

    Nestia

    NestJS Helper + AI Chatbot Development

    Nestia is a high-performance toolkit and ecosystem built on top of NestJS that enhances backend development by introducing strongly typed APIs, automated SDK generation, and advanced tooling for scalable server applications. It is designed to eliminate much of the boilerplate typically associated with API development by leveraging pure TypeScript types to automatically generate validation logic, API documentation, and client SDKs. One of its defining advantages is its focus on performance...
    Downloads: 10 This Week
    Last Update:
    See Project
  • 18
    Basic Memory

    Basic Memory

    Persistent AI memory using local Markdown knowledge graphs

    Basic Memory is an open source knowledge system that turns AI conversations into persistent, structured knowledge you control. Instead of losing context after each chat, it stores information as simple Markdown files on your device, allowing both you and AI to read and write to the same knowledge base. It uses the Model Context Protocol (MCP) so compatible AI tools can access, update, and build on your notes across sessions. Basic Memory creates a semantic knowledge graph by linking related...
    Downloads: 17 This Week
    Last Update:
    See Project
  • 19
    PasteGuard

    PasteGuard

    Masks sensitive data and secrets before they reach AI

    ...PasteGuard supports two primary modes: mask mode, which anonymizes data and still uses external APIs; and route mode, which forwards sensitive requests to a local LLM inference engine while sending the rest to the cloud. It can be self-hosted via Docker, works with a wide range of SDKs and tools, and includes a browser extension for automatic protection in everyday AI chats.
    Downloads: 8 This Week
    Last Update:
    See Project
  • 20
    CodeMachine

    CodeMachine

    CLI tool for multi-agent workflows and automated code generation

    CodeMachine CLI is a command-line orchestration engine designed to run coordinated multi-agent workflows locally. It enables developers to transform high-level specifications into production-ready code by managing planning, architecture, implementation, testing, and validation within a unified environment. CodeMachine CLI supports parallel execution through multiple specialized agents, allowing faster development cycles and scalable automation.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 21
    Memori

    Memori

    SQL-native memory layer enabling persistent context for AI agents

    Memori is an open source SQL-native memory engine designed to add persistent memory capabilities to AI applications, large language models, and multi-agent systems. It provides a memory layer that automatically captures conversations and interactions between users and AI models, allowing systems to retain knowledge across sessions instead of operating statelessly. It extracts structured information such as facts, preferences, rules, and summaries from interactions and stores them in standard SQL databases for later retrieval. ...
    Downloads: 10 This Week
    Last Update:
    See Project
  • 22
    Mantic.sh

    Mantic.sh

    A structural code search engine for Al agents

    Mantic.sh is a context-aware, structural code search engine designed specifically for use with AI coding agents and developers who need deep, semantically relevant search across large codebases. Unlike traditional text-based search tools that mainly match keywords, Mantic.sh understands code structure and meaning by combining syntactic heuristics with neural semantic reranking to produce results that reflect conceptual relevance, which helps find functions, definitions, and patterns that literal search might miss. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    AntV Infographic

    AntV Infographic

    Declarative engine for generating AI-powered infographic visuals

    AntV Infographic is a declarative infographic generation and rendering framework designed to transform structured data into visually rich infographic outputs. It provides a custom domain-specific language that allows developers and AI systems to describe infographic layouts in a concise and human-readable syntax. It focuses on simplifying data storytelling by enabling fast creation of professional-quality visuals without requiring complex design workflows. It includes a rendering pipeline...
    Downloads: 8 This Week
    Last Update:
    See Project
  • 24
    wllama

    wllama

    WebAssembly binding for llama.cpp - Enabling on-browser LLM inference

    wllama is a WebAssembly-based library that enables large language model inference directly inside a web browser. Built as a binding for the llama.cpp inference engine, the project allows developers to run LLM models locally without requiring a server backend or dedicated GPU hardware. The library leverages WebAssembly SIMD capabilities to achieve efficient execution within modern browsers while maintaining compatibility across platforms. By running models locally on the user’s device, wllama enables privacy-preserving AI applications that do not require sending data to remote servers. ...
    Downloads: 6 This Week
    Last Update:
    See Project
  • 25
    MemFree

    MemFree

    Hybrid AI Search Engine & AI Page Generator

    memfree is an open source hybrid AI search engine and page generation platform designed to help users retrieve information from both personal knowledge bases and the public web through a unified interface. The project combines retrieval-augmented search with AI summarization to deliver concise answers instead of forcing users to manually sift through multiple sources. It supports multiple AI models and search providers, enabling flexible configuration depending on cost, performance, or privacy requirements. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • Next
MongoDB Logo MongoDB