• Orchestrate Your AI Agents with Zenflow Icon
    Orchestrate Your AI Agents with Zenflow

    The multi-agent workflow engine for modern teams. Zenflow executes coding, testing, and verification with deep repo awareness

    Zenflow orchestrates AI agents like a real engineering system. With parallel execution, spec-driven workflows, and deep multi-repo understanding, agents plan, implement, test, and verify end-to-end. Upgrade to AI workflows that work the way your team does.
    Try free now
  • Our Free Plans just got better! | Auth0 Icon
    Our Free Plans just got better! | Auth0

    With up to 25k MAUs and unlimited Okta connections, our Free Plan lets you focus on what you do best—building great apps.

    You asked, we delivered! Auth0 is excited to expand our Free and Paid plans to include more options so you can focus on building, deploying, and scaling applications without having to worry about your security. Auth0 now, thank yourself later.
    Try free now
  • 1
    Letta

    Letta

    Letta (formerly MemGPT) is a framework for creating LLM services

    Letta is an AI-powered task automation framework designed to handle workflow automation, natural language commands, and AI-driven decision-making.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 2
    LangGraph

    LangGraph

    Build resilient language agents as graphs

    LangGraph is a library for building stateful, multi-actor applications with LLMs, used to create agent and multi-agent workflows. Compared to other LLM frameworks, it offers these core benefits: cycles, controllability, and persistence. LangGraph allows you to define flows that involve cycles, essential for most agentic architectures, differentiating it from DAG-based solutions. As a very low-level framework, it provides fine-grained control over both the flow and state of your application, crucial for creating reliable agents. Additionally, LangGraph includes built-in persistence, enabling advanced human-in-the-loop and memory features.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 3
    Phidata

    Phidata

    Build multi-modal Agents with memory, knowledge, tools and reasoning

    Phidata is an open source platform for building, deploying, and monitoring AI agents. It enables users to create domain-specific agents with memory, knowledge, and external tools, enhancing AI capabilities for various tasks. The platform supports a range of large language models and integrates seamlessly with different databases, vector stores, and APIs. Phidata offers pre-configured templates to accelerate development and deployment, allowing users to quickly go from building agents to shipping them into production. It includes features like real-time monitoring, agent evaluations, and performance optimization tools, ensuring the reliability and scalability of AI solutions. Phidata also allows developers to bring their own cloud infrastructure, offering flexibility for custom setups. The platform provides robust support for enterprises, including security features, agent guardrails, and automated DevOps for smoother deployment processes.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 4
    Self-Operating Computer

    Self-Operating Computer

    A framework to enable multimodal models to operate a computer

    The Self-Operating Computer Framework is an innovative system that enables multimodal models to autonomously operate a computer by interpreting the screen and executing mouse and keyboard actions to achieve specified objectives. This framework is compatible with various multimodal models and currently integrates with GPT-4o, o1, Gemini Pro Vision, Claude 3, and LLaVa. Notably, it was the first known project to implement a multimodal model capable of viewing and controlling a computer screen. The framework supports features like Optical Character Recognition (OCR) and Set-of-Mark (SoM) prompting to enhance visual grounding capabilities. It is designed to be compatible with macOS, Windows, and Linux (with X server installed), and is released under the MIT license.
    Downloads: 5 This Week
    Last Update:
    See Project
  • Build Securely on AWS with Proven Frameworks Icon
    Build Securely on AWS with Proven Frameworks

    Lay a foundation for success with Tested Reference Architectures developed by Fortinet’s experts. Learn more in this white paper.

    Moving to the cloud brings new challenges. How can you manage a larger attack surface while ensuring great network performance? Turn to Fortinet’s Tested Reference Architectures, blueprints for designing and securing cloud environments built by cybersecurity experts. Learn more and explore use cases in this white paper.
    Download Now
  • 5
    c/ua

    c/ua

    c/ua is the Docker Container for Computer-Use AI Agents

    Cua is a Docker-based framework that facilitates the deployment and management of computer-use AI agents. It provides a sandboxed environment where agents can perform tasks on macOS and Linux virtual machines, supporting various AI models and ensuring safe execution. Cua is particularly useful for developers looking to test and run AI agents in controlled settings.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 6
    gpt-engineer

    gpt-engineer

    Full stack AI software engineer

    gpt-engineer is an open-source platform designed to help developers automate the software development process using natural language. The platform allows users to specify software requirements in plain language, and the AI generates and executes the corresponding code. It can also handle improvements and iterative development, giving users more control over the software they’re building. Built with a terminal-based interface, gpt-engineer is customizable, enabling developers to experiment with AI-assisted programming and refine their development process. It is especially useful for automating the coding and iterative feedback loop in software development.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 7
    AskUI Vision Agent

    AskUI Vision Agent

    Enable AI to control your desktop, mobile and HMI devices

    AskUI’s Vision Agent is an automation framework that allows you—and AI agents—to control real desktops, mobile devices, and HMI systems by perceiving the UI and performing actions like clicking, typing, scrolling, and drag-and-drop. It is designed for multi-platform compatibility and supports multiple AI models so you can tailor perception and decision-making to your workload. The repository presents a feature overview, sample media, and frequent release notes, which show ongoing improvements such as CORS checks and other operational tweaks. The broader AskUI documentation covers the Python Vision Agent along with suite services and inference APIs, indicating a productized ecosystem rather than a single library. Community-curated lists also recognize Vision Agent as part of the broader “GUI agents” landscape, placing it among other computer-use agents.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 8
    Atomic Agents

    Atomic Agents

    Building AI agents, atomically

    The Atomic Agents framework is designed around the concept of atomicity to be an extremely lightweight and modular framework for building Agentic AI pipelines and applications without sacrificing developer experience and maintainability. The framework provides a set of tools and agents that can be combined to create powerful applications. It is built on top of Instructor and leverages the power of Pydantic for data and schema validation and serialization. All logic and control flows are written in Python, enabling developers to apply familiar best practices and workflows from traditional software development without compromising flexibility or clarity.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 9
    CopilotKit

    CopilotKit

    Build in-app AI chatbots, and AI-powered Textareas

    A bridge between your copilot and your app. A programmable 2-way bridge between your copilot, and your application state (client & cloud). Supports 3rd party integrations (e.g. Salesforce, Zendesk, etc.). Plug-and-play, fully customizable, copilot infrastructure. Build in-app AI chatbots that can "see" the current app state + take action inside your app. The AI chatbot can talk to your app frontend & backend, and to 3rd party services (Salesforce, Dropbox, etc.) via plugins. Autocompletion + AI editing + generate from scratch. Indexed on your users' content.
    Downloads: 4 This Week
    Last Update:
    See Project
  • Desktop and Mobile Device Management Software Icon
    Desktop and Mobile Device Management Software

    It's a modern take on desktop management that can be scaled as per organizational needs.

    Desktop Central is a unified endpoint management (UEM) solution that helps in managing servers, laptops, desktops, smartphones, and tablets from a central location.
    Learn More
  • 10
    Diplomacy Cicero

    Diplomacy Cicero

    Code for Cicero, an AI agent that plays the game of Diplomacy

    The project is the codebase for an AI agent named Cicero developed by Facebook Research. It is designed to play the board game Diplomacy by combining open-domain natural language negotiation with strategic planning. The repository includes training code, model checkpoints, and infrastructure for both language modelling (via the ParlAI framework) and reinforcement learning for strategy agents. It supports two variants: Cicero (which handles full “press” negotiation) and Diplodocus (a variant focused on no-press diplomacy) as described in the README. The codebase is implemented primarily in Python with performance-critical components in C++ (via pybind11 bindings) and is configured to run in a high‐GPU cluster environment. Configuration is managed via protobuf files to define tasks such as self-play, benchmark agent comparisons, and RL training. The project is now archived and read-only, reflecting that it is no longer actively developed but remains publicly available for research use.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 11
    Gemini Fullstack LangGraph Quickstart

    Gemini Fullstack LangGraph Quickstart

    Get started w/ building Fullstack Agents using Gemini 2.5 & LangGraph

    gemini-fullstack-langgraph-quickstart is a fullstack reference application from Google DeepMind’s Gemini team that demonstrates how to build a research-augmented conversational AI system using LangGraph and Google Gemini models. The project features a React (Vite) frontend and a LangGraph/FastAPI backend designed to work together seamlessly for real-time research and reasoning tasks. The backend agent dynamically generates search queries based on user input, retrieves information via the Google Search API, and performs reflective reasoning to identify knowledge gaps. It then iteratively refines its search until it produces a comprehensive, well-cited answer synthesized by the Gemini model. The repository provides both a browser-based chat interface and a command-line script (cli_research.py) for executing research queries directly. For production deployment, the backend integrates with Redis and PostgreSQL to manage persistent memory, streaming outputs, & background task coordination.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 12
    Olares

    Olares

    Olares: An Open-Source Sovereign Cloud OS for Local AI

    Olares is an AI-powered chatbot framework designed to support real-time natural language understanding and response generation.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 13
    Open Interface

    Open Interface

    Control Any Computer Using LLMs

    Open Interface is a cross-platform application that allows users to control their computers using large language models (LLMs). By sending user requests to an LLM backend, it determines the necessary steps and executes them by simulating keyboard and mouse inputs. The system can adjust its actions based on real-time feedback, providing a self-driving computer experience.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 14
    AIChat

    AIChat

    All-in-one LLM CLI tool featuring Shell Assistant

    AIChat is a lightweight terminal-based chatbot powered by GPT models, enabling AI-driven conversations directly from the command line.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 15
    Alan AI

    Alan AI

    In-App assistant SDK to build a multimodal conversational UX websites

    Quickly add voice to your app with the Alan Platform. Create an in-app voice assistant to enable human-like conversations and provide a personalized voice experience for every user. Alan is a conversational voice AI platform that lets you create an intelligent voice assistant for your app. It offers all the necessary tools to design, embed, and host your voice solutions. A powerful web-based IDE where you can write, test and debug dialog scenarios for your voice assistant or chatbot. Alan's AI-backend powered by the industry’s best Automatic Speech Recognition (ASR), Natural Language Understanding (NLU) and Speech Synthesis. The Alan Cloud provisions and handles the infrastructure required to maintain your voice deployments and perform all the voice processing tasks. To voice enable your app, you only need to get the Alan Client SDK and drop it to your app. No need to plan for, deploy and maintain any infrastructure or speech components - the Alan Platform does the bulk of the work.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 16
    AutoGen

    AutoGen

    An Open-Source Programming Framework for Agentic AI

    AutoGen is an open-source programming framework for building AI agents and facilitating cooperation among multiple agents to solve tasks. AutoGen aims to provide an easy-to-use and flexible framework for accelerating development and research on agentic AI, like PyTorch for Deep Learning. It offers features such as agents that can converse with other agents, LLM and tool use support, autonomous and human-in-the-loop workflows, and multi-agent conversation patterns. AutoGen provides multi-agent conversation framework as a high-level abstraction. With this framework, one can conveniently build LLM workflows. AutoGen offers a collection of working systems spanning a wide range of applications from various domains and complexities. AutoGen supports enhanced LLM inference APIs, which can be used to improve inference performance and reduce cost.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 17
    Ax

    Ax

    Build LLM powered Agents and "Agentic workflows"

    Build intelligent agents quickly — inspired by the power of "Agentic workflows" and the Stanford DSPy paper. Seamlessly integrates with multiple LLMs and VectorDBs to build RAG pipelines or collaborative agents that can solve complex problems. Advanced features streaming validation, multi-modal DSPy, etc. We've renamed from "llmclient" to "ax" to highlight our focus on powering agentic workflows. We agree with many experts like "Andrew Ng" that agentic workflows are the key to unlocking the true power of large language models and what can be achieved with in-context learning. Also, we are big fans of the Stanford DSPy paper, and this library is the result of all of this coming together to build a powerful framework for you to build with.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 18
    CAMEL AI

    CAMEL AI

    Finding the Scaling Law of Agents. A multi-agent framework

    The rapid advancement of conversational and chat-based language models has led to remarkable progress in complex task-solving. However, their success heavily relies on human input to guide the conversation, which can be challenging and time-consuming. This paper explores the potential of building scalable techniques to facilitate autonomous cooperation among communicative agents and provide insight into their "cognitive" processes. To address the challenges of achieving autonomous cooperation, we propose a novel communicative agent framework named role-playing. Our approach involves using inception prompting to guide chat agents toward task completion while maintaining consistency with human intentions. We showcase how role-playing can be used to generate conversational data for studying the behaviors and capabilities of chat agents, providing a valuable resource for investigating conversational language models.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 19
    Dendrite

    Dendrite

    Tools to build web AI agents that can authenticate

    Dendrite Python SDK is a toolkit for building web AI agents that can authenticate, interact with, and extract data from any website, facilitating web automation tasks.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 20
    IntentKit

    IntentKit

    An open and fair framework for everyone to build AI agents

    IntentKit is a natural language understanding (NLU) library focused on intent recognition and entity extraction, enabling developers to build conversational AI applications.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 21
    Magentic UI

    Magentic UI

    A research prototype of a human-centered web agent

    Magentic-UI is a research prototype developed by Microsoft that serves as a human-centered interface powered by a multi-agent system. It enables users to automate complex web tasks, such as browsing, form filling, and data analysis, while maintaining control over the process. The system emphasizes transparency and user involvement, making it suitable for tasks requiring both automation and human oversight.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 22
    Multi-Agent Particle Envs

    Multi-Agent Particle Envs

    Code for a multi-agent particle environment used in a paper

    Multiagent Particle Environments is a lightweight framework for simulating multi-agent reinforcement learning tasks in a continuous observation space with discrete action settings. It was originally developed by OpenAI and used in the influential paper Multi-Agent Actor-Critic for Mixed Cooperative-Competitive Environments. The environment provides simple particle-based worlds with simulated physics, where agents can move, communicate, and interact with each other. Scenarios are designed to model cooperative, competitive, and mixed interactions among agents, making it useful for testing algorithms in multi-agent settings. The project includes built-in scenarios such as navigation to landmarks, cooperative tasks, and adversarial setups. Although archived, its concepts and code structure remain foundational for more advanced libraries like PettingZoo, which extended and maintained this environment.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 23
    Nanobrowser

    Nanobrowser

    Open-Source Chrome extension for AI-powered web automation

    Nanobrowser is an open-source AI web automation tool that runs in your browser. A free alternative to OpenAI Operator with flexible LLM options and a multi-agent system. Nanobrowser, as a chrome extension, delivers premium web automation capabilities while keeping you in complete control. No subscription fees or hidden costs. Just install and use your own API keys, and you only pay what you use with your own API keys. Everything runs in your local browser. Your credentials stay with you, never shared with any cloud service. Connect to your preferred LLM providers with the freedom to choose different models for different agents.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 24
    OpenAI Assistants Quickstart

    OpenAI Assistants Quickstart

    OpenAI Assistants API quickstart with Next.js

    openai-assistants-quickstart is a template for using the Assistants API in a Next.js app, demonstrating streaming, tool use, and function calling in one place. The repository includes multiple example pages that each showcase specific capabilities, while all examples share the same underlying assistant with all capabilities enabled. The primary chat logic lives in the Chat component at app/components/chat.tsx, which manages rendering, streaming, and forwarding function calls. Server handlers for threads are provided under api/assistants/threads/..., giving a reference for wiring the API into Next.js routes. The Chat component can be copied directly into other projects, along with its styles from app/components/chat.module.css. Example pages include a basic chat, a function calling demo, a file search demo, and a full-featured example, allowing developers to explore each feature in isolation or together.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 25
    OpenAI Realtime Agents

    OpenAI Realtime Agents

    This is a simple demonstration of more advanced, agentic patterns

    This repository demonstrates how to build low-latency, streaming “voice + chat” agents using OpenAI’s Realtime API combined with the OpenAI Agents SDK. The demo shows patterns for connecting a realtime voice stream (audio in/out) with agents that can use tools, maintain state, and orchestrate multi-agent workflows. The SDK offers abstractions such as agent orchestration, event handling, handoffs, state management, and guardrails, tailored to support realtime, conversational systems. The demo includes a Next.js frontend for browser interaction and likely a backend component to orchestrate realtime sessions and agent logic. It also supports a “Chat-Supervisor” pattern where a lightweight realtime chat agent handles user interactions and delegates more complex reasoning or tool usage to a stronger textual model (e.g. GPT-4). Because realtime agents are still a beta feature, the code and API surface are subject to changes and may evolve.
    Downloads: 3 This Week
    Last Update:
    See Project