Showing 4 open source projects for "tool chains"

View related business solutions
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • Ship AI Apps Faster with Vertex AI Icon
    Ship AI Apps Faster with Vertex AI

    Go from idea to deployed AI app without managing infrastructure. Vertex AI offers one platform for the entire AI development lifecycle.

    Ship AI apps and features faster with Vertex AI—your end-to-end AI platform. Access Gemini 3 and 200+ foundation models, fine-tune for your needs, and deploy with enterprise-grade MLOps. Build chatbots, agents, or custom models. New customers get $300 in free credit.
    Try Vertex AI Free
  • 1
    AutoGPT

    AutoGPT

    Powerful tool that lets you create and run intelligent agents

    AutoGPT is an experimental open-source application showcasing the capabilities of the GPT-4 language model. This program, driven by GPT-4, chains together LLM "thoughts", to autonomously achieve whatever goal you set. As one of the first examples of GPT-4 running fully autonomously, AutoGPT pushes the boundaries of what is possible with AI.
    Downloads: 13 This Week
    Last Update:
    See Project
  • 2
    Code-Mode

    Code-Mode

    Plug-and-play library to enable agents to call MCP and UTCP tools

    Code-Mode is a plug-and-play library that lets AI agents call tools by executing TypeScript (or via a Python wrapper) instead of making many individual function calls. Its core philosophy is that language models are very good at writing code, so rather than exposing hundreds of separate tool endpoints, you give the model a single “code execution” tool that has access to your full toolkit through code. This approach can dramatically reduce the number of tool-call iterations needed in complex workflows, turning multi-step call chains into a single code execution with internal branching and loops. The repository contains both TypeScript and Python libraries, plus a code-mode-mcp component for integrating with MCP and UTCP ecosystems. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    GLM-4.6V

    GLM-4.6V

    GLM-4.6V/4.5V/4.1V-Thinking, towards versatile multimodal reasoning

    GLM-4.6V represents the latest generation of the GLM-V family and marks a major step forward in multimodal AI by combining advanced vision-language understanding with native “tool-call” capabilities, long-context reasoning, and strong generalization across domains. Unlike many vision-language models that treat images and text separately or require intermediate conversions, GLM-4.6V allows inputs such as images, screenshots or document pages directly as part of its reasoning pipeline — and...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 4
    n8n-MCP

    n8n-MCP

    A MCP for Claude Desktop / Claude Code / Windsurf / Cursor

    ...It exposes structured knowledge of n8n nodes and operations so an agent can reason about workflows, parameters, and executions without scraping docs or guessing API shapes. The server focuses on making Claude Desktop (and other MCP-capable clients) “n8n-literate,” enabling tasks such as inspecting existing workflows, proposing node chains, and validating configuration before runs. It ships with organized resources and tool definitions that map cleanly to n8n’s ecosystem, improving reliability compared with ad-hoc prompt patterns. The project targets practical agent ops: safer mutations, better error reporting, and predictable behavior when automating or refactoring automations. ...
    Downloads: 3 This Week
    Last Update:
    See Project
  • Easily Host LLMs and Web Apps on Cloud Run Icon
    Easily Host LLMs and Web Apps on Cloud Run

    Run everything from popular models with on-demand NVIDIA L4 GPUs to web apps without infrastructure management.

    Run frontend and backend services, batch jobs, host LLMs, and queue processing workloads without the need to manage infrastructure. Cloud Run gives you on-demand GPU access for hosting LLMs and running real-time AI—with 5-second cold starts and automatic scale-to-zero so you only pay for actual usage. New customers get $300 in free credit to start.
    Try Cloud Run Free
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB