Showing 10 open source projects for "remote execution"

View related business solutions
  • Build on Google Cloud with $300 in Free Credit Icon
    Build on Google Cloud with $300 in Free Credit

    New to Google Cloud? Get $300 in free credit to explore Compute Engine, BigQuery, Cloud Run, Vertex AI, and 150+ other products.

    Start your next project with $300 in free Google Cloud credit. Spin up VMs, run containers, query exabytes in BigQuery, or build AI apps with Vertex AI and Gemini. Once your credits are used, keep building with 20+ products with free monthly usage, including Compute Engine, Cloud Storage, GKE, and Cloud Run functions. Sign up to start building right away.
    Start Free Trial
  • Cut Cloud Costs with Google Compute Engine Icon
    Cut Cloud Costs with Google Compute Engine

    Save up to 91% with Spot VMs and get automatic sustained-use discounts. One free VM per month, plus $300 in credits.

    Save on compute costs with Compute Engine. Reduce your batch jobs and workload bill 60-91% with Spot VMs. Compute Engine's committed use offers customers up to 70% savings through sustained use discounts. Plus, you get one free e2-micro VM monthly and $300 credit to start.
    Try Compute Engine
  • 1
    MCP Shell Server

    MCP Shell Server

    Shell command execution server implementing the Model Context Protocol

    A secure shell command execution server implementing the Model Context Protocol (MCP), allowing remote execution of whitelisted shell commands with support for standard input. ​
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    Sled

    Sled

    Teleport Claude Code, Codex or Gemini CLI to your phone

    Sled under the layercodedev account appears to be a lightweight web and mobile UI aimed at interacting with local coding agents, likely including AI-assisted coding models or remote execution integrations; it’s designed in TypeScript/JavaScript and intended to let developers use or control coding agents from various devices. Although specific details in the repository are limited without direct project documentation, context and related online mentions indicate it functions as a local interface layer that abstracts development agent workflows and Teleport-style interactions, bringing parts of modern assistant capabilities to phone or web UIs. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    Hamilton DAGWorks

    Hamilton DAGWorks

    Helps scientists define testable, modular, self-documenting dataflow

    ...Your DAG is portable; it runs anywhere Python runs, whether it's a script, notebook, Airflow pipeline, FastAPI server, etc. Your DAG is expressive; Hamilton has extensive features to define and modify the execution of a DAG (e.g., data validation, experiment tracking, remote execution). To create a DAG, write regular Python functions that specify their dependencies with their parameters. As shown below, it results in readable code that can always be visualized. Hamilton loads that definition and automatically builds the DAG for you. Hamilton brings modularity and structure to any Python application moving data: ETL pipelines, ML workflows, LLM applications, RAG systems, BI dashboards, and the Hamilton UI allows you to automatically visualize, catalog, and monitor execution.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    Maestro AI Orchestration

    Maestro AI Orchestration

    Agent Orchestration Command Center

    ...It provides a high-performance experience for running multiple agent sessions in parallel, integrating with tools such as Claude Code, OpenAI Codex, and other agent tooling to automate tasks, perform unattended execution, and organize long-running work flows. Users can collaborate with AI to draft specifications, break tasks into playbooks, and run them sequentially or concurrently in isolated contexts, each with clean session history. Maestro includes advanced features like Git worktrees to isolate sub-agent tasks, command-line interfaces for integration into CI/CD pipelines, remote control via phone, and comprehensive session analytics.
    Downloads: 5 This Week
    Last Update:
    See Project
  • Easily Host LLMs and Web Apps on Cloud Run Icon
    Easily Host LLMs and Web Apps on Cloud Run

    Run everything from popular models with on-demand NVIDIA L4 GPUs to web apps without infrastructure management.

    Run frontend and backend services, batch jobs, host LLMs, and queue processing workloads without the need to manage infrastructure. Cloud Run gives you on-demand GPU access for hosting LLMs and running real-time AI—with 5-second cold starts and automatic scale-to-zero so you only pay for actual usage. New customers get $300 in free credit to start.
    Try Cloud Run Free
  • 5
    Open Interpreter

    Open Interpreter

    A natural language interface for computers

    Open Interpreter is an open-source tool that provides a natural-language interface for interacting with your computer. It lets large language models (LLMs) run code locally (Python, JavaScript, shell, etc.), enabling you to ask your computer to do tasks like data analysis, file manipulation, browsing, etc. in human terms (“chat with your computer”), with safeguards. Runs locally or via configured remote LLM servers/inference backends, giving flexibility to use models you trust or have...
    Downloads: 7 This Week
    Last Update:
    See Project
  • 6
    Omnara

    Omnara

    Talk to Your AI Agents from Anywhere

    Omnara is an open-source agent control platform that empowers developers to turn autonomous AI tools (e.g., Claude Code, Cursor, GitHub Copilot) into collaborative teammates by offering real-time dashboards, push notifications, and remote guidance across terminals, web, and mobile. Omnara transforms your AI agents (Claude Code, Codex CLI, n8n, and more) from silent workers into communicative teammates. Get real-time visibility into what your agents are doing, and respond to their questions...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    Constellation

    Constellation

    Constellation is the first Confidential Kubernetes

    Constellation is a distributed confidential computing platform developed by Edgeless Systems. It allows developers to run Kubernetes-native applications in secure enclaves across multiple machines, ensuring end-to-end encryption and trusted execution for workloads. Built on top of Kubernetes and using technologies like Intel SGX and Gramine, Constellation guarantees that not even infrastructure operators can access data or code, making it ideal for privacy-sensitive workloads and multi-party...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    Elyra

    Elyra

    Elyra extends JupyterLab with an AI centric approach

    Elyra is a set of AI-centric extensions to JupyterLab Notebooks. The Elyra Getting Started Guide includes more details on these features. A version-specific summary of new features is located on the releases page.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    DBHub

    DBHub

    Universal database MCP server connecting to MySQL, PostgreSQL

    DBHub is a universal database gateway that implements the MCP server interface so assistants and IDEs can explore and query databases through typed tools. It supports multiple transports—stdio for desktop clients and HTTP for networked scenarios—making it flexible to embed or deploy. Configuration is environment-variable driven, with a DSN and per-engine settings covering Postgres, MySQL, MariaDB, SQL Server, and SQLite. Operational flags include read-only mode, row limits, and even SSH...
    Downloads: 3 This Week
    Last Update:
    See Project
  • Build AI Apps with Gemini 3 on Vertex AI Icon
    Build AI Apps with Gemini 3 on Vertex AI

    Access Google’s most capable multimodal models. Train, test, and deploy AI with 200+ foundation models on one platform.

    Vertex AI gives developers access to Gemini 3—Google’s most advanced reasoning and coding model—plus 200+ foundation models including Claude, Llama, and Gemma. Build generative AI apps with Vertex AI Studio, customize with fine-tuning, and deploy to production with enterprise-grade MLOps. New customers get $300 in free credits.
    Try Vertex AI Free
  • 10
    OAGI Python SDK

    OAGI Python SDK

    Python SDK for the Computer Use model Lux, developed by OpenAGI

    OAGI Python SDK is a Python client library for the Lux computer-use model that turns Lux into a programmable automation layer for operating human-facing software via vision and actions. It exposes the OAGI API in an ergonomic way, letting you trigger Lux in three main modes: Tasker for precise scripted sequences, Actor for fast one-shot tasks, and Thinker for open-ended, multi-step objectives. The SDK is designed around “computer use” as a paradigm, where the AI actually navigates...
    Downloads: 3 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB