Showing 4 open source projects for "windows command processor"

View related business solutions
  • Easily Host LLMs and Web Apps on Cloud Run Icon
    Easily Host LLMs and Web Apps on Cloud Run

    Run everything from popular models with on-demand NVIDIA L4 GPUs to web apps without infrastructure management.

    Run frontend and backend services, batch jobs, host LLMs, and queue processing workloads without the need to manage infrastructure. Cloud Run gives you on-demand GPU access for hosting LLMs and running real-time AI—with 5-second cold starts and automatic scale-to-zero so you only pay for actual usage. New customers get $300 in free credit to start.
    Try Cloud Run Free
  • Go from Data Warehouse to Data and AI platform with BigQuery Icon
    Go from Data Warehouse to Data and AI platform with BigQuery

    Build, train, and run ML models with simple SQL. Automate data prep, analysis, and predictions with built-in AI assistance from Gemini.

    BigQuery is more than a data warehouse—it's an autonomous data-to-AI platform. Use familiar SQL to train ML models, run time-series forecasts, and generate AI-powered insights with native Gemini integration. Built-in agents handle data engineering and data science workflows automatically. Get $300 in free credit, query 1 TB, and store 10 GB free monthly.
    Try BigQuery Free
  • 1
    Ouroboros

    Ouroboros

    Automatically update running docker containers

    Ouroboros will monitor (all or specified) running docker containers and update them to the (latest or tagged) available image in the remote registry. The updated container uses the same tag and parameters that were used when the container was first created such as volume/bind mounts, docker network connections, environment variables, restart policies, entrypoints, commands, etc. Push your image to your registry and simply wait your defined interval for ouroboros to find the new image and...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 2
    Nebula docs

    Nebula docs

    Documentation repo of nebula orchestration system

    Nebula is a open source distributed Docker orchestrator designed for massive scales (tens of thousands of servers/worker devices), unlike Mesos/Swarm/Kubernetes it has the ability to have workers distributed on high latency connections (such as the internet) yet have the pods(containers) be managed centrally with changes taking affect (almost) immediately, this makes Nebula ideal for managing a vast cluster of servers\devices across the globe, some example use cases are appliances\virtual...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    Nebula reporter

    Nebula reporter

    The optional reporter container which reads nebula reports from Kafka

    Nebula is an open source-distributed Docker orchestrator designed for massive scales (tens of thousands of servers/worker devices), unlike Mesos/Swarm/Kubernetes it has the ability to have workers distributed on high latency connections (such as the internet) yet have the pods(containers) be managed centrally with changes taking effect (almost) immediately, this makes Nebula ideal for managing a vast cluster of servers\devices across the globe. Ever wandered how your going to push an update...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    Kubernetes shell

    Kubernetes shell

    Kubernetes shell: An integrated shell for working with the Kubernetes

    Kube-shell: An integrated shell for working with the Kubernetes CLI. Under the hood kube-shell still calls kubectl. Kube-shell aims to provide ease-of-use of kubectl and increase productivity. You can use up-arrow and down-arrow to walk through the history of commands executed. Also, up-arrow partial string matching is possible. For e.g. enter 'kubectl get' and use up-arrow and down-arrow to browse through all kubectl get commands. You could also use CTRL+r to search from the history of...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Run Any Workload on Compute Engine VMs Icon
    Run Any Workload on Compute Engine VMs

    From dev environments to AI training, choose preset or custom VMs with 1–96 vCPUs and industry-leading 99.95% uptime SLA.

    Compute Engine delivers high-performance virtual machines for web apps, databases, containers, and AI workloads. Choose from general-purpose, compute-optimized, or GPU/TPU-accelerated machine types—or build custom VMs to match your exact specs. With live migration and automatic failover, your workloads stay online. New customers get $300 in free credits.
    Try Compute Engine
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB