Open Source Linux Container Management Software - Page 2

Container Management Software for Linux

View 41 business solutions
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • Earn up to 16% annual interest with Nexo. Icon
    Earn up to 16% annual interest with Nexo.

    Let your crypto work for you

    Put idle assets to work with competitive interest rates, borrow without selling, and trade with precision. All in one platform. Geographic restrictions, eligibility, and terms apply.
    Get started with Nexo.
  • 1
    Hetzner k3s

    Hetzner k3s

    A CLI tool to install and manage Kubernetes clusters in Hetzner Cloud

    This is a CLI tool to quickly create and manage Kubernetes clusters in Hetzner Cloud using the lightweight Kubernetes distribution k3s from Rancher. Hetzner Cloud is an awesome cloud provider which offers a truly great service with the best performance/cost ratio in the market. With Hetzner's Cloud Controller Manager and CSI driver you can provision load balancers and persistent volumes very easily. k3s is my favorite Kubernetes distribution now because it uses much less memory and CPU, leaving more resources to workloads. It is also super quick to deploy because it's a single binary. Using this tool, creating a highly available k3s cluster with 3 masters for the control plane and 3 worker nodes takes a few minutes only. The tool assigns the label cluster to each server it creates for static node pools (this doesn't apply to autoscaled node pools), with the cluster name you specify in the config file, as the value.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 2
    Kaniko

    Kaniko

    Build Container Images In Kubernetes

    kaniko is a tool to build container images from a Dockerfile, inside a container or Kubernetes cluster. kaniko doesn't depend on a Docker daemon and executes each command within a Dockerfile completely in userspace. This enables building container images in environments that can't easily or securely run a Docker daemon, such as a standard Kubernetes cluster.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 3
    Kapitan

    Kapitan

    Generic templated configuration management for Kubernetes

    Generic templated configuration management for Kubernetes, Terraform, and other things. Kapitan aims to be your one-stop configuration management solution to help you manage the ever-growing complexity of your configurations by enabling Platform Engineering and GitOps workflows. It streamlines complex deployments across heterogeneous environments while providing a secure and adaptable framework for managing infrastructure configurations. Kapitan's inventory-driven model, powerful templating capabilities, and native secret management tools offer granular control, fostering consistency, reducing errors, and safeguarding sensitive data. Empower your team to make changes to your infrastructure whilst maintaining full control, with a GitOps approach and full transparency.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 4
    Kubeconform

    Kubeconform

    A FAST Kubernetes manifests validator

    Kubeconform is a Kubernetes manifest validation tool. Incorporate it into your CI, or use it locally to validate your Kubernetes configuration.
    Downloads: 3 This Week
    Last Update:
    See Project
  • Go From AI Idea to AI App Fast Icon
    Go From AI Idea to AI App Fast

    One platform to build, fine-tune, and deploy ML models. No MLOps team required.

    Access Gemini 3 and 200+ models. Build chatbots, agents, or custom models with built-in monitoring and scaling.
    Try Free
  • 5
    Kubernetes

    Kubernetes

    System to automate management of containerized applications

    Kubernetes is an open-source container orchestration platform designed to automate the deployment, scaling, and management of containerized applications. Also known as K8s, it groups application containers into logical units for efficient management and service discovery. Built on Google’s extensive experience running production workloads at massive scale, Kubernetes delivers proven reliability and performance. It supports automated rollouts, self-healing, load balancing, and flexible storage orchestration. Kubernetes runs consistently across on-premises, hybrid, and public cloud environments, giving teams complete infrastructure freedom.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 6
    LINKERD

    LINKERD

    Ultralight, security-first service mesh for Kubernetes

    Enterprise power without enterprise complexity. Linkerd adds security, observability, and reliability to any Kubernetes cluster. 100% open source, CNCF graduated, and written in Rust. Instantly add latency-aware load balancing, request retries, timeouts, and blue-green deploys to keep your applications resilient. Incredibly small and blazing fast Linkerd2-proxy micro-proxy written in Rust for security and performance. Self-contained control plane, incrementally deployable data plane, and lots and lots of diagnostics and debugging tools. Transparently add mutual TLS to any on-cluster TCP communication with no configuration. Designed by engineers, for engineers.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 7
    Rancher

    Rancher

    Complete container management platform

    From datacenter to cloud to edge, Rancher lets you deliver Kubernetes-as-a-Service. Rancher is a complete software stack for teams adopting containers. It addresses the operational and security challenges of managing multiple Kubernetes clusters, while providing DevOps teams with integrated tools for running containerized workloads. From datacenter to cloud to edge, Rancher's open source software lets you run Kubernetes everywhere. You don’t need to figure Kubernetes out all on your own. Rancher is open source software, with an enormous community of users. Managing Kubernetes installed in your local or remote development environment is so much easier with Rancher. Now with full support for Windows containers, Istio service mesh, and enhanced security for cloud-native workloads, Rancher helps developers innovate faster and with greater confidence.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 8
    dhall-haskell

    dhall-haskell

    Maintainable configuration files

    Maintainable configuration files. Navigate to each package's directory for their respective READMEs. You can download pre-built binaries for Windows, OS X and Linux on the release page. You can then click the "Help" button in the bottom right corner, which will show you a nix-env command that you can run to install the prebuilt executable. You will probably want to use the shared caches hosted at cache.dhall-lang.org and dhall.cachix.org when doing Nix development. This is not required, but this will save you a lot of time so that you don't have to build as many dependencies from scratch the first time. If you prefer installing the binaries locally in a nix shell environment instead, just run nix-shell in the top-level directory. This option provides additional flexibility with respect to overriding some of the default parameters (e.g. the compiler version), which makes it particularly useful for developers.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 9
    kail

    kail

    Kubernetes log viewer

    Kubernetes tail. Streams logs from all containers of all matched pods. Match pods by service, replicaset, deployment, and others. Adjusts to a changing cluster - pods are added and removed from logging as they fall in or out of the selection. With no arguments, kail matches all pods in the cluster. You can control the matching pods with arguments that select pods based on various criteria.
    Downloads: 3 This Week
    Last Update:
    See Project
  • Full-stack observability with actually useful AI | Grafana Cloud Icon
    Full-stack observability with actually useful AI | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • 10
    Colima

    Colima

    Container runtimes on macOS (and Linux) with minimal setup

    Container runtimes on macOS (and Linux) with minimal setup.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 11
    Gatekeeper

    Gatekeeper

    Policy Controller for Kubernetes

    A customizable cloud-native policy controller that helps enforce policies and strengthen governance. Gatekeeper makes managing policies on top of Kubernetes easy. Policies can be enforced at admission time or at runtime via the audit functionality. Gatekeeper is powered by the Open Policy Agent (OPA) project. Using OPA allows you to write policies that are powerful, flexible, and portable. Browse the policy library to find existing policies that fit your use case. Each policy in the library can be extended and customized to fit your needs.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 12
    KServe

    KServe

    Standardized Serverless ML Inference Platform on Kubernetes

    KServe provides a Kubernetes Custom Resource Definition for serving machine learning (ML) models on arbitrary frameworks. It aims to solve production model serving use cases by providing performant, high abstraction interfaces for common ML frameworks like Tensorflow, XGBoost, ScikitLearn, PyTorch, and ONNX. It encapsulates the complexity of autoscaling, networking, health checking, and server configuration to bring cutting edge serving features like GPU Autoscaling, Scale to Zero, and Canary Rollouts to your ML deployments. It enables a simple, pluggable, and complete story for Production ML Serving including prediction, pre-processing, post-processing and explainability. KServe is being used across various organizations.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 13
    Loggie

    Loggie

    A lightweight, cloud-native data transfer agent and aggregator

    Loggie is a lightweight, high-performance, cloud-native agent and aggregator based on Golang. Loggie includes LogConfig/ClusterLogConfig/Interceptor/Sink CRDs, allowing for the creation of data collection, transfer, processing, and sending pipelines through simple YAML file creation. Supports deployment as an independent intermediate machine, which can receive aggregated data sent by Loggie Agent and can also be used to consume and process various data sources. Configure Filebeat and Loggie to collect logs, and send them to a Kafka topic without using client compression, with the Kafka topic partition configured as 3. With sufficient resources for the Agent specification, modify the number of files collected, the concurrency of the sending client (configure Filebeat worker and Loggie parallelism), and observe their respective CPU, memory, and pod network card transmission rates.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 14
    NVIDIA AI Cluster Runtime (AICR)

    NVIDIA AI Cluster Runtime (AICR)

    Tooling for optimized and reproducible GPU-accelerated AI runtime

    NVIDIA AI Cluster Runtime (AICR) is an emerging project within NVIDIA’s AI infrastructure ecosystem focused on enabling advanced AI compute and runtime workflows, though publicly available documentation remains limited. Based on its positioning within NVIDIA’s repositories, it is designed to support scalable AI runtime environments, potentially addressing challenges related to orchestration, resource management, or reproducible AI execution. The project likely aligns with NVIDIA’s broader strategy of building modular infrastructure layers that integrate with GPU-accelerated workloads and cloud-native systems. It appears to emphasize automation, consistency, and performance optimization across AI pipelines, potentially targeting enterprise and research use cases. Given NVIDIA’s ecosystem, it may also integrate with containerized environments, Kubernetes, or other orchestration frameworks.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 15
    Robusta KRR

    Robusta KRR

    Prometheus-based Kubernetes Resource Recommendations

    Robusta KRR (Kubernetes Resource Recommender) is a CLI tool for optimizing resource allocation in Kubernetes clusters. It gathers pod usage data from Prometheus and recommends requests and limits for CPU and memory. This reduces costs and improves performance.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 16
    Trivy Operator

    Trivy Operator

    Kubernetes-native security toolkit

    The Trivy Operator leverages Trivy to continuously scan your Kubernetes cluster for security issues. The scans are summarised in security reports as Kubernetes Custom Resource Definitions, which become accessible through the Kubernetes API. The Operator does this by watching Kubernetes for state changes and automatically triggering security scans in response. For example, a vulnerability scan is initiated when a new Pod is created. This way, users can find and view the risks that relate to different resources in a Kubernetes-native way. The Trivy Operator automatically generates and updates security reports. These reports are generated in response to new workload and other changes on a Kubernetes cluster.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 17
    cri-dockerd

    cri-dockerd

    dockerd as a compliant Container Runtime Interface for Kubernetes

    This adapter provides a shim for Docker Engine that lets you control Docker via the Kubernetes Container Runtime Interface. Mirantis and Docker have agreed to partner to maintain the shim code standalone outside Kubernetes, as a conformant CRI interface for the Docker Engine API. For Mirantis customers, that means that Docker Engine’s commercially supported version, Mirantis Container Runtime (MCR), will be CRI compliant. This means that you can continue to build Kubernetes based on the Docker Engine as before, just switching from the built-in docker shim to the external one. Mirantis and Docker intend to work together to make sure it continues to work as well as before and that it passes all the conformance tests and continues to work just like the built-in version did. Mirantis will be using this in Mirantis Kubernetes Engine, and Docker will continue to ship this shim in Docker Desktop.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 18
    Argo CD Image Updater

    Argo CD Image Updater

    Automatic container image update for Argo CD

    Argo CD Image Updater is a tool to automatically update the container images of Kubernetes workloads which are managed by Argo CD. In a nutshell, it will track image versions specified by annotations on the Argo CD Application resources and update them by setting parameter overrides using the Argo CD API. Currently, it will only work with applications that are built using Kustomize or Helm tooling. Applications built from plain YAML or custom tools are not supported yet (and maybe never will).
    Downloads: 1 This Week
    Last Update:
    See Project
  • 19
    Bank of Anthos

    Bank of Anthos

    Retail banking sample application showcasing Kubernetes

    Bank of Anthos is a sample HTTP-based web app that simulates a bank's payment processing network, allowing users to create artificial bank accounts and complete transactions. Google uses this application to demonstrate how developers can modernize enterprise applications using Google Cloud products, including: Google Kubernetes Engine (GKE), Anthos Service Mesh (ASM), Anthos Config Management (ACM), Migrate to Containers, Spring Cloud GCP, Cloud Operations, Cloud SQL, Cloud Build, and Cloud Deploy. This application works on any Kubernetes cluster.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 20
    Docker Pi-hole

    Docker Pi-hole

    Pi-hole in a docker container

    You can run Pi-hole in a container, or deploy it directly to a supported operating system via our automated installer. Our intelligent, automated installer asks you a few questions and then sets everything up for you. Once complete, move onto step 3. Configure your router’s DHCP options to force clients to use Pi-hole as their DNS server, or manually configure each device​ to use the Pi-hole as their DNS server. By pairing your Pi-hole with a VPN, you can have ad blocking on your cellular devices, helping with limited bandwidth data plans. Instead of browser plugins or other software on each computer, install Pi-hole in one place and your entire network is protected. Network-level blocking allows you to block ads in non-traditional places such as mobile apps and smart TVs, regardless of hardware or OS. Since advertisements are blocked before they are downloaded, network performance is improved and will feel faster.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 21
    Docker-OSX

    Docker-OSX

    Run macOS VM in a Docker! Run near native OSX-KVM in Docker

    Run Mac OS X in Docker with near-native performance! X11 Forwarding. iMessage security research! iPhone USB working! macOS in a Docker container.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 22
    Gloo Gateway

    Gloo Gateway

    The Feature-rich, Kubernetes-native, Next-Generation API Gateway

    Gloo Gateway is a powerful Kubernetes-native ingress controller and API gateway that is based on the Kubernetes Gateway API. It excels in function-level routing, supports legacy apps, microservices and serverless, offers robust discovery capabilities, integrates seamlessly with open-source projects, and is designed to support hybrid applications with various technologies, architectures, protocols, and clouds.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 23
    KT-Connect

    KT-Connect

    A toolkit for Integrating with your kubernetes dev environment

    KtConnect ("Kt" is short for "Kubernetes Toolkit") is a utility tool to help you work with the Kubernetes dev environment more efficiently. Directly Access a remote Kubernetes cluster. KtConnect use ssh-vpn or socks-proxy to access remote Kubernetes cluster networks. Developers can exchange the workload to redirect the requests to a local app. You can create a mesh version service in the local host, and redirect specified workload requests to your local. Expose a local running app to the Kubernetes cluster as a common service, all requests to that service are redirected to the local app.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 24
    KUDO

    KUDO

    Kubernetes Universal Declarative Operator (KUDO)

    Kubernetes Universal Declarative Operator (KUDO) provides a declarative approach to building production-grade Kubernetes Operators covering the entire application lifecycle. The Kubernetes Universal Declarative Operator (KUDO) is a highly productive toolkit for writing Kubernetes Operators. Using KUDO you can deploy your applications, have the tools needed to operate them, and understand how they're behaving, all without a Ph.D. in Kubernetes. KUDO lets you configure an Operator’s entire lifecycle using a declarative spec, including things like backup/restore. You don’t have to write Go unless you want to.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 25
    Karmada

    Karmada

    Open, Multi-Cloud, Multi-Cluster Kubernetes Orchestration

    Karmada (Kubernetes Armada) is a Kubernetes management system that enables you to run your cloud-native applications across multiple Kubernetes clusters and clouds, with no changes to your applications. By speaking Kubernetes-native APIs and providing advanced scheduling capabilities, Karmada enables truly open, multi-cloud Kubernetes. Karmada aims to provide turnkey automation for multi-cluster application management in multi-cloud and hybrid cloud scenarios, with key features such as centralized multi-cloud management, high availability, failure recovery, and traffic scheduling.
    Downloads: 1 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB