Alternatives to Yandex Serverless Containers

Compare Yandex Serverless Containers alternatives for your business or organization using the curated list below. SourceForge ranks the best alternatives to Yandex Serverless Containers in 2024. Compare features, ratings, user reviews, pricing, and more from Yandex Serverless Containers competitors and alternatives in order to make an informed decision for your business.

  • 1
    Google Cloud Run
    Cloud Run is a fully-managed compute platform that lets you run your code in a container directly on top of Google's scalable infrastructure. We’ve intentionally designed Cloud Run to make developers more productive - you get to focus on writing your code, using your favorite language, and Cloud Run takes care of operating your service. Fully managed compute platform for deploying and scaling containerized applications quickly and securely. Write code your way using your favorite languages (Go, Python, Java, Ruby, Node.js, and more). Abstract away all infrastructure management for a simple developer experience. Build applications in your favorite language, with your favorite dependencies and tools, and deploy them in seconds. Cloud Run abstracts away all infrastructure management by automatically scaling up and down from zero almost instantaneously—depending on traffic. Cloud Run only charges you for the exact resources you use. Cloud Run makes app development & deployment simpler.
    Compare vs. Yandex Serverless Containers View Software
    Visit Website
  • 2
    Telepresence

    Telepresence

    Ambassador Labs

    Telepresence streamlines your local development process, enabling immediate feedback. You can launch your local environment on your laptop, equipped with your preferred tools, while Telepresence seamlessly connects them to the microservices and test databases they rely on. It simplifies and expedites collaborative development, debugging, and testing within Kubernetes environments by establishing a seamless connection between your local machine and shared remote Kubernetes clusters. Why Telepresence: Faster feedback loops: Spend less time building, containerizing, and deploying code. Get immediate feedback on code changes by running your service in the cloud from your local machine. Shift testing left: Create a remote-to-local debugging experience. Catch bugs pre-production without the configuration headache of remote debugging. Deliver better, faster user experience: Get new features and applications into the hands of users faster and more frequently.
    Compare vs. Yandex Serverless Containers View Software
    Visit Website
  • 3
    Portainer Business
    Portainer is an intuitive container management platform for Docker, Kubernetes, and Edge-based environments. With a smart UI, Portainer enables you to build, deploy, manage, and secure your containerized environments with ease. It makes container adoption easier for the whole team and reduces time-to-value on Kubernetes and Docker/Swarm. With a simple GUI and a comprehensive API, the product makes it easy for engineers to deploy and manage container-based apps, triage issues, automate CI/CD workflows and set up CaaS (container-as-a-service) environments regardless of hosting environment or K8s distro. Portainer Business is designed to be used in a team environment with multiple users and clusters. The product includes a range of security features, including RBAC, OAuth integration, and logging - making it suitable for use in complex production environments. Portainer also allows you to set up GitOps automation for deployment of your apps to Docker and K8s based on Git repos.
  • 4
    AWS Fargate

    AWS Fargate

    Amazon

    AWS Fargate is a serverless compute engine for containers that works with both Amazon Elastic Container Service (ECS) and Amazon Elastic Kubernetes Service (EKS). Fargate makes it easy for you to focus on building your applications. Fargate removes the need to provision and manage servers, lets you specify and pay for resources per application, and improves security through application isolation by design. Fargate allocates the right amount of compute, eliminating the need to choose instances and scale cluster capacity. You only pay for the resources required to run your containers, so there is no over-provisioning and paying for additional servers. Fargate runs each task or pod in its own kernel providing the tasks and pods their own isolated compute environment. This enables your application to have workload isolation and improved security by design.
  • 5
    Amazon EKS

    Amazon EKS

    Amazon

    Amazon Elastic Kubernetes Service (Amazon EKS) is a fully managed Kubernetes service. Customers such as Intel, Snap, Intuit, GoDaddy, and Autodesk trust EKS to run their most sensitive and mission-critical applications because of its security, reliability, and scalability. EKS is the best place to run Kubernetes for several reasons. First, you can choose to run your EKS clusters using AWS Fargate, which is serverless compute for containers. Fargate removes the need to provision and manage servers, lets you specify and pay for resources per application, and improves security through application isolation by design. Second, EKS is deeply integrated with services such as Amazon CloudWatch, Auto Scaling Groups, AWS Identity and Access Management (IAM), and Amazon Virtual Private Cloud (VPC), providing you a seamless experience to monitor, scale, and load-balance your applications.
  • 6
    Red Hat OpenShift
    The Kubernetes platform for big ideas. Empower developers to innovate and ship faster with the leading hybrid cloud, enterprise container platform. Red Hat OpenShift offers automated installation, upgrades, and lifecycle management throughout the container stack—the operating system, Kubernetes and cluster services, and applications—on any cloud. Red Hat OpenShift helps teams build with speed, agility, confidence, and choice. Code in production mode anywhere you choose to build. Get back to doing work that matters. Red Hat OpenShift is focused on security at every level of the container stack and throughout the application lifecycle. It includes long-term, enterprise support from one of the leading Kubernetes contributors and open source software companies. Support the most demanding workloads including AI/ML, Java, data analytics, databases, and more. Automate deployment and life-cycle management with our vast ecosystem of technology partners.
  • 7
    Google Kubernetes Engine (GKE)
    Run advanced apps on a secured and managed Kubernetes service. GKE is an enterprise-grade platform for containerized applications, including stateful and stateless, AI and ML, Linux and Windows, complex and simple web apps, API, and backend services. Leverage industry-first features like four-way auto-scaling and no-stress management. Optimize GPU and TPU provisioning, use integrated developer tools, and get multi-cluster support from SREs. Start quickly with single-click clusters. Leverage a high-availability control plane including multi-zonal and regional clusters. Eliminate operational overhead with auto-repair, auto-upgrade, and release channels. Secure by default, including vulnerability scanning of container images and data encryption. Integrated Cloud Monitoring with infrastructure, application, and Kubernetes-specific views. Speed up app development without sacrificing security.
  • 8
    Docker

    Docker

    Docker

    Docker takes away repetitive, mundane configuration tasks and is used throughout the development lifecycle for fast, easy and portable application development, desktop and cloud. Docker’s comprehensive end-to-end platform includes UIs, CLIs, APIs and security that are engineered to work together across the entire application delivery lifecycle. Get a head start on your coding by leveraging Docker images to efficiently develop your own unique applications on Windows and Mac. Create your multi-container application using Docker Compose. Integrate with your favorite tools throughout your development pipeline, Docker works with all development tools you use including VS Code, CircleCI and GitHub. Package applications as portable container images to run in any environment consistently from on-premises Kubernetes to AWS ECS, Azure ACI, Google GKE and more. Leverage Docker Trusted Content, including Docker Official Images and images from Docker Verified Publishers.
  • 9
    Spot Ocean

    Spot Ocean

    Spot by NetApp

    Spot Ocean lets you reap the benefits of Kubernetes without worrying about infrastructure while gaining deep cluster visibility and dramatically reducing costs. The key question is how to use containers without the operational overhead of managing the underlying VMs while also take advantage of the cost benefits associated with Spot Instances and multi-cloud. Spot Ocean is built to solve this problem by managing containers in a “Serverless” environment. Ocean provides an abstraction on top of virtual machines allowing to deploy Kubernetes clusters without the need to manage the underlying VMs. Ocean takes advantage of multiple compute purchasing options like Reserved and Spot instance pricing and failover to On-Demand instances whenever necessary, providing 80% reduction in infrastructure costs. Spot Ocean is a Serverless Compute Engine that abstracts the provisioning (launching), auto-scaling, and management of worker nodes in Kubernetes clusters.
  • 10
    Yandex Cloud Functions
    Run code as a function in a secure, fault-tolerant, and automatically scalable environment without creating or maintaining VMs. As the number of function calls increases, the service automatically creates additional instances of your function. All functions run in parallel. The runtime environment is hosted in three availability zones, ensuring availability even if one zone fails. Configure and prepare instances of functions always ready to process loads. This mode allows you to avoid cold starts and quickly process loads of any size. Give functions access to your VPC to accelerate interactions with private resources, database clusters, virtual machines, Kubernetes nodes, etc. Serverless Functions tracks and logs information about function calls and analyzes execution flow and performance. You can also describe logging mechanisms in your function code. Launch cloud functions in synchronized mode and delayed execution mode.
    Starting Price: $0.012240 per GB
  • 11
    Cloudflare Workers
    You write code. We handle the rest. Deploy serverless code instantly across the globe to give it exceptional performance, reliability, and scale. No more configuring auto-scaling, load balancers, or paying for capacity you don’t use. Traffic is automatically routed and load balanced across thousands of servers. Sleep well as your code scales effortlessly. Every deploy is made to a network of data centers running V8 isolates. Your code is powered by Cloudflare’s network which is milliseconds away from virtually every Internet user. Choose from a template in your language to kickstart building an app, creating a function, or writing an API. We have templates, tutorials, and a CLI to get you up and running in no time. Most serverless platforms experience a cold start every time you deploy or your service increases in popularity. Workers can run your code instantly, without cold starts. The first 100,000 requests each day are free and paid plans start at just $5/10 million requests.
    Starting Price: $5 per 10 million requests
  • 12
    Azure Kubernetes Service (AKS)
    The fully managed Azure Kubernetes Service (AKS) makes deploying and managing containerized applications easy. It offers serverless Kubernetes, an integrated continuous integration and continuous delivery (CI/CD) experience, and enterprise-grade security and governance. Unite your development and operations teams on a single platform to rapidly build, deliver, and scale applications with confidence. Elastic provisioning of additional capacity without the need to manage the infrastructure. Add event-driven autoscaling and triggers through KEDA. Faster end-to-end development experience with Azure Dev Spaces including integration with Visual Studio Code Kubernetes tools, Azure DevOps, and Azure Monitor. Advanced identity and access management using Azure Active Directory, and dynamic rules enforcement across multiple clusters with Azure Policy. Available in more regions than any other cloud providers.
  • 13
    Serverless Application Engine (SAE)
    Network isolation with sandboxed containers and virtual private cloud (VPC) ensures the security of application runtimes. SAE provides high availability solutions for large-scale events that require precise capacity handling, high scalability, and service throttling and degradation. Fully-managed IaaS with Kubernetes clusters provide low-cost solutions for your business. SAE scales within seconds and improves the efficiency of runtimes and Java application startup. One-Stop PaaS with seamlessly integrated basic services, microservices, and DevOps products. SAE provides full-lifecycle application management. You can implement different release policies, such as phased release and canary release. The traffic-ratio-based canary release model is also supported. The release process is fully observable and can be rolled back.
  • 14
    Knative

    Knative

    Google

    Knative, created originally by Google with contributions from over 50 different companies, delivers an essential set of components to build and run serverless applications on Kubernetes. Knative offers features like scale-to-zero, autoscaling, in-cluster builds, and eventing framework for cloud-native applications on Kubernetes. Whether on-premises, in the cloud, or in a third-party data center, Knative codifies the best practices shared by successful real-world Kubernetes-based frameworks. Most importantly, Knative enables developers to focus on writing code without the need to worry about the “boring but difficult” parts of building, deploying, and managing their application.
  • 15
    Neon

    Neon

    Neon

    The fully managed multi-cloud Postgres with a generous free tier. We separated storage and computing to offer autoscaling, branching, and bottomless storage. We separated storage and computing to make on-demand scalability possible. Compute activates on an incoming connection and scales to zero when not in use. Neon storage uses the "copy-on-write" technique to deliver data branching, online checkpointing, and point-in-time restore. This eliminates expensive full-data backup and restore operations required with traditional database-as-a-service systems. Neon allows you to instantly branch your Postgres database to support modern development workflows. You can create branches for test environments and for every deployment in your CI/CD pipeline. Our serverless architecture reduces computing and storage expenses. Specifically, Neon's autoscaling capabilities prevent over-provisioning and paying for under-utilized instances.
  • 16
    Google App Engine
    Scale your applications from zero to planet scale without having to manage infrastructure. Scale your applications from zero to planet scale without having to manage infrastructure. Stay agile with support for popular development languages and a range of developer tools. Build and deploy apps quickly using popular languages or bring your own language runtimes and frameworks. You can also manage resources from the command line, debug source code, and run API back ends easily. Focus on writing code without having to manage underlying infrastructure. Protect your apps from security threats using firewall capabilities, IAM rules, and managed SSL/ TLS certificates. Operate in a serverless environment without worrying about over or under provisioning. App Engine automatically scales depending on your app traffic and consumes resources only when your code is running.
  • 17
    kpt

    kpt

    kpt

    kpt is a package-centric toolchain that enables a WYSIWYG configuration authoring, automation, and delivery experience, which simplifies managing Kubernetes platforms and KRM-driven infrastructure at scale by manipulating declarative configuration as data, separated from the code that transforms it. Most Kubernetes users either manage their resources using conventional imperative graphical user interfaces, command-line tools (kubectl), and automation (e.g., operators) that operate directly against Kubernetes APIs, or declarative configuration tools, such as Helm, Terraform, cdk8s, or one of the dozens of other tools. At a small scale, this is largely driven by preference and familiarity. As companies expand the number of Kubernetes development and production clusters they use, creating and enforcing consistent configurations and security policies across a growing environment becomes difficult.
  • 18
    Azure Container Instances
    Develop apps fast without managing virtual machines or having to learn new tools—it's just your application, in a container, running in the cloud. By running your workloads in Azure Container Instances (ACI), you can focus on designing and building your applications instead of managing the infrastructure that runs them. Deploy containers to the cloud with unprecedented simplicity and speed—with a single command. Use ACI to provision additional compute for demanding workloads whenever you need. For example, with the Virtual Kubelet, use ACI to elastically burst from your Azure Kubernetes Service (AKS) cluster when traffic comes in spikes. Gain the security of virtual machines for your container workloads, while preserving the efficiency of lightweight containers. ACI provides hypervisor isolation for each container group to ensure containers run in isolation without sharing a kernel.
  • 19
    Tencent Container Registry
    Tencent Container Registry (TCR) offers secure, dedicated, and high-performance container image hosting and distribution service. You can create dedicated instances in multiple regions across the globe and pull container images from the nearest region to reduce pulling time and bandwidth costs. To guarantee data security, TCR features granular permission management and access control. It also supports P2P accelerated distribution to break through the performance bottleneck due to concurrent pulling of large images by large-scale clusters, helping you quickly expand and update businesses. You can customize image synchronization rules and triggers, and use TCR flexibly with your existing CI/CD workflow to quickly implement container DevOps. TCR instance adopts containerized deployment. You can dynamically adjust the service capability based on actual usage to manage sudden surges in business traffic.
  • 20
    Percona Kubernetes Operator
    The Percona Kubernetes Operator for Percona XtraDB Cluster or Percona Server for MongoDB automates the creation, alteration, or deletion of members in your Percona XtraDB Cluster or Percona Server for MongoDB environment. It can be used to instantiate a new Percona XtraDB Cluster or Percona Server for MongoDB replica set, or to scale an existing environment. The Operator contains all necessary Kubernetes settings to provide a proper and consistent Percona XtraDB Cluster or Percona Server for MongoDB instance. The Percona Kubernetes Operators are based on best practices for configuration and setup of a Percona XtraDB Cluster or Percona Server for MongoDB replica set. The benefits of the Operator are many but saving time and delivering a consistent and vetted environment is key.
  • 21
    IronWorker

    IronWorker

    Iron.io

    Container based workloads, full GPU support, autoscaling, custom built solutions. Let Worker take care of your jobs so you can focus on your application. A hosted background job solution that lets you run your containers with dynamic scale, detailed analytics and world class customer support. Run short lived containers quickly, or even containers needing to work across multiple days. Whether it's a one-off job, or one that needs 1000's of workers running in parallel, we've got you covered. Containerize your background tasks with confidence. Run your containers on our shared infrastructure. Dedicated hardware for your workloads. A great option when you need consistent performance and throughput. Our autoscale technology helps you scale up and down based on your usage. You run workers on your own hardware, and we take care of the authentication, scheduling, and everything else. If you have access to your own infrastructure and/or have higher security needs, this is often a great fit.
  • 22
    Macrometa

    Macrometa

    Macrometa

    We deliver a geo-distributed real-time database, stream processing and compute runtime for event-driven applications across up to 175 worldwide edge data centers. App & API builders love our platform because we solve the hardest problems of sharing mutable state across 100s of global locations, with strong consistency & low latency. Macrometa enables you to surgically extend your existing infrastructure to bring part of or your entire application closer to your end users. This allows you to improve performance, user experience, and comply with global data governance laws. Macrometa is a serverless, streaming NoSQL database, with integrated pub/sub and stream data processing and compute engine. Create stateful data infrastructure, stateful functions & containers for long running workloads, and process data streams in real time. You do the code, we do all the ops and orchestration.
  • 23
    Replex

    Replex

    Replex

    Configure policies to manage and govern cloud-native environments without impacting agility or speed. Allocate budgets to individual teams or projects, keep track of costs, govern resource usage and generate real-time alerts for cost overruns. Track the complete asset life cycle from ownership and creation to modification and termination. Understand detailed resource consumption patterns and costs associated with decentralized development teams while engaging developers in creating value with each and every deployment. Ensure microservices, containers, pods, and Kubernetes clusters have the most efficient resource footprint possible without compromising reliability, availability, or performance. Replex allows you to right size Kubernetes nodes and cloud instances based on historical and real-time utilization data and is a single source of truth for all performance-critical metrics.
  • 24
    VMware Tanzu Kubernetes Grid
    Power your modern applications with VMware Tanzu Kubernetes Grid. Run the same K8s across data center, public cloud and edge for a consistent, secure experience for all development teams. Keep your workloads properly isolated and secure. Get a complete, easy-to-upgrade Kubernetes runtime with preintegrated and validated components. Deploy and scale all clusters without downtime. Apply security fixes fast. Run your containerized applications on a certified Kubernetes distribution, bolstered by the global Kubernetes community. Use your existing data center tools and workflows to give developers secure, self-serve access to conformant Kubernetes clusters in your VMware private cloud, and extend the same consistent Kubernetes runtime across your public cloud and edge environments. Simplify operations of large-scale, multicluster Kubernetes environments, and keep your workloads properly isolated. Automate lifecycle management to reduce your risk and shift your focus to more strategic work.
  • 25
    DBOS

    DBOS

    DBOS

    A simpler, more secure way to build fault-tolerant cloud applications, powered by the revolutionary cloud-native DBOS operating system. Based on 3 years of joint MIT-Stanford open source R&D, DBOS revolutionizes cloud-native architecture. DBOS is a cloud-native OS that builds on a relational database to radically simplify today's complex cloud application stacks. DBOS powers DBOS Cloud, a transactional serverless platform that provides fault-tolerance, observability, cyber-resilience, and easy cloud deployment to stateful TypeScript applications. OS services are implemented on top of a distributed DBMS. Built-in transactional, fault-tolerant state management that simplifies the stack, with no need for containers, cluster management, or workflow orchestration. Seamless scaling, high performance, and high availability. Metrics, logs, and traces are stored in SQL-accessible tables. Smaller cyber attack surface, cyberattack self-detection, and cyber-resilience.
  • 26
    Quix

    Quix

    Quix

    Building real-time apps and services require lots of components running in concert: Kafka, VPC hosting, infrastructure as code, container orchestration, observability, CI/CD, persistent volumes, databases, and much more. The Quix platform takes care of all the moving parts. You just connect your data and start building. That’s it. No provisioning clusters or configuring resources. Use Quix connectors to ingest transaction messages streamed from your financial processing systems in a virtual private cloud or on-premise data center. All data in transit is encrypted end-to-end and compressed with G-Zip and Protobuf for security and efficiency. Detect fraudulent patterns with machine learning models or rule-based algorithms. Create fraud warning messages as troubleshooting tickets or display them in support dashboards.
  • 27
    Cloud Foundry

    Cloud Foundry

    Cloud Foundry

    Cloud Foundry makes it faster and easier to build, test, deploy and scale applications, providing a choice of clouds, developer frameworks, and application services. It is an open source project and is available through a variety of private cloud distributions and public cloud instances. Cloud Foundry has a container-based architecture that runs apps in any programming language. Deploy apps to CF using your existing tools and with zero modification to the code. Instantiate, deploy, and manage high-availability Kubernetes clusters with CF BOSH on any cloud. By decoupling applications from infrastructure, you can make individual decisions about where to host workloads – on premise, in public clouds, or in managed infrastructures – and move those workloads as necessary in minutes, with no changes to the app.
  • 28
    OpenShift Cloud Functions
    Red Hat Openshift Cloud Functions (OCF) is a FaaS - Function as a Service that can be deployed on Openshift and is based out of Knative a FaaS project in the Kubernetes community. It will enable developers to run code without needing to know anything about the underlying platform specifics. Developers need access to services faster, deploying backend services, platforms or applications can be time-consuming and tedious. This also means developers should not be restricted to any certain language or framework but also to create business value quickly and enhance services by way of FaaS i.e. Function as a Service that can scale a small unit of custom code while depending on other third-party/backend services. Serverless is an architectural model that provides an event-driven way to implement distributed applications that can auto-scale on demand.
  • 29
    Podman

    Podman

    Containers

    What is Podman? Podman is a daemonless container engine for developing, managing, and running OCI Containers on your Linux System. Containers can either be run as root or in rootless mode. Simply put: alias docker=podman. Manage pods, containers, and container images. Supporting docker swarm. We believe that Kubernetes is the defacto standard for composing Pods and for orchestrating containers, making Kubernetes YAML a defacto standard file format. Hence, Podman allows the creation and execution of Pods from a Kubernetes YAML file (see podman-play-kube). Podman can also generate Kubernetes YAML based on a container or Pod (see podman-generate-kube), which allows for an easy transition from a local development environment to a production Kubernetes cluster.
  • 30
    IBM Cloud Kubernetes Service
    IBM Cloud® Kubernetes Service is a certified, managed Kubernetes solution, built for creating a cluster of compute hosts to deploy and manage containerized apps on IBM Cloud®. It provides intelligent scheduling, self-healing, horizontal scaling and securely manages the resources that you need to quickly deploy, update and scale applications. IBM Cloud Kubernetes Service manages the master, freeing you from having to manage the host OS, container runtime and Kubernetes version-update process.
  • 31
    Azure Container Registry
    Build, store, secure, scan, replicate, and manage container images and artifacts with a fully managed, geo-replicated instance of OCI distribution. Connect across environments, including Azure Kubernetes Service and Azure Red Hat OpenShift, and across Azure services like App Service, Machine Learning, and Batch. Geo-replication to efficiently manage a single registry across multiple regions. OCI artifact repository for adding helm charts, singularity support, and new OCI artifact-supported formats. Automated container building and patching including base image updates and task scheduling. Integrated security with Azure Active Directory (Azure AD) authentication, role-based access control, Docker content trust, and virtual network integration. Streamline building, testing, pushing, and deploying images to Azure with Azure Container Registry Tasks.
  • 32
    Azure Web App for Containers
    It has never been easier to deploy container-based web apps. Just pull container images from Docker Hub or a private Azure Container Registry, and Web App for Containers will deploy the containerized app with your preferred dependencies to production in seconds. The platform automatically takes care of OS patching, capacity provisioning, and load balancing. Automatically scale vertically and horizontally based on application needs. Granular scaling rules are available to handle peaks in workload automatically while minimizing costs during off-peak times. Deploy data and host services across multiple locations with just few mouse clicks.
  • 33
    Azure Red Hat OpenShift
    Azure Red Hat OpenShift provides highly available, fully managed OpenShift clusters on demand, monitored and operated jointly by Microsoft and Red Hat. Kubernetes is at the core of Red Hat OpenShift. OpenShift brings added-value features to complement Kubernetes, making it a turnkey container platform as a service (PaaS) with a significantly improved developer and operator experience. Highly available, fully managed public and private clusters, automated operations, and over-the-air platform upgrades. Take advantage of the enhanced user interface for application topology and builds in the web console to build, deploy, configure, and visualize containerized applications and cluster resources more easily.
  • 34
    Oracle Container Cloud Service
    Oracle Container Cloud Service (also known as Oracle Cloud Infrastructure Container Service Classic) offers Development and Operations teams the benefits of easy and secure Docker containerization when building and deploying applications. Provides an easy-to-use interface to manage the Docker environment. Provides out-of-the-box examples of containerized services and application stacks that can be deployed in one click. Enables developers to easily connect to their private Docker registries (so they can ‘bring their own containers’). Enables developers to focus on building containerized application images and Continuous Integration/Continuous Delivery (CI/CD) pipelines, not on learning complex orchestration technologies.
  • 35
    Otomi Container Platform
    Red Kubes is a Dutch start-up founded in 2019 by Sander Rodenhuis and Maurice Faber. After building and operating Kubernetes clusters for years, we noticed organizations are having difficulty keeping up with the increasing complexity of Kubernetes. To make Kubernetes easy and fun, we developed our first product called Otomi Container Platform, a value-added layer on top of Kubernetes to shorten time to market and speed up agility and innovation. One web UI to access all integrated applications and self-service features. A complete and out-of-the-box platform experience for Kubernetes. A suite of integrated applications for Kubernetes combined with automation. An overview of all supported Cloud/Infrastructure providers. Self-hosted Platform-as-a-Service for Kubernetes. Stop reinventing the wheel and get a full platform experience out-of-the-box.
  • 36
    Rancher

    Rancher

    Rancher Labs

    From datacenter to cloud to edge, Rancher lets you deliver Kubernetes-as-a-Service. Rancher is a complete software stack for teams adopting containers. It addresses the operational and security challenges of managing multiple Kubernetes clusters, while providing DevOps teams with integrated tools for running containerized workloads. From datacenter to cloud to edge, Rancher's open source software lets you run Kubernetes everywhere. Compare Rancher with other leading Kubernetes management platforms in how they deliver. You don’t need to figure Kubernetes out all on your own. Rancher is open source software, with an enormous community of users. Rancher Labs builds software that helps enterprises deliver Kubernetes-as-a-Service across any infrastructure. When running Kubernetes workloads in mission-critical environments, our community knows that they can turn to us for world-class support.
  • 37
    Tencent Cloud EKS
    EKS is community-driven and supports the latest Kubernetes version as well as native Kubernetes cluster management. It is ready-to-use in the form of a plugin to support Tencent Cloud products for storage, networking, load balancing, and more. EKS is built on Tencent Cloud's well-developed virtualization technology and network architecture, providing 99.95% service availability. Tencent Cloud ensures the virtual and network isolation of EKS clusters between users. You can configure network policies for specific products using security groups, network ACL, etc. The serverless framework of EKS ensures higher resource utilization and lower OPS costs. Flexible and efficient auto scaling ensures that EKS only consumes the amount of resources required by the current load. EKS provides solutions that meet different business needs and can be integrated with most Tencent Cloud services, such as CBS, CFS, COS, TencentDB products, VPC and more.
  • 38
    Tencent Kubernetes Engine
    TKE is fully compatible with the entire range of Kubernetes capabilities and has been adapted to Tencent Cloud's fundamental IaaS capabilities such as CVM and CBS. In addition, Tencent Cloud’s Kubernetes-based cloud products such as CBS and CLB support one-click deployment to container clusters for a variety of open source applications, greatly improving deployment efficiency. Thanks to TKE, you can simplify the management of large-scale clusters and management and OPS of distributed applications without having to use cluster management software or design fault-tolerant cluster architecture. Simply launch TKE and specify the tasks you want to run, and then TKE will take care of all of the cluster management tasks, allowing you to focus on developing Dockerized applications.
  • 39
    Azure App Service
    Quickly build, deploy, and scale web apps and APIs on your terms. Work with .NET, .NET Core, Node.js, Java, Python or PHP, in containers or running on Windows or Linux. Meet rigorous, enterprise-grade performance, security and compliance requirements used a trusted, fully managed platform that handles over 40 billion requests per day. Fully managed platform with built-in infrastructure maintenance, security patching, and scaling. Built-in CI/CD integration and zero-downtime deployments. Rigorous security and compliance, including SOC and PCI, for seamless deployments across public cloud, Azure Government, and on-premises environments. Bring your code or container using the framework language of your choice. Increase developer productivity with tight integration of Visual Studio Code and Visual Studio. Streamline CI/CD with Git, GitHub, GitHub Actions, Atlassian Bitbucket, Azure DevOps, Docker Hub, and Azure Container Registry.
  • 40
    Anthos

    Anthos

    Google

    Anthos lets you build, deploy, and manage applications anywhere in a secure, consistent manner. You can modernize existing applications running on virtual machines while deploying cloud-native apps on containers in an increasingly hybrid and multi-cloud world. Our application platform provides a consistent development and operations experience across all your deployments while reducing operational overhead and improving developer productivity. Anthos GKE: Enterprise-grade container orchestration and management service for running Kubernetes clusters anywhere, in both cloud and on-premises environments. Anthos Config Management: Define, automate, and enforce policies across environments in order to meet your organization’s unique security and compliance requirements. Anthos Service Mesh: Anthos unburdens operations and development teams by empowering them to manage and secure traffic between services while monitoring, troubleshooting, and improving application performance.
  • 41
    Chkk

    Chkk

    Chkk

    Prioritize your top business-critical risks with clear and actionable insights. Continuously harden your Kubernetes availability. Learn from others and avoid repeating their mistakes. Eliminate risks before they cause incidents. Stay up to date with visibility across all your infrastructure layers. Catalog containers, clusters, add-ons, and dependencies. Consolidate insights across clouds, on-prem, and more. Get alerted about all EOL and incompatible versions. Never use spreadsheets or custom scripts again. Chkk’s mission is to enable developers to proactively prevent incidents from happening by learning from others and not repeating known mistakes. Chkk's collective learning technology mines and curates known errors, failures, and disruptions that the Kubernetes community (comprising users/operators, cloud providers, and vendors) has encountered, ensuring that past mistakes are not repeated.
  • 42
    SUSE Rancher
    SUSE Rancher addresses the needs of DevOps teams deploying applications with Kubernetes and IT operations delivering enterprise-critical services. SUSE Rancher supports any CNCF-certified Kubernetes distribution. For on-premises workloads, we offer the RKE. We support all the public cloud distributions, including EKS, AKS, and GKE. At the edge, we offer K3s. SUSE Rancher provides simple, consistent cluster operations, including provisioning, version management, visibility and diagnostics, monitoring and alerting, and centralized audit. SUSE Rancher lets you automate processes and applies a consistent set of user access and security policies for all your clusters, no matter where they’re running. SUSE Rancher provides a rich catalogue of services for building, deploying, and scaling containerized applications, including app packaging, CI/CD, logging, monitoring, and service mesh.
  • 43
    Amazon EKS Anywhere
    Amazon EKS Anywhere is a new deployment option for Amazon EKS that enables you to easily create and operate Kubernetes clusters on-premises, including on your own virtual machines (VMs) and bare metal servers. EKS Anywhere provides an installable software package for creating and operating Kubernetes clusters on-premises and automation tooling for cluster lifecycle support. EKS Anywhere brings a consistent AWS management experience to your data center, building on the strengths of Amazon EKS Distro (the same Kubernetes that powers EKS on AWS.) EKS Anywhere saves you the complexity of buying or building your own management tooling to create EKS Distro clusters, configure the operating environment, update software, and handle backup and recovery. EKS Anywhere enables you to automate cluster management, reduce support costs, and eliminate the redundant effort of using multiple open source or 3rd party tools for operating Kubernetes clusters. EKS Anywhere is fully supported by AWS.
  • 44
    HPE Ezmeral

    HPE Ezmeral

    Hewlett Packard Enterprise

    Run, manage, control and secure the apps, data and IT that run your business, from edge to cloud. HPE Ezmeral advances digital transformation initiatives by shifting time and resources from IT operations to innovations. Modernize your apps. Simplify your Ops. And harness data to go from insights to impact. Accelerate time-to-value by deploying Kubernetes at scale with integrated persistent data storage for app modernization on bare metal or VMs, in your data center, on any cloud or at the edge. Harness data and get insights faster by operationalizing the end-to-end process to build data pipelines. Bring DevOps agility to the machine learning lifecycle, and deliver a unified data fabric. Boost efficiency and agility in IT Ops with automation and advanced artificial intelligence. And provide security and control to eliminate risk and reduce costs. HPE Ezmeral Container Platform provides an enterprise-grade platform to deploy Kubernetes at scale for a wide range of use cases.
  • 45
    Swarm

    Swarm

    Docker

    Current versions of Docker include swarm mode for natively managing a cluster of Docker Engines called a swarm. Use the Docker CLI to create a swarm, deploy application services to a swarm, and manage swarm behavior. Cluster management integrated with Docker Engine: Use the Docker Engine CLI to create a swarm of Docker Engines where you can deploy application services. You don’t need additional orchestration software to create or manage a swarm. Decentralized design: Instead of handling differentiation between node roles at deployment time, the Docker Engine handles any specialization at runtime. You can deploy both kinds of nodes, managers and workers, using the Docker Engine. This means you can build an entire swarm from a single disk image. Declarative service model: Docker Engine uses a declarative approach to let you define the desired state of the various services in your application stack.
  • 46
    Scalingo

    Scalingo

    Scalingo

    Automagic cloud hosting for awesome teams. From prototypes to production-ready applications, Scalingo empowers software teams to manage entire application lifecycles without sysadmin knowledge. Just push your code, your application stack and dependencies are automagically detected and packaged. Instant deployment, instant availability, instant work makes continuous delivery possible. Without DevOps. Start small, grow later, from a prototype to a production-grade infrastructure, instantly, billed by the minute. From 1 to 100 containers in seconds, to handle a peak of traffic or absorb your usual growth, with end-to-end automated processes taking care of continuous deployment. Dedicated resources are allocated for your application. All your containers are isolated from others with security built-in: we take care of OS and software stacks security fixes.
    Starting Price: €7.20 per month
  • 47
    Serverless

    Serverless

    Serverless

    Declare AWS Lambda functions and their triggers through simple abstract syntax in YAML. AWS Lambda functions, triggers & code will be deployed and wired together in the cloud, automatically. Install thousands of Serverless Framework Plugins to build new serverless use-cases on AWS and integrate with other tools. Track usage, performance, errors on your serverless apps with instant, powerful metrics. See all of your serverless applications and their resources in one place, regardless of AWS account or region. Easily share secrets, outputs from your serverless applications and delegate AWS account access. You can deploy many familiar use-cases instantly with the Serverless Framework. From REST APIs on Node.js, Python, Go, Java, to GraphQL APIs, scheduled tasks, Express.js applications, and front-end applications.
  • 48
    ArvanCloud

    ArvanCloud

    ArvanCloud

    ArvanCloud CDN comprises tens of PoP sites in important locations all around the world to deliver online content to the users, from the closest geographical point at the highest speed and quality. With ArvanCloud Cloud Computing infrastructure, you are just a few clicks away from creating unlimited cloud servers on demand. You can create several cloud storage disks for each server and easily manage your cloud data center communications using Firewall and private and public networks. ArvanCloud allows you to save any kind of data on Cloud Storage in a completely encrypted format. You can have stable access to a reliable storage system from all around the world, without worrying about data loss. With ArvanCloud container-based Platform as a Service, which conforms to Kubernetes standards, you are just a few commands in CLI away from an operational product.
  • 49
    Alibaba Function Compute
    Alibaba Cloud Function Compute is a fully managed, event-driven compute service. Function Compute allows you to focus on writing and uploading code without having to manage infrastructure such as servers. Function Compute provides compute resources to run code flexibly and reliably. Additionally, Function Compute provides a generous amount of free resources. No fees are incurred for up to 1,000,000 invocations and 400,000 CU-second compute resources per month.
  • 50
    K3s

    K3s

    K3s

    K3s is a highly available, certified Kubernetes distribution designed for production workloads in unattended, resource-constrained, remote locations or inside IoT appliances. Both ARM64 and ARMv7 are supported with binaries and multiarch images available for both. K3s works great from something as small as a Raspberry Pi to an AWS a1.4xlarge 32GiB server. Lightweight storage backend based on sqlite3 as the default storage mechanism. etcd3, MySQL, Postgres also still available. Secure by default with reasonable defaults for lightweight environments. Simple but powerful “batteries-included” features have been added, such as: a local storage provider, a service load balancer, a Helm controller, and the Traefik ingress controller. Operation of all Kubernetes control plane components is encapsulated in a single binary and process. This allows K3s to automate and manage complex cluster operations like distributing certificates.