Compare the Top Function as a Service (FaaS) Providers that integrate with Docker as of November 2025

This a list of Function as a Service (FaaS) providers that integrate with Docker. Use the filters on the left to add additional filters for products that have integrations with Docker. View the products that work with Docker in the table below.

What are Function as a Service (FaaS) Providers for Docker?

Function as a Service (FaaS) providers are cloud computing services that allow developers to execute small units of code, or functions, in response to events without managing the underlying infrastructure. These providers offer a serverless architecture where the scaling, monitoring, and management of infrastructure are handled automatically, allowing developers to focus on writing code. FaaS platforms typically charge based on the number of executions and the duration of the function, making it a cost-efficient option for event-driven applications, microservices, and real-time data processing. Compare and read user reviews of the best Function as a Service (FaaS) providers for Docker currently available using the table below. This list is updated regularly.

  • 1
    Google Cloud Run
    While Cloud Run is primarily a CaaS platform, it also offers functionality similar to Function as a Service (FaaS), allowing developers to deploy single functions in containers and have them run in response to HTTP requests. This serverless model means that developers only pay for the actual compute time their functions use, without worrying about managing servers or infrastructure. Cloud Run automatically handles scaling, ensuring that functions are spun up and down based on demand, allowing for efficient resource allocation. The $300 in free credits provided to new customers is an excellent opportunity to experiment with Cloud Run’s FaaS capabilities, helping developers understand how it can simplify their application architecture. Cloud Run’s flexibility in supporting both containerized applications and serverless functions makes it a versatile tool for building scalable solutions. Whether for microservices or serverless functions, Cloud Run offers a unified environment.
    Starting Price: Free (2 mil requests/month)
    View Provider
    Visit Website
  • 2
    RunPod

    RunPod

    RunPod

    RunPod offers a cloud-based platform designed for running AI workloads, focusing on providing scalable, on-demand GPU resources to accelerate machine learning (ML) model training and inference. With its diverse selection of powerful GPUs like the NVIDIA A100, RTX 3090, and H100, RunPod supports a wide range of AI applications, from deep learning to data processing. The platform is designed to minimize startup time, providing near-instant access to GPU pods, and ensures scalability with autoscaling capabilities for real-time AI model deployment. RunPod also offers serverless functionality, job queuing, and real-time analytics, making it an ideal solution for businesses needing flexible, cost-effective GPU resources without the hassle of managing infrastructure.
    Starting Price: $0.40 per hour
    View Provider
    Visit Website
  • 3
    Fly.io

    Fly.io

    Fly.io

    We run physical servers in cities close to your users. As close to the metal as you can get without paying shipping. We run physical servers in cities close to your users. As close to the metal as you can get without paying shipping. Provision exactly what you need to make your apps fly. Pay only for what you use. The plumbing you need to sleep at night. Pretty graphs for morning coffee. Zero configuration private networking and global load balancing. Add certificates for your own purposes, or a million for your customers. We'll be using the standard web application generated by Remix. This is a bare-bones app with no database. If you're feeling ambitious, check out the Remix tutorial for building and deploying an SQLite-backed app on Fly.io.
    Starting Price: $0.02 per GB
  • 4
    Oracle Cloud Functions
    ​Oracle Cloud Infrastructure (OCI) Functions is a serverless computing service that enables developers to create, run, and scale applications without managing infrastructure. Built on the open source Fn Project, it supports multiple programming languages, including Python, Go, Java, Node.js, and C#, allowing for flexible function development. Developers can deploy code directly, with OCI handling automatic provisioning and scaling of resources. It offers provisioned concurrency to maintain low-latency execution, ensuring functions are ready to accept calls instantly. A catalog of prebuilt functions is available, enabling rapid deployment of common tasks without the need to write code from scratch. Functions are packaged as Docker images, and advanced users can utilize Dockerfiles to customize runtime environments. Integration with Oracle Identity and Access Management provides fine-grained access control, while OCI Vault securely stores sensitive configuration data.
    Starting Price: $0.0000002 per month
  • 5
    OpenFaaS

    OpenFaaS

    OpenFaaS

    Serverless functions, made simple. OpenFaaS® makes it simple to deploy both functions and existing code to Kubernetes. Avoid lock-in through the use of Docker. Run-on any public or private cloud. Build both microservices & functions in any language. Legacy code and binaries. Auto-scale for demand or to zero when idle. Bring your laptop, your own on-prem hardware, or create a cluster in the cloud. Let Kubernetes do the heavy lifting enabling you to build a scalable, fault-tolerant event-driven serverless platform for your applications. You can try out OpenFaaS in 60 seconds or write and deploy your first Python function in around 10-15 minutes. From there you can take the OpenFaaS workshop, a series of tried-and-tested self-paced labs which teach you everything you need to know about functions, and more. An ecosystem for sharing, reusing, and collaborating on functions. Reduce boilerplate code, and share code in the templates store.
  • 6
    NVIDIA Cloud Functions
    NVIDIA Cloud Functions (NVCF) is a serverless API designed to deploy and manage AI workloads on GPUs, providing security, scalability, and reliability. It supports HTTP polling, HTTP streaming, and gRPC protocols for accessing workloads. Cloud Functions is primarily suited for shorter-running, preemptable workloads such as inferencing and fine-tuning. It supports two function types: "Container" and "Helm Chart", allowing users to define functions based on their specific requirements. Workloads are ephemeral and preemptable, so users should ensure they save their work appropriately. Models, containers, helm charts, and additional resources are hosted and pulled from the NGC Private Registry. To get started, users can follow the functions quickstart guide, which provides an end-to-end workflow for creating and deploying a container-based function using the fastapi_echo_sample container.
  • Previous
  • You're on page 1
  • Next