Best Artificial Intelligence Software for Apache Airflow

Compare the Top Artificial Intelligence Software that integrates with Apache Airflow as of June 2025

This a list of Artificial Intelligence software that integrates with Apache Airflow. Use the filters on the left to add additional filters for products that have integrations with Apache Airflow. View the products that work with Apache Airflow in the table below.

What is Artificial Intelligence Software for Apache Airflow?

Artificial Intelligence (AI) software is computer technology designed to simulate human intelligence. It can be used to perform tasks that require cognitive abilities, such as problem-solving, data analysis, visual perception and language translation. AI applications range from voice recognition and virtual assistants to autonomous vehicles and medical diagnostics. Compare and read user reviews of the best Artificial Intelligence software for Apache Airflow currently available using the table below. This list is updated regularly.

  • 1
    DataBuck

    DataBuck

    FirstEigen

    DataBuck is an AI-powered data validation platform that automates risk detection across dynamic, high-volume, and evolving data environments. DataBuck empowers your teams to: βœ… Enhance trust in analytics and reports, ensuring they are built on accurate and reliable data. βœ… Reduce maintenance costs by minimizing manual intervention. βœ… Scale operations 10x faster compared to traditional tools, enabling seamless adaptability in ever-changing data ecosystems. By proactively addressing system risks and improving data accuracy, DataBuck ensures your decision-making is driven by dependable insights. Proudly recognized in Gartner’s 2024 Market Guide for #DataObservability, DataBuck goes beyond traditional observability practices with its AI/ML innovations to deliver autonomous Data Trustabilityβ€”empowering you to lead with confidence in today’s data-driven world.
    View Software
    Visit Website
  • 2
    Coursebox AI

    Coursebox AI

    Coursebox

    Transform your content into engaging eLearning experiences with Coursebox, the #1 AI-powered eLearning authoring tool. Our platform automates the course creation process, allowing you to design a structured course in seconds. Simply make edits, add any missing elements, and your course is ready to go. Whether you want to publish privately, share publicly, sell your course, or export it to your LMS, Coursebox has you covered. With a focus on mobile-first learning, Coursebox ensures your learners are more engaged and motivated through interactive and visual content, including videos, quizzes, and more. Take advantage of our branded learning management system with native mobile apps, offering you the flexibility to use custom hosting and a custom domain. Coursebox is the ultimate solution for organizations and individuals looking to rapidly scale their training and assessment programs.
    Starting Price: $99 per month
    View Software
    Visit Website
  • 3
    Netdata

    Netdata

    Netdata, Inc.

    The open-source observability platform everyone needs! Netdata collects metrics per second and presents them in beautiful low-latency dashboards. It is designed to run on all of your physical and virtual servers, cloud deployments, Kubernetes clusters, and edge/IoT devices, to monitor your systems, containers, and applications. It scales nicely from just a single server to thousands of servers, even in complex multi/mixed/hybrid cloud environments, and given enough disk space it can keep your metrics for years. KEY FEATURES: πŸ’₯ Collects metrics from 800+ integrations πŸ’ͺ Real-Time, Low-Latency, High-Resolution πŸ˜Άβ€πŸŒ«οΈ Unsupervised Anomaly Detection πŸ”₯ Powerful Visualization πŸ”” Out of box Alerts πŸ“– systemd Journal Logs Explorer 😎 Low Maintenance ⭐ Open and Extensible Try Netdata today and feel the pulse of your infrastructure, with high-resolution metrics, journal logs and real-time visualizations.
    Leader badge
    Starting Price: Free
  • 4
    Microsoft Purview
    Microsoft Purview is a unified data governance service that helps you manage and govern your on-premises, multicloud, and software-as-a-service (SaaS) data. Easily create a holistic, up-to-date map of your data landscape with automated data discovery, sensitive data classification, and end-to-end data lineage. Empower data consumers to find valuable, trustworthy data. Automated data discovery, lineage identification, and data classification across on-premises, multicloud, and SaaS sources. Unified map of your data assets and their relationships for more effective governance. Semantic search enables data discovery using business or technical terms. Insight into the location and movement of sensitive data across your hybrid data landscape. Establish the foundation for effective data usage and governance with Purview Data Map. Automate and manage metadata from hybrid sources. Classify data using built-in and custom classifiers and Microsoft Information Protection sensitivity labels.
    Starting Price: $0.342
  • 5
    Ray

    Ray

    Anyscale

    Develop on your laptop and then scale the same Python code elastically across hundreds of nodes or GPUs on any cloud, with no changes. Ray translates existing Python concepts to the distributed setting, allowing any serial application to be easily parallelized with minimal code changes. Easily scale compute-heavy machine learning workloads like deep learning, model serving, and hyperparameter tuning with a strong ecosystem of distributed libraries. Scale existing workloads (for eg. Pytorch) on Ray with minimal effort by tapping into integrations. Native Ray libraries, such as Ray Tune and Ray Serve, lower the effort to scale the most compute-intensive machine learning workloads, such as hyperparameter tuning, training deep learning models, and reinforcement learning. For example, get started with distributed hyperparameter tuning in just 10 lines of code. Creating distributed apps is hard. Ray handles all aspects of distributed execution.
    Starting Price: Free
  • 6
    Dagster

    Dagster

    Dagster Labs

    Dagster is a next-generation orchestration platform for the development, production, and observation of data assets. Unlike other data orchestration solutions, Dagster provides you with an end-to-end development lifecycle. Dagster gives you control over your disparate data tools and empowers you to build, test, deploy, run, and iterate on your data pipelines. It makes you and your data teams more productive, your operations more robust, and puts you in complete control of your data processes as you scale. Dagster brings a declarative approach to the engineering of data pipelines. Your team defines the data assets required, quickly assessing their status and resolving any discrepancies. An assets-based model is clearer than a tasks-based one and becomes a unifying abstraction across the whole workflow.
    Starting Price: $0
  • 7
    Prophecy

    Prophecy

    Prophecy

    Prophecy enables many more users - including visual ETL developers and Data Analysts. All you need to do is point-and-click and write a few SQL expressions to create your pipelines. As you use the Low-Code designer to build your workflows - you are developing high quality, readable code for Spark and Airflow that is committed to your Git. Prophecy gives you a gem builder - for you to quickly develop and rollout your own Frameworks. Examples are Data Quality, Encryption, new Sources and Targets that extend the built-in ones. Prophecy provides best practices and infrastructure as managed services – making your life and operations simple! With Prophecy, your workflows are high performance and use scale-out performance & scalability of the cloud.
    Starting Price: $299 per month
  • 8
    BentoML

    BentoML

    BentoML

    Serve your ML model in any cloud in minutes. Unified model packaging format enabling both online and offline serving on any platform. 100x the throughput of your regular flask-based model server, thanks to our advanced micro-batching mechanism. Deliver high-quality prediction services that speak the DevOps language and integrate perfectly with common infrastructure tools. Unified format for deployment. High-performance model serving. DevOps best practices baked in. The service uses the BERT model trained with the TensorFlow framework to predict movie reviews' sentiment. DevOps-free BentoML workflow, from prediction service registry, deployment automation, to endpoint monitoring, all configured automatically for your team. A solid foundation for running serious ML workloads in production. Keep all your team's models, deployments, and changes highly visible and control access via SSO, RBAC, client authentication, and auditing logs.
    Starting Price: Free
  • 9
    ZenML

    ZenML

    ZenML

    Simplify your MLOps pipelines. Manage, deploy, and scale on any infrastructure with ZenML. ZenML is completely free and open-source. See the magic with just two simple commands. Set up ZenML in a matter of minutes, and start with all the tools you already use. ZenML standard interfaces ensure that your tools work together seamlessly. Gradually scale up your MLOps stack by switching out components whenever your training or deployment requirements change. Keep up with the latest changes in the MLOps world and easily integrate any new developments. Define simple and clear ML workflows without wasting time on boilerplate tooling or infrastructure code. Write portable ML code and switch from experimentation to production in seconds. Manage all your favorite MLOps tools in one place with ZenML's plug-and-play integrations. Prevent vendor lock-in by writing extensible, tooling-agnostic, and infrastructure-agnostic code.
    Starting Price: Free
  • 10
    DataHub

    DataHub

    DataHub

    DataHub is an open source metadata platform designed to streamline data discovery, observability, and governance across diverse data ecosystems. It enables organizations to effortlessly discover trustworthy data, with experiences tailored for each person and eliminates breaking changes with detailed cross-platform and column-level lineage. DataHub builds confidence in your data by providing a comprehensive view of business, operational, and technical context, all in one place. The platform offers automated data quality checks and AI-driven anomaly detection, notifying teams when issues arise and centralizing incident tracking. With detailed lineage, documentation, and ownership information, DataHub facilitates swift issue resolution. It also automates governance programs by classifying assets as they evolve, minimizing manual work through GenAI documentation, AI-driven classification, and smart propagation. DataHub's extensible architecture supports over 70 native integrations.
    Starting Price: Free
  • 11
    Chalk

    Chalk

    Chalk

    Powerful data engineering workflows, without the infrastructure headaches. Complex streaming, scheduling, and data backfill pipelines, are all defined in simple, composable Python. Make ETL a thing of the past, fetch all of your data in real-time, no matter how complex. Incorporate deep learning and LLMs into decisions alongside structured business data. Make better predictions with fresher data, don’t pay vendors to pre-fetch data you don’t use, and query data just in time for online predictions. Experiment in Jupyter, then deploy to production. Prevent train-serve skew and create new data workflows in milliseconds. Instantly monitor all of your data workflows in real-time; track usage, and data quality effortlessly. Know everything you computed and data replay anything. Integrate with the tools you already use and deploy to your own infrastructure. Decide and enforce withdrawal limits with custom hold times.
    Starting Price: Free
  • 12
    Acryl Data

    Acryl Data

    Acryl Data

    No more data catalog ghost towns. Acryl Cloud drives fast time-to-value via Shift Left practices for data producers and an intuitive UI for data consumers. Continuously detect data quality incidents in real-time, automate anomaly detection to prevent breakages, and drive fast resolution when they do occur. Acryl Cloud supports both push-based and pull-based metadata ingestion for easy maintenance, ensuring information is trustworthy, up-to-date, and definitive. Data should be operational. Go beyond simple visibility and use automated Metadata Tests to continuously expose data insights and surface new areas for improvement. Reduce confusion and accelerate resolution with clear asset ownership, automatic detection, streamlined alerts, and time-based lineage for tracing root causes.
  • 13
    Determined AI

    Determined AI

    Determined AI

    Distributed training without changing your model code, determined takes care of provisioning machines, networking, data loading, and fault tolerance. Our open source deep learning platform enables you to train models in hours and minutes, not days and weeks. Instead of arduous tasks like manual hyperparameter tuning, re-running faulty jobs, and worrying about hardware resources. Our distributed training implementation outperforms the industry standard, requires no code changes, and is fully integrated with our state-of-the-art training platform. With built-in experiment tracking and visualization, Determined records metrics automatically, makes your ML projects reproducible and allows your team to collaborate more easily. Your researchers will be able to build on the progress of their team and innovate in their domain, instead of fretting over errors and infrastructure.
  • 14
    Cake AI

    Cake AI

    Cake AI

    Cake AI is a comprehensive AI infrastructure platform that enables teams to build and deploy AI applications using hundreds of pre-integrated open source components, offering complete visibility and control. It provides a curated, end-to-end selection of fully managed, best-in-class commercial and open source AI tools, with pre-built integrations across the full breadth of components needed to move an AI application into production. Cake supports dynamic autoscaling, comprehensive security measures including role-based access control and encryption, advanced monitoring, and infrastructure flexibility across various environments, including Kubernetes clusters and cloud services such as AWS. Its data layer equips teams with tools for data ingestion, transformation, and analytics, leveraging tools like Airflow, DBT, Prefect, Metabase, and Superset. For AI operations, Cake integrates with model catalogs like Hugging Face and supports modular workflows using LangChain, LlamaIndex, and more.
  • Previous
  • You're on page 1
  • Next