Compare the Top Data Pipeline Software that integrates with DataHub as of October 2025

This a list of Data Pipeline software that integrates with DataHub. Use the filters on the left to add additional filters for products that have integrations with DataHub. View the products that work with DataHub in the table below.

What is Data Pipeline Software for DataHub?

Data pipeline software helps businesses automate the movement, transformation, and storage of data from various sources to destinations such as data warehouses, lakes, or analytic platforms. These platforms provide tools for extracting data from multiple sources, processing it in real-time or batch, and loading it into target systems for analysis or reporting (ETL: Extract, Transform, Load). Data pipeline software often includes features for data monitoring, error handling, scheduling, and integration with other software tools, making it easier for organizations to ensure data consistency, accuracy, and flow. By using this software, businesses can streamline data workflows, improve decision-making, and ensure that data is readily available for analysis. Compare and read user reviews of the best Data Pipeline software for DataHub currently available using the table below. This list is updated regularly.

  • 1
    dbt

    dbt

    dbt Labs

    dbt Labs powers the transformation layer of modern data pipelines. Once data has been ingested into a warehouse or lakehouse, dbt enables teams to clean, model, and document it so it’s ready for analytics and AI. With dbt, teams can: - Transform raw data at scale with SQL and Jinja. - Orchestrate pipelines with built-in dependency management and scheduling. - Ensure trust with automated testing and continuous integration. - Visualize lineage across models for faster impact analysis. By embedding software engineering practices into pipeline development, dbt Labs helps data teams build reliable, production-grade pipelines — reducing data debt and accelerating time to insight.
    Starting Price: $100 per user per user/ month
    View Software
    Visit Website
  • 2
    Apache Kafka

    Apache Kafka

    The Apache Software Foundation

    Apache Kafka® is an open-source, distributed streaming platform. Scale production clusters up to a thousand brokers, trillions of messages per day, petabytes of data, hundreds of thousands of partitions. Elastically expand and contract storage and processing. Stretch clusters efficiently over availability zones or connect separate clusters across geographic regions. Process streams of events with joins, aggregations, filters, transformations, and more, using event-time and exactly-once processing. Kafka’s out-of-the-box Connect interface integrates with hundreds of event sources and event sinks including Postgres, JMS, Elasticsearch, AWS S3, and more. Read, write, and process streams of events in a vast array of programming languages.
  • 3
    Dagster

    Dagster

    Dagster Labs

    Dagster is a next-generation orchestration platform for the development, production, and observation of data assets. Unlike other data orchestration solutions, Dagster provides you with an end-to-end development lifecycle. Dagster gives you control over your disparate data tools and empowers you to build, test, deploy, run, and iterate on your data pipelines. It makes you and your data teams more productive, your operations more robust, and puts you in complete control of your data processes as you scale. Dagster brings a declarative approach to the engineering of data pipelines. Your team defines the data assets required, quickly assessing their status and resolving any discrepancies. An assets-based model is clearer than a tasks-based one and becomes a unifying abstraction across the whole workflow.
    Starting Price: $0
  • 4
    Prefect

    Prefect

    Prefect

    Prefect Cloud is a command center for your workflows. Deploy from Prefect core and instantly gain complete oversight and control. Cloud's beautiful UI lets you keep an eye on the health of your infrastructure. Stream realtime state updates and logs, kick off new runs, and receive critical information exactly when you need it. With Prefect's Hybrid Model, your code and data remain on-prem while Prefect Cloud's managed orchestration keeps everything running smoothly. The Cloud scheduler service runs asynchronously to ensure your runs start on time, every time. Advanced scheduling options allow for scheduled parameter value changes as well as the execution environment for each run! Configure custom notifications and actions when your workflows change state. Monitor the health of all agents connected to your cloud instance and receive custom alerts when an agent goes offline.
    Starting Price: $0.0025 per successful task
  • 5
    Databricks Data Intelligence Platform
    The Databricks Data Intelligence Platform allows your entire organization to use data and AI. It’s built on a lakehouse to provide an open, unified foundation for all data and governance, and is powered by a Data Intelligence Engine that understands the uniqueness of your data. The winners in every industry will be data and AI companies. From ETL to data warehousing to generative AI, Databricks helps you simplify and accelerate your data and AI goals. Databricks combines generative AI with the unification benefits of a lakehouse to power a Data Intelligence Engine that understands the unique semantics of your data. This allows the Databricks Platform to automatically optimize performance and manage infrastructure in ways unique to your business. The Data Intelligence Engine understands your organization’s language, so search and discovery of new data is as easy as asking a question like you would to a coworker.
  • 6
    Apache Airflow

    Apache Airflow

    The Apache Software Foundation

    Airflow is a platform created by the community to programmatically author, schedule and monitor workflows. Airflow has a modular architecture and uses a message queue to orchestrate an arbitrary number of workers. Airflow is ready to scale to infinity. Airflow pipelines are defined in Python, allowing for dynamic pipeline generation. This allows for writing code that instantiates pipelines dynamically. Easily define your own operators and extend libraries to fit the level of abstraction that suits your environment. Airflow pipelines are lean and explicit. Parametrization is built into its core using the powerful Jinja templating engine. No more command-line or XML black-magic! Use standard Python features to create your workflows, including date time formats for scheduling and loops to dynamically generate tasks. This allows you to maintain full flexibility when building your workflows.
  • Previous
  • You're on page 1
  • Next