Dataform
Dataform enables data analysts and data engineers to develop and operationalize scalable data transformation pipelines in BigQuery using only SQL from a single, unified environment. Its open source core language lets teams define table schemas, configure dependencies, add column descriptions, and set up data quality assertions within a shared code repository while applying software development best practices, version control, environments, testing, and documentation. A fully managed, serverless orchestration layer automatically handles workflow dependencies, tracks lineage, and executes SQL pipelines on demand or via schedules in Cloud Composer, Workflows, BigQuery Studio, or third-party services. In the browser-based development interface, users get real-time error feedback, visualize dependency graphs, connect to GitHub or GitLab for commits and code reviews, and launch production-grade pipelines in minutes without leaving BigQuery Studio.
Learn more
Kestra
Kestra is an open-source, event-driven orchestrator that simplifies data operations and improves collaboration between engineers and business users. By bringing Infrastructure as Code best practices to data pipelines, Kestra allows you to build reliable workflows and manage them with confidence.
Thanks to the declarative YAML interface for defining orchestration logic, everyone who benefits from analytics can participate in the data pipeline creation process. The UI automatically adjusts the YAML definition any time you make changes to a workflow from the UI or via an API call. Therefore, the orchestration logic is defined declaratively in code, even if some workflow components are modified in other ways.
Learn more
Datavolo
Capture all your unstructured data for all your LLM needs. Datavolo replaces single-use, point-to-point code with fast, flexible, reusable pipelines, freeing you to focus on what matters most, doing incredible work. Datavolo is the dataflow infrastructure that gives you a competitive edge. Get fast, unencumbered access to all of your data, including the unstructured files that LLMs rely on, and power up your generative AI. Get pipelines that grow with you, in minutes, not days, without custom coding. Instantly configure from any source to any destination at any time. Trust your data because lineage is built into every
pipeline. Make single-use pipelines and expensive configurations a thing of the past. Harness your unstructured data and unleash AI innovation with Datavolo, powered by Apache NiFi and built specifically for unstructured data. Our founders have spent a lifetime helping organizations make the most of their data.
Learn more
Cribl Stream
Cribl Stream allows you to implement an observability pipeline which helps you parse, restructure, and enrich data in flight - before you pay to analyze it. Get the right data, where you want, in the formats you need. Route data to the best tool for the job - or all the tools for the job - by translating and formatting data into any tooling schema you require. Let different departments choose different analytics environments without having to deploy new agents or forwarders. As much as 50% of log and metric data goes unused – null fields, duplicate data, and fields that offer zero analytical value. With Cribl Stream, you can trim wasted data streams and analyze only what you need. Cribl Stream is the best way to get multiple data formats into the tools you trust for your Security and IT efforts. Use the Cribl Stream universal receiver to collect from any machine data source - and even to schedule batch collection from REST APIs, Kinesis Firehose, Raw HTTP, and Microsoft Office 365 APIs
Learn more