Best Data Management Software for Databricks Data Intelligence Platform - Page 5

Compare the Top Data Management Software that integrates with Databricks Data Intelligence Platform as of October 2025 - Page 5

This a list of Data Management software that integrates with Databricks Data Intelligence Platform. Use the filters on the left to add additional filters for products that have integrations with Databricks Data Intelligence Platform. View the products that work with Databricks Data Intelligence Platform in the table below.

  • 1
    Mimic

    Mimic

    Facteus

    Advanced technology and services to safely transform and enhance sensitive data into actionable insights, help drive innovation, and open new revenue streams. Using the Mimic synthetic data engine, companies can safely synthesize their data assets, protecting consumer privacy information from being exposed, while still maintaining the statistical relevancy of the data. The synthetic data can then be used for internal initiatives like analytics, machine learning and AI, marketing and segmentation activities, and new revenue streams through external data monetization. Mimic enables you to safely move statistically-relevant synthetic data to the cloud ecosystem of your choice to get the most out of your data. Analytics, insights, product development, testing, and third-party data sharing can all be done in the cloud with the enhanced synthetic data, which has been certified to be compliant with regulatory and privacy laws.
  • 2
    Qualytics

    Qualytics

    Qualytics

    Helping enterprises proactively manage their full data quality lifecycle through contextual data quality checks, anomaly detection and remediation. Expose anomalies and metadata to help teams take corrective actions. Automatically trigger remediation workflows to resolve errors quickly and efficiently. Maintain high data quality and prevent errors from affecting business decisions. The SLA chart provides an overview of SLA, including the total number of SLA monitoring that have been performed and any violations that have occurred. This chart can help you identify areas of your data that may require further investigation or improvement.
  • 3
    LlamaIndex

    LlamaIndex

    LlamaIndex

    LlamaIndex is a “data framework” to help you build LLM apps. Connect semi-structured data from API's like Slack, Salesforce, Notion, etc. LlamaIndex is a simple, flexible data framework for connecting custom data sources to large language models. LlamaIndex provides the key tools to augment your LLM applications with data. Connect your existing data sources and data formats (API's, PDF's, documents, SQL, etc.) to use with a large language model application. Store and index your data for different use cases. Integrate with downstream vector store and database providers. LlamaIndex provides a query interface that accepts any input prompt over your data and returns a knowledge-augmented response. Connect unstructured sources such as documents, raw text files, PDF's, videos, images, etc. Easily integrate structured data sources from Excel, SQL, etc. Provides ways to structure your data (indices, graphs) so that this data can be easily used with LLMs.
  • 4
    Acryl Data

    Acryl Data

    Acryl Data

    No more data catalog ghost towns. Acryl Cloud drives fast time-to-value via Shift Left practices for data producers and an intuitive UI for data consumers. Continuously detect data quality incidents in real-time, automate anomaly detection to prevent breakages, and drive fast resolution when they do occur. Acryl Cloud supports both push-based and pull-based metadata ingestion for easy maintenance, ensuring information is trustworthy, up-to-date, and definitive. Data should be operational. Go beyond simple visibility and use automated Metadata Tests to continuously expose data insights and surface new areas for improvement. Reduce confusion and accelerate resolution with clear asset ownership, automatic detection, streamlined alerts, and time-based lineage for tracing root causes.
  • 5
    Demyst

    Demyst

    Demyst

    External data is the next frontier of business impact, powering competitive advantages across industries, but businesses struggle with the complexity of implementation. Demyst provides the end-to-end tools you need to discover, onboard, and ingest the right external data, with our experts working closely with you every step of the way. Browse and instantly deploy the right data from Demyst’s catalog of data sources, or our expert team will recommend and onboard something new for you from any external data provider around the globe. Demyst's data provider certification program means we procure and diligence data for your use, all covered under our contract. Demyst removes the "compliance versus speed" trade-off, performing ongoing legal, privacy and security due diligence for your safe and compliant data access, whilst typically onboarding new data in 4 weeks or less. Demyst performs the last mile. Deploy and monitor the data you need with consistently formatted APIs or files.
  • 6
    Unstructured

    Unstructured

    Unstructured

    80% of enterprise data exists in difficult-to-use formats like HTML, PDF, CSV, PNG, PPTX, and more. Unstructured effortlessly extracts and transforms complex data for use with every major vector database and LLM framework. Unstructured allows data scientists to pre-process data at scale so they spend less time collecting and cleaning, and more time modeling and analyzing. Our enterprise-grade connectors capture data wherever it lives, so we can transform it into AI-friendly JSON files for companies who are eager to fold AI into their business. You can count on Unstructured to deliver data that's curated, clean of artifacts, and most importantly, LLM-ready.
  • 7
    APERIO DataWise
    Data is used in every aspect of a processing plant or facility, it is underlying most operational processes, most business decisions, and most environmental events. Failures are often attributed to this same data, in terms of operator error, bad sensors, safety or environmental events, or poor analytics. This is where APERIO can alleviate these problems. Data integrity is a key element of Industry 4.0; the foundation upon which more advanced applications, such as predictive models, process optimization, and custom AI tools are developed. APERIO DataWise is the industry-leading provider of reliable, trusted data. Automate the quality of your PI data or digital twins continuously and at scale. Ensure validated data across the enterprise to improve asset reliability. Empower the operator to make better decisions. Detect threats made to operational data to ensure operational resilience. Accurately monitor & report sustainability metrics.
  • 8
    Virtualitics

    Virtualitics

    Virtualitics

    Embedded AI and rich 3D visualizations empower analysts to deliver transformational business strategies, never miss the critical insights in your data again. Virtualitics’ Intelligent Exploration empowers analysts with embedded AI-guided exploration, automatically surfacing insights that drive transformative action. Understand what you’re seeing with AI-guided exploration, explained in plain language so nothing gets missed. Drill into all the relevant data, no matter type or complexity, to discover key relationships in seconds. Increase engagement and understanding with rich 3D visualizations that bring data stories to life. Analyze data from new angles with 3D and VR data visualizations that make deciphering complex findings easier. Share strategic insight with annotated discoveries and clear explanations for all stakeholders.
  • 9
    Kestra

    Kestra

    Kestra

    Kestra is an open-source, event-driven orchestrator that simplifies data operations and improves collaboration between engineers and business users. By bringing Infrastructure as Code best practices to data pipelines, Kestra allows you to build reliable workflows and manage them with confidence. Thanks to the declarative YAML interface for defining orchestration logic, everyone who benefits from analytics can participate in the data pipeline creation process. The UI automatically adjusts the YAML definition any time you make changes to a workflow from the UI or via an API call. Therefore, the orchestration logic is defined declaratively in code, even if some workflow components are modified in other ways.
  • 10
    Pantomath

    Pantomath

    Pantomath

    Organizations continuously strive to be more data-driven, building dashboards, analytics, and data pipelines across the modern data stack. Unfortunately, most organizations struggle with data reliability issues leading to poor business decisions and lack of trust in data as an organization, directly impacting their bottom line. Resolving complex data issues is a manual and time-consuming process involving multiple teams all relying on tribal knowledge to manually reverse engineer complex data pipelines across different platforms to identify root-cause and understand the impact. Pantomath is a data pipeline observability and traceability platform for automating data operations. It continuously monitors datasets and jobs across the enterprise data ecosystem providing context to complex data pipelines by creating automated cross-platform technical pipeline lineage.
  • 11
    Validio

    Validio

    Validio

    See how your data assets are used: popularity, utilization, and schema coverage. Get important insights about your data assets such as popularity, utilization, quality, and schema coverage. Find and filter the data you need based on metadata tags and descriptions. Get important insights about your data assets such as popularity, utilization, quality, and schema coverage. Drive data governance and ownership across your organization. Stream-lake-warehouse lineage to facilitate data ownership and collaboration. Automatically generated field-level lineage map to understand the entire data ecosystem. Anomaly detection learns from your data and seasonality patterns, with automatic backfill from historical data. Machine learning-based thresholds are trained per data segment, trained on actual data instead of metadata only.
  • 12
    DataForge

    DataForge

    DataForge

    DataForge is the only framework to cover all three major data components of data development By combining unique elements within each component with an overall approach, DataForge creates the best foundation for any data design. DataForge Cloud (DFC) is a fully-featured data platform management service built around DataForge It translates the framework into automated developer workflows and provides tools to execute processing on platforms like Databricks and Snowflake. Through a combination of code structure and an event-based workflow engine, DFC fully automates the defining of required processing steps and dependency management. With standardized processing steps comes predictable infrastructure sizing. DFC leverages dynamic allocation and the best cluster/warehouse option at every step to ensure the best performance per cost possible.
    Starting Price: $2.50 per process
  • 13
    Tarsal

    Tarsal

    Tarsal

    Tarsal's infinite scalability means as your organization grows, Tarsal grows with you. Tarsal makes it easy for you to switch where you're sending data - today's SIEM data is tomorrow's data lake data; all with one click. Keep your SIEM and gradually migrate analytics over to a data lake. You don't have to rip anything out to use Tarsal. Some analytics just won't run on your SIEM. Use Tarsal to have query-ready data on a data lake. Your SIEM is one of the biggest line items in your budget. Use Tarsal to send some of that data to your data lake. Tarsal is the first highly scalable ETL data pipeline built for security teams. Easily exfil terabytes of data in just just a few clicks, with instant normalization, and route that data to your desired destination.
  • 14
    Velotix

    Velotix

    Velotix

    Velotix empowers organizations to maximize the value of their data while ensuring security and compliance in a rapidly evolving regulatory landscape. The Velotix Data Security Platform offers automated policy management, dynamic access controls, and comprehensive data discovery, all driven by advanced AI. With seamless integration across multi-cloud environments, Velotix enables secure, self-service data access, optimizing data utilization without compromising on governance. Trusted by leading enterprises across financial services, healthcare, telecommunications, and more, Velotix is reshaping data governance for the ‘need to share’ era.
  • 15
    GrowthLoop

    GrowthLoop

    GrowthLoop

    GrowthLoop is the company behind the #1 G2 rated AI-powered composable CDP. The world’s most advanced marketing teams rely on GrowthLoop’s cloud-native platform to unleash the power of their first-party data to grow revenue, increase conversion rates, and improve retention & loyalty. Enterprises get the benefit of compounding growth by activating a self-reinforcing marketing loop: AI agents that build audiences on the data cloud, omni-channel journey orchestration, and results that are automatically turned into actionable suggestions for their next campaign. Built for marketers, GrowthLoop is trusted by brands like Indeed, Mercari, NASCAR, and Express.
  • 16
    Decentriq

    Decentriq

    Decentriq

    Privacy-minded organizations work with Decentriq. With the latest advancements in encryption and privacy-enhancing technologies such as synthetic data, differential privacy, and confidential computing, your data stays under your control at all times. End-to-end encryption keeps your data private to all other parties. Decentriq cannot see or access your data. Remote attestation gives you verification that your data is encrypted and only approved analyses are running. Built-in partnership with market-leading hardware and infrastructure providers. Designed to handle even advanced AI and machine learning models, the platform keeps your data inaccessible no matter the challenge. With processing speeds approaching typical cloud levels, you don’t have to sacrifice scalability for excellent data protection. Our growing network of data connectors supports more streamlined workflows across leading data platforms.
  • 17
    Timbr.ai

    Timbr.ai

    Timbr.ai

    The smart semantic layer integrates data with business meaning and relationships, unifies metrics, and accelerates the delivery of data products with 90% shorter SQL queries. Easily model data using business terms to give it common meaning and align business metrics. Define semantic relationships that substitute JOINs so queries become much simpler. Use hierarchies and classifications to better understand data. Automatically map data to the semantic model. Join multiple data sources with a powerful distributed SQL engine to query data at scale. Consume data as a connected semantic graph. Boost performance and save compute costs with an intelligent cache engine and materialized views. Benefit from advanced query optimizations. Connect to most clouds, datalakes, data warehouses, databases, and any file format. Timbr empowers you to work with your data sources seamlessly. When a query is run, Timbr optimizes the query and pushes it down to the backend.
  • 18
    Salesforce Data Cloud
    Salesforce Data Cloud is a real-time data platform designed to unify and manage customer data from multiple sources across an organization, enabling a single, comprehensive view of each customer. It allows businesses to collect, harmonize, and analyze data in real time, creating a 360-degree customer profile that can be leveraged across Salesforce’s various applications, such as Marketing Cloud, Sales Cloud, and Service Cloud. This platform enables faster, more personalized customer interactions by integrating data from online and offline channels, including CRM data, transactional data, and third-party data sources. Salesforce Data Cloud also offers advanced AI gents and analytics capabilities, helping organizations gain deeper insights into customer behavior and predict future needs. By centralizing and refining data for actionable use, Salesforce Data Cloud supports enhanced customer experiences, targeted marketing, and efficient, data-driven decision-making across departments.
  • 19
    Adaptive

    Adaptive

    Adaptive

    Adaptive is a data security platform designed to prevent sensitive data exposure across all human and non-human entities. It offers a secure control plane to protect and access data, featuring an agentless architecture that requires zero network reconfiguration and can be deployed in the cloud or on-premises. The platform enables organizations to share privileged access to data sources without sharing actual credentials, enhancing security posture. It supports just-in-time access to various data sources, including databases, cloud infrastructure resources, data warehouses, and web services. Adaptive also facilitates non-human data access by connecting third-party tools or ETL pipelines through a central interface without exposing data source credentials. To minimize data exposure, the platform provides data masking and tokenization for non-privileged users without altering access workflows. Comprehensive audibility is achieved through identity-based audit trails across all resources.
  • 20
    OneTrust Data & AI Governance
    OneTrust's Data & AI Governance solution is an integrated platform designed to establish data and AI policies by consolidating insights from data, metadata, models, and risk assessments, providing comprehensive visibility into data products and AI development. It accelerates data-driven innovation by increasing the speed of approval for data products and AI systems. The solution enhances business continuity through continuous monitoring of data and AI systems, ensuring regulatory compliance, effective risk management, and reduced application downtime. It simplifies compliance by centrally defining, orchestrating, and natively enforcing data policies. Key features include consistent scanning, classification, and tagging of sensitive data to ensure the reliable application of data governance policies across structured and unstructured sources. It promotes responsible data usage by enforcing role-based access within a robust data governance framework.
  • 21
    Agile Data Engine

    Agile Data Engine

    Agile Data Engine

    Agile Data Engine is a comprehensive DataOps platform designed to streamline the development, deployment, and operation of cloud-based data warehouses. It integrates data modeling, transformations, continuous deployment, workflow orchestration, monitoring, and API connectivity within a single SaaS solution. The platform's metadata-driven approach automates SQL code generation and data load workflows, enhancing productivity and agility in data operations. Supporting multiple cloud database platforms, including Snowflake, Databricks SQL, Amazon Redshift, Microsoft Fabric (Warehouse), Azure Synapse SQL, Azure SQL Database, and Google BigQuery, Agile Data Engine offers flexibility in cloud environments. Its modular data product framework and out-of-the-box CI/CD pipelines facilitate seamless integration and continuous delivery, enabling data teams to adapt swiftly to changing business requirements. The platform also provides insights and statistics on data platform performance.
  • 22
    Unity Catalog

    Unity Catalog

    Databricks

    Databricks Unity Catalog is the industry’s only unified and open governance solution for data and AI, built into the Databricks Data Intelligence Platform. With Unity Catalog, organizations can seamlessly govern both structured and unstructured data in any format, as well as machine learning models, notebooks, dashboards, and files across any cloud or platform. Data scientists, analysts, and engineers can securely discover, access, and collaborate on trusted data and AI assets across platforms, leveraging AI to boost productivity and unlock the full potential of the lakehouse environment. This unified and open approach to governance promotes interoperability and accelerates data and AI initiatives while simplifying regulatory compliance. Easily discover and classify both structured and unstructured data in any format, including machine learning models, notebooks, dashboards, and files across all cloud platforms.
  • 23
    Google Cloud Analytics Hub
    Google Cloud's Analytics Hub is a data exchange platform that enables organizations to efficiently and securely share data assets across organizational boundaries, addressing challenges related to data reliability and cost. Built on the scalability and flexibility of BigQuery, it allows users to curate a library of internal and external assets, including unique datasets like Google Trends. Analytics Hub facilitates the publication, discovery, and subscription to data exchanges without the need to move data, streamlining the accessibility of data and analytics assets. It also provides privacy-safe, secure data sharing with governance, incorporating in-depth governance, encryption, and security features from BigQuery, Cloud IAM, and VPC Security Controls. By leveraging Analytics Hub, organizations can increase the return on investment of data initiatives by exchanging data. Analytics Hub is based on the scalability and flexibility of BigQuery.
  • 24
    TROCCO

    TROCCO

    primeNumber Inc

    TROCCO is a fully managed modern data platform that enables users to integrate, transform, orchestrate, and manage their data from a single interface. It supports a wide range of connectors, including advertising platforms like Google Ads and Facebook Ads, cloud services such as AWS Cost Explorer and Google Analytics 4, various databases like MySQL and PostgreSQL, and data warehouses including Amazon Redshift and Google BigQuery. The platform offers features like Managed ETL, which allows for bulk importing of data sources and centralized ETL configuration management, eliminating the need to manually create ETL configurations individually. Additionally, TROCCO provides a data catalog that automatically retrieves metadata from data analysis infrastructure, generating a comprehensive catalog to promote data utilization. Users can also define workflows to create a series of tasks, setting the order and combination to streamline data processing.
  • 25
    SAP Business Data Cloud
    SAP Business Data Cloud is a fully managed SaaS solution that unifies and governs all SAP data while seamlessly connecting with third-party data, providing line-of-business leaders with the context needed to make impactful decisions. It offers mission-critical data products, granting access to SAP data across essential business processes in a deeply contextual and governed manner, thereby eliminating the high costs associated with data extraction and replication. As a leading data platform, it enables the connection of all SAP and third-party data through a fully managed SaaS solution in collaboration with Databricks. The platform delivers powerful insight applications, facilitating transformational insights for advanced analytics and planning across various lines of business. By harmonizing all mission-critical data within an open data ecosystem and leveraging a robust semantic layer, SAP Business Data Cloud provides unparalleled business understanding.
  • 26
    Lumi AI

    Lumi AI

    Lumi AI

    ​Lumi AI is an enterprise analytics platform that enables users to explore data and extract custom insights through natural language queries, eliminating the need for SQL or Python expertise. It offers self-service analytics, conversational analytics, customizable visualizations, knowledge management, seamless integrations, and robust security features. Lumi AI supports diverse teams, including data analysis, supply chain management, procurement, sales, merchandising, and financial planning, by providing actionable insights tailored to unique business terms and metrics. Its agentic workflows address simple to complex queries, uncover root causes, and facilitate complex analyses, all while interpreting business-specific language. Lumi AI integrates effortlessly with various data sources, ensuring enterprise-grade security by processing data within the client's network and offering advanced user permissions and query controls. ​
  • 27
    Observo AI

    Observo AI

    Observo AI

    ​Observo AI is an AI-native data pipeline platform designed to address the challenges of managing vast amounts of telemetry data in security and DevOps operations. By leveraging machine learning and agentic AI, Observo AI automates data optimization, enabling enterprises to process AI-generated data more efficiently, securely, and cost-effectively. It reduces data processing costs by over 50% and accelerates incident response times by more than 40%. Observo AI's features include intelligent data deduplication and compression, real-time anomaly detection, and dynamic data routing to appropriate storage or analysis tools. It also enriches data streams with contextual information to enhance threat detection accuracy while minimizing false positives. Observo AI offers a searchable cloud data lake for efficient data storage and retrieval.
  • 28
    DataBahn

    DataBahn

    DataBahn

    DataBahn.ai is redefining how enterprises manage the explosion of security and operational data in the AI era. Our AI-powered data pipeline and fabric platform helps organizations securely collect, enrich, orchestrate, and optimize enterprise data—including security, application, observability, and IoT/OT telemetry—for analytics, automation, and AI. With native support for over 400 integrations and built-in enrichment capabilities, DataBahn streamlines fragmented data workflows and reduces SIEM and infrastructure costs from day one. The platform requires no specialist training, enabling security and IT teams to extract insights in real time and adapt quickly to new demands. We've helped Fortune 500 and Global 2000 companies reduce data processing costs by over 50% and automate more than 80% of their data engineering workloads.
  • 29
    Borneo

    Borneo

    Borneo

    Borneo is a real-time data security and privacy observability platform designed to help organizations discover, remediate, and govern data risks while ensuring privacy and compliance. It enables users to discover where health data, financial data, and PII are stored across unstructured data, SaaS apps, and public cloud environments. Borneo's risk correlation engine identifies data that violates security frameworks and privacy regulations, prompting immediate action. It offers automatic remediation through data masking, access changes, and encryption, and continuously monitors changes across the data landscape to maintain compliance and eliminate regulatory risk. Built by security practitioners from Uber, Facebook, and Yahoo, Borneo is crafted to handle data at scale. It features a powerful connector framework to integrate across diverse data landscapes, supports flexible and modular deployment, and ensures that data never leaves the user's cloud environment.
  • 30
    Unravel

    Unravel

    Unravel Data

    Unravel makes data work anywhere: on Azure, AWS, GCP or in your own data center– Optimizing performance, automating troubleshooting and keeping costs in check. Unravel helps you monitor, manage, and improve your data pipelines in the cloud and on-premises – to drive more reliable performance in the applications that power your business. Get a unified view of your entire data stack. Unravel collects performance data from every platform, system, and application on any cloud then uses agentless technologies and machine learning to model your data pipelines from end to end. Explore, correlate, and analyze everything in your modern data and cloud environment. Unravel’s data model reveals dependencies, issues, and opportunities, how apps and resources are being used, what’s working and what’s not. Don’t just monitor performance – quickly troubleshoot and rapidly remediate issues. Leverage AI-powered recommendations to automate performance improvements, lower costs, and prepare.