Business Software for Apache Hive - Page 3

Top Software that integrates with Apache Hive as of July 2025 - Page 3

  • 1
    Progress DataDirect

    Progress DataDirect

    Progress Software

    Empowering applications with enterprise data is our passion here at Progress DataDirect. We offer cloud and on-premises data connectivity solutions across relational, NoSQL, Big Data, and SaaS data sources. Performance, reliability, and security are at the heart of everything we design for thousands of enterprises and the leading vendors in analytics, BI, and data management. Minimize your development costs with our portfolio of high-value connectors for a variety of data sources. Enjoy 24/7 world-class support and security for greater peace of mind. Connect with affordable, easy-to-use, and time-saving drivers for faster SQL access to your data. As a leader in data connectivity, keeping up with the evolving trends in space is our mission. But if we haven’t built the connector you need yet, reach out and we’ll help you develop the right solution. Embed connectivity in an application or service.
  • 2
    jethro

    jethro

    jethro

    Data-driven decision-making has unleashed a surge of business data and a rise in user demand to analyze it. This trend drives IT departments to migrate off expensive Enterprise Data Warehouses (EDW) toward cost-effective Big Data platforms like Hadoop or AWS. These new platforms come with a Total Cost of Ownership (TCO) that is about 10 times lower. They are not ideal for interactive BI applications, however, as they fail to match the high performance and user concurrency of legacy EDWs. For this exact reason, we developed Jethro. Customers use Jethro for interactive BI on Big Data. Jethro is a transparent middle tier that requires no changes to existing apps or data. It is self-driving with no maintenance required. Jethro is compatible with BI tools like Tableau, Qlik, and Microstrategy and is data source agnostic. Jethro delivers on the demands of business users allowing for thousands of concurrent users to run complicated queries over billions of records.
  • 3
    Baidu Sugar

    Baidu Sugar

    Baidu AI Cloud

    Sugar will charge fees according to the organization. A user can belong to multiple organizations, and there are multiple users in an organization. Multiple spaces can be created under the organization. Generally, it is recommended to divide spaces according to projects or teams. Data between spaces is not shared. Each space has its own independent permission management. When you use Sugar to analyze and visualize data, you need to specify the data source of the original data. Data source is the place where data is stored. Generally, it refers to the connection address (host, port, user name, password, etc.) of the database. A dashboard is a kind of visual page type, that mainly reflects cool visual effect, and is generally used to put on the large screen for real-time data visualization.
    Starting Price: $0.33 per year
  • 4
    Foundational

    Foundational

    Foundational

    Identify code and optimization issues in real-time, prevent data incidents pre-deploy, and govern data-impacting code changes end to end—from the operational database to the user-facing dashboard. Automated, column-level data lineage, from the operational database all the way to the reporting layer, ensures every dependency is analyzed. Foundational automates data contract enforcement by analyzing every repository from upstream to downstream, directly from source code. Use Foundational to proactively identify code and data issues, find and prevent issues, and create controls and guardrails. Foundational can be set up in minutes with no code changes required.
  • 5
    IBM watsonx.data
    Put your data to work, wherever it resides, with the open, hybrid data lakehouse for AI and analytics. Connect your data from anywhere, in any format, and access through a single point of entry with a shared metadata layer. Optimize workloads for price and performance by pairing the right workloads with the right query engine. Embed natural-language semantic search without the need for SQL, so you can unlock generative AI insights faster. Manage and prepare trusted data to improve the relevance and precision of your AI applications. Use all your data, everywhere. With the speed of a data warehouse, the flexibility of a data lake, and special features to support AI, watsonx.data can help you scale AI and analytics across your business. Choose the right engines for your workloads. Flexibly manage cost, performance, and capability with access to multiple open engines including Presto, Presto C++, Spark Milvus, and more.
  • 6
    TapData

    TapData

    TapData

    CDC-based live data platform for heterogeneous database replication, real-time data integration, or building a real-time data warehouse. By using CDC to sync production line data stored in DB2 and Oracle to the modern database, TapData enabled an AI-augmented real-time dispatch software to optimize the semiconductor production line process. The real-time data made instant decision-making in the RTD software a possibility, leading to faster turnaround times and improved yield. As one of the largest telcos, customer has many regional systems that cater to the local customers. By syncing and aggregating data from various sources and locations into a centralized data store, customers were able to build an order center where the collective orders from many applications can now be aggregated. TapData seamlessly integrates inventory data from 500+ stores, providing real-time insights into stock levels and customer preferences, enhancing supply chain efficiency.
  • 7
    eQube®-DaaS

    eQube®-DaaS

    eQ Technologic

    Our platform establishes a data fabric with a connected network of integrated data, applications, and devices that puts the power of analytics in the hands of end users leading to actionable insight. Data from any source can be aggregated using eQube's data virtualization layer and exposed as a web service, REST service, OData service, or API. Efficiently and rapidly integrate many legacy systems and new COTS (Commercial off-the-shelf) systems. Responsibly retire legacy systems in an orderly manner without disrupting the business. Provide on-demand 'visibility' across the business processes with analytics and business intelligence (A/BI) capabilities. eQube®-MI-based application integration infrastructure can be readily extended for secure, scalable, and robust information collaboration across networks, partners, suppliers, and customers that are geographically dispersed.
  • 8
    Airtool

    Airtool

    Airtool

    Airtool is a next-generation low-code platform purpose-built to support the demands of large-scale, mission-critical enterprise systems. Developed by Deister Software, Airtool was born from the need to create a powerful ERP solution capable of handling massive datasets without sacrificing performance. Unlike traditional low-code tools focused on simple apps, Airtool offers unmatched scalability, making it ideal for building robust, data-intensive applications. Engineered with a unified development framework, Airtool enforces best practices and standardizes workflows, reducing complexity and making onboarding seamless. Its ERP-focused design ensures precision and control in managing complex business operations, while integrated data management allows real-time interaction with millions of records. Airtool empowers teams to reduce reliance on specialized developers with reusable components and intuitive tools, accelerating development while maintaining maintainability and consistency.
    Starting Price: $50/month
  • 9
    Data Virtuality

    Data Virtuality

    Data Virtuality

    Connect and centralize data. Transform your existing data landscape into a flexible data powerhouse. Data Virtuality is a data integration platform for instant data access, easy data centralization and data governance. Our Logical Data Warehouse solution combines data virtualization and materialization for the highest possible performance. Build your single source of data truth with a virtual layer on top of your existing data environment for high data quality, data governance, and fast time-to-market. Hosted in the cloud or on-premises. Data Virtuality has 3 modules: Pipes, Pipes Professional, and Logical Data Warehouse. Cut down your development time by up to 80%. Access any data in minutes and automate data workflows using SQL. Use Rapid BI Prototyping for significantly faster time-to-market. Ensure data quality for accurate, complete, and consistent data. Use metadata repositories to improve master data management.
  • 10
    Mode

    Mode

    Mode Analytics

    Understand how users are interacting with your product and identify opportunity areas to inform your product decisions. Mode empowers one Stitch analyst to do the work of a full data team through speed, flexibility, and collaboration. Build dashboards for annual revenue, then use chart visualizations to identify anomalies quickly. Create polished, investor-ready reports or share analysis with teams for collaboration. Connect your entire tech stack to Mode and identify upstream issues to improve performance. Speed up workflows across teams with APIs and webhooks. Understand how users are interacting with your product and identify opportunity areas to inform your product decisions. Leverage marketing and product data to fix weak spots in your funnel, improve landing-page performance, and understand churn before it happens.
  • 11
    Astro by Astronomer
    For data teams looking to increase the availability of trusted data, Astronomer provides Astro, a modern data orchestration platform, powered by Apache Airflow, that enables the entire data team to build, run, and observe data pipelines-as-code. Astronomer is the commercial developer of Airflow, the de facto standard for expressing data flows as code, used by hundreds of thousands of teams across the world.
  • 12
    Nucleon Database Master

    Nucleon Database Master

    Nucleon Software

    Nucleon Database Master is a modern, powerful, intuitive and easy to use database query, administration, and management software with a consistent and modern user interface. Database Master simplifies managing, monitoring, querying, editing, visualizing, designing relational and NoSQL DBMS. Database Master allows you to execute extended SQL, JQL and C# (Linq) query scripts, provides all database objects such as tables, views, procedures, packages, columns, indexes, relationships (constraints), collections, triggers and other database objects.
    Starting Price: $99 one-time payment
  • 13
    ActionIQ

    ActionIQ

    ActionIQ

    The ActionIQ Customer Data Platform enables you to align your people, technology, and processes to deliver exceptional customer experiences across every touchpoint. How to separate the CDP posers from the true players. Download ActionIQ’s guide to save yourself months of frustrating research and get to the truth on the confusing CDP landscape. In today’s experience economy, consumers expect brands to know them and consistently deliver authentic, helpful experiences. The ActionIQ CDP enables large organizations to solve chronic customer data fragmentation and gain the customer intelligence needed to orchestrate experiences across all their brand touchpoints. Build an omnichannel “smart hub” stack that unifies data and empowers teams with the real-time insights they need. Gain deep customer understanding that enables you to deliver trust-building, profitable customer experiences at scale.
  • 14
    Apache Spark

    Apache Spark

    Apache Software Foundation

    Apache Spark™ is a unified analytics engine for large-scale data processing. Apache Spark achieves high performance for both batch and streaming data, using a state-of-the-art DAG scheduler, a query optimizer, and a physical execution engine. Spark offers over 80 high-level operators that make it easy to build parallel apps. And you can use it interactively from the Scala, Python, R, and SQL shells. Spark powers a stack of libraries including SQL and DataFrames, MLlib for machine learning, GraphX, and Spark Streaming. You can combine these libraries seamlessly in the same application. Spark runs on Hadoop, Apache Mesos, Kubernetes, standalone, or in the cloud. It can access diverse data sources. You can run Spark using its standalone cluster mode, on EC2, on Hadoop YARN, on Mesos, or on Kubernetes. Access data in HDFS, Alluxio, Apache Cassandra, Apache HBase, Apache Hive, and hundreds of other data sources.
  • 15
    Amazon EMR
    Amazon EMR is the industry-leading cloud big data platform for processing vast amounts of data using open-source tools such as Apache Spark, Apache Hive, Apache HBase, Apache Flink, Apache Hudi, and Presto. With EMR you can run Petabyte-scale analysis at less than half of the cost of traditional on-premises solutions and over 3x faster than standard Apache Spark. For short-running jobs, you can spin up and spin down clusters and pay per second for the instances used. For long-running workloads, you can create highly available clusters that automatically scale to meet demand. If you have existing on-premises deployments of open-source tools such as Apache Spark and Apache Hive, you can also run EMR clusters on AWS Outposts. Analyze data using open-source ML frameworks such as Apache Spark MLlib, TensorFlow, and Apache MXNet. Connect to Amazon SageMaker Studio for large-scale model training, analysis, and reporting.
  • 16
    Nightfall

    Nightfall

    Nightfall

    Discover, classify, and protect your sensitive data. Nightfall™ uses machine learning to identify business-critical data, like customer PII, across your SaaS, APIs, and data infrastructure, so you can manage & protect it. Integrate in minutes with cloud services via APIs to monitor data without agents. Machine learning classifies your sensitive data & PII with high accuracy, so nothing gets missed. Setup automated workflows for quarantines, deletions, alerts, and more - saving you time and keeping your business safe. Nightfall integrates directly with all your SaaS, APIs, and data infrastructure. Start building with Nightfall’s APIs for sensitive data classification & protection for free. Via REST API, programmatically get structured results from Nightfall’s deep learning-based detectors for things like credit card numbers, API keys, and more. Integrate with just a few lines of code. Seamlessly add data classification to your applications & workflows using Nightfall's REST API.
  • 17
    TiMi

    TiMi

    TIMi

    With TIMi, companies can capitalize on their corporate data to develop new ideas and make critical business decisions faster and easier than ever before. The heart of TIMi’s Integrated Platform. TIMi’s ultimate real-time AUTO-ML engine. 3D VR segmentation and visualization. Unlimited self service business Intelligence. TIMi is several orders of magnitude faster than any other solution to do the 2 most important analytical tasks: the handling of datasets (data cleaning, feature engineering, creation of KPIs) and predictive modeling. TIMi is an “ethical solution”: no “lock-in” situation, just excellence. We guarantee you a work in all serenity and without unexpected extra costs. Thanks to an original & unique software infrastructure, TIMi is optimized to offer you the greatest flexibility for the exploration phase and the highest reliability during the production phase. TIMi is the ultimate “playground” that allows your analysts to test the craziest ideas!
  • 18
    Truedat

    Truedat

    Bluetab Solutions

    Truedat is an open source data governance business solution tool developed by Bluetab Solutions in order to help our clients become data-driven companies. We help to define business processes, roles & responsibilities. We also help putting processes into practice. Integration and customization of truedat´s open source components to support the data governance processes. We guarantee the support and maintenance of the process & software of our solution modules installed by us. Based on our experience in, we have developed a solution that covers the need for Data Governance, allowing to manage and control highly complex and changing data architectures. The highly increasing migration of enterprise IT platforms to cloud, multi-cloud and hybrid architectures, increases the sources, complexity and types of data and therefore rises the need for truedat. Our solution comes from more than 8 years of experience in Data Governance consulting and development projects.
  • 19
    Privacera

    Privacera

    Privacera

    At the intersection of data governance, privacy, and security, Privacera’s unified data access governance platform maximizes the value of data by providing secure data access control and governance across hybrid- and multi-cloud environments. The hybrid platform centralizes access and natively enforces policies across multiple cloud services—AWS, Azure, Google Cloud, Databricks, Snowflake, Starburst and more—to democratize trusted data enterprise-wide without compromising compliance with regulations such as GDPR, CCPA, LGPD, or HIPAA. Trusted by Fortune 500 customers across finance, insurance, retail, healthcare, media, public and the federal sector, Privacera is the industry’s leading data access governance platform that delivers unmatched scalability, elasticity, and performance. Headquartered in Fremont, California, Privacera was founded in 2016 to manage cloud data privacy and security by the creators of Apache Ranger™ and Apache Atlas™.
  • 20
    Microsoft Power Query
    Power Query is the easiest way to connect, extract, transform and load data from a wide range of sources. Power Query is a data transformation and data preparation engine. Power Query comes with a graphical interface for getting data from sources and a Power Query Editor for applying transformations. Because the engine is available in many products and services, the destination where the data will be stored depends on where Power Query was used. Using Power Query, you can perform the extract, transform, and load (ETL) processing of data. Microsoft’s Data Connectivity and Data Preparation technology that lets you seamlessly access data stored in hundreds of sources and reshape it to fit your needs—all with an easy to use, engaging, no-code experience. Power Query supports hundreds of data sources with built-in connectors, generic interfaces (such as REST APIs, ODBC, OLE, DB and OData) and the Power Query SDK to build your own connectors.
  • 21
    Apache Knox

    Apache Knox

    Apache Software Foundation

    The Knox API Gateway is designed as a reverse proxy with consideration for pluggability in the areas of policy enforcement, through providers and the backend services for which it proxies requests. Policy enforcement ranges from authentication/federation, authorization, audit, dispatch, hostmapping and content rewrite rules. Policy is enforced through a chain of providers that are defined within the topology deployment descriptor for each Apache Hadoop cluster gated by Knox. The cluster definition is also defined within the topology deployment descriptor and provides the Knox Gateway with the layout of the cluster for purposes of routing and translation between user facing URLs and cluster internals. Each Apache Hadoop cluster that is protected by Knox has its set of REST APIs represented by a single cluster specific application context path. This allows the Knox Gateway to both protect multiple clusters and present the REST API consumer with a single endpoint.
  • 22
    Mage Static Data Masking
    Mage™ Static Data Masking (SDM) and Test data Management (TDM) capabilities fully integrate with Imperva’s Data Security Fabric (DSF) delivering complete protection for all sensitive or regulated data while simultaneously integrating seamlessly with an organization’s existing IT framework and existing application development, testing and data flows without the requirement for any additional architectural changes.
  • 23
    Mage Dynamic Data Masking
    Mage™ Dynamic Data Masking module of the Mage data security platform has been designed with the end customer needs taken into consideration. Mage™ Dynamic Data Masking has been developed working alongside our customers, to address the specific needs and requirements they have. As a result, this product has evolved in a way to meet all the use cases that an enterprise could possibly have. Most other solutions in the market are either a part of an acquisition or are developed to meet only a specific use case. Mage™ Dynamic Data Masking has been designed to deliver adequate protection to sensitive data in production to application and database users while simultaneously integrating seamlessly with an organization's existing IT framework without the requirement of any additional architectural changes.​
  • 24
    Okera

    Okera

    Okera

    Okera, the Universal Data Authorization company, helps modern, data-driven enterprises accelerate innovation, minimize data security risks, and demonstrate regulatory compliance. The Okera Dynamic Access Platform automatically enforces universal fine-grained access control policies. This allows employees, customers, and partners to use data responsibly, while protecting them from inappropriately accessing data that is confidential, personally identifiable, or regulated. Okera’s robust audit capabilities and data usage intelligence deliver the real-time and historical information that data security, compliance, and data delivery teams need to respond quickly to incidents, optimize processes, and analyze the performance of enterprise data initiatives. Okera began development in 2016 and now dynamically authorizes access to hundreds of petabytes of sensitive data for the world’s most demanding F100 companies and regulatory agencies. The company is headquartered in San Francisco.
  • 25
    Acceldata

    Acceldata

    Acceldata

    Acceldata is an Agentic Data Management company helping enterprises manage complex data systems with AI-powered automation. Its unified platform brings together data quality, governance, lineage, and infrastructure monitoring to deliver trusted, actionable insights across the business. Acceldata’s Agentic Data Management platform uses intelligent AI agents to detect, understand, and resolve data issues in real time. Designed for modern data environments, it replaces fragmented tools with a self-learning system that ensures data is accurate, governed, and ready for AI and analytics.
  • 26
    Apache Sentry

    Apache Sentry

    Apache Software Foundation

    Apache Sentry™ is a system for enforcing fine grained role based authorization to data and metadata stored on a Hadoop cluster. Apache Sentry has successfully graduated from the Incubator in March of 2016 and is now a Top-Level Apache project. Apache Sentry is a granular, role-based authorization module for Hadoop. Sentry provides the ability to control and enforce precise levels of privileges on data for authenticated users and applications on a Hadoop cluster. Sentry currently works out of the box with Apache Hive, Hive Metastore/HCatalog, Apache Solr, Impala and HDFS (limited to Hive table data). Sentry is designed to be a pluggable authorization engine for Hadoop components. It allows you to define authorization rules to validate a user or application’s access requests for Hadoop resources. Sentry is highly modular and can support authorization for a wide variety of data models in Hadoop.
  • 27
    lakeFS

    lakeFS

    Treeverse

    lakeFS enables you to manage your data lake the way you manage your code. Run parallel pipelines for experimentation and CI/CD for your data. Simplifying the lives of engineers, data scientists and analysts who are transforming the world with data. lakeFS is an open source platform that delivers resilience and manageability to object-storage based data lakes. With lakeFS you can build repeatable, atomic and versioned data lake operations, from complex ETL jobs to data science and analytics. lakeFS supports AWS S3, Azure Blob Storage and Google Cloud Storage (GCS) as its underlying storage service. It is API compatible with S3 and works seamlessly with all modern data frameworks such as Spark, Hive, AWS Athena, Presto, etc. lakeFS provides a Git-like branching and committing model that scales to exabytes of data by utilizing S3, GCS, or Azure Blob for storage.
  • 28
    Amundsen

    Amundsen

    Amundsen

    Discover & trust data for your analysis and models. Be more productive by breaking silos. Get immediate context into the data and see how others are using it. Search for data within your organization by a simple text search. A PageRank-inspired search algorithm recommends results based on names, descriptions, tags, and querying/viewing activity on the table/dashboard. Build trust in data using automated and curated metadata, descriptions of tables and columns, other frequent users, when the table was last updated, statistics, a preview of the data if permitted, etc. Easy triage by linking the ETL job and code that generated the data. Update tables and columns with descriptions, reduce unnecessary back and forth about which table to use and what a column contains. See what data fellow co-workers frequently use, own or have bookmarked. Learn what most common queries for a table look like by seeing dashboards built on a given table.
  • 29
    Apache Kylin

    Apache Kylin

    Apache Software Foundation

    Apache Kylin™ is an open source, distributed Analytical Data Warehouse for Big Data; it was designed to provide OLAP (Online Analytical Processing) capability in the big data era. By renovating the multi-dimensional cube and precalculation technology on Hadoop and Spark, Kylin is able to achieve near constant query speed regardless of the ever-growing data volume. Reducing query latency from minutes to sub-second, Kylin brings online analytics back to big data. Kylin can analyze 10+ billions of rows in less than a second. No more waiting on reports for critical decisions. Kylin connects data on Hadoop to BI tools like Tableau, PowerBI/Excel, MSTR, QlikSense, Hue and SuperSet, making the BI on Hadoop faster than ever. As an Analytical Data Warehouse, Kylin offers ANSI SQL on Hadoop/Spark and supports most ANSI SQL query functions. Kylin can support thousands of interactive queries at the same time, thanks to the low resource consumption of each query.
  • 30
    Apache Zeppelin
    Web-based notebook that enables data-driven, interactive data analytics and collaborative documents with SQL, Scala and more. IPython interpreter provides comparable user experience like Jupyter Notebook. This release includes Note level dynamic form, note revision comparator and ability to run paragraph sequentially, instead of simultaneous paragraph execution in previous releases. Interpreter lifecycle manager automatically terminate interpreter process on idle timeout. So resources are released when they're not in use.