Business Software for Hadoop - Page 5

Top Software that integrates with Hadoop as of August 2025 - Page 5

Hadoop Clear Filters
  • 1
    Apache Kylin

    Apache Kylin

    Apache Software Foundation

    Apache Kylin™ is an open source, distributed Analytical Data Warehouse for Big Data; it was designed to provide OLAP (Online Analytical Processing) capability in the big data era. By renovating the multi-dimensional cube and precalculation technology on Hadoop and Spark, Kylin is able to achieve near constant query speed regardless of the ever-growing data volume. Reducing query latency from minutes to sub-second, Kylin brings online analytics back to big data. Kylin can analyze 10+ billions of rows in less than a second. No more waiting on reports for critical decisions. Kylin connects data on Hadoop to BI tools like Tableau, PowerBI/Excel, MSTR, QlikSense, Hue and SuperSet, making the BI on Hadoop faster than ever. As an Analytical Data Warehouse, Kylin offers ANSI SQL on Hadoop/Spark and supports most ANSI SQL query functions. Kylin can support thousands of interactive queries at the same time, thanks to the low resource consumption of each query.
  • 2
    Apache Zeppelin
    Web-based notebook that enables data-driven, interactive data analytics and collaborative documents with SQL, Scala and more. IPython interpreter provides comparable user experience like Jupyter Notebook. This release includes Note level dynamic form, note revision comparator and ability to run paragraph sequentially, instead of simultaneous paragraph execution in previous releases. Interpreter lifecycle manager automatically terminate interpreter process on idle timeout. So resources are released when they're not in use.
  • 3
    SOLIXCloud CDP

    SOLIXCloud CDP

    Solix Technologies

    SOLIXCloud CDP delivers cloud data management as-a-service for modern data-driven enterprises. Built on opensource, cloud native technologies SOLIXCloud CDP helps companies manage and process all of their structured, semi-structured and unstructured data for advanced anaytics, compliance, infrastructure optimization and data security. With features such as Solix Connect for data ingestion, Solix Data Governance, Solix Metadata Management and Solix Search, SOLIXCloud CDP offers a comprehensive cloud data management application framework to build and run data-driven applications such as SQL data warehouse, machine learning and artifitial intelligience while fulfilling the ever growing data management requirements of complex data regulations, data retention and consumer data privacy.
  • 4
    SOLIXCloud

    SOLIXCloud

    Solix Technologies

    Data volume keeps growing, but not all data has equal value. Cloud data management enables forward thinking companies to reduce the cost of managing enterprise data and still provide security, compliance, performance and easy access. As content ages, it loses value, but organizations can still monetize their less current data through modern SaaS-based solutions. SOLIXCloud delivers all of the capabilities required to strike the perfect balance between historical and current data management. With a complete suite of compliance features for structured, unstructured, and semi-structured data, SOLIXCloud offers a fully managed service for all enterprise data. Solix metadata management is an end-to-end framework to explore all enterprise metadata and lineage from a centralized repository and business glossary.
  • 5
    Quantexa

    Quantexa

    Quantexa

    Uncover hidden risk and reveal new, unexpected opportunities with graph analytics across the customer lifecycle. Standard MDM solutions are not built for high volumes of distributed, disparate data, that is generated by various applications and external sources. Traditional MDM probabilistic matching doesn’t work well with siloed data sources. It misses connections, losing context, leads to decision-making inaccuracy, and leaves business value on the table. An ineffective MDM solution affects everything from customer experience to operational performance. Without on-demand visibility of holistic payment patterns, trends and risk, your team can’t make the right decisions quickly, compliance costs escalate, and you can’t increase coverage fast enough. Your data isn’t connected – so customers suffer fragmented experiences across channels, business lines and geographies. Attempts at personalized engagement fall short as these are based on partial, often outdated data.
  • 6
    witboost

    witboost

    Agile Lab

    witboost is a modular, scalable, fast, efficient data management system for your company to truly become data driven, reduce time-to-market, it expenditures and overheads. witboost comprises a series of modules. These are building blocks that can work as standalone solutions to address and solve a single need or problem, or they can be combined to create the perfect data management ecosystem for your company. Each module improves a specific data engineering function and they can be combined to create the perfect solution to answer your specific needs, guaranteeing a blazingly fact and smooth implementation, thus dramatically reducing time-to-market, time-to-value and consequently the TCO of your data engineering infrastructure. Smart Cities need digital twins to predict needs and avoid unforeseen problems, gathering data from thousands of sources and managing ever more complex telematics.
  • 7
    ScriptString

    ScriptString

    ScriptString

    Optimize your document knowledge and make critical decisions with confidence. Tired of manual processing, time constraints, budget pressures and shifting compliance requirements? Hassle free collection and integration of your cloud spend data in half the time at half the cost. Recommended cost savings and guidance to save more than 50% of total spend. Gain 360° visibility of your entire cloud spend with KPI tracking, real-time insights and recommendations. Built-in peace of mind with security and compliance protection to meet any standards. Gather data via portal, email, API, repository, table, data lake or 3rd party data source. Automated AI powered intelligent document processing eliminates manual effort. Intelligent review of document knowledge identifies anomalies, duplicates and errors. Find the needle in the haystack with ScriptString's Knowledge Relationship Indexing.
  • 8
    Occubee

    Occubee

    3SOFT

    Occubee platform automatically converts large amount of receipt data, information on thousands of products and dozens of retail-specific factors into valuable sales and demand forecasts. In stores, Occubee forecasts sales individually for each product and generates replenishment commands. In warehouses, Occubee optimizes the availability of goods and allocated capital, and generates orders for suppliers. In the head office, Occubee provides real-time monitoring of sales processes and generates anomaly alerts and reports. Modern technologies for data collection and processing ensure automation of key business processes in the retail industry. Occubee fully responds to the needs of modern retail and fits in with the global megatrends related to the use of data in business.
  • 9
    Acxiom InfoBase
    Acxiom enables you to leverage comprehensive data for premium audiences and insights across the globe. Better understand, identify, and target ideal audiences by engaging and personalizing experiences across digital and offline channels. With marketing technology, identity resolution and digital connectivity converging in a “borderless digital world,” brands can now quickly locate data attributes, service availability and the digital footprint across the globe to fuel informed decisions. Acxiom is the global data leader with thousands of data attributes in more than 60 countries helping brands improve millions of customer experiences every day through meaningful data-driven insights, all while protecting consumer privacy. Understand, reach and engage audiences everywhere, maximize your media investments and power more personalized experiences. Reach audiences around the globe and deliver experiences that matter with Acxiom data.
  • 10
    Deeplearning4j

    Deeplearning4j

    Deeplearning4j

    DL4J takes advantage of the latest distributed computing frameworks including Apache Spark and Hadoop to accelerate training. On multi-GPUs, it is equal to Caffe in performance. The libraries are completely open-source, Apache 2.0, and maintained by the developer community and Konduit team. Deeplearning4j is written in Java and is compatible with any JVM language, such as Scala, Clojure, or Kotlin. The underlying computations are written in C, C++, and Cuda. Keras will serve as the Python API. Eclipse Deeplearning4j is the first commercial-grade, open-source, distributed deep-learning library written for Java and Scala. Integrated with Hadoop and Apache Spark, DL4J brings AI to business environments for use on distributed GPUs and CPUs. There are a lot of parameters to adjust when you're training a deep-learning network. We've done our best to explain them, so that Deeplearning4j can serve as a DIY tool for Java, Scala, Clojure, and Kotlin programmers.
  • 11
    Span Global Services

    Span Global Services

    Span Global Services

    Span Global Services is the powerhouse for digital and data-driven marketing services. We put targeted insight into every campaign; fueling your B2B sales and marketing results with data and insights across a plethora of industries: technology, healthcare, manufacturing, retail, telecommunication and more. Over 90 Million multi-verified contacts, business firmographics, business entity relationships, business intelligence, active social profile details, our customized databases can fulfill data requirements of large enterprises and SMEs simultaneously. We acquire and validate data through technology, public records and the human element, people contacting people. Our sales and marketing clients enjoy higher MQL and conversions, data quality guarantees, custom appending and profiling services, marketing automation and industry’s best subject matter expertise.
  • 12
    Apache Kudu

    Apache Kudu

    The Apache Software Foundation

    A Kudu cluster stores tables that look just like tables you're used to from relational (SQL) databases. A table can be as simple as a binary key and value, or as complex as a few hundred different strongly-typed attributes. Just like SQL, every table has a primary key made up of one or more columns. This might be a single column like a unique user identifier, or a compound key such as a (host, metric, timestamp) tuple for a machine time-series database. Rows can be efficiently read, updated, or deleted by their primary key. Kudu's simple data model makes it a breeze to port legacy applications or build new ones, no need to worry about how to encode your data into binary blobs or make sense of a huge database full of hard-to-interpret JSON. Tables are self-describing, so you can use standard tools like SQL engines or Spark to analyze your data. Kudu's APIs are designed to be easy to use.
  • 13
    Apache Parquet

    Apache Parquet

    The Apache Software Foundation

    We created Parquet to make the advantages of compressed, efficient columnar data representation available to any project in the Hadoop ecosystem. Parquet is built from the ground up with complex nested data structures in mind, and uses the record shredding and assembly algorithm described in the Dremel paper. We believe this approach is superior to simple flattening of nested namespaces. Parquet is built to support very efficient compression and encoding schemes. Multiple projects have demonstrated the performance impact of applying the right compression and encoding scheme to the data. Parquet allows compression schemes to be specified on a per-column level, and is future-proofed to allow adding more encodings as they are invented and implemented. Parquet is built to be used by anyone. The Hadoop ecosystem is rich with data processing frameworks, and we are not interested in playing favorites.
  • 14
    Hypertable

    Hypertable

    Hypertable

    Hypertable delivers scalable database capacity at maximum performance to speed up your big data application and reduce your hardware footprint. Hypertable delivers maximum efficiency and superior performance over the competition which translates into major cost savings. A proven scalable design that powers hundreds of Google services. All the benefits of open source with a strong and thriving community. C++ implementation for optimum performance. 24/7/365 support for your business-critical big data application. Unparalleled access to Hypertable brain power by the employer of all core Hypertable developers. Hypertable was designed for the express purpose of solving the scalability problem, a problem that is not handled well by a traditional RDBMS. Hypertable is based on a design developed by Google to meet their scalability requirements and solves the scale problem better than any of the other NoSQL solutions out there.
  • 15
    Apache Pinot

    Apache Pinot

    Apache Corporation

    Pinot is designed to answer OLAP queries with low latency on immutable data. Pluggable indexing technologies - Sorted Index, Bitmap Index, Inverted Index. Joins are currently not supported, but this problem can be overcome by using Trino or PrestoDB for querying. SQL like language that supports selection, aggregation, filtering, group by, order by, distinct queries on data. Consist of of both offline and real-time table. Use real-time table only to cover segments for which offline data may not be available yet. Detect the right anomalies by customizing anomaly detect flow and notification flow.
  • 16
    Apache Hudi

    Apache Hudi

    Apache Corporation

    Hudi is a rich platform to build streaming data lakes with incremental data pipelines on a self-managing database layer, while being optimized for lake engines and regular batch processing. Hudi maintains a timeline of all actions performed on the table at different instants of time that helps provide instantaneous views of the table, while also efficiently supporting retrieval of data in the order of arrival. A Hudi instant consists of the following components. Hudi provides efficient upserts, by mapping a given hoodie key consistently to a file id, via an indexing mechanism. This mapping between record key and file group/file id, never changes once the first version of a record has been written to a file. In short, the mapped file group contains all versions of a group of records.
  • 17
    Azure HDInsight
    Run popular open-source frameworks—including Apache Hadoop, Spark, Hive, Kafka, and more—using Azure HDInsight, a customizable, enterprise-grade service for open-source analytics. Effortlessly process massive amounts of data and get all the benefits of the broad open-source project ecosystem with the global scale of Azure. Easily migrate your big data workloads and processing to the cloud. Open-source projects and clusters are easy to spin up quickly without the need to install hardware or manage infrastructure. Big data clusters reduce costs through autoscaling and pricing tiers that allow you to pay for only what you use. Enterprise-grade security and industry-leading compliance with more than 30 certifications helps protect your data. Optimized components for open-source technologies such as Hadoop and Spark keep you up to date.
  • 18
    Cloudera Data Platform
    Unlock the potential of private and public clouds with the only hybrid data platform for modern data architectures with data anywhere. Cloudera is a hybrid data platform designed for unmatched freedom to choose—any cloud, any analytics, any data. Cloudera delivers faster and easier data management and data analytics for data anywhere, with optimal performance, scalability, and security. With Cloudera you get all the advantages of private cloud and public cloud for faster time to value and increased IT control. Cloudera provides the freedom to securely move data, applications, and users bi-directionally between the data center and multiple data clouds, regardless of where your data lives.
  • 19
    Datametica

    Datametica

    Datametica

    At Datametica, our birds with unprecedented capabilities help eliminate business risks, cost, time, frustration, and anxiety from the entire process of data warehouse migration to the cloud. Migration of existing data warehouse, data lake, ETL, and Enterprise business intelligence to the cloud environment of your choice using Datametica automated product suite. Architecting an end-to-end migration strategy, with workload discovery, assessment, planning, and cloud optimization. Starting from discovery and assessment of your existing data warehouse to planning the migration strategy – Eagle gives clarity on what’s needed to be migrated and in what sequence, how the process can be streamlined, and what are the timelines and costs. The holistic view of the workloads and planning reduces the migration risk without impacting the business.
  • 20
    IBM Intelligent Operations Center for Emergency Mgmt
    An incident and emergency management solution for daily operations, emergency and crisis situations. This command, control and communication (C3) solution uses data analytic technologies coupled with social and mobile technology to streamline and integrate preparation, response, recovery and mitigation of daily incidents, emergencies and disasters. IBM works with governments and public safety organizations worldwide to implement public safety technology solutions. Proven preparation techniques use the same technology to manage day-to-day community incidents when responding to crises situations. This familiarity helps ensure first responders and C3 staff can engage immediately and naturally in response, recovery and mitigation without needing access to special documentation and systems. This incident and emergency management solution integrates and correlates information sources to create a dynamic, near real-time geospatial framework for a common operating picture.
  • 21
    Red Hat JBoss Data Virtualization
    Red Hat JBoss Data Virtualization is a lean, virtual data integration solution that unlocks trapped data and delivers it as easily consumable, unified, and actionable information. Red Hat JBoss Data Virtualization makes data spread across physically diverse systems, such as multiple databases, XML files, and Hadoop systems, appear as a set of tables in a local database. Provides standards-based read/write access to heterogeneous data stores in real-time. Speeds application development and integration by simplifying access to distributed data. Integrate and transform data semantics based on data consumer requirements. Provides centralized access control, and auditing through robust security infrastructure. Turn fragmented data into actionable information at the speed your business needs. Red Hat offers support and maintenance over stated time periods for the major versions of JBoss products.
  • 22
    Value Innovation Labs Marketing Automation Platform
    Track your user behavior with power analytics. Segment users based on their behavior. Create engagement strategies with powerful AI. OS/Device level restrictions by certain handset makers restrict push notification delivery. With our product, you can bypass those restrictions to reach and engage an additional 20% of users. We ensure higher inbox reach with email consultants and industry experts to help you with the best practices. Avoid sending blast messages that end up in spam, or taint your domain and brand reputation. Localize the communication based on language, seamlessly. Our platform supports multilingual architecture and you can reach out to your customers in the local language for a local touch. Target users with acquisition source, uninstall data and more. Segment users just the way you want. Initiate conversation, reduce churn and do much more with powerful insights.
  • 23
    Value Innovation Labs Enterprise HRMS
    Assign, track, execute tasks, track productivity with powerful insight. Automate over 100+ tasks related and amplify human interactions with bots, group chat and more. Actionable insights that help Line Managers, HR Professionals & CXO achieve more. Define organizational structure, assign roles & permissions, grant access rights. Manage your employee life cycle from onboarding to exit, publish letters. Run error-free payroll, manage loans & reimbursements, meet statutory norms. Real-time attendance for managing attendance, holiday calendar, shifts and integration. Meet organizational goals & improve performance with 360-degree feedback. Boost employee morale & improve employee engagement using engagement tools. Real-time attendance for managing attendance, holiday calendar, shifts and integration. Meet organizational goals & improve performance with 360-degree feedback. Boost employee morale & improve employee engagement using engagement tools.
  • 24
    doolytic

    doolytic

    doolytic

    doolytic is leading the way in big data discovery, the convergence of data discovery, advanced analytics, and big data. doolytic is rallying expert BI users to the revolution in self-service exploration of big data, revealing the data scientist in all of us. doolytic is an enterprise software solution for native discovery on big data. doolytic is based on best-of-breed, scalable, open-source technologies. Lightening performance on billions of records and petabytes of data. Structured, unstructured and real-time data from any source. Sophisticated advanced query capabilities for expert users, Integration with R for advanced and predictive applications. Search, analyze, and visualize data from any format, any source in real-time with the flexibility of Elastic. Leverage the power of Hadoop data lakes with no latency and concurrency issues. doolytic solves common BI problems and enables big data discovery without clumsy and inefficient workarounds.
  • 25
    IBM InfoSphere Optim Data Privacy
    IBM InfoSphere® Optim™ Data Privacy provides extensive capabilities to effectively mask sensitive data across non-production environments, such as development, testing, QA or training. To protect confidential data this single offering provides a variety of transformation techniques that substitute sensitive information with realistic, fully functional masked data. Examples of masking techniques include substrings, arithmetic expressions, random or sequential number generation, date aging, and concatenation. The contextually accurate masking capabilities help masked data retain a similar format to the original information. Apply a range of masking techniques on-demand to transform personally-identifying information and confidential corporate data in applications, databases and reports. Data masking features help you to prevent misuse of information by masking, obfuscating, and privatizing personal information that is disseminated across non-production environments.
  • 26
    Pavilion HyperOS
    Powering the most performant, dense, scalable, and flexible storage platform in the universe. Pavilion HyperParallel File System™ provides the ability to scale across an unlimited number of Pavilion HyperParallel Flash Arrays™, providing 1.2 TB/s read, and 900 GB/s write bandwidth with 200M IOPS at 25µs latency per rack. Uniquely capable of providing independent, linear scalability of both capacity and performance, the Pavilion HyperOS 3 now provides global namespace support for both NFS and S3, enabling unlimited, linear scale across an unlimited number of Pavilion HyperParallel Flash Array systems. Take advantage of the power of the Pavilion HyperParallel Flash Array to enjoy unrivaled levels of performance and availability. The Pavilion HyperOS includes patent-pending technology to ensure that your data is always available, with performant access that legacy arrays cannot match.
  • 27
    Invenis

    Invenis

    Invenis

    Invenis is a data analysis and mining platform. Clean, aggregate and analyze your data in a simple way and scale up to improve your decision making. Data harmonization, preparation and cleansing, data enrichment, and aggregation. Prediction, segmentation, recommendation. Invenis connects to all your data sources, MySQL, Oracle, Postgres SQL, HDFS (Hadoop), and allows you to analyze all your files, CSV, JSON, etc. Make predictions on all your data, without code and without the need for a team of experts. The best algorithms are automatically chosen according to your data and use cases. Repetitive tasks and your recurring analyses are automated. Save time to exploit the full potential of your data! You can work as a team, with the other analysts in your team, but also with all teams. This makes decision-making more efficient and information is disseminated to all levels of the company.
  • 28
    Integration Eye
    Integration Eye® is a modular product, which streamlines system integrations, infrastructure, and business. It consists of 3 modules: proxy module IPM, logging module ILM, and the security module ISM, which can be used independently or combined. It is based on the widely used, secure, and platform-independent Java language (why choose Java?) and it runs on the lightweight integration engine Mule™. Using individual Integration Eye® modules, you can monitor your APIs and systems, create statistics on and analyze calls (logging with the ILM module), and be alerted to any problems, downtime, or slow responses of specific APIs and systems. You can secure your APIs and systems using roles (authorization and authentication with the ISM module) based on Keycloak SSO we deliver or your existing Auth server. You can extend or proxy service calls (both internal and external) with mutual SSL, headers, etc. (proxy with IPM) you can also monitor and analyze these calls.
  • 29
    Apache Gobblin

    Apache Gobblin

    Apache Software Foundation

    A distributed data integration framework that simplifies common aspects of Big Data integration such as data ingestion, replication, organization, and lifecycle management for both streaming and batch data ecosystems. Runs as a standalone application on a single box. Also supports embedded mode. Runs as an mapreduce application on multiple Hadoop versions. Also supports Azkaban for launching mapreduce jobs. Runs as a standalone cluster with primary and worker nodes. This mode supports high availability and can run on bare metals as well. Runs as an elastic cluster on public cloud. This mode supports high availability. Gobblin as it exists today is a framework that can be used to build different data integration applications like ingest, replication, etc. Each of these applications is typically configured as a separate job and executed through a scheduler like Azkaban.
  • 30
    Integrate.io

    Integrate.io

    Integrate.io

    Unify Your Data Stack: Experience the first no-code data pipeline platform and power enlightened decision making. Integrate.io is the only complete set of data solutions & connectors for easy building and managing of clean, secure data pipelines. Increase your data team's output with all of the simple, powerful tools & connectors you’ll ever need in one no-code data integration platform. Empower any size team to consistently deliver projects on-time & under budget. We ensure your success by partnering with you to truly understand your needs & desired outcomes. Our only goal is to help you overachieve yours. Integrate.io's Platform includes: -No-Code ETL & Reverse ETL: Drag & drop no-code data pipelines with 220+ out-of-the-box data transformations -Easy ELT & CDC :The Fastest Data Replication On The Market -Automated API Generation: Build Automated, Secure APIs in Minutes - Data Warehouse Monitoring: Finally Understand Your Warehouse Spend - FREE Data Observability: Custom