Business Software for Hadoop - Page 4

Top Software that integrates with Hadoop as of August 2025 - Page 4

Hadoop Clear Filters
  • 1
    FairCom DB

    FairCom DB

    FairCom Corporation

    FairCom DB is ideal for large-scale, mission-critical, core-business applications that require performance, reliability and scalability that cannot be achieved by other databases. FairCom DB delivers predictable high-velocity transactions and massively parallel big data analytics. It empowers developers with NoSQL APIs for processing binary data at machine speed and ANSI SQL for easy queries and analytics over the same binary data. Among the companies that take advantage of the flexibility of FairCom DB is Verizon, who recently chose FairCom DB as an in-memory database for its Verizon Intelligent Network Control Platform Transaction Server Migration. FairCom DB is an advanced database engine that gives you a Continuum of Control to achieve unprecedented performance with the lowest total cost of ownership (TCO). You do not conform to FairCom DB…FairCom DB conforms to you. With FairCom DB, you are not forced to conform your needs to meet the limitations of the database.
  • 2
    Apache Spark

    Apache Spark

    Apache Software Foundation

    Apache Spark™ is a unified analytics engine for large-scale data processing. Apache Spark achieves high performance for both batch and streaming data, using a state-of-the-art DAG scheduler, a query optimizer, and a physical execution engine. Spark offers over 80 high-level operators that make it easy to build parallel apps. And you can use it interactively from the Scala, Python, R, and SQL shells. Spark powers a stack of libraries including SQL and DataFrames, MLlib for machine learning, GraphX, and Spark Streaming. You can combine these libraries seamlessly in the same application. Spark runs on Hadoop, Apache Mesos, Kubernetes, standalone, or in the cloud. It can access diverse data sources. You can run Spark using its standalone cluster mode, on EC2, on Hadoop YARN, on Mesos, or on Kubernetes. Access data in HDFS, Alluxio, Apache Cassandra, Apache HBase, Apache Hive, and hundreds of other data sources.
  • 3
    Amazon EMR
    Amazon EMR is the industry-leading cloud big data platform for processing vast amounts of data using open-source tools such as Apache Spark, Apache Hive, Apache HBase, Apache Flink, Apache Hudi, and Presto. With EMR you can run Petabyte-scale analysis at less than half of the cost of traditional on-premises solutions and over 3x faster than standard Apache Spark. For short-running jobs, you can spin up and spin down clusters and pay per second for the instances used. For long-running workloads, you can create highly available clusters that automatically scale to meet demand. If you have existing on-premises deployments of open-source tools such as Apache Spark and Apache Hive, you can also run EMR clusters on AWS Outposts. Analyze data using open-source ML frameworks such as Apache Spark MLlib, TensorFlow, and Apache MXNet. Connect to Amazon SageMaker Studio for large-scale model training, analysis, and reporting.
  • 4
    Google Cloud Bigtable
    Google Cloud Bigtable is a fully managed, scalable NoSQL database service for large analytical and operational workloads. Fast and performant: Use Cloud Bigtable as the storage engine that grows with you from your first gigabyte to petabyte-scale for low-latency applications as well as high-throughput data processing and analytics. Seamless scaling and replication: Start with a single node per cluster, and seamlessly scale to hundreds of nodes dynamically supporting peak demand. Replication also adds high availability and workload isolation for live serving apps. Simple and integrated: Fully managed service that integrates easily with big data tools like Hadoop, Dataflow, and Dataproc. Plus, support for the open source HBase API standard makes it easy for development teams to get started.
  • 5
    Nightfall

    Nightfall

    Nightfall

    Discover, classify, and protect your sensitive data. Nightfall™ uses machine learning to identify business-critical data, like customer PII, across your SaaS, APIs, and data infrastructure, so you can manage & protect it. Integrate in minutes with cloud services via APIs to monitor data without agents. Machine learning classifies your sensitive data & PII with high accuracy, so nothing gets missed. Setup automated workflows for quarantines, deletions, alerts, and more - saving you time and keeping your business safe. Nightfall integrates directly with all your SaaS, APIs, and data infrastructure. Start building with Nightfall’s APIs for sensitive data classification & protection for free. Via REST API, programmatically get structured results from Nightfall’s deep learning-based detectors for things like credit card numbers, API keys, and more. Integrate with just a few lines of code. Seamlessly add data classification to your applications & workflows using Nightfall's REST API.
  • 6
    AutoSys Workload Automation
    Organizations need to effectively manage large volumes of complex, business-critical workloads across multiple applications and platforms. In such complex environments, there are number of business challenges you have to address. Availability of critical business services. A single workload failure can have a significant impact on an organization’s capability to deliver services. Respond to real time business events. Today’s on-demand business world requires real-time automation to efficiently respond to business events. Improve IT efficiency. Reducing IT costs continues to be a key requirement for organizations, at the same time IT is expected to improve service delivery. AutoSys Workload Automation enhances visibility and control of complex workloads across platforms, ERP systems, and the cloud. It helps to reduce the cost and complexity of managing mission critical business processes, ensuring consistent and reliable service delivery.
  • 7
    ER/Studio Enterprise Team Edition

    ER/Studio Enterprise Team Edition

    IDERA, an Idera, Inc. company

    ER/Studio Enterprise Team Edition helps data modelers and architects to design and share data models and metadata across the enterprise. Unlike its competition, it provides a complete solution for enterprise architecture and data governance, extensive model change management, unique incorporation of true enterprise data dictionaries, unique linking of constructs across models, integrated visual data lineage import, and integrated data and business process modeling.
  • 8
    Kylo

    Kylo

    Teradata

    Kylo is an open source enterprise-ready data lake management software platform for self-service data ingest and data preparation with integrated metadata management, governance, security and best practices inspired by Think Big's 150+ big data implementation projects. Self-service data ingest with data cleansing, validation, and automatic profiling. Wrangle data with visual sql and an interactive transform through a simple user interface. Search and explore data and metadata, view lineage, and profile statistics. Monitor health of feeds and services in the data lake. Track SLAs and troubleshoot performance. Design batch or streaming pipeline templates in Apache NiFi and register with Kylo to enable user self-service. Organizations can expend significant engineering effort moving data into Hadoop yet struggle to maintain governance and data quality. Kylo dramatically simplifies data ingest by shifting ingest to data owners through a simple guided UI.
  • 9
    Apache Atlas

    Apache Atlas

    Apache Software Foundation

    Atlas is a scalable and extensible set of core foundational governance services – enabling enterprises to effectively and efficiently meet their compliance requirements within Hadoop and allows integration with the whole enterprise data ecosystem. Apache Atlas provides open metadata management and governance capabilities for organizations to build a catalog of their data assets, classify and govern these assets and provide collaboration capabilities around these data assets for data scientists, analysts and the data governance team. Pre-defined types for various Hadoop and non-Hadoop metadata. Ability to define new types for the metadata to be managed. Types can have primitive attributes, complex attributes, object references; can inherit from other types. Instances of types, called entities, capture metadata object details and their relationships. REST APIs to work with types and instances allow easier integration.
  • 10
    Microsoft Power Query
    Power Query is the easiest way to connect, extract, transform and load data from a wide range of sources. Power Query is a data transformation and data preparation engine. Power Query comes with a graphical interface for getting data from sources and a Power Query Editor for applying transformations. Because the engine is available in many products and services, the destination where the data will be stored depends on where Power Query was used. Using Power Query, you can perform the extract, transform, and load (ETL) processing of data. Microsoft’s Data Connectivity and Data Preparation technology that lets you seamlessly access data stored in hundreds of sources and reshape it to fit your needs—all with an easy to use, engaging, no-code experience. Power Query supports hundreds of data sources with built-in connectors, generic interfaces (such as REST APIs, ODBC, OLE, DB and OData) and the Power Query SDK to build your own connectors.
  • 11
    SAS Data Loader for Hadoop
    Load your data into or out of Hadoop and data lakes. Prep it so it's ready for reports, visualizations or advanced analytics – all inside the data lakes. And do it all yourself, quickly and easily. Makes it easy to access, transform and manage data stored in Hadoop or data lakes with a web-based interface that reduces training requirements. Built from the ground up to manage big data on Hadoop or in data lakes; not repurposed from existing IT-focused tools. Lets you group multiple directives to run simultaneously or one after the other. Schedule and automate directives using the exposed Public API. Enables you to share and secure directives. Call them from SAS Data Integration Studio, uniting technical and nontechnical user activities. Includes built-in directives – casing, gender and pattern analysis, field extraction, match-merge and cluster-survive. Profiling runs in-parallel on the Hadoop cluster for better performance.
  • 12
    SAS MDM
    Integrate master data management technologies with those in SAS 9.4. SAS MDM is a web-based application that is accessed through the SAS Data Management Console. It provides a single, accurate and unified view of corporate data, integrating information from various data sources into one master record. SAS® Data Remediation and SAS® Task Manager work together with SAS MDM and as well as with other software offerings, such as SAS® Data Management and SAS® Data Quality. SAS Data Remediation enables users to manage and correct issues triggered by business rules in SAS MDM batch jobs and real-time processes. SAS Task Manager is a complementary application to others that integrate with SAS Workflow technologies giving users direct access to a workflow that might have been initiated from another SAS application. Users can start, stop, and transition workflows that have been uploaded to the SAS Workflow server environment.
  • 13
    Apache Knox

    Apache Knox

    Apache Software Foundation

    The Knox API Gateway is designed as a reverse proxy with consideration for pluggability in the areas of policy enforcement, through providers and the backend services for which it proxies requests. Policy enforcement ranges from authentication/federation, authorization, audit, dispatch, hostmapping and content rewrite rules. Policy is enforced through a chain of providers that are defined within the topology deployment descriptor for each Apache Hadoop cluster gated by Knox. The cluster definition is also defined within the topology deployment descriptor and provides the Knox Gateway with the layout of the cluster for purposes of routing and translation between user facing URLs and cluster internals. Each Apache Hadoop cluster that is protected by Knox has its set of REST APIs represented by a single cluster specific application context path. This allows the Knox Gateway to both protect multiple clusters and present the REST API consumer with a single endpoint.
  • 14
    The Respond Analyst
    Accelerate investigations and improve analyst productivity with a XDR Cybersecurity Solution. The Respond Analyst™, an XDR Engine, automates the discovery of security incidents by turning resource-intensive monitoring and initial analysis into thorough and consistent investigations. Unlike other XDR solutions, the Respond Analyst connects disparate evidence using probabilistic mathematics and integrated reasoning to determine the likelihood that events are malicious and actionable. The Respond Analyst augments security operations teams by significantly reducing the need to chase false positives resulting in more time for threat hunting. The Respond Analyst allows you to choose best-of-breed controls to modernize your sensor grid. The Respond Analyst integrates with the leading security vendor offerings across important categories such as EDR, IPS, Web Filtering, EPP, Vulnerability Scanning, Authentication, and more.
  • 15
    Gurucul

    Gurucul

    Gurucul

    Data science driven security controls to automate advanced threat detection, remediation and response. Gurucul’s Unified Security and Risk Analytics platform answers the question: Is anomalous behavior risky? This is our competitive advantage and why we’re different than everyone else in this space. We don’t waste your time with alerts on anomalous activity that isn’t risky. We use context to determine whether behavior is risky. Context is critical. Telling you what’s happening is not helpful. Telling you when something bad is happening is the Gurucul difference. That’s information you can act on. We put your data to work. We are the only security analytics company that can consume all your data out-of-the-box. We can ingest data from any source – SIEMs, CRMs, electronic medical records, identity and access management systems, end points – you name it, we ingest it into our enterprise risk engine.
  • 16
    OpenText Voltage SecureData
    Secure sensitive data wherever it flows—on premises, in the cloud, and in big data analytic platforms. Voltage encryption delivers data privacy protection, neutralizes data breach, and drives business value through secure data use. Data protection builds customer trust and enables compliance to global regulations, including GDPR, CCPA, and HIPAA. Privacy regulations recommend encryption, pseudonymization, and anonymization to protect personal data. Voltage SecureData enables enterprises to de-identify sensitive structured data and support the use of data in its protect state to safely drive business value. Ensure that applications operate on secure data flowing through the enterprise with no gaps, no decryption, and no performance overhead. SecureData supports the broadest range of platforms and encrypts data in any language. Structured Data Manager integrates SecureData so that businesses can easily and continuously protect data throughout the lifecycle, from discovery to encryption.
  • 17
    Mage Static Data Masking
    Mage™ Static Data Masking (SDM) and Test data Management (TDM) capabilities fully integrate with Imperva’s Data Security Fabric (DSF) delivering complete protection for all sensitive or regulated data while simultaneously integrating seamlessly with an organization’s existing IT framework and existing application development, testing and data flows without the requirement for any additional architectural changes.
  • 18
    Mage Dynamic Data Masking
    Mage™ Dynamic Data Masking module of the Mage data security platform has been designed with the end customer needs taken into consideration. Mage™ Dynamic Data Masking has been developed working alongside our customers, to address the specific needs and requirements they have. As a result, this product has evolved in a way to meet all the use cases that an enterprise could possibly have. Most other solutions in the market are either a part of an acquisition or are developed to meet only a specific use case. Mage™ Dynamic Data Masking has been designed to deliver adequate protection to sensitive data in production to application and database users while simultaneously integrating seamlessly with an organization's existing IT framework without the requirement of any additional architectural changes.​
  • 19
    Acxiom Real Identity
    Real Identity™ delivers sub second decisions to power relevant messages in real time. Real Identity enables the world’s biggest brands to accurately identify and ethically connect with people anytime, anywhere to create relevant experiences. Engage people with reach, scale and precision across every interaction. Manage and maintain identity across your enterprise by leveraging 50 years of data and identity expertise combined with the latest artificial intelligence and machine learning techniques. The adtech environment requires speed and access to identity and data to enable personalization and decisioning use cases. In a cookieless world, first-party data signals will drive these functions while the conversation continues to be between people, the brands, and the publishers. By delivering experiences that matter, across all channels, you can wow your customer and prospects while staying ahead of regulations and ahead of your competition.
  • 20
    Okera

    Okera

    Okera

    Okera, the Universal Data Authorization company, helps modern, data-driven enterprises accelerate innovation, minimize data security risks, and demonstrate regulatory compliance. The Okera Dynamic Access Platform automatically enforces universal fine-grained access control policies. This allows employees, customers, and partners to use data responsibly, while protecting them from inappropriately accessing data that is confidential, personally identifiable, or regulated. Okera’s robust audit capabilities and data usage intelligence deliver the real-time and historical information that data security, compliance, and data delivery teams need to respond quickly to incidents, optimize processes, and analyze the performance of enterprise data initiatives. Okera began development in 2016 and now dynamically authorizes access to hundreds of petabytes of sensitive data for the world’s most demanding F100 companies and regulatory agencies. The company is headquartered in San Francisco.
  • 21
    Apache Sentry

    Apache Sentry

    Apache Software Foundation

    Apache Sentry™ is a system for enforcing fine grained role based authorization to data and metadata stored on a Hadoop cluster. Apache Sentry has successfully graduated from the Incubator in March of 2016 and is now a Top-Level Apache project. Apache Sentry is a granular, role-based authorization module for Hadoop. Sentry provides the ability to control and enforce precise levels of privileges on data for authenticated users and applications on a Hadoop cluster. Sentry currently works out of the box with Apache Hive, Hive Metastore/HCatalog, Apache Solr, Impala and HDFS (limited to Hive table data). Sentry is designed to be a pluggable authorization engine for Hadoop components. It allows you to define authorization rules to validate a user or application’s access requests for Hadoop resources. Sentry is highly modular and can support authorization for a wide variety of data models in Hadoop.
  • 22
    Apache Bigtop

    Apache Bigtop

    Apache Software Foundation

    Bigtop is an Apache Foundation project for Infrastructure Engineers and Data Scientists looking for comprehensive packaging, testing, and configuration of the leading open source big data components. Bigtop supports a wide range of components/projects, including, but not limited to, Hadoop, HBase and Spark. Bigtop packages Hadoop RPMs and DEBs, so that you can manage and maintain your Hadoop cluster. Bigtop provides an integrated smoke testing framework, alongside a suite of over 50 test files. Bigtop provides vagrant recipes, raw images, and (work-in-progress) docker recipes for deploying Hadoop from zero. Bigtop support many Operating Systems, including Debian, Ubuntu, CentOS, Fedora, openSUSE and many others. Bigtop includes tools and a framework for testing at various levels (packaging, platform, runtime, etc.) for both initial deployments as well as upgrade scenarios for the entire data platform, not just the individual components.
  • 23
    Secuvy AI
    Secuvy is a next-generation cloud platform to automate data security, privacy compliance and governance via AI-driven workflows. Best in class data intelligence especially for unstructured data. Secuvy is a next-generation cloud platform to automate data security, privacy compliance and governance via ai-driven workflows. Best in class data intelligence especially for unstructured data. Automated data discovery, customizable subject access requests, user validations, data maps & workflows for privacy regulations such as ccpa, gdpr, lgpd, pipeda and other global privacy laws. Data intelligence to find sensitive and privacy information across multiple data stores at rest and in motion. In a world where data is growing exponentially, our mission is to help organizations to protect their brand, automate processes, and improve trust with customers. With ever-expanding data sprawls we wish to reduce human efforts, costs & errors for handling Sensitive Data.
  • 24
    iFinder

    iFinder

    IntraFind Software

    IntraFind's Enterprise Search solution iFinder is a central search platform for all of your company's data. iFinder can be connected to all of the data sources within your company. Are your data pools constantly growing? With iFinder you are well equipped for the future: our product is based on Elasticsearch technology and can therefore adapt to any volume of data with ease. It also improves search results by deploying artificial intelligence to deliver smart enterprise search functionality. iFinder helps you find important data and documents, whether they are located on a company drive, on the intranet, in wikis, or in e-mail systems. Take the next step in your company’s digital transformation by centralizing access to all company data with our enterprise search application. iFinder improves search results by deploying artificial intelligence to deliver smart enterprise search functionality.
  • 25
    NVMesh

    NVMesh

    Excelero

    Excelero delivers low-latency distributed block storage for web-scale applications. NVMesh enables shared NVMe across any network and supports any local or distributed file system. The solution features an intelligent management layer that abstracts underlying hardware with CPU offload, creates logical volumes with redundancy, and provides centralized, intelligent management and monitoring. Applications can enjoy the latency, throughput and IOPs of a local NVMe device with the convenience of centralized storage while avoiding proprietary hardware lock-in and reducing the overall storage TCO. NVMesh features a distributed block layer that allows unmodified applications to utilize pooled NVMe storage devices across a network at local speeds and latencies. Distributed NVMe storage resources are pooled with the ability to create arbitrary, dynamic block volumes that can be utilized by any host running the NVMesh block client.
  • 26
    lakeFS

    lakeFS

    Treeverse

    lakeFS enables you to manage your data lake the way you manage your code. Run parallel pipelines for experimentation and CI/CD for your data. Simplifying the lives of engineers, data scientists and analysts who are transforming the world with data. lakeFS is an open source platform that delivers resilience and manageability to object-storage based data lakes. With lakeFS you can build repeatable, atomic and versioned data lake operations, from complex ETL jobs to data science and analytics. lakeFS supports AWS S3, Azure Blob Storage and Google Cloud Storage (GCS) as its underlying storage service. It is API compatible with S3 and works seamlessly with all modern data frameworks such as Spark, Hive, AWS Athena, Presto, etc. lakeFS provides a Git-like branching and committing model that scales to exabytes of data by utilizing S3, GCS, or Azure Blob for storage.
  • 27
    Prodea

    Prodea

    Prodea

    Launch secure, scalable and globally compliant connected products with services within six months. Prodea provides the only IoT platform-as-a-service (PaaS) that was specifically designed for manufacturers of mass-market consumer home products. It is comprised of three main services. IoT Service X-Change Platform, for quickly launching connected products with services across global markets requiring minimal development. Insight™ Data Services, to gain key insights from user and product usage data. And EcoAdaptor™ Service, to enhance product value through cloud-to-cloud integration and interoperability with other products and services. Prodea has helped its global brand customers launch 100+ connected products, in less than six months on average, across six continents. This was made possible by using the Prodea X5 Program which was designed to work with our three main cloud services to help brands evolve their systems.
  • 28
    GO+

    GO+

    GO+

    GO+ offers development tools for companies service-providers. The platform allows to develop additional services to its business customers. Through algorithms developed by our decision to support a load with a huge number of devices at the same time. The platform will allow the service-provider does not think about the problem of creating new services for their customers. The core of the platform is an analytical decision-making engine. Granular Computing based analytical engine provides data processing and analysis with complex event processing. We use cloud-based technologies that is designed to endure business logic from real devices directly to the cloud. Scalability allows us to provide solutions in lower cost. Platform scripting engine provide developers with full-stack of dev tools to develop a highly customized IoT services regardless of industry application. The cloud-based IoT platform GO+ is built using the most advanced technology of cloud computing.
  • 29
    Foghub

    Foghub

    Foghub

    Simplified IT/OT Integration, Data Engineering & Real-Time Edge Intelligence. Easy to use, cross-platform, open architecture, edge computing for industrial time-series data. Foghub offers the Critical-Path to IT/OT convergence, connecting Operations (Sensors, Devices, and Systems) with Business (People, Processes, and Applications), enabling automated data acquisition, data engineering, transformations, advanced analytics and ML. Handle large variety, volume, and velocity of industrial data with out-of-the-box support for all data types, most popular industrial network protocols, OT/lab systems, and databases. Easily automate the collection of data about your production runs, batches, parts, cycle-times, process parameters, asset condition, performance, health, utilities, consumables as well as operators and their performance. Designed for scale, Foghub offers a comprehensive set of capabilities to handle large volumes and velocity of data.
  • 30
    Brainwave GRC

    Brainwave GRC

    Radiant Logic

    Brainwave is reinventing the way you analyze your user accesses! You will now be able to thoroughly analyze access risk thanks to a new user interface, predictive controls and risk-scoring functionality. With Autonomous Identity, you can engage your teams and improve their efficiency with a market-approved, ergonomic tool that accelerates your identity management program (IGA). Enable the business to review and make decisions about access to shared files and folders. Inventory, classify, review access and demonstrate compliance regardless of the location, file servers, NAS, Sharepoint, Office 365 and others. Our core product, Brainwave Identity GRC, provides a wealth of analytical capabilities to leverage the inventory of all access. Obtain full visibility at all time, on all resources. Brainwave’s inventory constitutes an entitlement catalog across infrastructure, business applications and data access.