Alternatives to TIBCO ComputeDB

Compare TIBCO ComputeDB alternatives for your business or organization using the curated list below. SourceForge ranks the best alternatives to TIBCO ComputeDB in 2024. Compare features, ratings, user reviews, pricing, and more from TIBCO ComputeDB competitors and alternatives in order to make an informed decision for your business.

  • 1
    Striim

    Striim

    Striim

    Data integration for your hybrid cloud. Modern, reliable data integration across your private and public cloud. All in real-time with change data capture and data streams. Built by the executive & technical team from GoldenGate Software, Striim brings decades of experience in mission-critical enterprise workloads. Striim scales out as a distributed platform in your environment or in the cloud. Scalability is fully configurable by your team. Striim is fully secure with HIPAA and GDPR compliance. Built ground up for modern enterprise workloads in the cloud or on-premise. Drag and drop to create data flows between your sources and targets. Process, enrich, and analyze your streaming data with real-time SQL queries.
  • 2
    IBM Cloud SQL Query
    Serverless, interactive querying for analyzing data in IBM Cloud Object Storage. Query your data directly where it is stored, there's no ETL, no databases, and no infrastructure to manage. IBM Cloud SQL Query uses Apache Spark, an open-source, fast, extensible, in-memory data processing engine optimized for low latency and ad hoc analysis of data. No ETL or schema definition needed to enable SQL queries. Analyze data where it sits in IBM Cloud Object Storage using our query editor and REST API. Run as many queries as you need; with pay-per-query pricing, you pay only for the data scan. Compress or partition data to drive savings and performance. IBM Cloud SQL Query is highly available and executes queries using compute resources across multiple facilities. IBM Cloud SQL Query supports a variety of data formats such as CSV, JSON and Parquet, and allows for standard ANSI SQL.
    Starting Price: $5.00/Terabyte-Month
  • 3
    Apache Druid
    Apache Druid is an open source distributed data store. Druid’s core design combines ideas from data warehouses, timeseries databases, and search systems to create a high performance real-time analytics database for a broad range of use cases. Druid merges key characteristics of each of the 3 systems into its ingestion layer, storage format, querying layer, and core architecture. Druid stores and compresses each column individually, and only needs to read the ones needed for a particular query, which supports fast scans, rankings, and groupBys. Druid creates inverted indexes for string values for fast search and filter. Out-of-the-box connectors for Apache Kafka, HDFS, AWS S3, stream processors, and more. Druid intelligently partitions data based on time and time-based queries are significantly faster than traditional databases. Scale up or down by just adding or removing servers, and Druid automatically rebalances. Fault-tolerant architecture routes around server failures.
  • 4
    TIBCO Cloud Messaging
    High-performing applications require a real-time exchange of information, exactly what TIBCO Cloud™ Messaging software provides. With multiple options, including Apache Kafka, Apache Pulsar, TIBCO Enterprise Message Service™, TIBCO FTL® and TIBCO eFTL™ software, we have you covered. Whether you need event-driven app integration; high-performance, low-latency messaging; connections for web apps to backend systems; or IoT integration, the robust yet simple TIBCO Cloud Messaging client API and cloud distribution network is the easiest way to connect. TIBCO Cloud Messaging software provides built-in fault tolerance, multi-zone data reliability, and scalable performance. It allows you to focus on application logic and not worry about deploying containers or infrastructure for your event streams. Simply connect your apps to the API and start publishing and subscribing to event messages immediately.
  • 5
    Spark Streaming

    Spark Streaming

    Apache Software Foundation

    Spark Streaming brings Apache Spark's language-integrated API to stream processing, letting you write streaming jobs the same way you write batch jobs. It supports Java, Scala and Python. Spark Streaming recovers both lost work and operator state (e.g. sliding windows) out of the box, without any extra code on your part. By running on Spark, Spark Streaming lets you reuse the same code for batch processing, join streams against historical data, or run ad-hoc queries on stream state. Build powerful interactive applications, not just analytics. Spark Streaming is developed as part of Apache Spark. It thus gets tested and updated with each Spark release. You can run Spark Streaming on Spark's standalone cluster mode or other supported cluster resource managers. It also includes a local run mode for development. In production, Spark Streaming uses ZooKeeper and HDFS for high availability.
  • 6
    Google Cloud Dataflow
    Unified stream and batch data processing that's serverless, fast, and cost-effective. Fully managed data processing service. Automated provisioning and management of processing resources. Horizontal autoscaling of worker resources to maximize resource utilization. OSS community-driven innovation with Apache Beam SDK. Reliable and consistent exactly-once processing. Streaming data analytics with speed. Dataflow enables fast, simplified streaming data pipeline development with lower data latency. Allow teams to focus on programming instead of managing server clusters as Dataflow’s serverless approach removes operational overhead from data engineering workloads. Allow teams to focus on programming instead of managing server clusters as Dataflow’s serverless approach removes operational overhead from data engineering workloads. Dataflow automates provisioning and management of processing resources to minimize latency and maximize utilization.
  • 7
    DeltaStream

    DeltaStream

    DeltaStream

    DeltaStream is a unified serverless stream processing platform that integrates with streaming storage services. Think about it as the compute layer on top of your streaming storage. It provides functionalities of streaming analytics(Stream processing) and streaming databases along with additional features to provide a complete platform to manage, process, secure and share streaming data. DeltaStream provides a SQL based interface where you can easily create stream processing applications such as streaming pipelines, materialized views, microservices and many more. It has a pluggable processing engine and currently uses Apache Flink as its primary stream processing engine. DeltaStream is more than just a query processing layer on top of Kafka or Kinesis. It brings relational database concepts to the data streaming world, including namespacing and role based access control enabling you to securely access, process and share your streaming data regardless of where they are stored.
  • 8
    HEAVY.AI

    HEAVY.AI

    HEAVY.AI

    HEAVY.AI is the pioneer in accelerated analytics. The HEAVY.AI platform is used in business and government to find insights in data beyond the limits of mainstream analytics tools. Harnessing the massive parallelism of modern CPU and GPU hardware, the platform is available in the cloud and on-premise. HEAVY.AI originated from research at Harvard and MIT Computer Science and Artificial Intelligence Laboratory (CSAIL). Expand beyond the limitations of traditional BI and GIS by leveraging the full power of modern GPU and CPU hardware so you can extract decision-quality information from your massive datasets without lag. Unify and explore your largest geospatial and time-series datasets to get the complete picture of the what, when, and where. Combine interactive visual analytics, hardware-accelerated SQL, and an advanced analytics & data science framework to find opportunity and risk hidden in your enterprise when you need to most.
  • 9
    Macrometa

    Macrometa

    Macrometa

    We deliver a geo-distributed real-time database, stream processing and compute runtime for event-driven applications across up to 175 worldwide edge data centers. App & API builders love our platform because we solve the hardest problems of sharing mutable state across 100s of global locations, with strong consistency & low latency. Macrometa enables you to surgically extend your existing infrastructure to bring part of or your entire application closer to your end users. This allows you to improve performance, user experience, and comply with global data governance laws. Macrometa is a serverless, streaming NoSQL database, with integrated pub/sub and stream data processing and compute engine. Create stateful data infrastructure, stateful functions & containers for long running workloads, and process data streams in real time. You do the code, we do all the ops and orchestration.
  • 10
    Hazelcast

    Hazelcast

    Hazelcast

    In-Memory Computing Platform. The digital world is different. Microseconds matter. That's why the world's largest organizations rely on us to power their most time-sensitive applications at scale. New data-enabled applications can deliver transformative business power – if they meet today’s requirement of immediacy. Hazelcast solutions complement virtually any database to deliver results that are significantly faster than a traditional system of record. Hazelcast’s distributed architecture provides redundancy for continuous cluster up-time and always available data to serve the most demanding applications. Capacity grows elastically with demand, without compromising performance or availability. The fastest in-memory data grid, combined with third-generation high-speed event processing, delivered through the cloud.
  • 11
    Oracle Database
    Oracle database products offer customers cost-optimized and high-performance versions of Oracle Database, the world's leading converged, multi-model database management system, as well as in-memory, NoSQL, and MySQL databases. Oracle Autonomous Database, available on-premises via Oracle Cloud@Customer or in the Oracle Cloud Infrastructure, enables customers to simplify relational database environments and reduce management workloads. Oracle Autonomous Database eliminates the complexity of operating and securing Oracle Database while giving customers the highest levels of performance, scalability, and availability. Oracle Database can be deployed on-premises when customers have data residency and network latency concerns. Customers with applications that are dependent on specific Oracle database versions have complete control over the versions they run and when those versions change.
  • 12
    SAP HANA
    SAP HANA in-memory database is for transactional and analytical workloads with any data type — on a single data copy. It breaks down the transactional and analytical silos in organizations, for quick decision-making, on premise and in the cloud. Innovate without boundaries on a database management system, where you can develop intelligent and live solutions for quick decision-making on a single data copy. And with advanced analytics, you can support next-generation transactional processing. Build data solutions with cloud-native scalability, speed, and performance. With the SAP HANA Cloud database, you can gain trusted, business-ready information from a single solution, while enabling security, privacy, and anonymization with proven enterprise reliability. An intelligent enterprise runs on insight from data – and more than ever, this insight must be delivered in real time.
  • 13
    Amazon MSK

    Amazon MSK

    Amazon

    Amazon MSK is a fully managed service that makes it easy for you to build and run applications that use Apache Kafka to process streaming data. Apache Kafka is an open-source platform for building real-time streaming data pipelines and applications. With Amazon MSK, you can use native Apache Kafka APIs to populate data lakes, stream changes to and from databases, and power machine learning and analytics applications. Apache Kafka clusters are challenging to setup, scale, and manage in production. When you run Apache Kafka on your own, you need to provision servers, configure Apache Kafka manually, replace servers when they fail, orchestrate server patches and upgrades, architect the cluster for high availability, ensure data is durably stored and secured, setup monitoring and alarms, and carefully plan scaling events to support load changes.
    Starting Price: $0.0543 per hour
  • 14
    SingleStore

    SingleStore

    SingleStore

    SingleStore (formerly MemSQL) is a distributed, highly-scalable SQL database that can run anywhere. We deliver maximum performance for transactional and analytical workloads with familiar relational models. SingleStore is a scalable SQL database that ingests data continuously to perform operational analytics for the front lines of your business. Ingest millions of events per second with ACID transactions while simultaneously analyzing billions of rows of data in relational SQL, JSON, geospatial, and full-text search formats. SingleStore delivers ultimate data ingestion performance at scale and supports built in batch loading and real time data pipelines. SingleStore lets you achieve ultra fast query response across both live and historical data using familiar ANSI SQL. Perform ad hoc analysis with business intelligence tools, run machine learning algorithms for real-time scoring, perform geoanalytic queries in real time.
    Starting Price: $0.69 per hour
  • 15
    Red Hat OpenShift Streams
    Red Hat® OpenShift® Streams for Apache Kafka is a managed cloud service that provides a streamlined developer experience for building, deploying, and scaling new cloud-native applications or modernizing existing systems. Red Hat OpenShift Streams for Apache Kafka makes it easy to create, discover, and connect to real-time data streams no matter where they are deployed. Streams are a key component for delivering event-driven and data analytics applications. The combination of seamless operations across distributed microservices, large data transfer volumes, and managed operations allows teams to focus on team strengths, speed up time to value, and lower operational costs. OpenShift Streams for Apache Kafka includes a Kafka ecosystem and is part of a family of cloud services—and the Red Hat OpenShift product family—which helps you build a wide range of data-driven solutions.
  • 16
    LeanXcale

    LeanXcale

    LeanXcale

    LeanXcale is a fast and scalable database that combines the characteristics of SQL and NoSQL. It is built to ingest massive batch and real-time data pipelines and make it available through SQL or GIS for any use, such as operational applications, analytics, dashboarding, or machine learning processing. No matter what stack you use, LeanXcale provides you both SQL and NoSQL interfaces. KiVi storage engine is a relational key-value data store. Users can access the data not only through the standard SQL API but also through a direct ACID key-value interface. This key-value interface allows users to perform data ingestion at very high rates and very efficiently by avoiding SQL processing overhead. Highly-scalable, efficient and distributed storage engine distributed data along the cluster to improve the performance and increase the reliability.
    Starting Price: $0.127 per GB per month
  • 17
    VoltDB

    VoltDB

    VoltDB

    Volt Active Data is a data platform built to make your entire tech stack leaner, faster, and less expensive, so that your applications (and your company) can scale seamlessly to meet the ultra-low latency SLAs of 5G, IoT, edge computing, and whatever comes next. Designed to augment your existing big data investments, such as NoSQL, Hadoop, Kubernetes, Kafka, and traditional databases or data warehouses, Volt Active Data replaces the various layers typically required to make contextual decisions on streaming data with a single, unified layer that can handle ingest to action in less than 10 milliseconds. The world is full of data that’s generated, stored, forgotten, and then deleted. “Active Data” is data that needs to be acted on immediately to gain business value from it. There are lots of traditional and NoSQL data storage products that you can use to keep such data. There’s also data that you can make money from, if only you can act on it fast enough to ‘influence the moment’.
  • 18
    Akka

    Akka

    Akka

    Akka is a toolkit for building highly concurrent, distributed, and resilient message-driven applications for Java and Scala. Akka Insights is intelligent monitoring and observability purpose-built for Akka. Actors and Streams let you build systems that scale up, using the resources of a server more efficiently, and out, using multiple servers. Building on the principles of The Reactive Manifesto Akka allows you to write systems that self-heal and stay responsive in the face of failures. Distributed systems without single points of failure. Load balancing and adaptive routing across nodes. Event Sourcing and CQRS with Cluster Sharding. Distributed Data for eventual consistency using CRDTs. Asynchronous non-blocking stream processing with backpressure. Fully async and streaming HTTP server and client provides a great platform for building microservices. Streaming integrations with Alpakka.
  • 19
    PubNub

    PubNub

    PubNub

    Innovate with Realtime Features: We take care of realtime communication infrastructure so you can focus on your app. Our Platform for Realtime Communication: A platform to build and operate real-time interactivity for web, mobile, AI/ML, IoT, and Edge computing applications Faster & Easier Deployments: SDK support for 50+ mobile, web, server, and IoT environments (PubNub and community supported) and more than 65 pre-built integrations with external and third-party APIs to give developers the features they need regardless of programming language or tech stack. Scalability: The industry’s most scalable platform capable of supporting millions of concurrent users and allows for rapid growth with low latency, high uptime, and without financial penalties. Security & Compliance: Enterprise-grade security and compliance with the most stringent regulations worldwide, including GDPR, SOC 2, HIPAA, ISO 27001, and CCPA.
  • 20
    Aiven for Apache Kafka
    Apache Kafka as a fully managed service, with zero vendor lock-in and a full set of capabilities to build your streaming pipeline. Set up fully managed Kafka in less than 10 minutes — directly from our web console or programmatically via our API, CLI, Terraform provider or Kubernetes operator. Easily connect it to your existing tech stack with over 30 connectors, and feel confident in your setup with logs and metrics available out of the box via the service integrations. A fully managed distributed data streaming platform, deployable in the cloud of your choice. Ideal for event-driven applications, near-real-time data transfer and pipelines, stream analytics, and any other case where you need to move a lot of data between applications — and quickly. With Aiven’s hosted and managed-for-you Apache Kafka, you can set up clusters, deploy new nodes, migrate clouds, and upgrade existing versions — in a single mouse click — and monitor them through a simple dashboard.
    Starting Price: $200 per month
  • 21
    Pathway

    Pathway

    Pathway

    Pathway is a Python ETL framework for stream processing, real-time analytics, LLM pipelines, and RAG. Pathway comes with an easy-to-use Python API, allowing you to seamlessly integrate your favorite Python ML libraries. Pathway code is versatile and robust: you can use it in both development and production environments, handling both batch and streaming data effectively. The same code can be used for local development, CI/CD tests, running batch jobs, handling stream replays, and processing data streams. Pathway is powered by a scalable Rust engine based on Differential Dataflow and performs incremental computation. Your Pathway code, despite being written in Python, is run by the Rust engine, enabling multithreading, multiprocessing, and distributed computations. All the pipeline is kept in memory and can be easily deployed with Docker and Kubernetes.
  • 22
    Radicalbit

    Radicalbit

    Radicalbit

    Radicalbit Natural Analytics (RNA) is a DataOps platform for Streaming Data Integration and Real-time Advanced Analytics. Choose the easiest way to deliver data at the right time in the right hands. RNA provides users with the latest technologies – in self service mode – for real-time data processing, taking advantage of Artificial Intelligence solutions for extracting value from data. It automates the labor-intensive process of data analysis and helps convey important findings and insights in understandable formats. Have a Real-time situational awareness and timely insights to respond quickly and appropriately. Achieve new levels of efficiency and optimization and guarantee collaboration between siloed teams. Manage and monitor the models from a centralized view, and deploy your evolving models in seconds. No downtime.
  • 23
    CockroachDB

    CockroachDB

    Cockroach Labs

    CockroachDB: Cloud-native, distributed SQL. Your cloud applications deserve a cloud-native database. Cloud-based apps and services deserve a database that scales across clouds, eases operational complexity, and improves reliability. CockroachDB delivers resilient, distributed SQL with ACID transactions and data partitioned by location. Automate operations for mission-critical applications by pairing CockroachDB with orchestration tools like Kubernetes and Mesosphere DC/OS. Every node can service both reads and writes so that you can scale query throughput and database capacity by simply adding more endpoints. Just add new nodes to CockroachDB, and it automatically rebalances data, completely removing the pain of manual sharding. As demand shifts, CockroachDB detects hotspots and intelligently distributes data to maintain performance. Tune your database at the row level so that data lives close to your users and you can minimize query latency.
  • 24
    PolarDB

    PolarDB

    Alibaba Cloud

    PolarDB is designed for business-critical database applications that require fast performance, high concurrency, and automatic scaling. You can scale up to millions of queries per second and 100 TB per database cluster with 15 low latency read replicas. PolarDB is six times faster than standard MySQL databases, and delivers the security, reliability, and availability of traditional commercial databases at 1/10 the cost. PolarDB embodies the proven database technology and best practices honed over the last decade that supported hyper-scale events such as the Alibaba Double 11 Global Shopping Festival. To support the developer community, we are introducing Always Free ApsaraDB for PolarDB (all three variations) when you use no more than 1 instance (2-core and 8GB of memory), and up to 50GB of storage. Register now and renew each month to continue this benefit. Regional resource availability is subject to change.
  • 25
    Dqlite

    Dqlite

    Canonical

    Dqlite is a fast, embedded, persistent SQL database with Raft consensus that is perfect for fault-tolerant IoT and Edge devices. Dqlite (“distributed SQLite”) extends SQLite across a cluster of machines, with automatic failover and high-availability to keep your application running. It uses C-Raft, an optimised Raft implementation in C, to gain high-performance transactional consensus and fault tolerance while preserving SQlite’s outstanding efficiency and tiny footprint. C-Raft is tuned to minimize transaction latency. C-Raft and dqlite are both written in C for maximum cross-platform portability. Published under the LGPLv3 license with a static linking exception for maximum compatibility. Includes common CLI pattern for database initialization and voting member joins and departures. Minimal, tunable delay for failover with automatic leader election. Disk-backed database with in-memory options and SQLite transactions.
  • 26
    MariaDB

    MariaDB

    MariaDB

    MariaDB Platform is a complete enterprise open source database solution. It has the versatility to support transactional, analytical and hybrid workloads as well as relational, JSON and hybrid data models. And it has the scalability to grow from standalone databases and data warehouses to fully distributed SQL for executing millions of transactions per second and performing interactive, ad hoc analytics on billions of rows. MariaDB can be deployed on prem on commodity hardware, is available on all major public clouds and through MariaDB SkySQL as a fully managed cloud database. To learn more, visit mariadb.com.
  • 27
    TIBCO Streaming
    Analyze, continuously query, and act on IoT and other streaming data at lightning fast speeds. Take real-time operations and analytics to the next level with intelligent applications that deploy quickly for taking action based on new decisions and models, all without extra overhead. TIBCO® Streaming software is enterprise-grade, cloud-ready streaming analytics for quickly building real-time applications at a fraction of the cost and risk of alternatives.
  • 28
    Amazon Aurora
    Amazon Aurora is a MySQL and PostgreSQL-compatible relational database built for the cloud, that combines the performance and availability of traditional enterprise databases with the simplicity and cost-effectiveness of open source databases. Amazon Aurora is up to five times faster than standard MySQL databases and three times faster than standard PostgreSQL databases. It provides the security, availability, and reliability of commercial databases at 1/10th the cost. Amazon Aurora is fully managed by Amazon Relational Database Service (RDS), which automates time-consuming administration tasks like hardware provisioning, database setup, patching, and backups. Amazon Aurora features a distributed, fault-tolerant, self-healing storage system that auto-scales up to 64TB per database instance. It delivers high performance and availability with up to 15 low-latency read replicas, point-in-time recovery, continuous backup to Amazon S3, and replication across three Availability Zones.
    Starting Price: $0.02 per month
  • 29
    Presto

    Presto

    Presto Foundation

    Presto is an open source distributed SQL query engine for running interactive analytic queries against data sources of all sizes ranging from gigabytes to petabytes. For data engineers who struggle with managing multiple query languages and interfaces to siloed databases and storage, Presto is the fast and reliable engine that provides one simple ANSI SQL interface for all your data analytics and your open lakehouse. Different engines for different workloads means you will have to re-platform down the road. With Presto, you get 1 familar ANSI SQL language and 1 engine for your data analytics so you don't need to graduate to another lakehouse engine. Presto can be used for interactive and batch workloads, small and large amounts of data, and scales from a few to thousands of users. Presto gives you one simple ANSI SQL interface for all of your data in various siloed data systems, helping you join your data ecosystem together.
  • 30
    Infobright DB

    Infobright DB

    IgniteTech

    Infobright DB is a high-performance enterprise database leveraging a columnar storage engine to enable business analysts to dissect data efficiently and more quickly obtain reports. InfoBright DB can be deployed on-premise or in the cloud. Store & analyze big data for interactive business intelligence and complex queries. Improve query performance, reduce storage cost and increase overall efficiency in business analytics and reporting. Easily store up to several hundred TB of data — traditionally not achievable with conventional databases. Run big data applications and eliminate indexing and partitioning — with zero administrative overhead. With the volumes of machine data exploding, IgniteTech’s Infobright DB is specifically designed to achieve high performance for large volumes of machine-generated data. Manage a complex ad hoc analytic environments without the database administration required by other products.
  • 31
    Pandio

    Pandio

    Pandio

    Connecting systems to scale AI initiatives is complex, expensive, and prone to fail. Pandio’s cloud-native managed solution simplifies your data pipelines to harness the power of AI. Access your data from anywhere at any time in order to query, analyze, and drive to insight. Big data analytics without the big cost. Enable data movement seamlessly. Streaming, queuing and pub-sub with unmatched throughput, latency, and durability. Design, train, and deploy machine learning models locally in less than 30 minutes. Accelerate your path to ML and democratize the process across your organization. And it doesn’t require months (or years) of disappointment. Pandio’s AI-driven architecture automatically orchestrates your models, data, and ML tools. Pandio works with your existing stack to accelerate your ML initiatives. Orchestrate your models and messages across your organization.
    Starting Price: $1.40 per hour
  • 32
    Informatica Data Engineering Streaming
    AI-powered Informatica Data Engineering Streaming enables data engineers to ingest, process, and analyze real-time streaming data for actionable insights. Advanced serverless deployment option​ with integrated metering dashboard cuts admin overhead. Rapidly build intelligent data pipelines with CLAIRE®-powered automation, including automatic change data capture (CDC). Ingest thousands of databases and millions of files, and streaming events. Efficiently ingest databases, files, and streaming data for real-time data replication and streaming analytics. Find and inventory all data assets throughout your organization. Intelligently discover and prepare trusted data for advanced analytics and AI/ML projects.
  • 33
    SAS Event Stream Processing
    Streaming data from operations, transactions, sensors and IoT devices is valuable – when it's well-understood. Event stream processing from SAS includes streaming data quality and analytics – and a vast array of SAS and open source machine learning and high-frequency analytics for connecting, deciphering, cleansing and understanding streaming data – in one solution. No matter how fast your data moves, how much data you have, or how many data sources you’re pulling from, it’s all under your control via a single, intuitive interface. You can define patterns and address scenarios from all aspects of your business, giving you the power to stay agile and tackle issues as they arise.
  • 34
    IBM Cloud Pak for Integration
    IBM Cloud Pak for Integration® is a hybrid integration platform with an automated, closed-loop approach that supports multiple styles of integration within a single, unified experience. Unlock business data and assets as APIs, connect cloud and on-premise applications, reliably move data with enterprise messaging, deliver real-time event interactions, transfer data across any cloud and deploy and scale with cloud-native architecture and shared foundational services, all with end-to-end enterprise-grade security and encryption. Achieve the best results from integration with an automated, closed-loop and multi-style approach. Apply targeted innovations to automate integrations, such as natural language–powered integration flows, AI-assisted mapping and RPA, and use company-specific operational data to continuously improve integrations, enhance API test generation, workload balancing and more.
    Starting Price: $934 per month
  • 35
    H2

    H2

    H2

    Welcome to H2, the Java SQL database. In embedded mode, an application opens a database from within the same JVM using JDBC. This is the fastest and easiest connection mode. The disadvantage is that a database may only be open in one virtual machine (and class loader) at any time. As in all modes, both persistent and in-memory databases are supported. There is no limit on the number of database open concurrently, or on the number of open connections. The mixed mode is a combination of the embedded and the server mode. The first application that connects to a database does that in embedded mode, but also starts a server so that other applications (running in different processes or virtual machines) can concurrently access the same data. The local connections are as fast as if the database is used in just the embedded mode, while the remote connections are a bit slower.
  • 36
    Vitess

    Vitess

    Vitess

    A database clustering system for horizontal scaling of MySQL. Vitess combines many important MySQL features with the scalability of a NoSQL database. Its built-in sharding features let you grow your database without adding sharding logic to your application. Vitess automatically rewrites queries that hurt database performance. It also uses caching mechanisms to mediate queries and prevent duplicate queries from simultaneously reaching your database. Vitess automatically handles functions like master failovers and backups. It uses a lock server to track and administer servers, letting your application be blissfully ignorant of database topology. Vitess eliminates the high-memory overhead of MySQL connections. Vitess servers easily handle thousands of connections at once. MySQL doesn’t natively support sharding, but you will likely need it as your database grows.
  • 37
    SAP Adaptive Server Enterprise (ASE)
    Modernize and accelerate your transaction-based applications on premise and in the cloud with SAP Adaptive Server Enterprise (SAP ASE). This high-performance SQL database server uses a relational management model to meet rising demand for performance, reliability, and efficiency in every industry. Simplify operations and reduce costs with workload analyzer and workload profiler features for built-in high availability, disaster recovery, and resource optimization. Scale transactions, data, and users and improve performance with added technology options, such as MemScale and XOLTP. Discover how reducing overhead on production databases enables Q-Free employees to focus on innovation and value-adding work. Deploy on-premise and on infrastructure as a service (IaaS) Deliver high performance and availability to process mission-critical transactions.
  • 38
    Altibase

    Altibase

    Altibase

    Altibase is an enterprise-grade, high-performance and relational open source database. A single database that delivers high-intensity data processing through an in-memory database portion and large storage capacity through an on-disk database portion. 10 times faster than conventional on-disk databases. Clients have consistently chosen Altibase over Oracle, IBM, Microsoft, and others. Altibase has replaced many traditional on-disk databases in various industries that require real time solutions since 1999. Altibase now has over 650 global enterprise clients including 8 Fortune Global 500 companies with thousands of mission-critical deployments worldwide. Product maturity rich with function and feature. Altibase is open source which includes its cutting-edge scale-out technology, sharding. No license costs with flexible and competitive subscription fees. 20 years’ accumulated know-how of dealing with over 6,000 mission-critical use cases.
  • 39
    Teradata Vantage
    As data volumes grow faster than ever, businesses struggle to get answers. Teradata Vantage™ solves this problem. Vantage uses 100 percent of available data to uncover real-time business intelligence at scale, powering the new era of Pervasive Data Intelligence. See all data from across the entire organization in one place, whenever it's needed, with preferred languages and tools. Start small and elastically scale compute or storage in areas that impact modern architecture. Vantage unifies analytics, Data Lakes, and Data Warehouses, all in the cloud to enable business intelligence. The importance of business intelligence increases. Frustration stems from four key challenges that arise when using existing data analytics platforms: Lack of proper tools and supportive environment needed to achieve quality results. Organizations do not authorize or provide proper accessibility to the necessary tools. Data preparation is difficult.
  • 40
    CA IDMS

    CA IDMS

    Broadcom Inc.

    Achieve superior, cost-effective database performance with maximum flexibility using CA's proven, Web-enabled, high-performance mainframe relational database management system. Rock-solid database with reliable, scalable, secure, high-performance and cost-effective database management. Let's you more easily participate in mobile-to-mainframe initiatives, cloud, web, and analytics using open capabilities. Highly scalable to handle ever-increasing business demands, designed for efficiency and to help reduce costs. IDMS users have the ability to quickly and easily access data in order to leverage solutions that analyze system issues, automate processes and alerts, and create intelligent recommendations for tuning. Next-gen users have the ability to execute and automate IDMS administration using common tools and scripting languages. Let’s take a journey in partnership to delve into your project needs.
  • 41
    Cubrid

    Cubrid

    Cubrid

    CUBRID is a relational DBMS optimized for online transaction processing (OLTP) that complies with ANSI SQL standards and provides MVCC support, High-Availability (HA) capabilities, and GUI-based tools for DB management/migration. It also provides Oracle/MySQL compatibility and supports a variety of interfaces, including JDBC. CUBRID provides ease of installation and native GUI-based administration tools for developers' convenience. Multi-threaded, multi-server architecture, native broker middleware, cost-based optimizer, and intensive caching techniques for your OLTP services. Very accurate predictable automatic fail-over built-in technology, based on the CUBRID Heartbeat native engine core. Multi-volume support, automatic volume expansion, and unlimited number and size of databases/ tables/indexes.
    Starting Price: $0.01/one-time/user
  • 42
    Astra Streaming
    Responsive applications keep users engaged and developers inspired. Rise to meet these ever-increasing expectations with the DataStax Astra Streaming service platform. DataStax Astra Streaming is a cloud-native messaging and event streaming platform powered by Apache Pulsar. Astra Streaming allows you to build streaming applications on top of an elastically scalable, multi-cloud messaging and event streaming platform. Astra Streaming is powered by Apache Pulsar, the next-generation event streaming platform which provides a unified solution for streaming, queuing, pub/sub, and stream processing. Astra Streaming is a natural complement to Astra DB. Using Astra Streaming, existing Astra DB users can easily build real-time data pipelines into and out of their Astra DB instances. With Astra Streaming, avoid vendor lock-in and deploy on any of the major public clouds (AWS, GCP, Azure) compatible with open-source Apache Pulsar.
  • 43
    IBM Event Streams
    Built on open-source Apache Kafka, IBM® Event Streams is an event-streaming platform that helps you build smart apps that can react to events as they happen. Event Streams is based on years of IBM operational expertise gained from running Apache Kafka event streams for enterprises. This makes Event Streams ideal for mission-critical workloads. With connectors to a wide range of core systems and a scalable REST API, you can extend the reach of your existing enterprise assets. Rich security and geo-replication aids disaster recovery. Take advantage of IBM productivity tools and use the CLI to ensure best practices. Replicate data between Event Streams deployments in a disaster-recovery situation.
  • 44
    TIBCO BusinessEvents
    In the emerging digital world, billions of people, systems, and devices will interact in real-time, suggesting new and disruptive competitive advantages. How do you play there? By building distributed, stateful, rule-based event-processing systems, experimenting, learning, and quickly evolving. Data is arriving in real-time from hundreds of internal and external sources—and it has shelf-life; its value diminishes over time. With TIBCO, you can jump-start big data processing initiatives that give you the ability to sense, reason, respond, and visualize—much different and much faster and smarter than the traditional store, analyze, report, act approach. With multiple rule authoring environments, you support collaboration between IT developers and business professionals.
  • 45
    Oracle Cloud Infrastructure Streaming
    Streaming service is a real-time, serverless, Apache Kafka-compatible event streaming platform for developers and data scientists. Streaming is tightly integrated with Oracle Cloud Infrastructure (OCI), Database, GoldenGate, and Integration Cloud. The service also provides out-of-the-box integrations for hundreds of third-party products across categories such as DevOps, databases, big data, and SaaS applications. Data engineers can easily set up and operate big data pipelines. Oracle handles all infrastructure and platform management for event streaming, including provisioning, scaling, and security patching. With the help of consumer groups, Streaming can provide state management for thousands of consumers. This helps developers easily build applications at scale.
  • 46
    Apache Kafka

    Apache Kafka

    The Apache Software Foundation

    Apache Kafka® is an open-source, distributed streaming platform. Scale production clusters up to a thousand brokers, trillions of messages per day, petabytes of data, hundreds of thousands of partitions. Elastically expand and contract storage and processing. Stretch clusters efficiently over availability zones or connect separate clusters across geographic regions. Process streams of events with joins, aggregations, filters, transformations, and more, using event-time and exactly-once processing. Kafka’s out-of-the-box Connect interface integrates with hundreds of event sources and event sinks including Postgres, JMS, Elasticsearch, AWS S3, and more. Read, write, and process streams of events in a vast array of programming languages.
  • 47
    DataStax

    DataStax

    DataStax

    The Open, Multi-Cloud Stack for Modern Data Apps. Built on open-source Apache Cassandra™. Global-scale and 100% uptime without vendor lock-in. Deploy on multi-cloud, on-prem, open-source, and Kubernetes. Elastic and pay-as-you-go for improved TCO. Start building faster with Stargate APIs for NoSQL, real-time, reactive, JSON, REST, and GraphQL. Skip the complexity of multiple OSS projects and APIs that don’t scale. Ideal for commerce, mobile, AI/ML, IoT, microservices, social, gaming, and richly interactive applications that must scale-up and scale-down with demand. Get building modern data applications with Astra, a database-as-a-service powered by Apache Cassandra™. Use REST, GraphQL, JSON with your favorite full-stack framework Richly interactive apps that are elastic and viral-ready from Day 1. Pay-as-you-go Apache Cassandra DBaaS that scales effortlessly and affordably.
  • 48
    StreamSets

    StreamSets

    StreamSets

    StreamSets DataOps Platform. The data integration platform to build, run, monitor and manage smart data pipelines that deliver continuous data for DataOps, and power modern analytics and hybrid integration. Only StreamSets provides a single design experience for all design patterns for 10x greater developer productivity; smart data pipelines that are resilient to change for 80% less breakages; and a single pane of glass for managing and monitoring all pipelines across hybrid and cloud architectures to eliminate blind spots and control gaps. With StreamSets, you can deliver the continuous data that drives the connected enterprise.
    Starting Price: $1000 per month
  • 49
    Cogility Cogynt

    Cogility Cogynt

    Cogility Software

    Deliver Continuous Intelligence solutions easier, faster, and cost-effectively - with less engineering effort. The Cogility Cogynt platform delivers cloud-scalable event stream processing software powered by advanced, Expert AI-based analytics. A complete, integrated toolset enables organizations to quickly, easily, and more efficiently deliver continuous intelligence solutions. The end-to-end platform streamlines deployment, constructing model logic, customizing data source intake, processing data streams, examining, visualizing and sharing intelligence findings, auditing and improving results, and integrating with other applications. Cogynt’s Authoring Tool provides a convenient, zero-code design environment for creating, updating, and deploying data models. Cogynt’s Data Management Tool makes it easy to publish your model to immediately apply to stream data processing while abstracting Flink job coding.
  • 50
    Arroyo

    Arroyo

    Arroyo

    Scale from zero to millions of events per second. Arroyo ships as a single, compact binary. Run locally on MacOS or Linux for development, and deploy to production with Docker or Kubernetes. Arroyo is a new kind of stream processing engine, built from the ground up to make real-time easier than batch. Arroyo was designed from the start so that anyone with SQL experience can build reliable, efficient, and correct streaming pipelines. Data scientists and engineers can build end-to-end real-time applications, models, and dashboards, without a separate team of streaming experts. Transform, filter, aggregate, and join data streams by writing SQL, with sub-second results. Your streaming pipelines shouldn't page someone just because Kubernetes decided to reschedule your pods. Arroyo is built to run in modern, elastic cloud environments, from simple container runtimes like Fargate to large, distributed deployments on the Kubernetes logo Kubernetes.