Alternatives to Axibase Time Series Database

Compare Axibase Time Series Database alternatives for your business or organization using the curated list below. SourceForge ranks the best alternatives to Axibase Time Series Database in 2026. Compare features, ratings, user reviews, pricing, and more from Axibase Time Series Database competitors and alternatives in order to make an informed decision for your business.

  • 1
    RaimaDB

    RaimaDB

    Raima

    RaimaDB is an embedded time series database for IoT and Edge devices that can run in-memory. It is an extremely powerful, lightweight and secure RDBMS. Field tested by over 20 000 developers worldwide and has more than 25 000 000 deployments. RaimaDB is a high-performance, cross-platform embedded database designed for mission-critical applications, particularly in the Internet of Things (IoT) and edge computing markets. It offers a small footprint, making it suitable for resource-constrained environments, and supports both in-memory and persistent storage configurations. RaimaDB provides developers with multiple data modeling options, including traditional relational models and direct relationships through network model sets. It ensures data integrity with ACID-compliant transactions and supports various indexing methods such as B+Tree, Hash Table, R-Tree, and AVL-Tree.
    Partner badge
    Compare vs. Axibase Time Series Database View Software
    Visit Website
  • 2
    StarTree

    StarTree

    StarTree

    StarTree, powered by Apache Pinot™, is a fully managed real-time analytics platform built for customer-facing applications that demand instant insights on the freshest data. Unlike traditional data warehouses or OLTP databases—optimized for back-office reporting or transactions—StarTree is engineered for real-time OLAP at true scale, meaning: - Data Volume: query performance sustained at petabyte scale - Ingest Rates: millions of events per second, continuously indexed for freshness - Concurrency: thousands to millions of simultaneous users served with sub-second latency With StarTree, businesses deliver always-fresh insights at interactive speed, enabling applications that personalize, monitor, and act in real time.
  • 3
    Snowflake

    Snowflake

    Snowflake

    Snowflake is a comprehensive AI Data Cloud platform designed to eliminate data silos and simplify data architectures, enabling organizations to get more value from their data. The platform offers interoperable storage that provides near-infinite scale and access to diverse data sources, both inside and outside Snowflake. Its elastic compute engine delivers high performance for any number of users, workloads, and data volumes with seamless scalability. Snowflake’s Cortex AI accelerates enterprise AI by providing secure access to leading large language models (LLMs) and data chat services. The platform’s cloud services automate complex resource management, ensuring reliability and cost efficiency. Trusted by over 11,000 global customers across industries, Snowflake helps businesses collaborate on data, build data applications, and maintain a competitive edge.
  • 4
    Quasar AI

    Quasar AI

    QuasarDB

    Quasar is a high-cardinality analytics infrastructure designed for handling large-scale numerical data. It is built to support modern AI systems that rely on telemetry, trades, sensors, and simulations. The platform replaces traditional data stacks with a single distributed system for improved performance. It eliminates latency caused by batch pipelines and multi-stage ETL processes. Quasar also reduces costs by avoiding repeated data scans and complex infrastructure layers. With deterministic query execution and numerical compression, it ensures fast and reliable analytics. Overall, Quasar provides predictable performance and stable costs for data-intensive environments.
  • 5
    Prometheus

    Prometheus

    Prometheus

    Power your metrics and alerting with a leading open-source monitoring solution. Prometheus fundamentally stores all data as time series: streams of timestamped values belonging to the same metric and the same set of labeled dimensions. Besides stored time series, Prometheus may generate temporary derived time series as the result of queries. Prometheus provides a functional query language called PromQL (Prometheus Query Language) that lets the user select and aggregate time series data in real time. The result of an expression can either be shown as a graph, viewed as tabular data in Prometheus's expression browser, or consumed by external systems via the HTTP API. Prometheus is configured via command-line flags and a configuration file. While the command-line flags configure immutable system parameters (such as storage locations, amount of data to keep on disk and in memory, etc.). Download: https://sourceforge.net/projects/prometheus.mirror/
  • 6
    Amazon Timestream
    Amazon Timestream is a fast, scalable, and serverless time series database service for IoT and operational applications that makes it easy to store and analyze trillions of events per day up to 1,000 times faster and at as little as 1/10th the cost of relational databases. Amazon Timestream saves you time and cost in managing the lifecycle of time series data by keeping recent data in memory and moving historical data to a cost optimized storage tier based upon user defined policies. Amazon Timestream’s purpose-built query engine lets you access and analyze recent and historical data together, without needing to specify explicitly in the query whether the data resides in the in-memory or cost-optimized tier. Amazon Timestream has built-in time series analytics functions, helping you identify trends and patterns in your data in near real-time.
  • 7
    QuestDB

    QuestDB

    QuestDB

    QuestDB is a relational column-oriented database designed for time series and event data. It uses SQL with extensions for time series to assist with real-time analytics. These pages cover core concepts of QuestDB, including setup steps, usage guides, and reference documentation for syntax, APIs and configuration. This section describes the architecture of QuestDB, how it stores and queries data, and introduces features and capabilities unique to the system. Designated timestamp is a core feature that enables time-oriented language capabilities and partitioning. Symbol type makes storing and retrieving repetitive strings efficient. Storage model describes how QuestDB stores records and partitions within tables. Indexes can be used for faster read access on specific columns. Partitions can be used for significant performance benefits on calculations and queries. SQL extensions allow performant time series analysis with a concise syntax.
  • 8
    JaguarDB

    JaguarDB

    JaguarDB

    JaguarDB enables fast ingestion of time series data, coupling location-based data. It also can index in both dimensions, space and time. Back-filling time series data is also fast (inserting large volumes of data in past time). Normally time series is a series of data points indexed in time order. In JaguarDB, the time series has a different meaning: it is both a sequence of data points and a series of tick tables holding aggregated data values at specified time spans. For example, a time series table in JaguarDB can have a base table storing data points in time order, and tick tables such as 5 minute, 15 minute, hourly, daily, weekly, monthly tables to store aggregated data within these time spans. The format for the RETENTION is the same as the TICK format, except that it can have any number of retention periods. The RETENTION specifies how long the data points in the base table should be kept.
  • 9
    Dremio

    Dremio

    Dremio

    Dremio delivers lightning-fast queries and a self-service semantic layer directly on your data lake storage. No moving data to proprietary data warehouses, no cubes, no aggregation tables or extracts. Just flexibility and control for data architects, and self-service for data consumers. Dremio technologies like Data Reflections, Columnar Cloud Cache (C3) and Predictive Pipelining work alongside Apache Arrow to make queries on your data lake storage very, very fast. An abstraction layer enables IT to apply security and business meaning, while enabling analysts and data scientists to explore data and derive new virtual datasets. Dremio’s semantic layer is an integrated, searchable catalog that indexes all of your metadata, so business users can easily make sense of your data. Virtual datasets and spaces make up the semantic layer, and are all indexed and searchable.
  • 10
    R2 SQL

    R2 SQL

    Cloudflare

    R2 SQL is Cloudflare’s serverless, distributed analytics query engine (currently in open beta) that enables you to run SQL queries over Apache Iceberg tables stored in R2 Data Catalog without needing to manage your own compute clusters. It is built to efficiently query large volumes of data by leveraging metadata pruning, partition-level statistics, file and row-group filtering, and Cloudflare’s globally distributed compute infrastructure to parallelize execution. The system works by integrating with R2 object storage and an Iceberg catalog layer, so you can ingest data via Cloudflare Pipelines into Iceberg tables, and then query that data with minimal overhead. Queries can be issued via the Wrangler CLI or HTTP API (with an API token granting permissions across R2 SQL, Data Catalog, and storage). During the open beta period, using R2 SQL itself is not billed, only storage and standard R2 operations incur charges.
  • 11
    VeloDB

    VeloDB

    VeloDB

    Powered by Apache Doris, VeloDB is a modern data warehouse for lightning-fast analytics on real-time data at scale. Push-based micro-batch and pull-based streaming data ingestion within seconds. Storage engine with real-time upsert、append and pre-aggregation. Unparalleled performance in both real-time data serving and interactive ad-hoc queries. Not just structured but also semi-structured data. Not just real-time analytics but also batch processing. Not just run queries against internal data but also work as a federate query engine to access external data lakes and databases. Distributed design to support linear scalability. Whether on-premise deployment or cloud service, separation or integration of storage and compute, resource usage can be flexibly and efficiently adjusted according to workload requirements. Built on and fully compatible with open source Apache Doris. Support MySQL protocol, functions, and SQL for easy integration with other data tools.
  • 12
    SSuite MonoBase Database

    SSuite MonoBase Database

    SSuite Office Software

    Create relational or flat file databases with unlimited tables, fields, and rows. Includes a custom report builder. Interface with ODBC compatible databases and create custom reports for them. Create your own personal and custom databases. Some Highlights: - Filter tables instantly - Ultra simple graphical-user-interface - One click table and data form creation - Open up to 5 databases simultaneously - Export your data to comma separated files - Create custom reports for all your databases - Full helpfile to assist in creating database reports - Print tables and queries directly from the data grid - Supports any SQL standard that your ODBC compatible database requires Please install and run this database application with full administrator rights for best performance and user experience. Requires: . 1024x768 Display Size . Windows 98 / XP / 7 / 8 / 10 - 32bit and 64bit No Java or DotNet required. Green Energy Software. Saving the planet one bit at a time...
  • 13
    Falcon LogScale

    Falcon LogScale

    CrowdStrike

    Rapidly shut down threats with real-time detection and blazing-fast search while reducing logging costs. Detect threats faster by processing incoming data in under a second. Find suspicious activity in a fraction of the time of traditional security logging tools. A powerful, index-free architecture lets you log all your data and retain it for years while avoiding ingestion bottlenecks. Collect more data for investigations, and threat hunting, and scale to over 1 PB of data ingestion per day with negligible performance impact. Falcon LogScale takes your searching, hunting, and troubleshooting capabilities to the next level with its powerful, intuitive query language. Dig deeper to gain additional context with filtering, aggregation, and regex support. Quickly scan all events with a free-text search. Live and historical dashboards let users instantly prioritize threats, monitor trends, and troubleshoot issues. Easily drill down from charts to search results.
  • 14
    ITTIA DB
    The ITTIA DB product family combines the best of time series, real-time data streaming, and analytics for embedded systems to reduce development time and costs. ITTIA DB IoT is a small-footprint embedded database for real-time resource-constrained 32-bit microcontrollers (MCUs), and ITTIA DB SQL is a high-performance time-series embedded database for single or multicore microprocessors (MPUs). Both ITTIA DB products enable devices to monitor, process, and store real-time data. ITTIA DB products are also built for the automotive industry Electronic Control Units (ECUs). ITTIA DB data security protocols offer data protection against malicious access with encryption, authentication, and DB SEAL. ITTIA SDL is conformant to the principles of IEC/ISO 62443. Embed ITTIA DB to collect, process, and enrich incoming real-time data streams in a purpose-built SDK for edge devices. Search, filter, join, and aggregate at the edge.
  • 15
    Trino

    Trino

    Trino

    Trino is a query engine that runs at ludicrous speed. Fast-distributed SQL query engine for big data analytics that helps you explore your data universe. Trino is a highly parallel and distributed query engine, that is built from the ground up for efficient, low-latency analytics. The largest organizations in the world use Trino to query exabyte-scale data lakes and massive data warehouses alike. Supports diverse use cases, ad-hoc analytics at interactive speeds, massive multi-hour batch queries, and high-volume apps that perform sub-second queries. Trino is an ANSI SQL-compliant query engine, that works with BI tools such as R, Tableau, Power BI, Superset, and many others. You can natively query data in Hadoop, S3, Cassandra, MySQL, and many others, without the need for complex, slow, and error-prone processes for copying the data. Access data from multiple systems within a single query.
  • 16
    LlamaIndex

    LlamaIndex

    LlamaIndex

    LlamaIndex is a “data framework” to help you build LLM apps. Connect semi-structured data from API's like Slack, Salesforce, Notion, etc. LlamaIndex is a simple, flexible data framework for connecting custom data sources to large language models. LlamaIndex provides the key tools to augment your LLM applications with data. Connect your existing data sources and data formats (API's, PDF's, documents, SQL, etc.) to use with a large language model application. Store and index your data for different use cases. Integrate with downstream vector store and database providers. LlamaIndex provides a query interface that accepts any input prompt over your data and returns a knowledge-augmented response. Connect unstructured sources such as documents, raw text files, PDF's, videos, images, etc. Easily integrate structured data sources from Excel, SQL, etc. Provides ways to structure your data (indices, graphs) so that this data can be easily used with LLMs.
  • 17
    Arroyo

    Arroyo

    Arroyo

    Scale from zero to millions of events per second. Arroyo ships as a single, compact binary. Run locally on MacOS or Linux for development, and deploy to production with Docker or Kubernetes. Arroyo is a new kind of stream processing engine, built from the ground up to make real-time easier than batch. Arroyo was designed from the start so that anyone with SQL experience can build reliable, efficient, and correct streaming pipelines. Data scientists and engineers can build end-to-end real-time applications, models, and dashboards, without a separate team of streaming experts. Transform, filter, aggregate, and join data streams by writing SQL, with sub-second results. Your streaming pipelines shouldn't page someone just because Kubernetes decided to reschedule your pods. Arroyo is built to run in modern, elastic cloud environments, from simple container runtimes like Fargate to large, distributed deployments on the Kubernetes logo Kubernetes.
  • 18
    Apache Spark

    Apache Spark

    Apache Software Foundation

    Apache Spark™ is a unified analytics engine for large-scale data processing. Apache Spark achieves high performance for both batch and streaming data, using a state-of-the-art DAG scheduler, a query optimizer, and a physical execution engine. Spark offers over 80 high-level operators that make it easy to build parallel apps. And you can use it interactively from the Scala, Python, R, and SQL shells. Spark powers a stack of libraries including SQL and DataFrames, MLlib for machine learning, GraphX, and Spark Streaming. You can combine these libraries seamlessly in the same application. Spark runs on Hadoop, Apache Mesos, Kubernetes, standalone, or in the cloud. It can access diverse data sources. You can run Spark using its standalone cluster mode, on EC2, on Hadoop YARN, on Mesos, or on Kubernetes. Access data in HDFS, Alluxio, Apache Cassandra, Apache HBase, Apache Hive, and hundreds of other data sources.
  • 19
    IBM Db2 Big SQL
    A hybrid SQL-on-Hadoop engine delivering advanced, security-rich data query across enterprise big data sources, including Hadoop, object storage and data warehouses. IBM Db2 Big SQL is an enterprise-grade, hybrid ANSI-compliant SQL-on-Hadoop engine, delivering massively parallel processing (MPP) and advanced data query. Db2 Big SQL offers a single database connection or query for disparate sources such as Hadoop HDFS and WebHDFS, RDMS, NoSQL databases, and object stores. Benefit from low latency, high performance, data security, SQL compatibility, and federation capabilities to do ad hoc and complex queries. Db2 Big SQL is now available in 2 variations. It can be integrated with Cloudera Data Platform, or accessed as a cloud-native service on the IBM Cloud Pak® for Data platform. Access and analyze data and perform queries on batch and real-time data across sources, like Hadoop, object stores and data warehouses.
  • 20
    Riemann

    Riemann

    Riemann

    Riemann aggregates events from your servers and applications with a powerful stream processing language. Send an email for every exception in your app. Track the latency distribution of your web app. See the top processes on any host, by memory and CPU. Combine statistics from every Riak node in your cluster and forward them to Graphite. Track user activity from second to second. Riemann provides a low-latency, transient shared state for systems with many moving parts. Riemann streams are just functions that accept an event. Since Riemann's configuration is a Clojure program, its syntax is concise, regular, and extendable. Configuration-as-code minimizes boilerplate and gives you the flexibility to adapt to complex situations. Riemann can tell you as much or as little as you want. Throttle or roll up multiple events into a single message. Get emails about exceptions in your code, provider downtime, or latency spikes. You can also integrate with PagerDuty for SMS or phone alerts.
  • 21
    Baidu Palo

    Baidu Palo

    Baidu AI Cloud

    Palo helps enterprises to create the PB-level MPP architecture data warehouse service within several minutes and import the massive data from RDS, BOS, and BMR. Thus, Palo can perform the multi-dimensional analytics of big data. Palo is compatible with mainstream BI tools. Data analysts can analyze and display the data visually and gain insights quickly to assist decision-making. It has the industry-leading MPP query engine, with column storage, intelligent index,and vector execution functions. It can also provide in-library analytics, window functions, and other advanced analytics functions. You can create a materialized view and change the table structure without the suspension of service. It supports flexible and efficient data recovery.
  • 22
    ksqlDB

    ksqlDB

    Confluent

    Now that your data is in motion, it’s time to make sense of it. Stream processing enables you to derive instant insights from your data streams, but setting up the infrastructure to support it can be complex. That’s why Confluent developed ksqlDB, the database purpose-built for stream processing applications. Make your data immediately actionable by continuously processing streams of data generated throughout your business. ksqlDB’s intuitive syntax lets you quickly access and augment data in Kafka, enabling development teams to seamlessly create real-time innovative customer experiences and fulfill data-driven operational needs. ksqlDB offers a single solution for collecting streams of data, enriching them, and serving queries on new derived streams and tables. That means less infrastructure to deploy, maintain, scale, and secure. With less moving parts in your data architecture, you can focus on what really matters -- innovation.
  • 23
    KairosDB

    KairosDB

    KairosDB

    Data can be pushed in KairosDB via multiple protocols like Telnet, Rest and Graphite. Other mechanisms such as plugins can also be used. KairosDB stores time series in Cassandra, the popular and performant NoSQL datastore. The schema consists of 3 column families. This API provides operations to list existing metric names, list tag names and values, store metric data points, and query for metric data points. With a default install, KairosDB serve up a query page whereby you can query data within the data store. It's designed primarily for development purposes. Aggregators perform an operation on data points and down samples. Standard functions like min, max, sum, count, mean and more are available. Import and export is available on the KairosDB server from the command line. Internal metrics to the data store can monitor the server’s performance.
  • 24
    OpenTSDB

    OpenTSDB

    OpenTSDB

    OpenTSDB consists of a Time Series Daemon (TSD) as well as set of command line utilities. Interaction with OpenTSDB is primarily achieved by running one or more of the independent TSDs. There is no master, no shared state so you can run as many TSDs as required to handle any load you throw at it. Each TSD uses the open source database HBase or hosted Google Bigtable service to store and retrieve time-series data. The data schema is highly optimized for fast aggregations of similar time series to minimize storage space. Users of the TSD never need to access the underlying store directly. You can communicate with the TSD via a simple telnet-style protocol, an HTTP API or a simple built-in GUI. The first step in using OpenTSDB is to send time series data to the TSDs. A number of tools exist to pull data from various sources into OpenTSDB.
  • 25
    Apache Hive

    Apache Hive

    Apache Software Foundation

    The Apache Hive data warehouse software facilitates reading, writing, and managing large datasets residing in distributed storage using SQL. Structure can be projected onto data already in storage. A command line tool and JDBC driver are provided to connect users to Hive. Apache Hive is an open source project run by volunteers at the Apache Software Foundation. Previously it was a subproject of Apache® Hadoop®, but has now graduated to become a top-level project of its own. We encourage you to learn about the project and contribute your expertise. Traditional SQL queries must be implemented in the MapReduce Java API to execute SQL applications and queries over distributed data. Hive provides the necessary SQL abstraction to integrate SQL-like queries (HiveQL) into the underlying Java without the need to implement queries in the low-level Java API.
  • 26
    Apache Druid
    Apache Druid is an open source distributed data store. Druid’s core design combines ideas from data warehouses, timeseries databases, and search systems to create a high performance real-time analytics database for a broad range of use cases. Druid merges key characteristics of each of the 3 systems into its ingestion layer, storage format, querying layer, and core architecture. Druid stores and compresses each column individually, and only needs to read the ones needed for a particular query, which supports fast scans, rankings, and groupBys. Druid creates inverted indexes for string values for fast search and filter. Out-of-the-box connectors for Apache Kafka, HDFS, AWS S3, stream processors, and more. Druid intelligently partitions data based on time and time-based queries are significantly faster than traditional databases. Scale up or down by just adding or removing servers, and Druid automatically rebalances. Fault-tolerant architecture routes around server failures.
  • 27
    Backtrace

    Backtrace

    Backtrace

    Don’t let app, device, or game crashes get in the way of a great experience. Backtrace takes all the manual labor out of cross-platform crash and exception management so you can focus on shipping. Cross-platform callstack and event aggregation and monitoring. Process errors from panics, core dumps, minidumps, and during runtime across your stack with a single system. Backtrace generates structured, searchable error reports from your data. Automated analysis cuts down on time to resolution by surfacing important signals that lead engineers to crash root cause. Never worry about missing a clue with rich integrations into dashboards, notification, and workflow systems. Answer the questions that matter to you with Backtrace’s rich query engine. View a high-level overview of error frequency, prioritization, and trends across all your projects. Search through key data points and your own custom data across all your errors.
  • 28
    Apache Impala
    Impala provides low latency and high concurrency for BI/analytic queries on the Hadoop ecosystem, including Iceberg, open data formats, and most cloud storage options. Impala also scales linearly, even in multitenant environments. Impala is integrated with native Hadoop security and Kerberos for authentication, and via the Ranger module, you can ensure that the right users and applications are authorized for the right data. Utilize the same file and data formats and metadata, security, and resource management frameworks as your Hadoop deployment, with no redundant infrastructure or data conversion/duplication. For Apache Hive users, Impala utilizes the same metadata and ODBC driver. Like Hive, Impala supports SQL, so you don't have to worry about reinventing the implementation wheel. With Impala, more users, whether using SQL queries or BI applications, can interact with more data through a single repository and metadata stored from source through analysis.
  • 29
    dbKoda

    dbKoda

    dbKoda

    dbKoda’s rich code editor allows you to compose MongoDB queries and scripts with syntax highlighting, auto-complete and code formatting. View output in foldable JSON, tabular format or as a chart. Generate visual explain plans with a single click. The explain plan viewer will suggest indexes that may improve performance, and provide one-button creation of those indexes. Build simple queries or complex aggregates using the graphical query builder. dbKoda lets you perform administrative activities using fil-in-the-blanks actions from the database tree. Manage configuration, user roles, logging, database storage, and data import/export without having to lookup obscure MongoDB syntax. We construct the appropriate commands as you fill-in-the-blanks, so you can re-use the command in admin scripts or tweak the syntax as desired. dbKoda includes an SSH terminal that allows command line access to the server OS, so you can issue OS commands directly from within the dbKoda product.
  • 30
    TimescaleDB

    TimescaleDB

    Tiger Data

    TimescaleDB is the leading time-series database built on PostgreSQL, designed to handle massive volumes of real-time data efficiently. It enables organizations to store, analyze, and query time-series data — such as IoT sensor data, financial transactions, or event logs — using standard SQL. With hypertables, TimescaleDB automatically partitions data by time and ID for fast ingestion and predictable query performance. Its compression engine reduces storage costs by up to 95%, while continuous aggregates make real-time dashboards instantly responsive. Fully compatible with PostgreSQL, it integrates seamlessly with existing tools and applications. TimescaleDB combines the simplicity of Postgres with the scalability and speed of a specialized analytical system.
  • 31
    ArcadeDB

    ArcadeDB

    ArcadeDB

    ArcadeDB is an open-source, next-generation multi-model database. Forget Polyglot Persistence — store graphs, documents, key-value pairs, search engine indexes, vectors, and time-series data all in one database with native support for every model. No translation layers, no performance penalties. Process over 10 million records per second. Traversal speed stays constant whether your database has hundreds or billions of records. Query in the language you prefer: SQL, Cypher, Gremlin, GraphQL, MongoDB API, or Java. Deploy ArcadeDB embedded in your JVM application, on a standalone server, or distributed across multiple nodes with Raft Consensus for high availability. Fully ACID-compliant. Super lightweight. Apache 2.0 licensed — free for production and commercial use.
  • 32
    Motif Analytics

    Motif Analytics

    Motif Analytics

    Rich interactive visualizations for identifying patterns in user and business flows, with full visibility into underlying computation. A small set of sequence operations providing full expressivity and fine-grained control in under 10 lines of code. An incremental query engine to seamlessly trade between query precision, speed and cost according to your needs. Currently Motif uses a tiny custom-built DSL called Sequence Operations Language (SOL), which we believe is more natural to use than SQL and more powerful than a drag-and-drop interface. We built a custom engine to optimize sequence queries and are also trading off precision, which goes unused in decision-making, for query speed.
  • 33
    Amazon FinSpace
    Amazon FinSpace simplifies running kdb Insights applications on AWS. Amazon FinSpace automates the undifferentiated tasks required to provision, integrate, and secure infrastructure for kdb Insights. In addition, Amazon FinSpace provides easy-to-use APIs so customers can configure and run new kdb Insights applications in just a few minutes. Amazon FinSpace gives customers the flexibility required to move existing kdb Insights applications to AWS and get the benefits of the cloud while eliminating the complex and costly work of self-managing the infrastructure. KX's kdb Insights is a high-performance analytics engine that is optimized for the analysis of real-time and multi-petabyte historical time-series data. Kdb Insights is commonly used by Capital Markets customers to power business-critical workloads, such as options pricing, transaction cost analysis, and backtesting. Eliminate the work to integrate more than 15 AWS services to deploy kdb.
  • 34
    Blueflood

    Blueflood

    Blueflood

    Blueflood is a high throughput, low latency, multi-tenant distributed metric processing system behind Rackspace Metrics, which is currently used in production by the Rackspace Monitoring team and Rackspace public cloud team to store metrics generated by their systems. In addition to Rackspace metrics, other large scale deployments of Blueflood can be found at community Wiki. Data from Blueflood can be used to construct dashboards, generate reports, graphs or for any other use involving time-series data. It focuses on near-realtime data, with data that is queryable mere milliseconds after ingestion. You send metrics to the ingestion service. You query your metrics from the Query service. And in the background, rollups are batch-processed offline so that queries for large time-periods are returned quickly.
  • 35
    Warp 10
    Warp 10 is a modular open source platform that collects, stores, and analyzes data from sensors. Shaped for the IoT with a flexible data model, Warp 10 provides a unique and powerful framework to simplify your processes from data collection to analysis and visualization, with the support of geolocated data in its core model (called Geo Time Series). Warp 10 is both a time series database and a powerful analytics environment, allowing you to make: statistics, extraction of characteristics for training models, filtering and cleaning of data, detection of patterns and anomalies, synchronization or even forecasts. The analysis environment can be implemented within a large ecosystem of software components such as Spark, Kafka Streams, Hadoop, Jupyter, Zeppelin and many more. It can also access data stored in many existing solutions, relational or NoSQL databases, search engines and S3 type object storage system.
  • 36
    Altinity

    Altinity

    Altinity

    Altinity's expert engineering team can implement everything from core ClickHouse features to Kubernetes operator behavior to client library improvements. A flexible docker-based GUI manager for ClickHouse that can do the following: Install ClickHouse clusters; Add, delete, and replace nodes; Monitor cluster status; Help with troubleshooting and diagnostics. 3rd party tools and software integrations: Ingest: Kafka, ClickTail; APIs: Python, Golang, ODBC, Java; Kubernetes; UI tools: Grafana, Superset, Tabix, Graphite; Databases: MySQL, PostgreSQL; BI tools: Tableau and many more. Altinity.Cloud incorporates lessons from helping hundreds of customers operate ClickHouse-based analytics. Altinity.Cloud has a Kubernetes-based architecture that delivers portability and user choice of where to operate. Designed from the beginning to run anywhere without lock-in. Cost management is critical for SaaS businesses.
  • 37
    KX Streaming Analytics
    KX Streaming Analytics provides the ability to ingest, store, process, and analyze historic and time series data to make analytics, insights, and visualizations instantly available. To help ensure your applications and users are productive quickly, the platform provides the full lifecycle of data services, including query processing, tiering, migration, archiving, data protection, and scaling. Our advanced analytics and visualization tools, used widely across finance and industry, enable you to define and perform queries, calculations, aggregations, machine learning and AI on any streaming and historical data. Deployable across multiple hardware environments, data can come from real-time business events and high-volume sources including sensors, clickstreams, radio-frequency identification, GPS systems, social networking sites, and mobile devices.
  • 38
    Alibaba Cloud TSDB
    Time Series Database (TSDB) supports high-speed data reading and writing. It offers high compression ratios for cost-efficient data storage. This service also supports visualization of precision reduction, interpolation, multi-metric aggregate computing, and query results. The TSDB service reduces storage costs and improves the efficiency of data writing, query, and analysis. This enables you to handle large amounts of data points and collect data more frequently. This service has been widely applied to systems in different industries, such as IoT monitoring systems, enterprise energy management systems (EMSs), production security monitoring systems, and power supply monitoring systems. Optimizes database architectures and algorithms. TSDB can read or write millions of data points within seconds. Applies an efficient compression algorithm to reduce the size of each data point to 2 bytes, saving more than 90% in storage costs.
  • 39
    Presto

    Presto

    Presto Foundation

    Presto is an open source distributed SQL query engine for running interactive analytic queries against data sources of all sizes ranging from gigabytes to petabytes. For data engineers who struggle with managing multiple query languages and interfaces to siloed databases and storage, Presto is the fast and reliable engine that provides one simple ANSI SQL interface for all your data analytics and your open lakehouse. Different engines for different workloads means you will have to re-platform down the road. With Presto, you get 1 familar ANSI SQL language and 1 engine for your data analytics so you don't need to graduate to another lakehouse engine. Presto can be used for interactive and batch workloads, small and large amounts of data, and scales from a few to thousands of users. Presto gives you one simple ANSI SQL interface for all of your data in various siloed data systems, helping you join your data ecosystem together.
  • 40
    RazorSQL

    RazorSQL

    RazorSQL

    RazorSQL is an SQL query tool, database browser, SQL editor, and database administration tool for Windows, macOS, Mac OS X, Linux, and Solaris. RazorSQL has been tested on over 40 databases, can connect to databases via either JDBC or ODBC. Browse database objects such as schemas, tables, columns, primary and foreign keys, views, indexes, procedures, functions, and more. Visual tools to create, alter, describe, execute, and drop database objects such as tables, views, indexes, stored procedures, functions, triggers, and more. Includes multi-tabular display of queries with options for filtering, sorting, searching, and much more. Import data from various formats such as delimited files, Excel spreadsheets, and fixed-width files. Includes a robust relational database (HSQLDB) that is up and running with no manual configuration out of the box.
    Starting Price: $99.95 one-time payment
  • 41
    Humio

    Humio

    Humio

    Log everything, answer anything in real-time. Modern log management with streaming observability and affordable Unlimited Plans. Humio is built to ingest and retain streaming data as quickly as it arrives, regardless of volume. Alerts, scripts, and dashboards are updated in real-time, and live tail and retained data searches have virtually no latency. Humio is index-free, and it works with any structured or unstructured data format. Because you don’t need to define fields upfront, you can ask any question with live or archived data, and experience blazing-fast response. Humio offers affordable licenses and industry-leading Unlimited Plans. Its advanced compression and use of bucket storage saves up to 70% on compute and storage costs. And Humio deploys in minutes and requires little maintenance. Humio ingests unlimited data at any throughput to provide the full data set needed to detect and respond to any incident.
  • 42
    Databricks

    Databricks

    Databricks

    The Databricks Data Intelligence Platform allows your entire organization to use data and AI. It’s built on a lakehouse to provide an open, unified foundation for all data and governance, and is powered by a Data Intelligence Engine that understands the uniqueness of your data. The winners in every industry will be data and AI companies. From ETL to data warehousing to generative AI, Databricks helps you simplify and accelerate your data and AI goals. Databricks combines generative AI with the unification benefits of a lakehouse to power a Data Intelligence Engine that understands the unique semantics of your data. This allows the Databricks Platform to automatically optimize performance and manage infrastructure in ways unique to your business. The Data Intelligence Engine understands your organization’s language, so search and discovery of new data is as easy as asking a question like you would to a coworker.
  • 43
    RTView

    RTView

    SL Corporation

    See application health state as a reflection of the entire application environment from physical infrastructure thru middleware to the end user experience. Consolidate health metrics across technologies. Proactively monitor stress for early warning. Correlate performance & application health. Share information with other teams. Still using the management console for each product to monitor your middleware platforms? It doesn’t have to be so complicated. See all your middleware technologies in one consolidated interface. Collect data without performance overhead. Correlate performance with hosts, networks, databases & app servers. Start small. Expand as needed. Monitor your applications and the technologies they run on in real-time using our packaged solutions. Build your own custom real-time monitoring system using this high-performance IDE.
  • 44
    Tabular

    Tabular

    Tabular

    Tabular is an open table store from the creators of Apache Iceberg. Connect multiple computing engines and frameworks. Decrease query time and storage costs by up to 50%. Centralize enforcement of data access (RBAC) policies. Connect any query engine or framework, including Athena, BigQuery, Redshift, Snowflake, Databricks, Trino, Spark, and Python. Smart compaction, clustering, and other automated data services reduce storage costs and query times by up to 50%. Unify data access at the database or table. RBAC controls are simple to manage, consistently enforced, and easy to audit. Centralize your security down to the table. Tabular is easy to use plus it features high-powered ingestion, performance, and RBAC under the hood. Tabular gives you the flexibility to work with multiple “best of breed” compute engines based on their strengths. Assign privileges at the data warehouse database, table, or column level.
  • 45
    ThinkSQL

    ThinkSQL

    ThinkSQL

    ThinkSQL is a powerful, cross-platform relational database management system designed for modern operating systems and hardware. It aims to comply fully with the ISO SQL standard to provide an open and flexible SQL server for your applications. Compliance with the SQL standard means it supports transactions, sub-selects, views, set operators (union, except, intersect), access privileges and comprehensive referential constraints to ensure the integrity of your data. ThinkSQL uses a multi-versioning engine to provide consistent views of data and concurrent access to rows without locking. The server also has stored procedures and functions (using the standard syntax), and has support for large multimedia objects. It's easy to install, runs on Windows ® and Linux ®, and has native ODBC, JDBC™, dbExpress™ and Python drivers so your applications can connect directly from any platform.
  • 46
    Hydra

    Hydra

    Hydra

    Hydra is an open source, column-oriented Postgres. Query billions of rows instantly, no code changes. Hydra parallelizes and vectorizes aggregates (COUNT, SUM, AVG) to deliver the speed you’ve always wanted on Postgres. Boost performance at every size! Set up Hydra in 5 minutes without changing your syntax, tools, data model, or extensions. Use Hydra Cloud for fully managed operations and smooth sailing. Different industries have different needs. Get better analytics with powerful Postgres extensions, custom functions, and take control. Built by you, for you. Hydra is the fastest Postgres in the market for analytics. Boost performance with columnar storage, vectorization, and query parallelization.
  • 47
    Polars

    Polars

    Polars

    Knowing of data wrangling habits, Polars exposes a complete Python API, including the full set of features to manipulate DataFrames using an expression language that will empower you to create readable and performant code. Polars is written in Rust, uncompromising in its choices to provide a feature-complete DataFrame API to the Rust ecosystem. Use it as a DataFrame library or as a query engine backend for your data models.
  • 48
    TheVWAP

    TheVWAP

    TheVWAP

    TheVWAP is an educational platform dedicated to enhancing traders' understanding and application of the Volume Weighted Average Price (VWAP) indicator, a tool often overlooked by retail traders but extensively utilized by institutional investors. The platform offers a comprehensive suite of resources, including an on-demand course comprising 16 hours of instruction that delves into various aspects of VWAP strategy. Additionally, TheVWAP provides custom indicators compatible with multiple charting platforms, facilitating the practical application of VWAP concepts. A vibrant community chatroom is available for discussing potential setups and broader market analysis, complemented by daily live streams where moderators execute trades in real time using VWAP strategies. The platform also features a VWAP chart gallery with over 500 examples showcasing daily setups and offers live chat support to address any user inquiries.
  • 49
    PySpark

    PySpark

    PySpark

    PySpark is an interface for Apache Spark in Python. It not only allows you to write Spark applications using Python APIs, but also provides the PySpark shell for interactively analyzing your data in a distributed environment. PySpark supports most of Spark’s features such as Spark SQL, DataFrame, Streaming, MLlib (Machine Learning) and Spark Core. Spark SQL is a Spark module for structured data processing. It provides a programming abstraction called DataFrame and can also act as distributed SQL query engine. Running on top of Spark, the streaming feature in Apache Spark enables powerful interactive and analytical applications across both streaming and historical data, while inheriting Spark’s ease of use and fault tolerance characteristics.
  • 50
    Qubole

    Qubole

    Qubole

    Qubole is a simple, open, and secure Data Lake Platform for machine learning, streaming, and ad-hoc analytics. Our platform provides end-to-end services that reduce the time and effort required to run Data pipelines, Streaming Analytics, and Machine Learning workloads on any cloud. No other platform offers the openness and data workload flexibility of Qubole while lowering cloud data lake costs by over 50 percent. Qubole delivers faster access to petabytes of secure, reliable and trusted datasets of structured and unstructured data for Analytics and Machine Learning. Users conduct ETL, analytics, and AI/ML workloads efficiently in end-to-end fashion across best-of-breed open source engines, multiple formats, libraries, and languages adapted to data volume, variety, SLAs and organizational policies.