Alternatives to Xeotek
Compare Xeotek alternatives for your business or organization using the curated list below. SourceForge ranks the best alternatives to Xeotek in 2024. Compare features, ratings, user reviews, pricing, and more from Xeotek competitors and alternatives in order to make an informed decision for your business.
-
1
StarTree
StarTree
StarTree Cloud is a fully-managed real-time analytics platform designed for OLAP at massive speed and scale for user-facing applications. Powered by Apache Pinot, StarTree Cloud provides enterprise-grade reliability and advanced capabilities such as tiered storage, scalable upserts, plus additional indexes and connectors. It integrates seamlessly with transactional databases and event streaming platforms, ingesting data at millions of events per second and indexing it for lightning-fast query responses. StarTree Cloud is available on your favorite public cloud or for private SaaS deployment. • Gain critical real-time insights to run your business • Seamlessly integrate data streaming and batch data • High performance in throughput and low-latency at petabyte scale • Fully-managed cloud service • Tiered storage to optimize cloud performance & spend • Fully-secure & enterprise-ready -
2
TreasuryPay
TreasuryPay
Instant™ Enterprise Data and Intelligence. Visibility into all transaction data, as it is happening, wherever it is happening worldwide. With just one network connection, organizations receive worldwide Accounting, Liquidity Management, FX, Marketing, and Supply Chain information — delivered in a single managed solution to empower enterprise intelligence. The TreasuryPay product set streams your global receivables information, delivering instant accountancy and cognitive services. It is, quite simply, the most advanced intelligence and insights platform currently available to global organizations. Instantly provide your organization with enriched information for your entire global enterprise. The change is easy. The Return on Investment, remarkable. Actionable intelligence and real-time global accountancy are now available at your fingertips with TreasuryPay Instant™. -
3
Streaming service is a real-time, serverless, Apache Kafka-compatible event streaming platform for developers and data scientists. Streaming is tightly integrated with Oracle Cloud Infrastructure (OCI), Database, GoldenGate, and Integration Cloud. The service also provides out-of-the-box integrations for hundreds of third-party products across categories such as DevOps, databases, big data, and SaaS applications. Data engineers can easily set up and operate big data pipelines. Oracle handles all infrastructure and platform management for event streaming, including provisioning, scaling, and security patching. With the help of consumer groups, Streaming can provide state management for thousands of consumers. This helps developers easily build applications at scale.
-
4
Materialize
Materialize
Materialize is a reactive database that delivers incremental view updates. We help developers easily build with streaming data using standard SQL. Materialize can connect to many different external sources of data without pre-processing. Connect directly to streaming sources like Kafka, Postgres databases, CDC, or historical sources of data like files or S3. Materialize allows you to query, join, and transform data sources in standard SQL - and presents the results as incrementally-updated Materialized views. Queries are maintained and continually updated as new data streams in. With incrementally-updated views, developers can easily build data visualizations or real-time applications. Building with streaming data can be as simple as writing a few lines of SQL.Starting Price: $0.98 per hour -
5
Rockset
Rockset
Real-Time Analytics on Raw Data. Live ingest from S3, Kafka, DynamoDB & more. Explore raw data as SQL tables. Build amazing data-driven applications & live dashboards in minutes. Rockset is a serverless search and analytics engine that powers real-time apps and live dashboards. Operate directly on raw data, including JSON, XML, CSV, Parquet, XLSX or PDF. Plug data from real-time streams, data lakes, databases, and data warehouses into Rockset. Ingest real-time data without building pipelines. Rockset continuously syncs new data as it lands in your data sources without the need for a fixed schema. Use familiar SQL, including joins, filters, and aggregations. It’s blazing fast, as Rockset automatically indexes all fields in your data. Serve fast queries that power the apps, microservices, live dashboards, and data science notebooks you build. Scale without worrying about servers, shards, or pagers.Starting Price: Free -
6
Axual
Axual
Axual is Kafka-as-a-Service for DevOps teams. Empower your team to unlock insights and drive decisions with our intuitive Kafka platform. Axual offers the ultimate solution for enterprises looking to seamlessly integrate data streaming into their core IT infrastructure. Our all-in-one Kafka platform is designed to eliminate the need for extensive technical knowledge or skills, and provides a ready-made solution that delivers all the benefits of event streaming without the hassle. The Axual Platform is a all-in-one solution, designed to help you simplify and enhance the deployment, management, and utilization of real-time data streaming with Apache Kafka. By providing an array of features that cater to the diverse needs of modern enterprises, the Axual Platform enables organizations to harness the full potential of data streaming while minimizing complexity and operational overhead. -
7
Redpanda
Redpanda Data
Breakthrough data streaming capabilities that let you deliver customer experiences never before possible. Kafka API and ecosystem are compatible. Redpanda BulletPredictable low latencies with zero data loss. Redpanda BulletUpto 10x faster than Kafka. Redpanda BulletEnterprise-grade support and hotfixes. Redpanda BulletAutomated backups to S3/GCS. Redpanda Bullet100% freedom from routine Kafka operations. Redpanda BulletSupport for AWS and GCP. Redpanda was designed from the ground up to be easily installed to get streaming up and running quickly. After you see its power, put Redpanda to the test in production. Use the more advanced Redpanda features. We manage provisioning, monitoring, and upgrades. Without any access to your cloud credentials. Sensitive data never leaves your environment. Provisioned, operated, and maintained for you. Configurable instance types. Expand cluster as your needs grow. -
8
Azure Event Hubs
Microsoft
Event Hubs is a fully managed, real-time data ingestion service that’s simple, trusted, and scalable. Stream millions of events per second from any source to build dynamic data pipelines and immediately respond to business challenges. Keep processing data during emergencies using the geo-disaster recovery and geo-replication features. Integrate seamlessly with other Azure services to unlock valuable insights. Allow existing Apache Kafka clients and applications to talk to Event Hubs without any code changes—you get a managed Kafka experience without having to manage your own clusters. Experience real-time data ingestion and microbatching on the same stream. Focus on drawing insights from your data instead of managing infrastructure. Build real-time big data pipelines and respond to business challenges right away.Starting Price: $0.03 per hour -
9
Azure Data Explorer
Microsoft
Azure Data Explorer is a fast, fully managed data analytics service for real-time analysis on large volumes of data streaming from applications, websites, IoT devices, and more. Ask questions and iteratively explore data on the fly to improve products, enhance customer experiences, monitor devices, and boost operations. Quickly identify patterns, anomalies, and trends in your data. Explore new questions and get answers in minutes. Run as many queries as you need, thanks to the optimized cost structure. Explore new possibilities with your data cost-effectively. Focus on insights, not infrastructure, with the easy-to-use, fully managed data analytics service. Respond quickly to fast-flowing and rapidly changing data. Azure Data Explorer simplifies analytics from all forms of streaming data.Starting Price: $0.11 per hour -
10
Esper Enterprise Edition
EsperTech Inc.
Esper Enterprise Edition is a distributable platform for linear and elastic horizontal scalability and fault-tolerant event processing. EPL editor and debugger; Hot deployment; Detailed metric and memory use reporting with break-down and summary per EPL. Data Push for multi-tier CEP-to-Browser delivery; Management of Logical and Physical Subscribers and Subscriptions. Web-based user interface for managing all aspects of multiple distributed engine instances with JavaScript and HTML 5. Composable, configurable and interactive displays of distributed event streams or series; Charts, Gauges, Timelines, Grids. JDBC-compliant client and server endpoints for interoperability. Esper Enterprise Edition is a closed-source commercial product by EsperTech. The source code is made available to support customers only. Esper Enterprise Edition is a distributable platform for linear and elastic horizontal scalability and fault-tolerant event processing. -
11
KX Streaming Analytics provides the ability to ingest, store, process, and analyze historic and time series data to make analytics, insights, and visualizations instantly available. To help ensure your applications and users are productive quickly, the platform provides the full lifecycle of data services, including query processing, tiering, migration, archiving, data protection, and scaling. Our advanced analytics and visualization tools, used widely across finance and industry, enable you to define and perform queries, calculations, aggregations, machine learning and AI on any streaming and historical data. Deployable across multiple hardware environments, data can come from real-time business events and high-volume sources including sensors, clickstreams, radio-frequency identification, GPS systems, social networking sites, and mobile devices.
-
12
Kapacitor
InfluxData
Kapacitor is a native data processing engine for InfluxDB 1.x and is an integrated component in the InfluxDB 2.0 platform. Kapacitor can process both stream and batch data from InfluxDB, acting on this data in real-time via its programming language TICKscript. Today’s modern applications require more than just dashboarding and operator alerts—they need the ability to trigger actions. Kapacitor’s alerting system follows a publish-subscribe design pattern. Alerts are published to topics and handlers subscribe to a topic. This pub/sub model and the ability for these to call User Defined Functions make Kapacitor very flexible to act as the control plane in your environment, performing tasks like auto-scaling, stock reordering, and IoT device control. Kapacitor provides a simple plugin architecture, or interface, that allows it to integrate with any anomaly detection engine.Starting Price: $0.002 per GB per hour -
13
GigaSpaces
GigaSpaces
Smart DIH is an operational data hub that powers real-time modern applications. It unleashes the power of customers’ data by transforming data silos into assets, turning organizations into data-driven enterprises. Smart DIH consolidates data from multiple heterogeneous systems into a highly performant data layer. Low code tools empower data professionals to deliver data microservices in hours, shortening developing cycles and ensuring data consistency across all digital channels. XAP Skyline is a cloud-native, in memory data grid (IMDG) and developer framework designed for mission critical, cloud-native apps. XAP Skyline delivers maximal throughput, microsecond latency and scale, while maintaining transactional consistency. It provides extreme performance, significantly reducing data access time, which is crucial for real-time decisioning, and transactional applications. XAP Skyline is used in financial services, retail, and other industries where speed and scalability are critical. -
14
Gretel
Gretel.ai
Privacy engineering tools delivered to you as APIs. Synthesize and transform data in minutes. Build trust with your users and community. Gretel’s APIs grant immediate access to creating anonymized or synthetic datasets so you can work safely with data while preserving privacy. Keeping the pace with development velocity requires faster access to data. Gretel is accelerating access to data with data privacy tools that bypass blockers and fuel Machine Learning and AI applications. Keep your data contained by running Gretel containers in your own environment or scale out workloads to the cloud in seconds with Gretel Cloud runners. Using our cloud GPUs makes it radically more effortless for developers to train and generate synthetic data. Scale workloads automatically with no infrastructure to set up and manage. Invite team members to collaborate on cloud projects and share data across teams. -
15
DeltaStream
DeltaStream
DeltaStream is a unified serverless stream processing platform that integrates with streaming storage services. Think about it as the compute layer on top of your streaming storage. It provides functionalities of streaming analytics(Stream processing) and streaming databases along with additional features to provide a complete platform to manage, process, secure and share streaming data. DeltaStream provides a SQL based interface where you can easily create stream processing applications such as streaming pipelines, materialized views, microservices and many more. It has a pluggable processing engine and currently uses Apache Flink as its primary stream processing engine. DeltaStream is more than just a query processing layer on top of Kafka or Kinesis. It brings relational database concepts to the data streaming world, including namespacing and role based access control enabling you to securely access, process and share your streaming data regardless of where they are stored. -
16
Amazon MSK
Amazon
Amazon MSK is a fully managed service that makes it easy for you to build and run applications that use Apache Kafka to process streaming data. Apache Kafka is an open-source platform for building real-time streaming data pipelines and applications. With Amazon MSK, you can use native Apache Kafka APIs to populate data lakes, stream changes to and from databases, and power machine learning and analytics applications. Apache Kafka clusters are challenging to setup, scale, and manage in production. When you run Apache Kafka on your own, you need to provision servers, configure Apache Kafka manually, replace servers when they fail, orchestrate server patches and upgrades, architect the cluster for high availability, ensure data is durably stored and secured, setup monitoring and alarms, and carefully plan scaling events to support load changes.Starting Price: $0.0543 per hour -
17
WarpStream
WarpStream
WarpStream is an Apache Kafka-compatible data streaming platform built directly on top of object storage, with no inter-AZ networking costs, no disks to manage, and infinitely scalable, all within your VPC. WarpStream is deployed as a stateless and auto-scaling agent binary in your VPC with no local disks to manage. Agents stream data directly to and from object storage with no buffering on local disks and no data tiering. Create new “virtual clusters” in our control plane instantly. Support different environments, teams, or projects without managing any dedicated infrastructure. WarpStream is protocol compatible with Apache Kafka, so you can keep using all your favorite tools and software. No need to rewrite your application or use a proprietary SDK. Just change the URL in your favorite Kafka client library and start streaming. Never again have to choose between reliability and your budget.Starting Price: $2,987 per month -
18
Digital Twin Streaming Service
ScaleOut Software
ScaleOut Digital Twin Streaming Service™ Easily build and deploy real-time digital twins for streaming analytics Connect to many data sources with Azure & AWS IoT hubs, Kafka, and more Maximize situational awareness with live, aggregate analytics. Introducing a breakthrough cloud service that simultaneously tracks telemetry from millions of data sources with “real-time” digital twins — enabling immediate, deep introspection with state-tracking and highly targeted, real-time feedback for thousands of devices. A powerful UI simplifies deployment and displays aggregate analytics in real time to maximize situational awareness. Ideal for a wide range of applications, including the Internet of Things (IoT), real-time intelligent monitoring, logistics, and financial services. Simplified pricing makes getting started fast and easy. Combined with the ScaleOut Digital Twin Builder software toolkit, the ScaleOut Digital Twin Streaming Service enables the next generation in stream processing. -
19
Confluent
Confluent
Infinite retention for Apache Kafka® with Confluent. Be infrastructure-enabled, not infrastructure-restricted Legacy technologies require you to choose between being real-time or highly-scalable. Event streaming enables you to innovate and win - by being both real-time and highly-scalable. Ever wonder how your rideshare app analyzes massive amounts of data from multiple sources to calculate real-time ETA? Ever wonder how your credit card company analyzes millions of credit card transactions across the globe and sends fraud notifications in real-time? The answer is event streaming. Move to microservices. Enable your hybrid strategy through a persistent bridge to cloud. Break down silos to demonstrate compliance. Gain real-time, persistent event transport. The list is endless. -
20
TIBCO Streaming
TIBCO
Analyze, continuously query, and act on IoT and other streaming data at lightning fast speeds. Take real-time operations and analytics to the next level with intelligent applications that deploy quickly for taking action based on new decisions and models, all without extra overhead. TIBCO® Streaming software is enterprise-grade, cloud-ready streaming analytics for quickly building real-time applications at a fraction of the cost and risk of alternatives. -
21
Apache Flink
Apache Software Foundation
Apache Flink is a framework and distributed processing engine for stateful computations over unbounded and bounded data streams. Flink has been designed to run in all common cluster environments, perform computations at in-memory speed and at any scale. Any kind of data is produced as a stream of events. Credit card transactions, sensor measurements, machine logs, or user interactions on a website or mobile application, all of these data are generated as a stream. Apache Flink excels at processing unbounded and bounded data sets. Precise control of time and state enable Flink’s runtime to run any kind of application on unbounded streams. Bounded streams are internally processed by algorithms and data structures that are specifically designed for fixed sized data sets, yielding excellent performance. Flink is designed to work well each of the previously listed resource managers. -
22
Oracle Stream Analytics
Oracle
Oracle Stream Analytics allows users to process and analyze large scale real-time information by using sophisticated correlation patterns, enrichment, and machine learning. It offers real-time actionable business insight on streaming data and automates action to drive today’s agile businesses. Visual GEOProcessing with GEOFence relationship spatial analytics. New Expressive Patterns Library, including Spatial, Statistical, General industry and Anomaly detection, streaming machine learning. Abstracted visual façade to interrogate live real time streaming data and perform intuitive in-memory real time business analytics. -
23
Lenses
Lenses.io
Enable everyone to discover and observe streaming data. Sharing, documenting and cataloging your data can increase productivity by up to 95%. Then from data, build apps for production use cases. Apply a data-centric security model to cover all the gaps of open source technology, and address data privacy. Provide secure and low-code data pipeline capabilities. Eliminate all darkness and offer unparalleled observability in data and apps. Unify your data mesh and data technologies and be confident with open source in production. Lenses is the highest rated product for real-time stream analytics according to independent third party reviews. With feedback from our community and thousands of engineering hours invested, we've built features that ensure you can focus on what drives value from your real time data. Deploy and run SQL-based real time applications over any Kafka Connect or Kubernetes infrastructure including AWS EKS.Starting Price: $49 per month -
24
Google Cloud Dataflow
Google
Unified stream and batch data processing that's serverless, fast, and cost-effective. Fully managed data processing service. Automated provisioning and management of processing resources. Horizontal autoscaling of worker resources to maximize resource utilization. OSS community-driven innovation with Apache Beam SDK. Reliable and consistent exactly-once processing. Streaming data analytics with speed. Dataflow enables fast, simplified streaming data pipeline development with lower data latency. Allow teams to focus on programming instead of managing server clusters as Dataflow’s serverless approach removes operational overhead from data engineering workloads. Allow teams to focus on programming instead of managing server clusters as Dataflow’s serverless approach removes operational overhead from data engineering workloads. Dataflow automates provisioning and management of processing resources to minimize latency and maximize utilization. -
25
Apama
Apama
Apama Streaming Analytics allows organizations to analyze and act on IoT and fast-moving data in real-time, responding to events intelligently the moment they happen. Apama Community Edition is a freemium version of Apama by Software AG that can be used to learn about, develop and put streaming analytics applications into production. The Software AG Data & Analytics Platform is an end-toend, modular and integrated set of world-class capabilities optimized for high-speed data management and analytics on real-time data and offering out-of-the-box integration and connectivity to all key enterprise data sources. Choose the capabilities you need: streaming, predictive and visual analytics along with messaging for easy integration with other enterprise apps and an in-memory data store for extremely fast access. Integrate historical and other data for comparison—ideal when building models or enriching customer and other vital data. -
26
Amazon Kinesis
Amazon
Easily collect, process, and analyze video and data streams in real time. Amazon Kinesis makes it easy to collect, process, and analyze real-time, streaming data so you can get timely insights and react quickly to new information. Amazon Kinesis offers key capabilities to cost-effectively process streaming data at any scale, along with the flexibility to choose the tools that best suit the requirements of your application. With Amazon Kinesis, you can ingest real-time data such as video, audio, application logs, website clickstreams, and IoT telemetry data for machine learning, analytics, and other applications. Amazon Kinesis enables you to process and analyze data as it arrives and respond instantly instead of having to wait until all your data is collected before the processing can begin. Amazon Kinesis enables you to ingest, buffer, and process streaming data in real-time, so you can derive insights in seconds or minutes instead of hours or days. -
27
SQLstream
Guavus, a Thales company
SQLstream ranks #1 for IoT stream processing & analytics (ABI Research). Used by Verizon, Walmart, Cisco, & Amazon, our technology powers applications across data centers, the cloud, & the edge. Thanks to sub-ms latency, SQLstream enables live dashboards, time-critical alerts, & real-time action. Smart cities can optimize traffic light timing or reroute ambulances & fire trucks. Security systems can shut down hackers & fraudsters right away. AI / ML models, trained by streaming sensor data, can predict equipment failures. With lightning performance, up to 13M rows / sec / CPU core, companies have drastically reduced their footprint & cost. Our efficient, in-memory processing permits operations at the edge that are otherwise impossible. Acquire, prepare, analyze, & act on data in any format from any source. Create pipelines in minutes not months with StreamLab, our interactive, low-code GUI dev environment. Export SQL scripts & deploy with the flexibility of Kubernetes. -
28
Apache Spark
Apache Software Foundation
Apache Spark™ is a unified analytics engine for large-scale data processing. Apache Spark achieves high performance for both batch and streaming data, using a state-of-the-art DAG scheduler, a query optimizer, and a physical execution engine. Spark offers over 80 high-level operators that make it easy to build parallel apps. And you can use it interactively from the Scala, Python, R, and SQL shells. Spark powers a stack of libraries including SQL and DataFrames, MLlib for machine learning, GraphX, and Spark Streaming. You can combine these libraries seamlessly in the same application. Spark runs on Hadoop, Apache Mesos, Kubernetes, standalone, or in the cloud. It can access diverse data sources. You can run Spark using its standalone cluster mode, on EC2, on Hadoop YARN, on Mesos, or on Kubernetes. Access data in HDFS, Alluxio, Apache Cassandra, Apache HBase, Apache Hive, and hundreds of other data sources. -
29
Solix Test Data Management
Solix Technologies
Accurate test data improves application development and testing quality, which is why the most demanding application development teams require that their test data be populated from production databases frequently. It is not unusual for a Test Data Management (TDM) program to maintain six to eight full clones/copies of the production database for use as test and development instances. Without proper automation, provisioning test data is not only inefficient, time-consuming and storage intensive but it could also potentially expose sensitive data to unauthorized personnel leading to compliance risks. Since the cloning process creates such resource drain and data governance challenges, test and development databases are often not refreshed frequently enough resulting in inaccurate test results or even test failures. And of course, the cost of application development increases as the errors are discovered later in the application development lifecycle. -
30
KX Insights
KX
KX Insights is a cloud-native platform for critical real-time performance and continuous actionable intelligence. Using complex event processing, high-speed analytics and machine learning interfaces, it enables fast decision-making and automated responses to events in fractions of a second. It’s not just storage and compute elasticity that have moved to the cloud. It’s everything: data, tools, development, security, connectivity, operations, maintenance. KX can help you leverage that power to make smarter, more insightful decisions by integrating real-time analytics into your business operations. KX Insights leverages industry standards to ensure openness and interoperability with other technologies in order to deliver insights faster and more cost-effectively. It operates a microservices-based architecture for capturing, storing and processing high-volume, high-velocity data using cloud standards, services, and protocols. -
31
Delphix
Delphix
Delphix is the industry leader in DataOps and provides an intelligent data platform that accelerates digital transformation for leading companies around the world. The Delphix DataOps Platform supports a broad spectrum of systems, from mainframes to Oracle databases, ERP applications, and Kubernetes containers. Delphix supports a comprehensive range of data operations to enable modern CI/CD workflows and automates data compliance for privacy regulations, including GDPR, CCPA, and the New York Privacy Act. In addition, Delphix helps companies sync data from private to public clouds, accelerating cloud migrations, customer experience transformation, and the adoption of disruptive AI technologies. Automate data for fast, quality software releases, cloud adoption, and legacy modernization. Source data from mainframe to cloud-native apps across SaaS, private, and public clouds. -
32
By managing data properly over its lifetime, organizations are better equipped to support business goals with less risk. Archive data from decommissioned applications and historical transaction records, while providing ongoing access to the data for query and reporting that is compliant with retention regulations. Scale data across applications, databases, operating systems and hardware platforms to help secure your test environments, accelerate release cycles and reduce costs. Lack of data archiving can impair the performance of mission-critical enterprise systems. Solve data growth problems at the source, improve efficiency and minimize the risks associated with managing structured data throughout its lifetime. Protect unstructured data in testing, development and analytics environments across the enterprise. Lack of data archiving can impair the performance of mission-critical enterprise systems. Solve data growth problems at the source, improve efficiency and minimize the risks.
-
33
Mockaroo
Mockaroo
It's hard to put together a meaningful UI prototype without making real requests to an API. By making real requests, you'll uncover problems with application flow, timing, and API design early, improving the quality of both the user experience and API. With Mockaroo, you can design your own mock APIs, You control the URLs, responses, and error conditions. Paralellize UI and API development and start delivering better applications faster today! There are plenty of great data mocking libraries available for almost every language and platform. But not everyone is a programmer or has time to learn a new framework. Mockaroo allows you to quickly and easily to download large amounts of randomly generated test data based on your own specs which you can then load directly into your test environment using SQL or CSV formats.Starting Price: $50 per year -
34
Informatica Data Engineering Streaming
Informatica
AI-powered Informatica Data Engineering Streaming enables data engineers to ingest, process, and analyze real-time streaming data for actionable insights. Advanced serverless deployment option with integrated metering dashboard cuts admin overhead. Rapidly build intelligent data pipelines with CLAIRE®-powered automation, including automatic change data capture (CDC). Ingest thousands of databases and millions of files, and streaming events. Efficiently ingest databases, files, and streaming data for real-time data replication and streaming analytics. Find and inventory all data assets throughout your organization. Intelligently discover and prepare trusted data for advanced analytics and AI/ML projects. -
35
IBM Streams
IBM
IBM Streams evaluates a broad range of streaming data — unstructured text, video, audio, geospatial and sensor — helping organizations spot opportunities and risks and make decisions in real-time. Make sense of your data, turning fast-moving volumes and varieties into insight with IBM® Streams. Streams evaluate a broad range of streaming data — unstructured text, video, audio, geospatial and sensor — helping organizations spot opportunities and risks as they happen. Combine Streams with other IBM Cloud Pak® for Data capabilities, built on an open, extensible architecture. Help enable data scientists to collaboratively build models to apply to stream flows, plus, analyze massive amounts of data in real-time. Acting upon your data and deriving true value is easier than ever. -
36
StreamSets
StreamSets
StreamSets DataOps Platform. The data integration platform to build, run, monitor and manage smart data pipelines that deliver continuous data for DataOps, and power modern analytics and hybrid integration. Only StreamSets provides a single design experience for all design patterns for 10x greater developer productivity; smart data pipelines that are resilient to change for 80% less breakages; and a single pane of glass for managing and monitoring all pipelines across hybrid and cloud architectures to eliminate blind spots and control gaps. With StreamSets, you can deliver the continuous data that drives the connected enterprise.Starting Price: $1000 per month -
37
SAS Event Stream Processing
SAS Institute
Streaming data from operations, transactions, sensors and IoT devices is valuable – when it's well-understood. Event stream processing from SAS includes streaming data quality and analytics – and a vast array of SAS and open source machine learning and high-frequency analytics for connecting, deciphering, cleansing and understanding streaming data – in one solution. No matter how fast your data moves, how much data you have, or how many data sources you’re pulling from, it’s all under your control via a single, intuitive interface. You can define patterns and address scenarios from all aspects of your business, giving you the power to stay agile and tackle issues as they arise. -
38
ERBuilder
Softbuilder
ERBuilder Data Modeler is a GUI data modeling tool that allows developers to visualize, design, and model databases by using entity relationship diagrams and automatically generates the most popular SQL databases. Generate and share the data Model documentation with your team. Optimize your data model by using advanced features such as test data generation, schema compare, and schema synchronization.Starting Price: $49 -
39
Embiot
Telchemy
Embiot® is a compact, high performance IoT analytics software agent for IoT gateway and smart sensor applications. This edge computing application is small enough to integrate directly into devices, smart sensors and gateways, but powerful enough to calculate complex analytics from large amounts of raw data at high speed. Internally, Embiot uses a stream processing model to enable it to handle sensor data that arrives at different rates and out of order. It has a simple intuitive configuration language and a rich set of math, stats and AI functions making it fast and easy to solve your analytics problems. Embiot supports a range of input methods including MODBUS, MQTT, REST/XML, REST/JSON, Name/Value and CSV. Embiot is able to send output reports to multiple destinations concurrently in REST, MQTT and custom text formats. For security, Embiot supports TLS selectively on any input or output stream, HTTP and MQTT authentication. -
40
Alibaba Cloud DataHub
Alibaba Cloud
DataHub supports various SDKs and APIs and provides multiple third-party plug-ins such as Flume and Logstash. You can import data to DataHub in an efficient manner. The DataConnector module can synchronize imported data to downstream storage and analysis systems in real time, such as MaxCompute, OSS, and Tablestore. You can import heterogeneous data that is generated by applications, websites, IoT devices, or databases to DataHub in real time. You can manage the data in a unified manner by using DataHub. You can also deliver the data to downstream systems such as analysis systems and archiving systems. This way, you can build a data streaming pipeline and extract more data value. -
41
DTM Data Generator
DTM Data Generator
Fast test data generation engine with about 70 built-in functions and expression processor enables users to define complex test data with dependencies, internal structure, and relationships. The product analyzes existing database schema and resolves master-detail key structure (relationships) automatically. Value Library is a predefined data sets: names, countries, cities, streets, currencies, companies, industries, departments, regions, etc. Variables and Named Generators features provide a way to share data generation properties to similar columns. Intelligent schema analyzer makes your data realistic without extra project modifications and "data by example" feature makes data more realistic without extra efforts. -
42
Kinetica
Kinetica
A scalable cloud database for real-time analysis on large and streaming datasets. Kinetica is designed to harness modern vectorized processors to be orders of magnitude faster and more efficient for real-time spatial and temporal workloads. Track and gain intelligence from billions of moving objects in real-time. Vectorization unlocks new levels of performance for analytics on spatial and time series data at scale. Ingest and query at the same time to act on real-time events. Kinetica's lockless architecture and distributed ingestion ensures data is available to query as soon as it lands. Vectorized processing enables you to do more with less. More power allows for simpler data structures, which lead to lower storage costs, more flexibility and less time engineering your data. Vectorized processing opens the door to amazingly fast analytics and detailed visualization of moving objects at scale. -
43
Gathr
Gathr
The only all-in-one data pipeline platform. Built ground-up for a cloud-first world, Gathr is the only platform to handle all your data integration and engineering needs - ingestion, ETL, ELT, CDC, streaming analytics, data preparation, machine learning, advanced analytics and more. With Gathr, anyone can build and deploy pipelines in minutes, irrespective of skill levels. Create Ingestion pipelines in minutes, not weeks. Ingest data from any source, deliver to any destination. Build applications quickly with a wizard-based approach. Replicate data in real-time using a templatized CDC app. Native integration for all sources and targets. Best-in-class capabilities with everything you need to succeed today and tomorrow. Choose between free, pay-per-use or customize as per your requirements. -
44
GenRocket
GenRocket
Enterprise synthetic test data solutions. In order to generate test data that accurately reflects the structure of your application or database, it must be easy to model and maintain each test data project as changes to the data model occur throughout the lifecycle of the application. Maintain referential integrity of parent/child/sibling relationships across the data domains within an application database or across multiple databases used by multiple applications. Ensure the consistency and integrity of synthetic data attributes across applications, data sources and targets. For example, a customer name must always match the same customer ID across multiple transactions simulated by real-time synthetic data generation. Customers want to quickly and accurately create their data model as a test data project. GenRocket offers 10 methods for data model setup. XTS, DDL, Scratchpad, Presets, XSD, CSV, YAML, JSON, Spark Schema, Salesforce. -
45
Solace PubSub+
Solace
Solace PubSub+ Platform helps enterprises design, deploy and manage event-driven systems across hybrid and multi-cloud and IoT environments so they can be more event-driven and operate in real-time. The PubSub+ Platform includes the powerful PubSub+ Event Brokers, event management capabilities with PubSub+ Event Portal, as well as monitoring and integration capabilities all available via a single cloud console. PubSub+ allows easy creation of an event mesh, an interconnected network of event brokers, allowing for seamless and dynamic data movement across highly distributed network environments. PubSub+ Event Brokers can be deployed as fully managed cloud services, self-managed software in private cloud or on-premises environments, or as turnkey hardware appliances for unparalleled performance and low TCO. PubSub+ Event Portal is a complimentary toolset for design and governance of event-driven systems including both Solace and Kafka-based event broker environments. -
46
Cumulocity IoT
Software AG
Cumulocity IoT is the #1 low-code, self-service IoT platform—the only one that comes pre-integrated with the tools you need for fast results: device connectivity and management, application enablement and integration, as well as streaming and predictive analytics. Free your business from proprietary technology stacks. Because you’ll be using the only completely open IoT platform, you can connect any “thing” today and tomorrow. Bring your own hardware and tools, and pick the components that best fit. Get up and running on the IoT in minutes. Connect a device and view its data. Create a real-time interactive dashboard. Define rules to monitor and act on events. Do all of this without calling on IT or writing any code! Easily integrate new IoT data with the core enterprise systems, applications and processes that have run your business for years—again, without coding—for a fluid flow of data. You’ll have more context to make better decisions. -
47
Cloudera DataFlow
Cloudera
Cloudera DataFlow for the Public Cloud (CDF-PC) is a cloud-native universal data distribution service powered by Apache NiFi that lets developers connect to any data source anywhere with any structure, process it, and deliver to any destination. CDF-PC offers a flow-based low-code development paradigm that aligns best with how developers design, develop, and test data distribution pipelines. With over 400+ connectors and processors across the ecosystem of hybrid cloud services—including data lakes, lakehouses, cloud warehouses, and on-premises sources—CDF-PC provides indiscriminate data distribution. These data distribution flows can then be version-controlled into a catalog where operators can self-serve deployments to different runtimes. -
48
Datanamic Data Generator
Datanamic
Datanamic Data Generator is a powerful data generator that allows developers to easily populate databases with thousands of rows of meaningful and syntactically correct test data for database testing purposes. An empty database is not useful for making sure your application will work as designed. You need test data. Writing your own test data generators or scripts is time consuming. Datanamic Data Generator will help you. The tool can be used by DBAs, developers, or testers, who need sample data to test a database-driven application. Datanamic Data Generator makes database test data generation easy and painless. It reads your database and displays tables and columns with their data generation settings. Only a few simple entries are necessary to generate comprehensive (realistic) test data. The tool can be used to generate test data from scratch or from existing data.Starting Price: €59 per month -
49
Tonic
Tonic
Tonic automatically creates mock data that preserves key characteristics of secure datasets so that developers, data scientists, and salespeople can work conveniently without breaching privacy. Tonic mimics your production data to create de-identified, realistic, and safe data for your test environments. With Tonic, your data is modeled from your production data to help you tell an identical story in your testing environments. Safe, useful data created to mimic your real-world data, at scale. Generate data that looks, acts, and feels just like your production data and safely share it across teams, businesses, and international borders. PII/PHI identification, obfuscation, and transformation. Proactively protect your sensitive data with automatic scanning, alerts, de-identification, and mathematical guarantees of data privacy. Advanced sub setting across diverse database types. Collaboration, compliance, and data workflows — perfectly automated. -
50
Qlik Gold Client
Qlik
Qlik Gold Client improves the efficiency, cost and security of managing test data in SAP environments. Qlik Gold Client is designed to eliminate development workarounds by easily moving configuration, master, and transactional data subsets into testing environments. Rapidly define, copy and synchronize transactional data from production to non-production targets. Identify, select, and delete non-production data. Manage extensive and powerful data transformations through a clean and easy-to-use interface. Automate data selection and enable hands-free test data refresh cycles, reducing time and effort for test data management. Qlik Gold Client provides several options to protect PII data in non-production environments via data masking. Data masking applies a set of rules to “scramble” your production data when it’s replicated to a non-production environment.