Alternatives to Lenses

Compare Lenses alternatives for your business or organization using the curated list below. SourceForge ranks the best alternatives to Lenses in 2024. Compare features, ratings, user reviews, pricing, and more from Lenses competitors and alternatives in order to make an informed decision for your business.

  • 1
    groundcover

    groundcover

    groundcover

    Cloud-based observability solution that helps businesses track and manage workload and performance on a unified dashboard. Monitor everything you run in your cloud without compromising on cost, granularity, or scale. groundcover is a full stack cloud-native APM platform designed to make observability effortless so that you can focus on building world-class products. By leveraging our proprietary sensor, groundcover unlocks unprecedented granularity on all your applications, eliminating the need for costly code changes and development cycles to ensure monitoring continuity. 100% visibility, all the time. Cover your entire Kubernetes stack instantly, with no code changes using the superpowers of eBPF instrumentation. Take control of your data, all in-cloud. groundcover’s unique inCloud architecture keeps your data private, secured and under your control without ever leaving your cloud premises.
    Compare vs. Lenses View Software
    Visit Website
  • 2
    DATPROF

    DATPROF

    DATPROF

    Test Data Management solutions like data masking, synthetic data generation, data subsetting, data discovery, database virtualization, data automation are our core business. We see and understand the struggles of software development teams with test data. Personally Identifiable Information? Too large environments? Long waiting times for a test data refresh? We envision to solve these issues: - Obfuscating, generating or masking databases and flat files; - Extracting or filtering specific data content with data subsetting; - Discovering, profiling and analysing solutions for understanding your test data, - Automating, integrating and orchestrating test data provisioning into your CI/CD pipelines and - Cloning, snapshotting and timetraveling throug your test data with database virtualization. We improve and innovate our test data software with the latest technologies every single day to support medium to large size organizations in their Test Data Management.
  • 3
    IRI Voracity

    IRI Voracity

    IRI, The CoSort Company

    Voracity is the only high-performance, all-in-one data management platform accelerating AND consolidating the key activities of data discovery, integration, migration, governance, and analytics. Voracity helps you control your data in every stage of the lifecycle, and extract maximum value from it. Only in Voracity can you: 1) CLASSIFY, profile and diagram enterprise data sources 2) Speed or LEAVE legacy sort and ETL tools 3) MIGRATE data to modernize and WRANGLE data to analyze 4) FIND PII everywhere and consistently MASK it for referential integrity 5) Score re-ID risk and ANONYMIZE quasi-identifiers 6) Create and manage DB subsets or intelligently synthesize TEST data 7) Package, protect and provision BIG data 8) Validate, scrub, enrich and unify data to improve its QUALITY 9) Manage metadata and MASTER data. Use Voracity to comply with data privacy laws, de-muck and govern the data lake, improve the reliability of your analytics, and create safe, smart test data
  • 4
    TreasuryPay

    TreasuryPay

    TreasuryPay

    Instant™ Enterprise Data and Intelligence. Visibility into all transaction data, as it is happening, wherever it is happening worldwide. With just one network connection, organizations receive worldwide Accounting, Liquidity Management, FX, Marketing, and Supply Chain information — delivered in a single managed solution to empower enterprise intelligence. The TreasuryPay product set streams your global receivables information, delivering instant accountancy and cognitive services. It is, quite simply, the most advanced intelligence and insights platform currently available to global organizations. Instantly provide your organization with enriched information for your entire global enterprise. The change is easy. The Return on Investment, remarkable. Actionable intelligence and real-time global accountancy are now available at your fingertips with TreasuryPay Instant™.
  • 5
    Striim

    Striim

    Striim

    Data integration for your hybrid cloud. Modern, reliable data integration across your private and public cloud. All in real-time with change data capture and data streams. Built by the executive & technical team from GoldenGate Software, Striim brings decades of experience in mission-critical enterprise workloads. Striim scales out as a distributed platform in your environment or in the cloud. Scalability is fully configurable by your team. Striim is fully secure with HIPAA and GDPR compliance. Built ground up for modern enterprise workloads in the cloud or on-premise. Drag and drop to create data flows between your sources and targets. Process, enrich, and analyze your streaming data with real-time SQL queries.
  • 6
    Amazon MSK
    Amazon MSK is a fully managed service that makes it easy for you to build and run applications that use Apache Kafka to process streaming data. Apache Kafka is an open-source platform for building real-time streaming data pipelines and applications. With Amazon MSK, you can use native Apache Kafka APIs to populate data lakes, stream changes to and from databases, and power machine learning and analytics applications. Apache Kafka clusters are challenging to setup, scale, and manage in production. When you run Apache Kafka on your own, you need to provision servers, configure Apache Kafka manually, replace servers when they fail, orchestrate server patches and upgrades, architect the cluster for high availability, ensure data is durably stored and secured, setup monitoring and alarms, and carefully plan scaling events to support load changes.
    Starting Price: $0.0543 per hour
  • 7
    Confluent

    Confluent

    Confluent

    Infinite retention for Apache Kafka® with Confluent. Be infrastructure-enabled, not infrastructure-restricted Legacy technologies require you to choose between being real-time or highly-scalable. Event streaming enables you to innovate and win - by being both real-time and highly-scalable. Ever wonder how your rideshare app analyzes massive amounts of data from multiple sources to calculate real-time ETA? Ever wonder how your credit card company analyzes millions of credit card transactions across the globe and sends fraud notifications in real-time? The answer is event streaming. Move to microservices. Enable your hybrid strategy through a persistent bridge to cloud. Break down silos to demonstrate compliance. Gain real-time, persistent event transport. The list is endless.
  • 8
    Azure Event Hubs
    Event Hubs is a fully managed, real-time data ingestion service that’s simple, trusted, and scalable. Stream millions of events per second from any source to build dynamic data pipelines and immediately respond to business challenges. Keep processing data during emergencies using the geo-disaster recovery and geo-replication features. Integrate seamlessly with other Azure services to unlock valuable insights. Allow existing Apache Kafka clients and applications to talk to Event Hubs without any code changes—you get a managed Kafka experience without having to manage your own clusters. Experience real-time data ingestion and microbatching on the same stream. Focus on drawing insights from your data instead of managing infrastructure. Build real-time big data pipelines and respond to business challenges right away.
    Starting Price: $0.03 per hour
  • 9
    Oracle Cloud Infrastructure Streaming
    Streaming service is a real-time, serverless, Apache Kafka-compatible event streaming platform for developers and data scientists. Streaming is tightly integrated with Oracle Cloud Infrastructure (OCI), Database, GoldenGate, and Integration Cloud. The service also provides out-of-the-box integrations for hundreds of third-party products across categories such as DevOps, databases, big data, and SaaS applications. Data engineers can easily set up and operate big data pipelines. Oracle handles all infrastructure and platform management for event streaming, including provisioning, scaling, and security patching. With the help of consumer groups, Streaming can provide state management for thousands of consumers. This helps developers easily build applications at scale.
  • 10
    Solace PubSub+
    Solace PubSub+ Platform helps enterprises design, deploy and manage event-driven systems across hybrid and multi-cloud and IoT environments so they can be more event-driven and operate in real-time. The PubSub+ Platform includes the powerful PubSub+ Event Brokers, event management capabilities with PubSub+ Event Portal, as well as monitoring and integration capabilities all available via a single cloud console. PubSub+ allows easy creation of an event mesh, an interconnected network of event brokers, allowing for seamless and dynamic data movement across highly distributed network environments. PubSub+ Event Brokers can be deployed as fully managed cloud services, self-managed software in private cloud or on-premises environments, or as turnkey hardware appliances for unparalleled performance and low TCO. PubSub+ Event Portal is a complimentary toolset for design and governance of event-driven systems including both Solace and Kafka-based event broker environments.
  • 11
    K2View

    K2View

    K2View

    At K2View, we believe that every enterprise should be able to leverage its data to become as disruptive and agile as the best companies in its industry. We make this possible through our patented Data Product Platform, which creates and manages a complete and compliant dataset for every business entity – on demand, and in real time. The dataset is always in sync with its underlying sources, adapts to changes in the source structures, and is instantly accessible to any authorized data consumer. Data Product Platform fuels many operational use cases, including customer 360, data masking and tokenization, test data management, data migration, legacy application modernization, data pipelining and more – to deliver business outcomes in less than half the time, and at half the cost, of any other alternative. The platform inherently supports modern data architectures – data mesh, data fabric, and data hub – and deploys in cloud, on-premise, or hybrid environments.
  • 12
    Informatica Data Engineering Streaming
    AI-powered Informatica Data Engineering Streaming enables data engineers to ingest, process, and analyze real-time streaming data for actionable insights. Advanced serverless deployment option​ with integrated metering dashboard cuts admin overhead. Rapidly build intelligent data pipelines with CLAIRE®-powered automation, including automatic change data capture (CDC). Ingest thousands of databases and millions of files, and streaming events. Efficiently ingest databases, files, and streaming data for real-time data replication and streaming analytics. Find and inventory all data assets throughout your organization. Intelligently discover and prepare trusted data for advanced analytics and AI/ML projects.
  • 13
    SAS Event Stream Processing
    Streaming data from operations, transactions, sensors and IoT devices is valuable – when it's well-understood. Event stream processing from SAS includes streaming data quality and analytics – and a vast array of SAS and open source machine learning and high-frequency analytics for connecting, deciphering, cleansing and understanding streaming data – in one solution. No matter how fast your data moves, how much data you have, or how many data sources you’re pulling from, it’s all under your control via a single, intuitive interface. You can define patterns and address scenarios from all aspects of your business, giving you the power to stay agile and tackle issues as they arise.
  • 14
    Cloudera DataFlow
    Cloudera DataFlow for the Public Cloud (CDF-PC) is a cloud-native universal data distribution service powered by Apache NiFi ​​that lets developers connect to any data source anywhere with any structure, process it, and deliver to any destination. CDF-PC offers a flow-based low-code development paradigm that aligns best with how developers design, develop, and test data distribution pipelines. With over 400+ connectors and processors across the ecosystem of hybrid cloud services—including data lakes, lakehouses, cloud warehouses, and on-premises sources—CDF-PC provides indiscriminate data distribution. These data distribution flows can then be version-controlled into a catalog where operators can self-serve deployments to different runtimes.
  • 15
    TIBCO Streaming
    Analyze, continuously query, and act on IoT and other streaming data at lightning fast speeds. Take real-time operations and analytics to the next level with intelligent applications that deploy quickly for taking action based on new decisions and models, all without extra overhead. TIBCO® Streaming software is enterprise-grade, cloud-ready streaming analytics for quickly building real-time applications at a fraction of the cost and risk of alternatives.
  • 16
    Amazon Kinesis
    Easily collect, process, and analyze video and data streams in real time. Amazon Kinesis makes it easy to collect, process, and analyze real-time, streaming data so you can get timely insights and react quickly to new information. Amazon Kinesis offers key capabilities to cost-effectively process streaming data at any scale, along with the flexibility to choose the tools that best suit the requirements of your application. With Amazon Kinesis, you can ingest real-time data such as video, audio, application logs, website clickstreams, and IoT telemetry data for machine learning, analytics, and other applications. Amazon Kinesis enables you to process and analyze data as it arrives and respond instantly instead of having to wait until all your data is collected before the processing can begin. Amazon Kinesis enables you to ingest, buffer, and process streaming data in real-time, so you can derive insights in seconds or minutes instead of hours or days.
  • 17
    TIBCO BusinessEvents
    In the emerging digital world, billions of people, systems, and devices will interact in real-time, suggesting new and disruptive competitive advantages. How do you play there? By building distributed, stateful, rule-based event-processing systems, experimenting, learning, and quickly evolving. Data is arriving in real-time from hundreds of internal and external sources—and it has shelf-life; its value diminishes over time. With TIBCO, you can jump-start big data processing initiatives that give you the ability to sense, reason, respond, and visualize—much different and much faster and smarter than the traditional store, analyze, report, act approach. With multiple rule authoring environments, you support collaboration between IT developers and business professionals.
  • 18
    WarpStream

    WarpStream

    WarpStream

    WarpStream is an Apache Kafka-compatible data streaming platform built directly on top of object storage, with no inter-AZ networking costs, no disks to manage, and infinitely scalable, all within your VPC. WarpStream is deployed as a stateless and auto-scaling agent binary in your VPC with no local disks to manage. Agents stream data directly to and from object storage with no buffering on local disks and no data tiering. Create new “virtual clusters” in our control plane instantly. Support different environments, teams, or projects without managing any dedicated infrastructure. WarpStream is protocol compatible with Apache Kafka, so you can keep using all your favorite tools and software. No need to rewrite your application or use a proprietary SDK. Just change the URL in your favorite Kafka client library and start streaming. Never again have to choose between reliability and your budget.
    Starting Price: $2,987 per month
  • 19
    DeltaStream

    DeltaStream

    DeltaStream

    DeltaStream is a unified serverless stream processing platform that integrates with streaming storage services. Think about it as the compute layer on top of your streaming storage. It provides functionalities of streaming analytics(Stream processing) and streaming databases along with additional features to provide a complete platform to manage, process, secure and share streaming data. DeltaStream provides a SQL based interface where you can easily create stream processing applications such as streaming pipelines, materialized views, microservices and many more. It has a pluggable processing engine and currently uses Apache Flink as its primary stream processing engine. DeltaStream is more than just a query processing layer on top of Kafka or Kinesis. It brings relational database concepts to the data streaming world, including namespacing and role based access control enabling you to securely access, process and share your streaming data regardless of where they are stored.
  • 20
    Visual KPI

    Visual KPI

    Transpara

    Real-time operations monitoring and data visualization, including KPIs, dashboards, trends, analytics, hierarchy and alerts. Aggregates all of your data sources (industrial, IoT, business, external, etc.) on the fly without moving the data, and displays it in real-time with obvious context on any device.
  • 21
    StreamSets

    StreamSets

    StreamSets

    StreamSets DataOps Platform. The data integration platform to build, run, monitor and manage smart data pipelines that deliver continuous data for DataOps, and power modern analytics and hybrid integration. Only StreamSets provides a single design experience for all design patterns for 10x greater developer productivity; smart data pipelines that are resilient to change for 80% less breakages; and a single pane of glass for managing and monitoring all pipelines across hybrid and cloud architectures to eliminate blind spots and control gaps. With StreamSets, you can deliver the continuous data that drives the connected enterprise.
    Starting Price: $1000 per month
  • 22
    TIBCO Cloud Events
    Digital businesses must design applications that allow the business to be proactive not reactive. This requires the ability to process huge amounts of data in real time to identify meaningful events while there is still time to influence the outcome—whether to mitigate risk or improve the customer experience or operations. TIBCO Cloud™ Events service is a cloud-native solution for improving business outcomes by detecting key events and automating the next best action. Build declarative rules without worrying about the underlying plumbing. Define and capture key business conditions, and automate actions. TIBCO Cloud Events service takes care of the execution and matching of related data.
    Starting Price: $450 per unit per month
  • 23
    SQLstream

    SQLstream

    Guavus, a Thales company

    SQLstream ranks #1 for IoT stream processing & analytics (ABI Research). Used by Verizon, Walmart, Cisco, & Amazon, our technology powers applications across data centers, the cloud, & the edge. Thanks to sub-ms latency, SQLstream enables live dashboards, time-critical alerts, & real-time action. Smart cities can optimize traffic light timing or reroute ambulances & fire trucks. Security systems can shut down hackers & fraudsters right away. AI / ML models, trained by streaming sensor data, can predict equipment failures. With lightning performance, up to 13M rows / sec / CPU core, companies have drastically reduced their footprint & cost. Our efficient, in-memory processing permits operations at the edge that are otherwise impossible. Acquire, prepare, analyze, & act on data in any format from any source. Create pipelines in minutes not months with StreamLab, our interactive, low-code GUI dev environment. Export SQL scripts & deploy with the flexibility of Kubernetes.
  • 24
    Red Hat OpenShift Streams
    Red Hat® OpenShift® Streams for Apache Kafka is a managed cloud service that provides a streamlined developer experience for building, deploying, and scaling new cloud-native applications or modernizing existing systems. Red Hat OpenShift Streams for Apache Kafka makes it easy to create, discover, and connect to real-time data streams no matter where they are deployed. Streams are a key component for delivering event-driven and data analytics applications. The combination of seamless operations across distributed microservices, large data transfer volumes, and managed operations allows teams to focus on team strengths, speed up time to value, and lower operational costs. OpenShift Streams for Apache Kafka includes a Kafka ecosystem and is part of a family of cloud services—and the Red Hat OpenShift product family—which helps you build a wide range of data-driven solutions.
  • 25
    Apache Kafka

    Apache Kafka

    The Apache Software Foundation

    Apache Kafka® is an open-source, distributed streaming platform. Scale production clusters up to a thousand brokers, trillions of messages per day, petabytes of data, hundreds of thousands of partitions. Elastically expand and contract storage and processing. Stretch clusters efficiently over availability zones or connect separate clusters across geographic regions. Process streams of events with joins, aggregations, filters, transformations, and more, using event-time and exactly-once processing. Kafka’s out-of-the-box Connect interface integrates with hundreds of event sources and event sinks including Postgres, JMS, Elasticsearch, AWS S3, and more. Read, write, and process streams of events in a vast array of programming languages.
  • 26
    Cumulocity IoT

    Cumulocity IoT

    Software AG

    Cumulocity IoT is the #1 low-code, self-service IoT platform—the only one that comes pre-integrated with the tools you need for fast results: device connectivity and management, application enablement and integration, as well as streaming and predictive analytics. Free your business from proprietary technology stacks. Because you’ll be using the only completely open IoT platform, you can connect any “thing” today and tomorrow. Bring your own hardware and tools, and pick the components that best fit. Get up and running on the IoT in minutes. Connect a device and view its data. Create a real-time interactive dashboard. Define rules to monitor and act on events. Do all of this without calling on IT or writing any code! Easily integrate new IoT data with the core enterprise systems, applications and processes that have run your business for years—again, without coding—for a fluid flow of data. You’ll have more context to make better decisions.
  • 27
    Google Cloud Dataflow
    Unified stream and batch data processing that's serverless, fast, and cost-effective. Fully managed data processing service. Automated provisioning and management of processing resources. Horizontal autoscaling of worker resources to maximize resource utilization. OSS community-driven innovation with Apache Beam SDK. Reliable and consistent exactly-once processing. Streaming data analytics with speed. Dataflow enables fast, simplified streaming data pipeline development with lower data latency. Allow teams to focus on programming instead of managing server clusters as Dataflow’s serverless approach removes operational overhead from data engineering workloads. Allow teams to focus on programming instead of managing server clusters as Dataflow’s serverless approach removes operational overhead from data engineering workloads. Dataflow automates provisioning and management of processing resources to minimize latency and maximize utilization.
  • 28
    Materialize

    Materialize

    Materialize

    Materialize is a reactive database that delivers incremental view updates. We help developers easily build with streaming data using standard SQL. Materialize can connect to many different external sources of data without pre-processing. Connect directly to streaming sources like Kafka, Postgres databases, CDC, or historical sources of data like files or S3. Materialize allows you to query, join, and transform data sources in standard SQL - and presents the results as incrementally-updated Materialized views. Queries are maintained and continually updated as new data streams in. With incrementally-updated views, developers can easily build data visualizations or real-time applications. Building with streaming data can be as simple as writing a few lines of SQL.
    Starting Price: $0.98 per hour
  • 29
    Digital Twin Streaming Service
    ScaleOut Digital Twin Streaming Service™ Easily build and deploy real-time digital twins for streaming analytics Connect to many data sources with Azure & AWS IoT hubs, Kafka, and more Maximize situational awareness with live, aggregate analytics. Introducing a breakthrough cloud service that simultaneously tracks telemetry from millions of data sources with “real-time” digital twins — enabling immediate, deep introspection with state-tracking and highly targeted, real-time feedback for thousands of devices. A powerful UI simplifies deployment and displays aggregate analytics in real time to maximize situational awareness. Ideal for a wide range of applications, including the Internet of Things (IoT), real-time intelligent monitoring, logistics, and financial services. Simplified pricing makes getting started fast and easy. Combined with the ScaleOut Digital Twin Builder software toolkit, the ScaleOut Digital Twin Streaming Service enables the next generation in stream processing.
  • 30
    Apache Atlas

    Apache Atlas

    Apache Software Foundation

    Atlas is a scalable and extensible set of core foundational governance services – enabling enterprises to effectively and efficiently meet their compliance requirements within Hadoop and allows integration with the whole enterprise data ecosystem. Apache Atlas provides open metadata management and governance capabilities for organizations to build a catalog of their data assets, classify and govern these assets and provide collaboration capabilities around these data assets for data scientists, analysts and the data governance team. Pre-defined types for various Hadoop and non-Hadoop metadata. Ability to define new types for the metadata to be managed. Types can have primitive attributes, complex attributes, object references; can inherit from other types. Instances of types, called entities, capture metadata object details and their relationships. REST APIs to work with types and instances allow easier integration.
  • 31
    PHEMI Health DataLab
    The PHEMI Trustworthy Health DataLab is a unique, cloud-based, integrated big data management system that allows healthcare organizations to enhance innovation and generate value from healthcare data by simplifying the ingestion and de-identification of data with NSA/military-grade governance, privacy, and security built-in. Conventional products simply lock down data, PHEMI goes further, solving privacy and security challenges and addressing the urgent need to secure, govern, curate, and control access to privacy-sensitive personal healthcare information (PHI). This improves data sharing and collaboration inside and outside of an enterprise—without compromising the privacy of sensitive information or increasing administrative burden. PHEMI Trustworthy Health DataLab can scale to any size of organization, is easy to deploy and manage, connects to hundreds of data sources, and integrates with popular data science and business analysis tools.
  • 32
    Okera

    Okera

    Okera

    Okera, the Universal Data Authorization company, helps modern, data-driven enterprises accelerate innovation, minimize data security risks, and demonstrate regulatory compliance. The Okera Dynamic Access Platform automatically enforces universal fine-grained access control policies. This allows employees, customers, and partners to use data responsibly, while protecting them from inappropriately accessing data that is confidential, personally identifiable, or regulated. Okera’s robust audit capabilities and data usage intelligence deliver the real-time and historical information that data security, compliance, and data delivery teams need to respond quickly to incidents, optimize processes, and analyze the performance of enterprise data initiatives. Okera began development in 2016 and now dynamically authorizes access to hundreds of petabytes of sensitive data for the world’s most demanding F100 companies and regulatory agencies. The company is headquartered in San Francisco.
  • 33
    Rockset

    Rockset

    Rockset

    Real-Time Analytics on Raw Data. Live ingest from S3, Kafka, DynamoDB & more. Explore raw data as SQL tables. Build amazing data-driven applications & live dashboards in minutes. Rockset is a serverless search and analytics engine that powers real-time apps and live dashboards. Operate directly on raw data, including JSON, XML, CSV, Parquet, XLSX or PDF. Plug data from real-time streams, data lakes, databases, and data warehouses into Rockset. Ingest real-time data without building pipelines. Rockset continuously syncs new data as it lands in your data sources without the need for a fixed schema. Use familiar SQL, including joins, filters, and aggregations. It’s blazing fast, as Rockset automatically indexes all fields in your data. Serve fast queries that power the apps, microservices, live dashboards, and data science notebooks you build. Scale without worrying about servers, shards, or pagers.
    Starting Price: Free
  • 34
    IBM Event Streams
    Built on open-source Apache Kafka, IBM® Event Streams is an event-streaming platform that helps you build smart apps that can react to events as they happen. Event Streams is based on years of IBM operational expertise gained from running Apache Kafka event streams for enterprises. This makes Event Streams ideal for mission-critical workloads. With connectors to a wide range of core systems and a scalable REST API, you can extend the reach of your existing enterprise assets. Rich security and geo-replication aids disaster recovery. Take advantage of IBM productivity tools and use the CLI to ensure best practices. Replicate data between Event Streams deployments in a disaster-recovery situation.
  • 35
    Aiven for Apache Kafka
    Apache Kafka as a fully managed service, with zero vendor lock-in and a full set of capabilities to build your streaming pipeline. Set up fully managed Kafka in less than 10 minutes — directly from our web console or programmatically via our API, CLI, Terraform provider or Kubernetes operator. Easily connect it to your existing tech stack with over 30 connectors, and feel confident in your setup with logs and metrics available out of the box via the service integrations. A fully managed distributed data streaming platform, deployable in the cloud of your choice. Ideal for event-driven applications, near-real-time data transfer and pipelines, stream analytics, and any other case where you need to move a lot of data between applications — and quickly. With Aiven’s hosted and managed-for-you Apache Kafka, you can set up clusters, deploy new nodes, migrate clouds, and upgrade existing versions — in a single mouse click — and monitor them through a simple dashboard.
    Starting Price: $200 per month
  • 36
    Apama

    Apama

    Apama

    Apama Streaming Analytics allows organizations to analyze and act on IoT and fast-moving data in real-time, responding to events intelligently the moment they happen. Apama Community Edition is a freemium version of Apama by Software AG that can be used to learn about, develop and put streaming analytics applications into production. The Software AG Data & Analytics Platform is an end-toend, modular and integrated set of world-class capabilities optimized for high-speed data management and analytics on real-time data and offering out-of-the-box integration and connectivity to all key enterprise data sources. Choose the capabilities you need: streaming, predictive and visual analytics along with messaging for easy integration with other enterprise apps and an in-memory data store for extremely fast access. Integrate historical and other data for comparison—ideal when building models or enriching customer and other vital data.
  • 37
    Nussknacker

    Nussknacker

    Nussknacker

    Nussknacker is a low-code visual tool for domain experts to define and run real-time decisioning algorithms instead of implementing them in the code. It serves where real-time actions on data have to be made: real-time marketing, fraud detection, Internet of Things, Customer 360, and Machine Learning inferring. An essential part of Nussknacker is a visual design tool for decision algorithms. It allows not-so-technical users – analysts or business people – to define decision logic in an imperative, easy-to-follow, and understandable way. Once authored, with a click of a button, scenarios are deployed for execution. And can be changed and redeployed anytime there’s a need. Nussknacker supports two processing modes: streaming and request-response. In streaming mode, it uses Kafka as its primary interface. It supports both stateful and stateless processing.
    Starting Price: 0
  • 38
    KX Streaming Analytics
    KX Streaming Analytics provides the ability to ingest, store, process, and analyze historic and time series data to make analytics, insights, and visualizations instantly available. To help ensure your applications and users are productive quickly, the platform provides the full lifecycle of data services, including query processing, tiering, migration, archiving, data protection, and scaling. Our advanced analytics and visualization tools, used widely across finance and industry, enable you to define and perform queries, calculations, aggregations, machine learning and AI on any streaming and historical data. Deployable across multiple hardware environments, data can come from real-time business events and high-volume sources including sensors, clickstreams, radio-frequency identification, GPS systems, social networking sites, and mobile devices.
  • 39
    Eclipse Streamsheets
    Build professional applications to automate workflows, continuously monitor operations, and control processes in real-time. Your solutions run 24/7 on servers in the cloud and on the edge. Thanks to the spreadsheet user interface, you do not have to be a software developer. Instead of writing program code, you drag-and-drop data, fill cells with formulas, and design charts in a way you already know. Find all necessary protocols on board that you need to connect to sensors, and machines like MQTT, REST, and OPC UA. Streamsheets is native to stream data processing like MQTT and kafka. Pick up a topic stream, transform it and blast it back out into the endless streaming world. REST opens you the world, Streamsheets let you connect to any web service or let them connect to you. Streamsheets run in the cloud, on your servers, but also on edge devices like a Raspberry Pi.
  • 40
    Astra Streaming
    Responsive applications keep users engaged and developers inspired. Rise to meet these ever-increasing expectations with the DataStax Astra Streaming service platform. DataStax Astra Streaming is a cloud-native messaging and event streaming platform powered by Apache Pulsar. Astra Streaming allows you to build streaming applications on top of an elastically scalable, multi-cloud messaging and event streaming platform. Astra Streaming is powered by Apache Pulsar, the next-generation event streaming platform which provides a unified solution for streaming, queuing, pub/sub, and stream processing. Astra Streaming is a natural complement to Astra DB. Using Astra Streaming, existing Astra DB users can easily build real-time data pipelines into and out of their Astra DB instances. With Astra Streaming, avoid vendor lock-in and deploy on any of the major public clouds (AWS, GCP, Azure) compatible with open-source Apache Pulsar.
  • 41
    Arroyo

    Arroyo

    Arroyo

    Scale from zero to millions of events per second. Arroyo ships as a single, compact binary. Run locally on MacOS or Linux for development, and deploy to production with Docker or Kubernetes. Arroyo is a new kind of stream processing engine, built from the ground up to make real-time easier than batch. Arroyo was designed from the start so that anyone with SQL experience can build reliable, efficient, and correct streaming pipelines. Data scientists and engineers can build end-to-end real-time applications, models, and dashboards, without a separate team of streaming experts. Transform, filter, aggregate, and join data streams by writing SQL, with sub-second results. Your streaming pipelines shouldn't page someone just because Kubernetes decided to reschedule your pods. Arroyo is built to run in modern, elastic cloud environments, from simple container runtimes like Fargate to large, distributed deployments on the Kubernetes logo Kubernetes.
  • 42
    Pathway

    Pathway

    Pathway

    Pathway is a Python ETL framework for stream processing, real-time analytics, LLM pipelines, and RAG. Pathway comes with an easy-to-use Python API, allowing you to seamlessly integrate your favorite Python ML libraries. Pathway code is versatile and robust: you can use it in both development and production environments, handling both batch and streaming data effectively. The same code can be used for local development, CI/CD tests, running batch jobs, handling stream replays, and processing data streams. Pathway is powered by a scalable Rust engine based on Differential Dataflow and performs incremental computation. Your Pathway code, despite being written in Python, is run by the Rust engine, enabling multithreading, multiprocessing, and distributed computations. All the pipeline is kept in memory and can be easily deployed with Docker and Kubernetes.
  • 43
    DataStax

    DataStax

    DataStax

    The Open, Multi-Cloud Stack for Modern Data Apps. Built on open-source Apache Cassandra™. Global-scale and 100% uptime without vendor lock-in. Deploy on multi-cloud, on-prem, open-source, and Kubernetes. Elastic and pay-as-you-go for improved TCO. Start building faster with Stargate APIs for NoSQL, real-time, reactive, JSON, REST, and GraphQL. Skip the complexity of multiple OSS projects and APIs that don’t scale. Ideal for commerce, mobile, AI/ML, IoT, microservices, social, gaming, and richly interactive applications that must scale-up and scale-down with demand. Get building modern data applications with Astra, a database-as-a-service powered by Apache Cassandra™. Use REST, GraphQL, JSON with your favorite full-stack framework Richly interactive apps that are elastic and viral-ready from Day 1. Pay-as-you-go Apache Cassandra DBaaS that scales effortlessly and affordably.
  • 44
    IBM Streams
    IBM Streams evaluates a broad range of streaming data — unstructured text, video, audio, geospatial and sensor — helping organizations spot opportunities and risks and make decisions in real-time. Make sense of your data, turning fast-moving volumes and varieties into insight with IBM® Streams. Streams evaluate a broad range of streaming data — unstructured text, video, audio, geospatial and sensor — helping organizations spot opportunities and risks as they happen. Combine Streams with other IBM Cloud Pak® for Data capabilities, built on an open, extensible architecture. Help enable data scientists to collaboratively build models to apply to stream flows, plus, analyze massive amounts of data in real-time. Acting upon your data and deriving true value is easier than ever.
  • 45
    Azure Data Explorer
    Azure Data Explorer is a fast, fully managed data analytics service for real-time analysis on large volumes of data streaming from applications, websites, IoT devices, and more. Ask questions and iteratively explore data on the fly to improve products, enhance customer experiences, monitor devices, and boost operations. Quickly identify patterns, anomalies, and trends in your data. Explore new questions and get answers in minutes. Run as many queries as you need, thanks to the optimized cost structure. Explore new possibilities with your data cost-effectively. Focus on insights, not infrastructure, with the easy-to-use, fully managed data analytics service. Respond quickly to fast-flowing and rapidly changing data. Azure Data Explorer simplifies analytics from all forms of streaming data.
    Starting Price: $0.11 per hour
  • 46
    Evam Continuous Intelligence Platform
    Evam's Continuous Intelligence Platform combines multiple products for processing and visualizing real-time data. It runs real-time machine learning models on streaming data, while enriching the real-time data with a smart in-memory caching mechanism. EVAM empowers telecommunications, financial services, retail, transportation and travel companies to maximize their business value. Through continuous intelligence platform with machine learning capabilities. EVAM processes real-time data and designs and orchestrates customer journeys visually with advanced analytical models, machine learning, and artificial intelligence algorithms. EVAM enables enterprises to engage their customers using their data across all channels, including legacy ones, in real-time. Collect billions of events and process them in real-time. Understand each customer's needs and attract, engage, and retain them more effectively.
  • 47
    Kinetica

    Kinetica

    Kinetica

    A scalable cloud database for real-time analysis on large and streaming datasets. Kinetica is designed to harness modern vectorized processors to be orders of magnitude faster and more efficient for real-time spatial and temporal workloads. Track and gain intelligence from billions of moving objects in real-time. Vectorization unlocks new levels of performance for analytics on spatial and time series data at scale. Ingest and query at the same time to act on real-time events. Kinetica's lockless architecture and distributed ingestion ensures data is available to query as soon as it lands. Vectorized processing enables you to do more with less. More power allows for simpler data structures, which lead to lower storage costs, more flexibility and less time engineering your data. Vectorized processing opens the door to amazingly fast analytics and detailed visualization of moving objects at scale.
  • 48
    Gathr

    Gathr

    Gathr

    The only all-in-one data pipeline platform. Built ground-up for a cloud-first world, Gathr is the only platform to handle all your data integration and engineering needs - ingestion, ETL, ELT, CDC, streaming analytics, data preparation, machine learning, advanced analytics and more. With Gathr, anyone can build and deploy pipelines in minutes, irrespective of skill levels. Create Ingestion pipelines in minutes, not weeks. Ingest data from any source, deliver to any destination. Build applications quickly with a wizard-based approach. Replicate data in real-time using a templatized CDC app. Native integration for all sources and targets. Best-in-class capabilities with everything you need to succeed today and tomorrow. Choose between free, pay-per-use or customize as per your requirements.
  • 49
    CipherTrust Data Security Platform
    Thales has pushed the innovation envelope with the CipherTrust Data Security Platform to remove complexity from data security, accelerate time to compliance, and secure cloud migrations. This next-generation platform is built on a modern micro-services architecture, is designed for the cloud, includes Data Discovery and Classification, and fuses together the best capabilities from the Vormetric Data Security Platform and KeySecure and connector products. CipherTrust Data Security Platform unifies data discovery, classification, data protection, and unprecedented granular access controls with centralized key management – all on a single platform. This results in less resources dedicated to data security operations, ubiquitous compliance controls, and significantly reduced risk across your business. The CipherTrust Data Security Platform is an integrated suite of data-centric security products and solutions that unify data discovery, protection and control in one platform.
  • 50
    KX Insights
    KX Insights is a cloud-native platform for critical real-time performance and continuous actionable intelligence. Using complex event processing, high-speed analytics and machine learning interfaces, it enables fast decision-making and automated responses to events in fractions of a second. It’s not just storage and compute elasticity that have moved to the cloud. It’s everything: data, tools, development, security, connectivity, operations, maintenance. KX can help you leverage that power to make smarter, more insightful decisions by integrating real-time analytics into your business operations. KX Insights leverages industry standards to ensure openness and interoperability with other technologies in order to deliver insights faster and more cost-effectively. It operates a microservices-based architecture for capturing, storing and processing high-volume, high-velocity data using cloud standards, services, and protocols.