Alternatives to Azure Databricks

Compare Azure Databricks alternatives for your business or organization using the curated list below. SourceForge ranks the best alternatives to Azure Databricks in 2024. Compare features, ratings, user reviews, pricing, and more from Azure Databricks competitors and alternatives in order to make an informed decision for your business.

  • 1
    Google Cloud Platform
    Google Cloud is a cloud-based service that allows you to create anything from simple websites to complex applications for businesses of all sizes. New customers get $300 in free credits to run, test, and deploy workloads. All customers can use 25+ products for free, up to monthly usage limits. Use Google's core infrastructure, data analytics & machine learning. Secure and fully featured for all enterprises. Tap into big data to find answers faster and build better products. Grow from prototype to production to planet-scale, without having to think about capacity, reliability or performance. From virtual machines with proven price/performance advantages to a fully managed app development platform. Scalable, resilient, high performance object storage and databases for your applications. State-of-the-art software-defined networking products on Google’s private fiber network. Fully managed data warehousing, batch and stream processing, data exploration, Hadoop/Spark, and messaging.
    Leader badge
    Compare vs. Azure Databricks View Software
    Visit Website
  • 2
    FlowWright
    Business Process Management Software (BPMS) & BPM Workflow Automation Tool. Companies need workflow, forms, compliance, and automation routing support. Our low-code options make creating + editing workflows simple. Our best-in-class forms capabilities, make it possible to rapidly build forms, forms logic, and workflows for forms-driven workflow processes. Companies have many existing systems in place that need to work with each other. Our business process integrations across systems are loosely-coupled + intelligently integrated. When you use FlowWright to automate your business, you gain access to standard metrics and metrics that you define. BPM analytics are a key part of any BPM workflow management software solution. FlowWright can be deployed as a cloud solution or deployed in an on-premise or .NET hosted environment (including AWS and Azure). It was built in .NET Foundation C# code and all tools are fully browser-based, requiring no plug-ins.
  • 3
    Azure Data Explorer
    Azure Data Explorer is a fast, fully managed data analytics service for real-time analysis on large volumes of data streaming from applications, websites, IoT devices, and more. Ask questions and iteratively explore data on the fly to improve products, enhance customer experiences, monitor devices, and boost operations. Quickly identify patterns, anomalies, and trends in your data. Explore new questions and get answers in minutes. Run as many queries as you need, thanks to the optimized cost structure. Explore new possibilities with your data cost-effectively. Focus on insights, not infrastructure, with the easy-to-use, fully managed data analytics service. Respond quickly to fast-flowing and rapidly changing data. Azure Data Explorer simplifies analytics from all forms of streaming data.
    Starting Price: $0.11 per hour
  • 4
    Databricks Data Intelligence Platform
    The Databricks Data Intelligence Platform allows your entire organization to use data and AI. It’s built on a lakehouse to provide an open, unified foundation for all data and governance, and is powered by a Data Intelligence Engine that understands the uniqueness of your data. The winners in every industry will be data and AI companies. From ETL to data warehousing to generative AI, Databricks helps you simplify and accelerate your data and AI goals. Databricks combines generative AI with the unification benefits of a lakehouse to power a Data Intelligence Engine that understands the unique semantics of your data. This allows the Databricks Platform to automatically optimize performance and manage infrastructure in ways unique to your business. The Data Intelligence Engine understands your organization’s language, so search and discovery of new data is as easy as asking a question like you would to a coworker.
  • 5
    Apache Spark

    Apache Spark

    Apache Software Foundation

    Apache Spark™ is a unified analytics engine for large-scale data processing. Apache Spark achieves high performance for both batch and streaming data, using a state-of-the-art DAG scheduler, a query optimizer, and a physical execution engine. Spark offers over 80 high-level operators that make it easy to build parallel apps. And you can use it interactively from the Scala, Python, R, and SQL shells. Spark powers a stack of libraries including SQL and DataFrames, MLlib for machine learning, GraphX, and Spark Streaming. You can combine these libraries seamlessly in the same application. Spark runs on Hadoop, Apache Mesos, Kubernetes, standalone, or in the cloud. It can access diverse data sources. You can run Spark using its standalone cluster mode, on EC2, on Hadoop YARN, on Mesos, or on Kubernetes. Access data in HDFS, Alluxio, Apache Cassandra, Apache HBase, Apache Hive, and hundreds of other data sources.
  • 6
    Amazon EMR
    Amazon EMR is the industry-leading cloud big data platform for processing vast amounts of data using open-source tools such as Apache Spark, Apache Hive, Apache HBase, Apache Flink, Apache Hudi, and Presto. With EMR you can run Petabyte-scale analysis at less than half of the cost of traditional on-premises solutions and over 3x faster than standard Apache Spark. For short-running jobs, you can spin up and spin down clusters and pay per second for the instances used. For long-running workloads, you can create highly available clusters that automatically scale to meet demand. If you have existing on-premises deployments of open-source tools such as Apache Spark and Apache Hive, you can also run EMR clusters on AWS Outposts. Analyze data using open-source ML frameworks such as Apache Spark MLlib, TensorFlow, and Apache MXNet. Connect to Amazon SageMaker Studio for large-scale model training, analysis, and reporting.
  • 7
    Google Cloud Dataproc
    Dataproc makes open source data and analytics processing fast, easy, and more secure in the cloud. Build custom OSS clusters on custom machines faster. Whether you need extra memory for Presto or GPUs for Apache Spark machine learning, Dataproc can help accelerate your data and analytics processing by spinning up a purpose-built cluster in 90 seconds. Easy and affordable cluster management. With autoscaling, idle cluster deletion, per-second pricing, and more, Dataproc can help reduce the total cost of ownership of OSS so you can focus your time and resources elsewhere. Security built in by default. Encryption by default helps ensure no piece of data is unprotected. With JobsAPI and Component Gateway, you can define permissions for Cloud IAM clusters, without having to set up networking or gateway nodes.
  • 8
    Azure HDInsight
    Run popular open-source frameworks—including Apache Hadoop, Spark, Hive, Kafka, and more—using Azure HDInsight, a customizable, enterprise-grade service for open-source analytics. Effortlessly process massive amounts of data and get all the benefits of the broad open-source project ecosystem with the global scale of Azure. Easily migrate your big data workloads and processing to the cloud. Open-source projects and clusters are easy to spin up quickly without the need to install hardware or manage infrastructure. Big data clusters reduce costs through autoscaling and pricing tiers that allow you to pay for only what you use. Enterprise-grade security and industry-leading compliance with more than 30 certifications helps protect your data. Optimized components for open-source technologies such as Hadoop and Spark keep you up to date.
  • 9
    Horovod

    Horovod

    Horovod

    Horovod was originally developed by Uber to make distributed deep learning fast and easy to use, bringing model training time down from days and weeks to hours and minutes. With Horovod, an existing training script can be scaled up to run on hundreds of GPUs in just a few lines of Python code. Horovod can be installed on-premise or run out-of-the-box in cloud platforms, including AWS, Azure, and Databricks. Horovod can additionally run on top of Apache Spark, making it possible to unify data processing and model training into a single pipeline. Once Horovod has been configured, the same infrastructure can be used to train models with any framework, making it easy to switch between TensorFlow, PyTorch, MXNet, and future frameworks as machine learning tech stacks continue to evolve.
  • 10
    E-MapReduce
    EMR is an all-in-one enterprise-ready big data platform that provides cluster, job, and data management services based on open-source ecosystems, such as Hadoop, Spark, Kafka, Flink, and Storm. Alibaba Cloud Elastic MapReduce (EMR) is a big data processing solution that runs on the Alibaba Cloud platform. EMR is built on Alibaba Cloud ECS instances and is based on open-source Apache Hadoop and Apache Spark. EMR allows you to use the Hadoop and Spark ecosystem components, such as Apache Hive, Apache Kafka, Flink, Druid, and TensorFlow, to analyze and process data. You can use EMR to process data stored on different Alibaba Cloud data storage service, such as Object Storage Service (OSS), Log Service (SLS), and Relational Database Service (RDS). You can quickly create clusters without the need to configure hardware and software. All maintenance operations are completed on its Web interface.
  • 11
    Privacera

    Privacera

    Privacera

    At the intersection of data governance, privacy, and security, Privacera’s unified data access governance platform maximizes the value of data by providing secure data access control and governance across hybrid- and multi-cloud environments. The hybrid platform centralizes access and natively enforces policies across multiple cloud services—AWS, Azure, Google Cloud, Databricks, Snowflake, Starburst and more—to democratize trusted data enterprise-wide without compromising compliance with regulations such as GDPR, CCPA, LGPD, or HIPAA. Trusted by Fortune 500 customers across finance, insurance, retail, healthcare, media, public and the federal sector, Privacera is the industry’s leading data access governance platform that delivers unmatched scalability, elasticity, and performance. Headquartered in Fremont, California, Privacera was founded in 2016 to manage cloud data privacy and security by the creators of Apache Ranger™ and Apache Atlas™.
  • 12
    WarpStream

    WarpStream

    WarpStream

    WarpStream is an Apache Kafka-compatible data streaming platform built directly on top of object storage, with no inter-AZ networking costs, no disks to manage, and infinitely scalable, all within your VPC. WarpStream is deployed as a stateless and auto-scaling agent binary in your VPC with no local disks to manage. Agents stream data directly to and from object storage with no buffering on local disks and no data tiering. Create new “virtual clusters” in our control plane instantly. Support different environments, teams, or projects without managing any dedicated infrastructure. WarpStream is protocol compatible with Apache Kafka, so you can keep using all your favorite tools and software. No need to rewrite your application or use a proprietary SDK. Just change the URL in your favorite Kafka client library and start streaming. Never again have to choose between reliability and your budget.
    Starting Price: $2,987 per month
  • 13
    GeoSpock

    GeoSpock

    GeoSpock

    GeoSpock enables data fusion for the connected world with GeoSpock DB – the space-time analytics database. GeoSpock DB is a unique, cloud-native database optimised for querying for real-world use cases, able to fuse multiple sources of Internet of Things (IoT) data together to unlock its full value, whilst simultaneously reducing complexity and cost. GeoSpock DB enables efficient storage, data fusion, and rapid programmatic access to data, and allows you to run ANSI SQL queries and connect to analytics tools via JDBC/ODBC connectors. Users are able to perform analysis and share insights using familiar toolsets, with support for common BI tools (such as Tableau™, Amazon QuickSight™, and Microsoft Power BI™), and Data Science and Machine Learning environments (including Python Notebooks and Apache Spark). The database can also be integrated with internal applications and web services – with compatibility for open-source and visualisation libraries such as Kepler and Cesium.js.
  • 14
    Saturn Cloud

    Saturn Cloud

    Saturn Cloud

    Saturn Cloud is an award-winning ML platform for any cloud with 100,000+ users, including NVIDIA, CFA Institute, Snowflake, Flatiron School, Nestle, and more. It is an all-in-one solution for data science & ML development, deployment, and data pipelines in the cloud. Users can spin up a notebook with 4TB of RAM, add a GPU, connect to a distributed cluster of workers, build large language models, and more in a completely hosted environment. Data professionals can use your preferred languages, IDEs, and machine-learning libraries in Saturn Cloud. We offer full Git integration, shared custom images, and secure credential storage, making scaling and building your team in the cloud easy. We support the entire machine learning lifecycle from experimentation to production with features like jobs and deployments. These features and built-in tools are easily shareable within teams, so time is saved and work is reproducible.
    Leader badge
    Starting Price: $0.005 per GB per hour
  • 15
    IBM Watson Studio
    Build, run and manage AI models, and optimize decisions at scale across any cloud. IBM Watson Studio empowers you to operationalize AI anywhere as part of IBM Cloud Pak® for Data, the IBM data and AI platform. Unite teams, simplify AI lifecycle management and accelerate time to value with an open, flexible multicloud architecture. Automate AI lifecycles with ModelOps pipelines. Speed data science development with AutoAI. Prepare and build models visually and programmatically. Deploy and run models through one-click integration. Promote AI governance with fair, explainable AI. Drive better business outcomes by optimizing decisions. Use open source frameworks like PyTorch, TensorFlow and scikit-learn. Bring together the development tools including popular IDEs, Jupyter notebooks, JupterLab and CLIs — or languages such as Python, R and Scala. IBM Watson Studio helps you build and scale AI with trust and transparency by automating AI lifecycle management.
  • 16
    Dataiku DSS
    Bring data analysts, engineers, and scientists together. Enable self-service analytics and operationalize machine learning. Get results today and build for tomorrow. Dataiku DSS is the collaborative data science software platform for teams of data scientists, data analysts, and engineers to explore, prototype, build, and deliver their own data products more efficiently. Use notebooks (Python, R, Spark, Scala, Hive, etc.) or a customizable drag-and-drop visual interface at any step of the predictive dataflow prototyping process – from wrangling to analysis to modeling. Profile the data visually at every step of the analysis. Interactively explore and chart your data using 25+ built-in charts. Prepare, enrich, blend, and clean data using 80+ built-in functions. Leverage Machine Learning technologies (Scikit-Learn, MLlib, TensorFlow, Keras, etc.) in a visual UI. Build & optimize models in Python or R and integrate any external ML library through code APIs.
  • 17
    Oracle Cloud Infrastructure Data Flow
    Oracle Cloud Infrastructure (OCI) Data Flow is a fully managed Apache Spark service to perform processing tasks on extremely large data sets without infrastructure to deploy or manage. This enables rapid application delivery because developers can focus on app development, not infrastructure management. OCI Data Flow handles infrastructure provisioning, network setup, and teardown when Spark jobs are complete. Storage and security are also managed, which means less work is required for creating and managing Spark applications for big data analysis. With OCI Data Flow, there are no clusters to install, patch, or upgrade, which saves time and operational costs for projects. OCI Data Flow runs each Spark job in private dedicated resources, eliminating the need for upfront capacity planning. With OCI Data Flow, IT only needs to pay for the infrastructure resources that Spark jobs use while they are running.
    Starting Price: $0.0085 per GB per hour
  • 18
    Oracle Big Data Service
    Oracle Big Data Service makes it easy for customers to deploy Hadoop clusters of all sizes, with VM shapes ranging from 1 OCPU to a dedicated bare metal environment. Customers choose between high-performance NVmE storage or cost-effective block storage, and can grow or shrink their clusters. Quickly create Hadoop-based data lakes to extend or complement customer data warehouses, and ensure that all data is both accessible and managed cost-effectively. Query, visualize and transform data so data scientists can build machine learning models using the included notebook with its R, Python and SQL support. Move customer-managed Hadoop clusters to a fully-managed cloud-based service, reducing management costs and improving resource utilization.
    Starting Price: $0.1344 per hour
  • 19
    Spark Streaming

    Spark Streaming

    Apache Software Foundation

    Spark Streaming brings Apache Spark's language-integrated API to stream processing, letting you write streaming jobs the same way you write batch jobs. It supports Java, Scala and Python. Spark Streaming recovers both lost work and operator state (e.g. sliding windows) out of the box, without any extra code on your part. By running on Spark, Spark Streaming lets you reuse the same code for batch processing, join streams against historical data, or run ad-hoc queries on stream state. Build powerful interactive applications, not just analytics. Spark Streaming is developed as part of Apache Spark. It thus gets tested and updated with each Spark release. You can run Spark Streaming on Spark's standalone cluster mode or other supported cluster resource managers. It also includes a local run mode for development. In production, Spark Streaming uses ZooKeeper and HDFS for high availability.
  • 20
    Google Cloud Deep Learning VM Image
    Provision a VM quickly with everything you need to get your deep learning project started on Google Cloud. Deep Learning VM Image makes it easy and fast to instantiate a VM image containing the most popular AI frameworks on a Google Compute Engine instance without worrying about software compatibility. You can launch Compute Engine instances pre-installed with TensorFlow, PyTorch, scikit-learn, and more. You can also easily add Cloud GPU and Cloud TPU support. Deep Learning VM Image supports the most popular and latest machine learning frameworks, like TensorFlow and PyTorch. To accelerate your model training and deployment, Deep Learning VM Images are optimized with the latest NVIDIA® CUDA-X AI libraries and drivers and the Intel® Math Kernel Library. Get started immediately with all the required frameworks, libraries, and drivers pre-installed and tested for compatibility. Deep Learning VM Image delivers a seamless notebook experience with integrated support for JupyterLab.
  • 21
    Azure Data Lake Analytics
    Easily develop and run massively parallel data transformation and processing programs in U-SQL, R, Python, and .NET over petabytes of data. With no infrastructure to manage, you can process data on demand, scale instantly, and only pay per job. Process big data jobs in seconds with Azure Data Lake Analytics. There is no infrastructure to worry about because there are no servers, virtual machines, or clusters to wait for, manage, or tune. Instantly scale the processing power, measured in Azure Data Lake Analytics Units (AU), from one to thousands for each job. You only pay for the processing that you use per job. Act on all of your data with optimized data virtualization of your relational sources such as Azure SQL Database and Azure Synapse Analytics. Your queries are automatically optimized by moving processing close to the source data without data movement, which maximizes performance and minimizes latency.
    Starting Price: $2 per hour
  • 22
    Hadoop

    Hadoop

    Apache Software Foundation

    The Apache Hadoop software library is a framework that allows for the distributed processing of large data sets across clusters of computers using simple programming models. It is designed to scale up from single servers to thousands of machines, each offering local computation and storage. Rather than rely on hardware to deliver high-availability, the library itself is designed to detect and handle failures at the application layer, so delivering a highly-available service on top of a cluster of computers, each of which may be prone to failures. A wide variety of companies and organizations use Hadoop for both research and production. Users are encouraged to add themselves to the Hadoop PoweredBy wiki page. Apache Hadoop 3.3.4 incorporates a number of significant enhancements over the previous major release line (hadoop-3.2).
  • 23
    IBM Watson Machine Learning
    IBM Watson Machine Learning is a full-service IBM Cloud offering that makes it easy for developers and data scientists to work together to integrate predictive capabilities with their applications. The Machine Learning service is a set of REST APIs that you can call from any programming language to develop applications that make smarter decisions, solve tough problems, and improve user outcomes. Take advantage of machine learning models management (continuous learning system) and deployment (online, batch, streaming). Select any of widely supported machine learning frameworks: TensorFlow, Keras, Caffe, PyTorch, Spark MLlib, scikit learn, xgboost and SPSS. Use the command-line interface and Python client to manage your artifacts. Extend your application with artificial intelligence through the Watson Machine Learning REST API.
    Starting Price: $0.575 per hour
  • 24
    Apache Arrow

    Apache Arrow

    The Apache Software Foundation

    Apache Arrow defines a language-independent columnar memory format for flat and hierarchical data, organized for efficient analytic operations on modern hardware like CPUs and GPUs. The Arrow memory format also supports zero-copy reads for lightning-fast data access without serialization overhead. Arrow's libraries implement the format and provide building blocks for a range of use cases, including high performance analytics. Many popular projects use Arrow to ship columnar data efficiently or as the basis for analytic engines. Apache Arrow is software created by and for the developer community. We are dedicated to open, kind communication and consensus decisionmaking. Our committers come from a range of organizations and backgrounds, and we welcome all to participate with us.
  • 25
    Instaclustr

    Instaclustr

    Instaclustr

    Instaclustr is the Open Source-as-a-Service company, delivering reliability at scale. We operate an automated, proven, and trusted managed environment, providing database, analytics, search, and messaging. We enable companies to focus internal development and operational resources on building cutting edge customer-facing applications. Instaclustr works with cloud providers including AWS, Heroku, Azure, IBM Cloud, and Google Cloud Platform. The company has SOC 2 certification and provides 24/7 customer support.
    Starting Price: $20 per node per month
  • 26
    Deeplearning4j

    Deeplearning4j

    Deeplearning4j

    DL4J takes advantage of the latest distributed computing frameworks including Apache Spark and Hadoop to accelerate training. On multi-GPUs, it is equal to Caffe in performance. The libraries are completely open-source, Apache 2.0, and maintained by the developer community and Konduit team. Deeplearning4j is written in Java and is compatible with any JVM language, such as Scala, Clojure, or Kotlin. The underlying computations are written in C, C++, and Cuda. Keras will serve as the Python API. Eclipse Deeplearning4j is the first commercial-grade, open-source, distributed deep-learning library written for Java and Scala. Integrated with Hadoop and Apache Spark, DL4J brings AI to business environments for use on distributed GPUs and CPUs. There are a lot of parameters to adjust when you're training a deep-learning network. We've done our best to explain them, so that Deeplearning4j can serve as a DIY tool for Java, Scala, Clojure, and Kotlin programmers.
  • 27
    GraphDB

    GraphDB

    Ontotext

    *GraphDB allows you to link diverse data, index it for semantic search and enrich it via text analysis to build big knowledge graphs.* GraphDB is a highly efficient and robust graph database with RDF and SPARQL support. The GraphDB database supports a highly available replication cluster, which has been proven in a number of enterprise use cases that required resilience in data loading and query answering. If you need a quick overview of GraphDB or a download link to its latest releases, please visit the GraphDB product section. GraphDB uses RDF4J as a library, utilizing its APIs for storage and querying, as well as the support for a wide variety of query languages (e.g., SPARQL and SeRQL) and RDF syntaxes (e.g., RDF/XML, N3, Turtle).
  • 28
    Hydrolix

    Hydrolix

    Hydrolix

    Hydrolix is a streaming data lake that combines decoupled storage, indexed search, and stream processing to deliver real-time query performance at terabyte-scale for a radically lower cost. CFOs love the 4x reduction in data retention costs. Product teams love 4x more data to work with. Spin up resources when you need them and scale to zero when you don’t. Fine-tune resource consumption and performance by workload to control costs. Imagine what you can build when you don’t have to sacrifice data because of budget. Ingest, enrich, and transform log data from multiple sources including Kafka, Kinesis, and HTTP. Return just the data you need, no matter how big your data is. Reduce latency and costs, eliminate timeouts, and brute force queries. Storage is decoupled from ingest and query, allowing each to independently scale to meet performance and budget targets. Hydrolix’s high-density compression (HDX) typically reduces 1TB of stored data to 55GB.
    Starting Price: $2,237 per month
  • 29
    Arcadia Data

    Arcadia Data

    Arcadia Data

    Arcadia Data provides the first visual analytics and BI platform native to Hadoop and cloud (big data) that delivers the scale, performance, and agility business users need for both real-time and historical insights. Its flagship product, Arcadia Enterprise, was built from inception for big data platforms such as Apache Hadoop, Apache Spark, Apache Kafka, and Apache Solr, in the cloud and/or on-premises. Using artificial intelligence (AI) and machine learning (ML), Arcadia Enterprise streamlines the self-service analytics process with search-based BI and visualization recommendations. It enables real-time, high-definition insights in use cases like data lakes, cybersecurity, connected IoT devices, and customer intelligence. Arcadia Enterprise is deployed by some of the world’s leading brands, including Procter & Gamble, Citibank, Nokia, Royal Bank of Canada, Kaiser Permanente, HPE, and Neustar.
  • 30
    Deepnote

    Deepnote

    Deepnote

    Deepnote is building the best data science notebook for teams. In the notebook, users can connect their data, explore, and analyze it with real-time collaboration and version control. Users can easily share project links with team collaborators, or with end-users to present polished assets. All of this is done through a powerful, browser-based UI that runs in the cloud. We built Deepnote because data scientists don't work alone. Features: - Sharing notebooks and projects via URL - Inviting others to view, comment and collaborate, with version control - Publishing notebooks with visualizations for presentations - Sharing datasets between projects - Set team permissions to decide who can edit vs view code - Full linux terminal access - Code completion - Automatic python package management - Importing from github - PostgreSQL DB connection
  • 31
    IntelliHub

    IntelliHub

    Spotflock

    We work closely with businesses to find out what are the common issues preventing companies from realising benefits. We design to open up opportunities that were previously not viable using conventional approaches Corporations -big and small, require an AI platform with complete empowerment and ownership. Tackle data privacy and adopt to AI platforms at a sustainable cost. Enhance the efficiency of businesses and augment the work humans do. We apply AI to gain control over repetitive or dangerous tasks and bypass human intervention, thereby expediting tasks with creativity and empathy. Machine Learning helps to give predictive capabilities to applications with ease. You can build classification and regression models. It can also do clustering and visualize different clusters. It supports multiple ML libraries like Weka, Scikit-Learn, H2O and Tensorflow. It includes around 22 different algorithms for building classification, regression and clustering models.
  • 32
    Delta Lake

    Delta Lake

    Delta Lake

    Delta Lake is an open-source storage layer that brings ACID transactions to Apache Spark™ and big data workloads. Data lakes typically have multiple data pipelines reading and writing data concurrently, and data engineers have to go through a tedious process to ensure data integrity, due to the lack of transactions. Delta Lake brings ACID transactions to your data lakes. It provides serializability, the strongest level of isolation level. Learn more at Diving into Delta Lake: Unpacking the Transaction Log. In big data, even the metadata itself can be "big data". Delta Lake treats metadata just like data, leveraging Spark's distributed processing power to handle all its metadata. As a result, Delta Lake can handle petabyte-scale tables with billions of partitions and files at ease. Delta Lake provides snapshots of data enabling developers to access and revert to earlier versions of data for audits, rollbacks or to reproduce experiments.
  • 33
    MotherDuck

    MotherDuck

    MotherDuck

    We’re MotherDuck, a software company founded by a passionate flock of experienced data geeks. We’ve worked as leaders for some of the greatest companies in data. Scale-out is expensive and slow, let’s scale up. Big Data is dead, long live easy data. Your laptop is faster than your data warehouse. Why wait for the cloud? DuckDB slaps, so let’s supercharge it. When we founded MotherDuck we recognized that DuckDB might just be the next major game changer thanks to its ease of use, portability, lightning-fast performance, and rapid pace of community-driven innovation. At MotherDuck, we want to help the community, the DuckDB Foundation, and DuckDB Labs build greater awareness and adoption of DuckDB, whether users are working locally or want a serverless always-on way to execute their SQL. We are a world-class team of engineers and leaders with experience working on databases and cloud services at AWS, Databricks, Elastic, Facebook, Firebolt, Google BigQuery, Neo4j, SingleStore, and more.
  • 34
    Tabular

    Tabular

    Tabular

    Tabular is an open table store from the creators of Apache Iceberg. Connect multiple computing engines and frameworks. Decrease query time and storage costs by up to 50%. Centralize enforcement of data access (RBAC) policies. Connect any query engine or framework, including Athena, BigQuery, Redshift, Snowflake, Databricks, Trino, Spark, and Python. Smart compaction, clustering, and other automated data services reduce storage costs and query times by up to 50%. Unify data access at the database or table. RBAC controls are simple to manage, consistently enforced, and easy to audit. Centralize your security down to the table. Tabular is easy to use plus it features high-powered ingestion, performance, and RBAC under the hood. Tabular gives you the flexibility to work with multiple “best of breed” compute engines based on their strengths. Assign privileges at the data warehouse database, table, or column level.
    Starting Price: $100 per month
  • 35
    Apache Mahout

    Apache Mahout

    Apache Software Foundation

    Apache Mahout is a powerful, scalable, and versatile machine learning library designed for distributed data processing. It offers a comprehensive set of algorithms for various tasks, including classification, clustering, recommendation, and pattern mining. Built on top of the Apache Hadoop ecosystem, Mahout leverages MapReduce and Spark to enable data processing on large-scale datasets. Apache Mahout(TM) is a distributed linear algebra framework and mathematically expressive Scala DSL designed to let mathematicians, statisticians, and data scientists quickly implement their own algorithms. Apache Spark is the recommended out-of-the-box distributed back-end or can be extended to other distributed backends. Matrix computations are a fundamental part of many scientific and engineering applications, including machine learning, computer vision, and data analysis. Apache Mahout is designed to handle large-scale data processing by leveraging the power of Hadoop and Spark.
  • 36
    GPUonCLOUD

    GPUonCLOUD

    GPUonCLOUD

    Traditionally, deep learning, 3D modeling, simulations, distributed analytics, and molecular modeling take days or weeks time. However, with GPUonCLOUD’s dedicated GPU servers, it's a matter of hours. You may want to opt for pre-configured systems or pre-built instances with GPUs featuring deep learning frameworks like TensorFlow, PyTorch, MXNet, TensorRT, libraries e.g. real-time computer vision library OpenCV, thereby accelerating your AI/ML model-building experience. Among the wide variety of GPUs available to us, some of the GPU servers are best fit for graphics workstations and multi-player accelerated gaming. Instant jumpstart frameworks increase the speed and agility of the AI/ML environment with effective and efficient environment lifecycle management.
    Starting Price: $1 per hour
  • 37
    SigView

    SigView

    Sigmoid

    Get access to granular data for effortless slice & dice on billions of rows, and ensure real-time reporting in seconds! Sigview is a plug-n-play real-time data analytics tool by Sigmoid to carry exploratory data analysis. Custom built on Apache Spark, Sigview is capable of drilling down into massive data sets within a few seconds. Used by around 30k users across the globe to analyze billions of ad impressions, Sigview is designed to give real-time access to your Programmatic and non-programmatic data by analyzing enormous data sets while creating real-time reports. Whether it is optimizing your ad campaigns or discovering new inventory or generating revenue opportunities with changing times, Sigview is your go-to platform for all your reporting needs. Connects to multiple data sources like DFP, Pixel Servers, Audience and viewability partners to ingest data in any format and location maintaining data latency of less than 15 minutes.
  • 38
    EspressReport ES

    EspressReport ES

    Quadbase Systems

    EspressRepot ES (Enterprise Server) is a web and desktop-based software that allows users to develop stunning and interactive data visualization and reporting. The platform offers full Java EE integration, to draw data from data sources such as Bid Data (Hadoop, Spark, and MongoDB), ad-hoc queries and reports, online map support, mobile compatibility, alert monitor, and many other amazing features.
  • 39
    IBM Analytics Engine
    IBM Analytics Engine provides an architecture for Hadoop clusters that decouples the compute and storage tiers. Instead of a permanent cluster formed of dual-purpose nodes, the Analytics Engine allows users to store data in an object storage layer such as IBM Cloud Object Storage and spins up clusters of computing notes when needed. Separating compute from storage helps to transform the flexibility, scalability and maintainability of big data analytics platforms. Build on an ODPi compliant stack with pioneering data science tools with the broader Apache Hadoop and Apache Spark ecosystem. Define clusters based on your application's requirements. Choose the appropriate software pack, version, and size of the cluster. Use as long as required and delete as soon as an application finishes jobs. Configure clusters with third-party analytics libraries and packages. Deploy workloads from IBM Cloud services like machine learning.
    Starting Price: $0.014 per hour
  • 40
    scikit-learn

    scikit-learn

    scikit-learn

    Scikit-learn provides simple and efficient tools for predictive data analysis. Scikit-learn is a robust, open source machine learning library for the Python programming language, designed to provide simple and efficient tools for data analysis and modeling. Built on the foundations of popular scientific libraries like NumPy, SciPy, and Matplotlib, scikit-learn offers a wide range of supervised and unsupervised learning algorithms, making it an essential toolkit for data scientists, machine learning engineers, and researchers. The library is organized into a consistent and flexible framework, where various components can be combined and customized to suit specific needs. This modularity makes it easy for users to build complex pipelines, automate repetitive tasks, and integrate scikit-learn into larger machine-learning workflows. Additionally, the library’s emphasis on interoperability ensures that it works seamlessly with other Python libraries, facilitating smooth data processing.
  • 41
    Astro

    Astro

    Astronomer

    For data teams looking to increase the availability of trusted data, Astronomer provides Astro, a modern data orchestration platform, powered by Apache Airflow, that enables the entire data team to build, run, and observe data pipelines-as-code. Astronomer is the commercial developer of Airflow, the de facto standard for expressing data flows as code, used by hundreds of thousands of teams across the world.
  • 42
    AWS Deep Learning AMIs
    AWS Deep Learning AMIs (DLAMI) provides ML practitioners and researchers with a curated and secure set of frameworks, dependencies, and tools to accelerate deep learning in the cloud. Built for Amazon Linux and Ubuntu, Amazon Machine Images (AMIs) come preconfigured with TensorFlow, PyTorch, Apache MXNet, Chainer, Microsoft Cognitive Toolkit (CNTK), Gluon, Horovod, and Keras, allowing you to quickly deploy and run these frameworks and tools at scale. Develop advanced ML models at scale to develop autonomous vehicle (AV) technology safely by validating models with millions of supported virtual tests. Accelerate the installation and configuration of AWS instances, and speed up experimentation and evaluation with up-to-date frameworks and libraries, including Hugging Face Transformers. Use advanced analytics, ML, and deep learning capabilities to identify trends and make predictions from raw, disparate health data.
  • 43
    Tencent Cloud Elastic MapReduce
    EMR enables you to scale the managed Hadoop clusters manually or automatically according to your business curves or monitoring metrics. EMR's storage-computation separation even allows you to terminate a cluster to maximize resource efficiency. EMR supports hot failover for CBS-based nodes. It features a primary/secondary disaster recovery mechanism where the secondary node starts within seconds when the primary node fails, ensuring the high availability of big data services. The metadata of its components such as Hive supports remote disaster recovery. Computation-storage separation ensures high data persistence for COS data storage. EMR is equipped with a comprehensive monitoring system that helps you quickly identify and locate cluster exceptions to ensure stable cluster operations. VPCs provide a convenient network isolation method that facilitates your network policy planning for managed Hadoop clusters.
  • 44
    Azure Data Share
    Share data, in any format and any size, from multiple sources with other organizations. Easily control what you share, who receives your data, and the terms of use. Data Share provides full visibility into your data-sharing relationships with a user-friendly interface. Share data in just a few clicks, or build your own application using the REST API. Serverless code-free data-sharing service that requires no infrastructure setup or management. Intuitive interface to govern all your data-sharing relationships. Automated data-sharing processes for productivity and predictability. Secure data-sharing service that uses underlying Azure security measures. Share structured and unstructured data from multiple Azure data stores with other organizations in just a few clicks. There’s no infrastructure to set up or manage, no SAS keys are required, and sharing is all code-free. You control data access and set terms of use aligned with your enterprise policies.
    Starting Price: $0.05 per dataset-snapshot
  • 45
    Robin.io

    Robin.io

    Robin.io

    ROBIN is the industry’s first hyper-converged Kubernetes platform for big data, databases, and AI/ML. The platform provides a self-service App-store experience for the deployment of any application, anywhere – runs on-premises in your private data center or in public-cloud (AWS, Azure, GCP) environments. Hyper-converged Kubernetes is a software-defined application orchestration framework that combines containerized storage, networking, compute (Kubernetes), and the application management layer into a single system. Our approach extends Kubernetes for data-intensive applications such as Hortonworks, Cloudera, Elastic stack, RDBMS, NoSQL databases, and AI/ML apps. Facilitates simpler and faster roll-out of critical Enterprise IT and LoB initiatives, such as containerization, cloud-migration, cost-consolidation, and productivity improvement. Solves the fundamental challenges of running big data and databases in Kubernetes.
  • 46
    Deequ

    Deequ

    Deequ

    Deequ is a library built on top of Apache Spark for defining "unit tests for data", which measure data quality in large datasets. We are happy to receive feedback and contributions. Deequ depends on Java 8. Deequ version 2.x only runs with Spark 3.1, and vice versa. If you rely on a previous Spark version, please use a Deequ 1.x version (legacy version is maintained in legacy-spark-3.0 branch). We provide legacy releases compatible with Apache Spark versions 2.2.x to 3.0.x. The Spark 2.2.x and 2.3.x releases depend on Scala 2.11 and the Spark 2.4.x, 3.0.x, and 3.1.x releases depend on Scala 2.12. Deequ's purpose is to "unit-test" data to find errors early, before the data gets fed to consuming systems or machine learning algorithms. In the following, we will walk you through a toy example to showcase the most basic usage of our library.
  • 47
    Unravel

    Unravel

    Unravel Data

    Unravel makes data work anywhere: on Azure, AWS, GCP or in your own data center– Optimizing performance, automating troubleshooting and keeping costs in check. Unravel helps you monitor, manage, and improve your data pipelines in the cloud and on-premises – to drive more reliable performance in the applications that power your business. Get a unified view of your entire data stack. Unravel collects performance data from every platform, system, and application on any cloud then uses agentless technologies and machine learning to model your data pipelines from end to end. Explore, correlate, and analyze everything in your modern data and cloud environment. Unravel’s data model reveals dependencies, issues, and opportunities, how apps and resources are being used, what’s working and what’s not. Don’t just monitor performance – quickly troubleshoot and rapidly remediate issues. Leverage AI-powered recommendations to automate performance improvements, lower costs, and prepare.
  • 48
    PySpark

    PySpark

    PySpark

    PySpark is an interface for Apache Spark in Python. It not only allows you to write Spark applications using Python APIs, but also provides the PySpark shell for interactively analyzing your data in a distributed environment. PySpark supports most of Spark’s features such as Spark SQL, DataFrame, Streaming, MLlib (Machine Learning) and Spark Core. Spark SQL is a Spark module for structured data processing. It provides a programming abstraction called DataFrame and can also act as distributed SQL query engine. Running on top of Spark, the streaming feature in Apache Spark enables powerful interactive and analytical applications across both streaming and historical data, while inheriting Spark’s ease of use and fault tolerance characteristics.
  • 49
    Hopsworks

    Hopsworks

    Logical Clocks

    Hopsworks is an open-source Enterprise platform for the development and operation of Machine Learning (ML) pipelines at scale, based around the industry’s first Feature Store for ML. You can easily progress from data exploration and model development in Python using Jupyter notebooks and conda to running production quality end-to-end ML pipelines, without having to learn how to manage a Kubernetes cluster. Hopsworks can ingest data from the datasources you use. Whether they are in the cloud, on‑premise, IoT networks, or from your Industry 4.0-solution. Deploy on‑premises on your own hardware or at your preferred cloud provider. Hopsworks will provide the same user experience in the cloud or in the most secure of air‑gapped deployments. Learn how to set up customized alerts in Hopsworks for different events that are triggered as part of the ingestion pipeline.
    Starting Price: $1 per month
  • 50
    Apache Gobblin

    Apache Gobblin

    Apache Software Foundation

    A distributed data integration framework that simplifies common aspects of Big Data integration such as data ingestion, replication, organization, and lifecycle management for both streaming and batch data ecosystems. Runs as a standalone application on a single box. Also supports embedded mode. Runs as an mapreduce application on multiple Hadoop versions. Also supports Azkaban for launching mapreduce jobs. Runs as a standalone cluster with primary and worker nodes. This mode supports high availability and can run on bare metals as well. Runs as an elastic cluster on public cloud. This mode supports high availability. Gobblin as it exists today is a framework that can be used to build different data integration applications like ingest, replication, etc. Each of these applications is typically configured as a separate job and executed through a scheduler like Azkaban.