Compare the Top Big Data Platforms in China as of April 2026 - Page 6

  • 1
    Micropole

    Micropole

    Micropole

    Micropole is a consulting and engineering company, with bases in Europe and Asia, specializing in the creation of added-value. Micropole partners its customers in the Performance Management, Digital Transformation and Data Governance fields. At Micropole Group, we are convinced that optimizing companies' data assets is the key to their performance. Every day our Innovative People detect trends and explore new territories. Their mission is to make companies data intelligent and help them transform themselves to prepare their future. A privileged partner of major international software vendors, our ambition is to boost the distinctiveness of your corporation through efficient business solutions and innovative cutting-edge technologies. Micropole is a consulting, engineering and training company specialized in the development and integration of decision-support, Performance Management, Digital Transformation and Data Governance solutions.
  • 2
    Oracle MDM
    To fully understand master data management (MDM), we must first define and explain master data and differentiate between it and master data management. Master data is the critical business information that supports and classifies the transactional and analytical data of an enterprise. It may also be referred to as “enterprise data” or “metadata,” and often includes application-specific metadata, alternative business perspectives, corporate dimensions, reference data, and master data assets. Examples of enterprise data include chart of accounts, organization or cost-center structures, market segments, product categories, and more. Master data management (MDM) does exactly what the name implies—manages master data. MDM is the combination of applications and technologies that consolidates, cleanses, and augments this master data and synchronizes it with applications, business processes, and analytical tools.
  • 3
    Datalytics

    Datalytics

    Datalytics

    Our team of talents has the necessary creativity, wit, and implementation capacity to help organizations to analyze and understand data, and turn it into intelligent decisions. Datalytics features over 10 years of experience in data integration, visualization, and mining, Big Data, predictive analytics, and data science. Our main goal is to become organizations’ strategic allies helping them to recognize the actual possibilities to analyze data. We offer our wit and analytical creativity for organizations to be able to make intelligent decisions and create new business opportunities. In our century, information is the most relevant asset. These are the services we offer in order to understand, analyze, and transform data. Data architecture, training, and technical assessment are some of the tactics we implement to provide the Big Data service.
  • 4
    NextGen Population Health

    NextGen Population Health

    NextGen Healthcare

    Meet the challenges of value based care—no matter your current EHR. Get a clear view into your patient population with aggregated multi-source data and an easy-to-navigate visual display. Use insights based in data to better manage chronic conditions and care transitions, prevent illness, lower costs, and implement care management. Facilitate care coordination with tools that encourage a proactive approach, including a pre-visit dashboard, risk stratification, and automated tracking of admission, discharge, and transfer events. Put care management in operation. Extend physician reach. Foster critical interactions with patients and valuable follow-up in between appointments. Identify patients with the greatest risk for high-cost utilization, using the Johns Hopkins ACG system for risk stratification. Accurately assign resources where intervention is needed most. Improve performance on quality measures. Participate successfully in value-based payment programs and optimize reimbursement.
  • 5
    Arundo Enterprise
    Arundo Enterprise is a modular, flexible software suite to create data products for people. We connect live data to machine learning and other analytical models, and model outputs to business decisions. Arundo Edge Agent enables industrial connectivity and analytics in rugged, remote, or disconnected environments. Arundo Composer allows data scientists to quickly and easily deploy desktop-based analytical models into the Arundo Fabric cloud environment with a single command. Composer also enables companies to create and manage live data streams and integrate such streams with deployed data models. Arundo Fabric is the cloud-based hub for deployed machine learning models, data streams, edge agent management, and quick navigation to extended applications. Arundo offers a portfolio of high ROI SaaS products. Each of these solutions comes with a core out-of-the-box functional capability that leverages the core strengths of Arundo Enterprise.
  • 6
    Peak DSP

    Peak DSP

    Peak DSP (by Edge 226)

    Edge 226 is a global provider of data-driven tech solutions, focused on providing its clients with smart tools for quality and transparent user acquisition. Edge’s leading product is Peak DSP, a Performance-Driven DSP that enables programmatic buying for quality user acquisition and re-engagement. Peak DSP offers: • An AI-driven algorithm optimizing and predicting install & post-install events: registrations, subscriptions, purchases or any other action • Data-based targeting with Lookalike Audiences, External User Data and Audience Match • Direct integrations: ­ Owned & operated and direct apps ­ Mobile device manufacturers & carrier-based supply ­ Over 35 of the world’s top SSPs • All verticals and environments: Gaming, shopping, utilities, sports (etc.) campaigns across in-app, mobile web and desktop • Multiple creative types: ­ Rewarded video ­ Playable ads ­ Banners, native ads & text ads ­ HTML/Rich Media ­ JavaScript tags
  • 7
    MX

    MX

    MX Technologies

    MX helps financial institutions and fintechs utilize their data more effectively to outperform the competition in a rapidly evolving industry. Our solutions enable clients to quickly and easily collect, enhance, analyze, present, and act on their financial data. MX puts a user’s data on center stage, molding it into a cohesive, intelligible, and interactive visualization. As a result, users engage more often and more deeply with your digital banking products. The Helios cross-platform framework gives MX clients the ability to offer mobile banking across a range of platforms and device types — all built from a single C++ codebase. This dramatically lowers maintenance costs and powers agile development.
  • 8
    Sigma

    Sigma

    Sigma Computing

    Sigma is a modern business intelligence (BI) and analytics application built for the cloud. Trusted by data-first companies, Sigma provides live access to cloud data warehouses using an intuitive spreadsheet interface empowering business experts to ask more of their data without writing a single line of code. With the full power of SQL, the cloud, and a familiar interface, business users have the freedom to analyze data in real time without limits. Sigma is self-service analytics as it was meant to be.
  • 9
    Pickaxe

    Pickaxe

    Pickaxe Foundry

    Give your business the power of hundreds of data scientists and analysts. AI powered data analytics that anyone can use and understand. Stop just spending all your time pulling data to explain what has happened, and instead focus on building a persuasive story of what you should do next. Pickaxe does it all for you, in realtime, with AI-powered dashboards and deep human insights. Your data platform can tell you ‘what’ is happening, but can it also tell you ‘so what’ and ‘now what’?
  • 10
    SafeGraph

    SafeGraph

    SafeGraph

    Unlock innovation with the most accurate Points-of-Interest (POI) data, business listings, & store visitor insights data for commercial places in the U.S. Business listing and building footprint dataset for every place people spend money in the U.S. (~5MM POIs). Covers locations for major retail chains, shopping malls, convenience stores, airports, & more. Store visitor analytics, foot-traffic counts, and demographic insights data for POI. Data can answers questions such as: how often do people visit stores, where did they come from, & where else do they shop? Seamlessly integrate your existing POI data with SafeGraph's enriched Places data. Business category, open hours, visit count, popular times and more are associated with each place. The top 5,000+ brands are mapped to over 1MM POI. Noisy locations are removed (ATMs, Red Box kiosks, etc.). Closed stores are filtered out. Irrelevant businesses (like home LLCs with no employees) are kept out.
  • 11
    BryteFlow

    BryteFlow

    BryteFlow

    BryteFlow builds the most efficient automated environments for analytics ever. It converts Amazon S3 into an awesome analytics platform by leveraging the AWS ecosystem intelligently to deliver data at lightning speeds. It complements AWS Lake Formation and automates the Modern Data Architecture providing performance and productivity. You can completely automate data ingestion with BryteFlow Ingest’s simple point-and-click interface while BryteFlow XL Ingest is great for the initial full ingest for very large datasets. No coding is needed! With BryteFlow Blend you can merge data from varied sources like Oracle, SQL Server, Salesforce and SAP etc. and transform it to make it ready for Analytics and Machine Learning. BryteFlow TruData reconciles the data at the destination with the source continually or at a frequency you select. If data is missing or incomplete you get an alert so you can fix the issue easily.
  • 12
    Edge Intelligence

    Edge Intelligence

    Edge Intelligence

    Start benefiting your business within minutes of installation. Learn how our system works. It's the fastest, easiest way to analyze vast amounts of geographically distributed data. A new approach to analytics. Overcome the architectural constraints associated with traditional big data warehouses, database design and edge computing architectures. Understand details within the platform that allow for centralized command & control, automated software installation & orchestration and geographically distributed data input & storage.
  • 13
    Intelligent Artifacts

    Intelligent Artifacts

    Intelligent Artifacts

    A new category of AI. Most current AI solutions are engineered through a statistical and purely mathematical lens. We took a different approach. With discoveries in information theory, the team at Intelligent Artifacts has built a new category of AI: a true AGI that eliminates current machine intelligence shortcomings. Our framework keeps the data and application layers separate from the intelligence layer allowing it to learn in real-time, and enabling it to explain predictions down to root cause. A true AGI demands a truly integrated platform. With Intelligent Artifacts, you'll model information, not data — predictions and decisions are real-time and transparent, and can be deployed across various domains without the need to rewrite code. And by combining specialized AI consultants with our dynamic platform, you'll get a customized solution that rapidly offers deep insights and greater outcomes from your data.
  • 14
    HEAVY.AI

    HEAVY.AI

    HEAVY.AI

    HEAVY.AI is the pioneer in accelerated analytics. The HEAVY.AI platform is used in business and government to find insights in data beyond the limits of mainstream analytics tools. Harnessing the massive parallelism of modern CPU and GPU hardware, the platform is available in the cloud and on-premise. HEAVY.AI originated from research at Harvard and MIT Computer Science and Artificial Intelligence Laboratory (CSAIL). Expand beyond the limitations of traditional BI and GIS by leveraging the full power of modern GPU and CPU hardware so you can extract decision-quality information from your massive datasets without lag. Unify and explore your largest geospatial and time-series datasets to get the complete picture of the what, when, and where. Combine interactive visual analytics, hardware-accelerated SQL, and an advanced analytics & data science framework to find opportunity and risk hidden in your enterprise when you need to most.
  • 15
    Hadoop

    Hadoop

    Apache Software Foundation

    The Apache Hadoop software library is a framework that allows for the distributed processing of large data sets across clusters of computers using simple programming models. It is designed to scale up from single servers to thousands of machines, each offering local computation and storage. Rather than rely on hardware to deliver high-availability, the library itself is designed to detect and handle failures at the application layer, so delivering a highly-available service on top of a cluster of computers, each of which may be prone to failures. A wide variety of companies and organizations use Hadoop for both research and production. Users are encouraged to add themselves to the Hadoop PoweredBy wiki page. Apache Hadoop 3.3.4 incorporates a number of significant enhancements over the previous major release line (hadoop-3.2).
  • 16
    Apache Spark

    Apache Spark

    Apache Software Foundation

    Apache Spark™ is a unified analytics engine for large-scale data processing. Apache Spark achieves high performance for both batch and streaming data, using a state-of-the-art DAG scheduler, a query optimizer, and a physical execution engine. Spark offers over 80 high-level operators that make it easy to build parallel apps. And you can use it interactively from the Scala, Python, R, and SQL shells. Spark powers a stack of libraries including SQL and DataFrames, MLlib for machine learning, GraphX, and Spark Streaming. You can combine these libraries seamlessly in the same application. Spark runs on Hadoop, Apache Mesos, Kubernetes, standalone, or in the cloud. It can access diverse data sources. You can run Spark using its standalone cluster mode, on EC2, on Hadoop YARN, on Mesos, or on Kubernetes. Access data in HDFS, Alluxio, Apache Cassandra, Apache HBase, Apache Hive, and hundreds of other data sources.
  • 17
    Incorta

    Incorta

    Incorta

    Direct is the shortest path from data to insight. Incorta empowers everyone in your business with a true self-service data experience and breakthrough performance for better decisions and incredible results. What if you could bypass fragile ETL and expensive data warehouses, and deliver data projects in days, instead of weeks or months? Our direct approach to analytics delivers true self-service in the cloud or on-premises with agility and performance. Incorta is used by the world’s largest brands to succeed where other analytics solutions fail. Across multiple industries and lines of business, we boast connectors and pre-built solutions for your enterprise applications and technologies. Game-changing innovation and customer success happen through Incorta’s partners including Microsoft, AWS, eCapital, and Wipro. Explore or join our thriving partner ecosystem.
  • 18
    Amazon EMR
    Amazon EMR is the industry-leading cloud big data platform for processing vast amounts of data using open-source tools such as Apache Spark, Apache Hive, Apache HBase, Apache Flink, Apache Hudi, and Presto. With EMR you can run Petabyte-scale analysis at less than half of the cost of traditional on-premises solutions and over 3x faster than standard Apache Spark. For short-running jobs, you can spin up and spin down clusters and pay per second for the instances used. For long-running workloads, you can create highly available clusters that automatically scale to meet demand. If you have existing on-premises deployments of open-source tools such as Apache Spark and Apache Hive, you can also run EMR clusters on AWS Outposts. Analyze data using open-source ML frameworks such as Apache Spark MLlib, TensorFlow, and Apache MXNet. Connect to Amazon SageMaker Studio for large-scale model training, analysis, and reporting.
  • 19
    Kraken

    Kraken

    Big Squid

    Kraken is for everyone from analysts to data scientists. Built to be the easiest-to-use, no-code automated machine learning platform. The Kraken no-code automated machine learning (AutoML) platform simplifies and automates data science tasks like data prep, data cleaning, algorithm selection, model training, and model deployment. Kraken was built with analysts and engineers in mind. If you've done data analysis before, you're ready! Kraken's no-code, easy-to-use interface and integrated SONAR© training make it easy to become a citizen data scientist. Advanced features allow data scientists to work faster and more efficiently. Whether you use Excel or flat files for day-to-day reporting or just ad-hoc analysis and exports, drag-and-drop CSV upload and the Amazon S3 connector in Kraken make it easy to start building models with a few clicks. Data Connectors in Kraken allow you to connect to your favorite data warehouse, business intelligence tools, and cloud storage.
    Starting Price: $100 per month
  • 20
    Scuba

    Scuba

    Scuba Analytics

    Self-service analytics at scale. Whether you’re a product manager, the head of a business unit, a chief experience officer, a data scientist, a business analyst, or an IT staffer - you’ll appreciate how simple Scuba makes it to access your data and immediately begin mining it for insights. Whether you’re trying to understand the behavior of your customers, your systems, your apps – or anything else associated with actions taken over time – Interana is the only analytics platform that lets you move beyond dashboards and static reports, to a mode where you and your team can interactively explore your data in real-time to see not just what is happening in your business, but why. With Scuba you're never waiting for your data. All of your data is always available, so you can ask questions as quickly as you can think of them. Scuba is designed for everyday business users, so there’s no need to code or know SQL.
  • 21
    INDICA Data Life Cycle Management
    One platform, four solutions. INDICA connects to all company applications and data sources. It indexes all live data and gives you grip on your complete data landscape. With its platform as a basis, INDICA offers four solutions. INDICA Enterprise Search enables access to all the corporate data sources through one interface. It indexes all structured and unstructured data and ranks the results to relevance. INDICA eDiscovery can be set up as a case by case platform and as a platform that will allow you to run fraud or compliance investigations on the fly. The INDICA Privacy Suite provides you with an extensive toolkit to allow your organization to comply to GDPR and CCPA laws and to remain compliant. INDICA Data Lifecycle Management allows you to take control of your corporate data, keep track of your data and clean or migrate your data. INDICA’s data platform consists of a broad set of features to get in control of your data.
  • 22
    eDrain

    eDrain

    Eclettica

    Planning. Innovating. developing. From need to solution. eDrain DATA CLOUD PLATFORM. eDrain is a tool specialized in data collection, monitoring and production of aggregate reporting. It is a system that operates in the BigData field, able to integrate, thanks to a driver oriented mechanism, the collection of heterogeneous data. The implemented driver engine allows you to integrate a large number of data streams and devices simultaneously. Features. Customizing the dashboard. Add views. Customized widget creation. Configuration of new devices. Configuration of new flows. Configuration of new sensors. Custom report configuration. Check of the sensor status. Realtime original data flow. Definition of the logic of flows. Definition of analysis rules. Definition of warning thresholds. Events configuration. Elaboration of actions. Creation of new devices. Configuration of new stations. Latching new data streams. Management and verification of alerts.
  • 23
    IBM DataStage
    Accelerate AI innovation with cloud-native data integration on IBM Cloud Pak for data. AI-powered data integration, anywhere. Your AI and analytics are only as good as the data that fuels them. With a modern container-based architecture, IBM® DataStage® for IBM Cloud Pak® for Data delivers that high-quality data. It combines industry-leading data integration with DataOps, governance and analytics on a single data and AI platform. Automation accelerates administrative tasks to help reduce TCO. AI-based design accelerators and out-of-the-box integration with DataOps and data science services speed AI innovation. Parallelism and multicloud integration let you deliver trusted data at scale across hybrid or multicloud environments. Manage the data and analytics lifecycle on the IBM Cloud Pak for Data platform. Services include data science, event messaging, data virtualization and data warehousing. Parallel engine and automated load balancing.
  • 24
    Delta Lake

    Delta Lake

    Delta Lake

    Delta Lake is an open-source storage layer that brings ACID transactions to Apache Spark™ and big data workloads. Data lakes typically have multiple data pipelines reading and writing data concurrently, and data engineers have to go through a tedious process to ensure data integrity, due to the lack of transactions. Delta Lake brings ACID transactions to your data lakes. It provides serializability, the strongest level of isolation level. Learn more at Diving into Delta Lake: Unpacking the Transaction Log. In big data, even the metadata itself can be "big data". Delta Lake treats metadata just like data, leveraging Spark's distributed processing power to handle all its metadata. As a result, Delta Lake can handle petabyte-scale tables with billions of partitions and files at ease. Delta Lake provides snapshots of data enabling developers to access and revert to earlier versions of data for audits, rollbacks or to reproduce experiments.
  • 25
    Privacera

    Privacera

    Privacera

    At the intersection of data governance, privacy, and security, Privacera’s unified data access governance platform maximizes the value of data by providing secure data access control and governance across hybrid- and multi-cloud environments. The hybrid platform centralizes access and natively enforces policies across multiple cloud services—AWS, Azure, Google Cloud, Databricks, Snowflake, Starburst and more—to democratize trusted data enterprise-wide without compromising compliance with regulations such as GDPR, CCPA, LGPD, or HIPAA. Trusted by Fortune 500 customers across finance, insurance, retail, healthcare, media, public and the federal sector, Privacera is the industry’s leading data access governance platform that delivers unmatched scalability, elasticity, and performance. Headquartered in Fremont, California, Privacera was founded in 2016 to manage cloud data privacy and security by the creators of Apache Ranger™ and Apache Atlas™.
  • 26
    Apache Storm

    Apache Storm

    Apache Software Foundation

    Apache Storm is a free and open source distributed realtime computation system. Apache Storm makes it easy to reliably process unbounded streams of data, doing for realtime processing what Hadoop did for batch processing. Apache Storm is simple, can be used with any programming language, and is a lot of fun to use! Apache Storm has many use cases: realtime analytics, online machine learning, continuous computation, distributed RPC, ETL, and more. Apache Storm is fast: a benchmark clocked it at over a million tuples processed per second per node. It is scalable, fault-tolerant, guarantees your data will be processed, and is easy to set up and operate. Apache Storm integrates with the queueing and database technologies you already use. An Apache Storm topology consumes streams of data and processes those streams in arbitrarily complex ways, repartitioning the streams between each stage of the computation however needed. Read more in the tutorial.
  • 27
    Wavo

    Wavo

    Wavo

    We’ve released a revolutionary big data platform that gathers all information about a music business, providing a single source of truth for decisions. Every music business has hundreds of data sources. But they are siloed and fragmented. Our platform identifies and connects them to build a foundation of quality data that can be applied to all daily music business operations. To work efficiently and securely—and to surface valuable insight no one else can—record labels and agencies require a sophisticated data management and governance system, so that data is available, relevant, and usable at all times. As data sources are ingested into Wavo’s Big Data Platform, machine learning is deployed to tag data based on personalized templates, making it easy to access and drill-down into important information. This enables everyone in a music business to activate and deliver business-ready data, backed up and organized for immediate value.
  • 28
    TEOCO SmartHub Analytics
    SmartHub Analytics is a dedicated telecom big-data analytics platform that enables financial and subscriber-based ROI-driven use cases. Designed to support and encourage data sharing and reuse, SmartHub Analytics optimizes business performance and delivers analytics at the speed of thought. SmartHub Analytics eliminates silos and can assess, validate and model vast amounts of data from across TEOCO’s solution portfolio, including: customers, planning, optimization, service assurance, geo-location, service quality and costs. As an added analytics layer residing on top of other existing OSS & BSS solutions, SmartHub Analytics provides a standalone analytics environment with a proven return on investment (ROI), saving operators billions. We consistently uncover significant cost savings for our customers, utilizing prediction-based machine learning algorithms. SmartHub Analytics remains at the forefront of technology, delivering accelerated data analyses.
  • 29
    Isima

    Isima

    Isima

    bi(OS)® delivers unparalleled speed to insight for data app builders in a unified manner. With bi(OS)®, the complete life-cycle of building data apps takes hours to days. This includes adding varied data sources, deriving real-time insights, and deploying to production. Join enterprise data teams across industries and become the data superhero your business deserves. The trifecta of Open Source, Cloud, and SaaS has failed to deliver the promised data-driven impact. All of the enterprises' investments have been in data movement and integration, which isn't sustainable. There is a dire need for a new approach to data, built with enterprise empathy in mind. bi(OS)® is built by reimagining first principles in enterprise data management, from ingest to insight. It serves API, AI, and BI builders in a unified manner, to achieve data-driven impact in days. Engineers build enduring moat as a symphony emerges between IT teams, tools, and processes.
  • 30
    Tencent Cloud Elastic MapReduce
    EMR enables you to scale the managed Hadoop clusters manually or automatically according to your business curves or monitoring metrics. EMR's storage-computation separation even allows you to terminate a cluster to maximize resource efficiency. EMR supports hot failover for CBS-based nodes. It features a primary/secondary disaster recovery mechanism where the secondary node starts within seconds when the primary node fails, ensuring the high availability of big data services. The metadata of its components such as Hive supports remote disaster recovery. Computation-storage separation ensures high data persistence for COS data storage. EMR is equipped with a comprehensive monitoring system that helps you quickly identify and locate cluster exceptions to ensure stable cluster operations. VPCs provide a convenient network isolation method that facilitates your network policy planning for managed Hadoop clusters.
MongoDB Logo MongoDB