Best Machine Learning Software for Amazon Web Services (AWS) - Page 2

Compare the Top Machine Learning Software that integrates with Amazon Web Services (AWS) as of July 2025 - Page 2

This a list of Machine Learning software that integrates with Amazon Web Services (AWS). Use the filters on the left to add additional filters for products that have integrations with Amazon Web Services (AWS). View the products that work with Amazon Web Services (AWS) in the table below.

  • 1
    Krista

    Krista

    Krista

    Krista is a nothing-like-code intelligent automation platform that orchestrates your people, apps, and AI so you can optimize business outcomes. Krista builds and integrates machine learning and apps more simply than you can imagine. Krista is purpose-built to automate business outcomes, not just back-office tasks. Optimizing outcomes requires spanning departments of people & apps, deploying AI/ML for autonomous decision-making, leveraging your existing task automation, and enabling constant change. By digitizing complete processes, Krista delivers organization-wide, bottom-line impact.Krista empowers your people to create and modify automations without programming. Democratizing automation increases business speed and keeps you from waiting in the dreaded IT backlog. Krista dramatically reduces TCO compared to your current automation platform.
  • 2
    Launchable

    Launchable

    Launchable

    You can have the best developers in the world, but every test is making them slower. 80% of your software tests are pointless. The problem is you don't know which 80%. We find the right 20% using your data so that you can ship faster. We have shrink-wrapped predictive test selection, a machine learning-based approach being used at companies like Facebook so that it can be used by any company. We support multiple languages, test runners, and CI systems. Just bring Git to the table. Launchable uses machine learning to analyze your test failures and source code. It doesn't rely on code syntax analysis. This means it's trivial for Launchable to add support for almost any file-based programming language. It also means we can scale across teams and projects with different languages and tools. Out of the box, we currently support Python, Ruby, Java, JavaScript, Go, C, and C++, and we regularly add support for new languages.
  • 3
    Mona

    Mona

    Mona

    Gain complete visibility into the performance of your data, models, and processes with the most flexible monitoring solution. Automatically surface and resolve performance issues within your AI/ML or intelligent automation processes to avoid negative impacts on both your business and customers. Learning how your data, models, and processes perform in the real world is critical to continuously improving your processes. Monitoring is the ‘eyes and ears' needed to observe your data and workflows to tell you if they’re performing well. Mona exhaustively analyzes your data to provide actionable insights based on advanced anomaly detection mechanisms, to alert you before your business KPIs are hurt. Take stock of any part of your production workflows and business processes, including models, pipelines, and business outcomes. Whatever datatype you work with, whether you have a batch or streaming real-time processes, and for the specific way in which you want to measure your performance.
  • 4
    Amazon DevOps Guru
    Amazon DevOps Guru is a machine learning (ML)-powered service designed to make it easy to improve the operational performance and availability of an application. DevOps Guru helps detect behaviors that deviate from normal operating patterns, so you can identify operational errors long before they affect your customers. DevOps Guru uses ML models with information collected over years by Amazon.com and AWS Operational Excellence to identify anomalous application behavior (for example, increased latency, error rates, resource limitations, etc.) and helps detect critical errors that could potentially cause service interruptions. When the DevOps Guru identifies a critical issue, it automatically sends an alert and provides a summary of related anomalies, the likely root cause, and context on when and where the issue occurred.
    Starting Price: $0.0028 per resource per hour
  • 5
    WEKA

    WEKA

    WEKA

    WEKA provides a high-performance data platform optimized for AI and machine learning, offering scalable solutions for businesses and research labs. With the ability to handle vast amounts of data across on-premises, cloud, and hybrid environments, WEKA accelerates data workflows, enabling faster AI training, inference, and high-performance computing (HPC). The platform features infinite scalability, simplifying data storage, and providing seamless access to data across multiple locations. WEKA's environmentally conscious approach minimizes energy consumption, making it ideal for organizations aiming for both performance and sustainability in AI-driven projects.
  • 6
    Chalk

    Chalk

    Chalk

    Powerful data engineering workflows, without the infrastructure headaches. Complex streaming, scheduling, and data backfill pipelines, are all defined in simple, composable Python. Make ETL a thing of the past, fetch all of your data in real-time, no matter how complex. Incorporate deep learning and LLMs into decisions alongside structured business data. Make better predictions with fresher data, don’t pay vendors to pre-fetch data you don’t use, and query data just in time for online predictions. Experiment in Jupyter, then deploy to production. Prevent train-serve skew and create new data workflows in milliseconds. Instantly monitor all of your data workflows in real-time; track usage, and data quality effortlessly. Know everything you computed and data replay anything. Integrate with the tools you already use and deploy to your own infrastructure. Decide and enforce withdrawal limits with custom hold times.
    Starting Price: Free
  • 7
    Zerve AI

    Zerve AI

    Zerve AI

    Merging the best of a notebook and an IDE into one integrated coding environment, experts can explore their data and write stable code at the same time with fully automated cloud infrastructure. Zerve’s data science development environment gives data science and ML teams a unified space to explore, collaborate, build, and deploy data science & AI projects like never before. Zerve offers true language interoperability, meaning that as well as being able to use Python, R, SQL, or Markdown all in the same canvas, users can connect these code blocks to each other. No more long-running code blocks or containers, with Zerve enjoying unlimited parallelization at any stage of the development journey. Analysis artifacts are automatically serialized, versioned, stored, and preserved for later use, meaning easily changing a step in the data flow without needing to rerun any preceding steps. Fine-grained selection of compute resources and extra memory for complex data transformation.
  • 8
    Amazon EC2 Trn1 Instances
    Amazon Elastic Compute Cloud (EC2) Trn1 instances, powered by AWS Trainium chips, are purpose-built for high-performance deep learning training of generative AI models, including large language models and latent diffusion models. Trn1 instances offer up to 50% cost-to-train savings over other comparable Amazon EC2 instances. You can use Trn1 instances to train 100B+ parameter DL and generative AI models across a broad set of applications, such as text summarization, code generation, question answering, image and video generation, recommendation, and fraud detection. The AWS Neuron SDK helps developers train models on AWS Trainium (and deploy models on the AWS Inferentia chips). It integrates natively with frameworks such as PyTorch and TensorFlow so that you can continue using your existing code and workflows to train models on Trn1 instances.
    Starting Price: $1.34 per hour
  • 9
    Amazon EC2 Inf1 Instances
    Amazon EC2 Inf1 instances are purpose-built to deliver high-performance and cost-effective machine learning inference. They provide up to 2.3 times higher throughput and up to 70% lower cost per inference compared to other Amazon EC2 instances. Powered by up to 16 AWS Inferentia chips, ML inference accelerators designed by AWS, Inf1 instances also feature 2nd generation Intel Xeon Scalable processors and offer up to 100 Gbps networking bandwidth to support large-scale ML applications. These instances are ideal for deploying applications such as search engines, recommendation systems, computer vision, speech recognition, natural language processing, personalization, and fraud detection. Developers can deploy their ML models on Inf1 instances using the AWS Neuron SDK, which integrates with popular ML frameworks like TensorFlow, PyTorch, and Apache MXNet, allowing for seamless migration with minimal code changes.
    Starting Price: $0.228 per hour
  • 10
    Amazon EC2 G5 Instances
    Amazon EC2 G5 instances are the latest generation of NVIDIA GPU-based instances that can be used for a wide range of graphics-intensive and machine-learning use cases. They deliver up to 3x better performance for graphics-intensive applications and machine learning inference and up to 3.3x higher performance for machine learning training compared to Amazon EC2 G4dn instances. Customers can use G5 instances for graphics-intensive applications such as remote workstations, video rendering, and gaming to produce high-fidelity graphics in real time. With G5 instances, machine learning customers get high-performance and cost-efficient infrastructure to train and deploy larger and more sophisticated models for natural language processing, computer vision, and recommender engine use cases. G5 instances deliver up to 3x higher graphics performance and up to 40% better price performance than G4dn instances. They have more ray tracing cores than any other GPU-based EC2 instance.
    Starting Price: $1.006 per hour
  • 11
    Databricks Data Intelligence Platform
    The Databricks Data Intelligence Platform allows your entire organization to use data and AI. It’s built on a lakehouse to provide an open, unified foundation for all data and governance, and is powered by a Data Intelligence Engine that understands the uniqueness of your data. The winners in every industry will be data and AI companies. From ETL to data warehousing to generative AI, Databricks helps you simplify and accelerate your data and AI goals. Databricks combines generative AI with the unification benefits of a lakehouse to power a Data Intelligence Engine that understands the unique semantics of your data. This allows the Databricks Platform to automatically optimize performance and manage infrastructure in ways unique to your business. The Data Intelligence Engine understands your organization’s language, so search and discovery of new data is as easy as asking a question like you would to a coworker.
  • 12
    Grace Enterprise AI Platform
    The Grace Enterprise AI Platform, the AI platform with full support for Governance, Risk & Compliance (GRC) for AI. Grace offers an efficient, secure, and robust AI implementation across any organization, standardizing processes, and workflows across all your AI projects. Grace covers the full range of rich functionality your organization needs to become fully AI proficient and helps ensure regulatory excellence for AI, to avoid compliance requirements slowing or stopping AI implementation. Grace lowers the entry barriers for AI users across all functional and operational roles in your organization, including technical, IT, project management, and compliance, while still offering efficient workflows for experienced data scientists and engineers. Ensuring that all activities are traced, explained, and enforced. This includes all areas within the data science model development, data used for model training and development, model bias, and more.
  • 13
    Privacera

    Privacera

    Privacera

    At the intersection of data governance, privacy, and security, Privacera’s unified data access governance platform maximizes the value of data by providing secure data access control and governance across hybrid- and multi-cloud environments. The hybrid platform centralizes access and natively enforces policies across multiple cloud services—AWS, Azure, Google Cloud, Databricks, Snowflake, Starburst and more—to democratize trusted data enterprise-wide without compromising compliance with regulations such as GDPR, CCPA, LGPD, or HIPAA. Trusted by Fortune 500 customers across finance, insurance, retail, healthcare, media, public and the federal sector, Privacera is the industry’s leading data access governance platform that delivers unmatched scalability, elasticity, and performance. Headquartered in Fremont, California, Privacera was founded in 2016 to manage cloud data privacy and security by the creators of Apache Ranger™ and Apache Atlas™.
  • 14
    Altair Knowledge Works
    It is undeniable that data and analytics are key drivers behind transformative business initiatives. More people across the enterprise are accessing data to answer complex questions. The demand for easy-to-use, low-code yet flexible data transformation and machine learning tools has never been greater. The use of multiple tools leads to inefficient data analytic processes, higher costs, and slows down decision making. Aging solutions with overlapping features threaten current data science projects as proprietary functions in closed vendor solutions become obsolete. Combining decades of experience in data preparation, machine learning and visualization with one unified interface Knowledge Works scales as data sizes grow, new open source features and functionalities are developed, and user profiles become more sophisticated. Data scientists and business analysts can efficiently operationalize data analytics applications using its low-code, cloud ready interface.
  • 15
    Xilinx

    Xilinx

    Xilinx

    The Xilinx’s AI development platform for AI inference on Xilinx hardware platforms consists of optimized IP, tools, libraries, models, and example designs. It is designed with high efficiency and ease-of-use in mind, unleashing the full potential of AI acceleration on Xilinx FPGA and ACAP. Supports mainstream frameworks and the latest models capable of diverse deep learning tasks. Provides a comprehensive set of pre-optimized models that are ready to deploy on Xilinx devices. You can find the closest model and start re-training for your applications! Provides a powerful open source quantizer that supports pruned and unpruned model quantization, calibration, and fine tuning. The AI profiler provides layer by layer analysis to help with bottlenecks. The AI library offers open source high-level C++ and Python APIs for maximum portability from edge to cloud. Efficient and scalable IP cores can be customized to meet your needs of many different applications.
  • 16
    Graviti

    Graviti

    Graviti

    Unstructured data is the future of AI. Unlock this future now and build an ML/AI pipeline that scales all of your unstructured data in one place. Use better data to deliver better models, only with Graviti. Get to know the data platform that enables AI developers with management, query, and version control features that are designed for unstructured data. Quality data is no longer a pricey dream. Manage your metadata, annotation, and predictions in one place. Customize filters and visualize filtering results to get you straight to the data that best match your needs. Utilize a Git-like structure to manage data versions and collaborate with your teammates. Role-based access control and visualization of version differences allows your team to work together safely and flexibly. Automate your data pipeline with Graviti’s built-in marketplace and workflow builder. Level-up to fast model iterations with no more grinding.
  • 17
    StreamFlux

    StreamFlux

    Fractal

    Data is crucial when it comes to building, streamlining and growing your business. However, getting the full value out of data can be a challenge, many organizations are faced with poor access to data, incompatible tools, spiraling costs and slow results. Simply put, leaders who can turn raw data into real results will thrive in today’s landscape. The key to this is empowering everyone across your business to be able to analyze, build and collaborate on end-to-end AI and machine learning solutions in one place, fast. Streamflux is a one-stop shop to meet your data analytics and AI challenges. Our self-serve platform allows you the freedom to build end-to-end data solutions, uses models to answer complex questions and assesses user behaviors. Whether you’re predicting customer churn and future revenue, or generating recommendations, you can go from raw data to genuine business impact in days, not months.
  • 18
    Polyaxon

    Polyaxon

    Polyaxon

    A Platform for reproducible and scalable Machine Learning and Deep Learning applications. Learn more about the suite of features and products that underpin today's most innovative platform for managing data science workflows. Polyaxon provides an interactive workspace with notebooks, tensorboards, visualizations,and dashboards. Collaborate with the rest of your team, share and compare experiments and results. Reproducible results with a built-in version control for code and experiments. Deploy Polyaxon in the cloud, on-premises or in hybrid environments, including single laptop, container management platforms, or on Kubernetes. Spin up or down, add more nodes, add more GPUs, and expand storage.
  • 19
    Fiddler AI

    Fiddler AI

    Fiddler AI

    Fiddler is a pioneer in Model Performance Management for responsible AI. The Fiddler platform’s unified environment provides a common language, centralized controls, and actionable insights to operationalize ML/AI with trust. Model monitoring, explainable AI, analytics, and fairness capabilities address the unique challenges of building in-house stable and secure MLOps systems at scale. Unlike observability solutions, Fiddler integrates deep XAI and analytics to help you grow into advanced capabilities over time and build a framework for responsible AI practices. Fortune 500 organizations use Fiddler across training and production models to accelerate AI time-to-value and scale, build trusted AI solutions, and increase revenue.
  • 20
    Amazon Monitron
    Detect machine issues before they occur with machine learning (ML), and take action. Start monitoring equipment in minutes with easy installation and automatic, secure analysis through the Amazon Monitron end-to-end system. Improve system accuracy continuously as Amazon Monitron learns from technician feedback entered in the mobile and web apps. Amazon Monitron is an end-to-end system that uses machine learning to detect abnormal conditions in industrial equipment and enable predictive maintenance. Save on costly repairs and prevent factory equipment downtime with easy-to-install hardware and the power of ML. Reduce unplanned equipment downtime with predictive maintenance and machine learning. Amazon Monitron uses machine learning on temperature and vibration data. Amazon Monitron can help you predict equipment downtime before it happens. Compare what it costs to get started with how much you could save.
  • 21
    Amazon SageMaker Debugger
    Optimize ML models by capturing training metrics in real-time and sending alerts when anomalies are detected. Automatically stop training processes when the desired accuracy is achieved to reduce the time and cost of training ML models. Automatically profile and monitor system resource utilization and send alerts when resource bottlenecks are identified to continuously improve resource utilization. Amazon SageMaker Debugger can reduce troubleshooting during training from days to minutes by automatically detecting and alerting you to remediate common training errors such as gradient values becoming too large or too small. Alerts can be viewed in Amazon SageMaker Studio or configured through Amazon CloudWatch. Additionally, the SageMaker Debugger SDK enables you to automatically detect new classes of model-specific errors such as data sampling, hyperparameter values, and out-of-bound values.
  • 22
    Amazon SageMaker Model Training
    Amazon SageMaker Model Training reduces the time and cost to train and tune machine learning (ML) models at scale without the need to manage infrastructure. You can take advantage of the highest-performing ML compute infrastructure currently available, and SageMaker can automatically scale infrastructure up or down, from one to thousands of GPUs. Since you pay only for what you use, you can manage your training costs more effectively. To train deep learning models faster, SageMaker distributed training libraries can automatically split large models and training datasets across AWS GPU instances, or you can use third-party libraries, such as DeepSpeed, Horovod, or Megatron. Efficiently manage system resources with a wide choice of GPUs and CPUs including P4d.24xl instances, which are the fastest training instances currently available in the cloud. Specify the location of data, indicate the type of SageMaker instances, and get started with a single click.
  • 23
    Amazon SageMaker Model Building
    Amazon SageMaker provides all the tools and libraries you need to build ML models, the process of iteratively trying different algorithms and evaluating their accuracy to find the best one for your use case. In Amazon SageMaker you can pick different algorithms, including over 15 that are built-in and optimized for SageMaker, and use over 150 pre-built models from popular model zoos available with a few clicks. SageMaker also offers a variety of model-building tools including Amazon SageMaker Studio Notebooks and RStudio where you can run ML models on a small scale to see results and view reports on their performance so you can come up with high-quality working prototypes. Amazon SageMaker Studio Notebooks help you build ML models faster and collaborate with your team. Amazon SageMaker Studio notebooks provide one-click Jupyter notebooks that you can start working within seconds. Amazon SageMaker also enables one-click sharing of notebooks.
  • 24
    Amazon SageMaker Studio
    Amazon SageMaker Studio is an integrated development environment (IDE) that provides a single web-based visual interface where you can access purpose-built tools to perform all machine learning (ML) development steps, from preparing data to building, training, and deploying your ML models, improving data science team productivity by up to 10x. You can quickly upload data, create new notebooks, train and tune models, move back and forth between steps to adjust experiments, collaborate seamlessly within your organization, and deploy models to production without leaving SageMaker Studio. Perform all ML development steps, from preparing raw data to deploying and monitoring ML models, with access to the most comprehensive set of tools in a single web-based visual interface. Amazon SageMaker Unified Studio is a comprehensive, AI and data development environment designed to streamline workflows and simplify the process of building and deploying machine learning models.
  • 25
    Amazon SageMaker Studio Lab
    Amazon SageMaker Studio Lab is a free machine learning (ML) development environment that provides the compute, storage (up to 15GB), and security, all at no cost, for anyone to learn and experiment with ML. All you need to get started is a valid email address, you don’t need to configure infrastructure or manage identity and access or even sign up for an AWS account. SageMaker Studio Lab accelerates model building through GitHub integration, and it comes preconfigured with the most popular ML tools, frameworks, and libraries to get you started immediately. SageMaker Studio Lab automatically saves your work so you don’t need to restart in between sessions. It’s as easy as closing your laptop and coming back later. Free machine learning development environment that provides the computing, storage, and security to learn and experiment with ML. GitHub integration and preconfigured with the most popular ML tools, frameworks, and libraries so you can get started immediately.
  • 26
    Amazon SageMaker Feature Store
    Amazon SageMaker Feature Store is a fully managed, purpose-built repository to store, share, and manage features for machine learning (ML) models. Features are inputs to ML models used during training and inference. For example, in an application that recommends a music playlist, features could include song ratings, listening duration, and listener demographics. Features are used repeatedly by multiple teams and feature quality is critical to ensure a highly accurate model. Also, when features used to train models offline in batch are made available for real-time inference, it’s hard to keep the two feature stores synchronized. SageMaker Feature Store provides a secured and unified store for feature use across the ML lifecycle. Store, share, and manage ML model features for training and inference to promote feature reuse across ML applications. Ingest features from any data source including streaming and batch such as application logs, service logs, clickstreams, sensors, etc.
  • 27
    Amazon SageMaker Data Wrangler
    Amazon SageMaker Data Wrangler reduces the time it takes to aggregate and prepare data for machine learning (ML) from weeks to minutes. With SageMaker Data Wrangler, you can simplify the process of data preparation and feature engineering, and complete each step of the data preparation workflow (including data selection, cleansing, exploration, visualization, and processing at scale) from a single visual interface. You can use SQL to select the data you want from a wide variety of data sources and import it quickly. Next, you can use the Data Quality and Insights report to automatically verify data quality and detect anomalies, such as duplicate rows and target leakage. SageMaker Data Wrangler contains over 300 built-in data transformations so you can quickly transform data without writing any code. Once you have completed your data preparation workflow, you can scale it to your full datasets using SageMaker data processing jobs; train, tune, and deploy models.
  • 28
    Amazon SageMaker Canvas
    Amazon SageMaker Canvas expands access to machine learning (ML) by providing business analysts with a visual interface that allows them to generate accurate ML predictions on their own, without requiring any ML experience or having to write a single line of code. Visual point-and-click interface to connect, prepare, analyze, and explore data for building ML models and generating accurate predictions. Automatically build ML models to run what-if analysis and generate single or bulk predictions with a few clicks. Boost collaboration between business analysts and data scientists by sharing, reviewing, and updating ML models across tools. Import ML models from anywhere and generate predictions directly in Amazon SageMaker Canvas. With Amazon SageMaker Canvas, you can import data from disparate sources, select values you want to predict, automatically prepare and explore data, and quickly and more easily build ML models. You can then analyze models and generate accurate predictions.
  • 29
    Amazon SageMaker Edge
    The SageMaker Edge Agent allows you to capture data and metadata based on triggers that you set so that you can retrain your existing models with real-world data or build new models. Additionally, this data can be used to conduct your own analysis, such as model drift analysis. We offer three options for deployment. GGv2 (~ size 100MB) is a fully integrated AWS IoT deployment mechanism. For those customers with a limited device capacity, we have a smaller built-in deployment mechanism within SageMaker Edge. For customers who have a preferred deployment mechanism, we support third party mechanisms that can be plugged into our user flow. Amazon SageMaker Edge Manager provides a dashboard so you can understand the performance of models running on each device across your fleet. The dashboard helps you visually understand overall fleet health and identify the problematic models through a dashboard in the console.
  • 30
    Amazon SageMaker Clarify
    Amazon SageMaker Clarify provides machine learning (ML) developers with purpose-built tools to gain greater insights into their ML training data and models. SageMaker Clarify detects and measures potential bias using a variety of metrics so that ML developers can address potential bias and explain model predictions. SageMaker Clarify can detect potential bias during data preparation, after model training, and in your deployed model. For instance, you can check for bias related to age in your dataset or in your trained model and receive a detailed report that quantifies different types of potential bias. SageMaker Clarify also includes feature importance scores that help you explain how your model makes predictions and produces explainability reports in bulk or real time through online explainability. You can use these reports to support customer or internal presentations or to identify potential issues with your model.