Business Software for Amazon SageMaker - Page 3

Top Software that integrates with Amazon SageMaker as of August 2025 - Page 3

  • 1
    Galileo

    Galileo

    Galileo

    Models can be opaque in understanding what data they didn’t perform well on and why. Galileo provides a host of tools for ML teams to inspect and find ML data errors 10x faster. Galileo sifts through your unlabeled data to automatically identify error patterns and data gaps in your model. We get it - ML experimentation is messy. It needs a lot of data and model changes across many runs. Track and compare your runs in one place and quickly share reports with your team. Galileo has been built to integrate with your ML ecosystem. Send a fixed dataset to your data store to retrain, send mislabeled data to your labelers, share a collaborative report, and a lot more! Galileo is purpose-built for ML teams to build better quality models, faster.
  • 2
    Fiddler AI

    Fiddler AI

    Fiddler AI

    Fiddler is a pioneer in Model Performance Management for responsible AI. The Fiddler platform’s unified environment provides a common language, centralized controls, and actionable insights to operationalize ML/AI with trust. Model monitoring, explainable AI, analytics, and fairness capabilities address the unique challenges of building in-house stable and secure MLOps systems at scale. Unlike observability solutions, Fiddler integrates deep XAI and analytics to help you grow into advanced capabilities over time and build a framework for responsible AI practices. Fortune 500 organizations use Fiddler across training and production models to accelerate AI time-to-value and scale, build trusted AI solutions, and increase revenue.
  • 3
    Wizata

    Wizata

    Wizata

    With Digital Twin & Data Explorer, AI Solutions Builder and Automation of Production features, the Wizata Platform empowers the manufacturing industry to drive its digital transformation and facilitates the development of AI solutions from proof of concept to real-time production recommendations for a complete loop process control through AI. This open architecture platform (SaaS– Software as a Service) acts as an orchestrator of your different assets (machines, sensors, AI, cloud, edge,) and ensures you gather and explore easily your data that stays under your sole control. Control resources invested into AI experiments step by step and prioritize your projects depending on how your AI solutions solve your business pains and improve production processes, their return of investment and according to data science best practices in metallurgy that we developed since 4 years around the world.
  • 4
    Mantium

    Mantium

    Mantium

    Mantium’s AI platform promotes knowledge sharing and alignment within organizations helping teams work towards the same goals. With large distributed teams, a company’s knowledge management systems (KMS) are the key to collaborating and learning about processes, meetings, events, and more. With Mantium, we enable enterprises to quickly and easily find knowledge within KMS, using AI to return the best answers to questions. If Mantium doesn’t have the answer to your question, your team can provide updated information, and the AI will improve for future instances. With Mantium, you can search systems holistically, with Natural Language Processing (NLP), ensuring your team finds the information they need quickly. With our Slackbot integration, you can ask a question on Slack, with no need to switch to a separate app to get the answers you need.
  • 5
    AWS HealthLake
    Extract meaning from unstructured data with integrated Amazon Comprehend Medical for easy search and querying. Make predictions on health data using Amazon Athena queries, Amazon SageMaker ML models, and Amazon QuickSight analytics. Support interoperable standards such as the Fast Healthcare Interoperability Resources (FHIR). Run medical imaging applications in the cloud to increase scale and reduce costs. AWS HealthLake is a HIPAA-eligible service offering healthcare and life sciences companies a chronological view of individual or patient population health data for query and analytics at scale. Analyze population health trends, predict outcomes, and manage costs with advanced analytics tools and ML models. Identify opportunities to close gaps in care and deliver targeted interventions with a longitudinal view of patient journeys. Apply advanced analytics and ML to newly structured data to optimize appointment scheduling and reduce unnecessary procedures.
  • 6
    NVIDIA AI Foundations
    Impacting virtually every industry, generative AI unlocks a new frontier of opportunities, for knowledge and creative workers, to solve today’s most important challenges. NVIDIA is powering generative AI through an impressive suite of cloud services, pre-trained foundation models, as well as cutting-edge frameworks, optimized inference engines, and APIs to bring intelligence to your enterprise applications. NVIDIA AI Foundations is a set of cloud services that advance enterprise-level generative AI and enable customization across use cases in areas such as text (NVIDIA NeMo™), visual content (NVIDIA Picasso), and biology (NVIDIA BioNeMo™). Unleash the full potential with NeMo, Picasso, and BioNeMo cloud services, powered by NVIDIA DGX™ Cloud, the AI supercomputer. Marketing copy, storyline creation, and global translation in many languages. For news, email, meeting minutes, and information synthesis.
  • 7
    Amazon SageMaker Debugger
    Optimize ML models by capturing training metrics in real-time and sending alerts when anomalies are detected. Automatically stop training processes when the desired accuracy is achieved to reduce the time and cost of training ML models. Automatically profile and monitor system resource utilization and send alerts when resource bottlenecks are identified to continuously improve resource utilization. Amazon SageMaker Debugger can reduce troubleshooting during training from days to minutes by automatically detecting and alerting you to remediate common training errors such as gradient values becoming too large or too small. Alerts can be viewed in Amazon SageMaker Studio or configured through Amazon CloudWatch. Additionally, the SageMaker Debugger SDK enables you to automatically detect new classes of model-specific errors such as data sampling, hyperparameter values, and out-of-bound values.
  • 8
    Amazon SageMaker Model Training
    Amazon SageMaker Model Training reduces the time and cost to train and tune machine learning (ML) models at scale without the need to manage infrastructure. You can take advantage of the highest-performing ML compute infrastructure currently available, and SageMaker can automatically scale infrastructure up or down, from one to thousands of GPUs. Since you pay only for what you use, you can manage your training costs more effectively. To train deep learning models faster, SageMaker distributed training libraries can automatically split large models and training datasets across AWS GPU instances, or you can use third-party libraries, such as DeepSpeed, Horovod, or Megatron. Efficiently manage system resources with a wide choice of GPUs and CPUs including P4d.24xl instances, which are the fastest training instances currently available in the cloud. Specify the location of data, indicate the type of SageMaker instances, and get started with a single click.
  • 9
    Amazon SageMaker Model Building
    Amazon SageMaker provides all the tools and libraries you need to build ML models, the process of iteratively trying different algorithms and evaluating their accuracy to find the best one for your use case. In Amazon SageMaker you can pick different algorithms, including over 15 that are built-in and optimized for SageMaker, and use over 150 pre-built models from popular model zoos available with a few clicks. SageMaker also offers a variety of model-building tools including Amazon SageMaker Studio Notebooks and RStudio where you can run ML models on a small scale to see results and view reports on their performance so you can come up with high-quality working prototypes. Amazon SageMaker Studio Notebooks help you build ML models faster and collaborate with your team. Amazon SageMaker Studio notebooks provide one-click Jupyter notebooks that you can start working within seconds. Amazon SageMaker also enables one-click sharing of notebooks.
  • 10
    Amazon SageMaker Studio
    Amazon SageMaker Studio is an integrated development environment (IDE) that provides a single web-based visual interface where you can access purpose-built tools to perform all machine learning (ML) development steps, from preparing data to building, training, and deploying your ML models, improving data science team productivity by up to 10x. You can quickly upload data, create new notebooks, train and tune models, move back and forth between steps to adjust experiments, collaborate seamlessly within your organization, and deploy models to production without leaving SageMaker Studio. Perform all ML development steps, from preparing raw data to deploying and monitoring ML models, with access to the most comprehensive set of tools in a single web-based visual interface. Amazon SageMaker Unified Studio is a comprehensive, AI and data development environment designed to streamline workflows and simplify the process of building and deploying machine learning models.
  • 11
    Amazon SageMaker Studio Lab
    Amazon SageMaker Studio Lab is a free machine learning (ML) development environment that provides the compute, storage (up to 15GB), and security, all at no cost, for anyone to learn and experiment with ML. All you need to get started is a valid email address, you don’t need to configure infrastructure or manage identity and access or even sign up for an AWS account. SageMaker Studio Lab accelerates model building through GitHub integration, and it comes preconfigured with the most popular ML tools, frameworks, and libraries to get you started immediately. SageMaker Studio Lab automatically saves your work so you don’t need to restart in between sessions. It’s as easy as closing your laptop and coming back later. Free machine learning development environment that provides the computing, storage, and security to learn and experiment with ML. GitHub integration and preconfigured with the most popular ML tools, frameworks, and libraries so you can get started immediately.
  • 12
    Amazon SageMaker Feature Store
    Amazon SageMaker Feature Store is a fully managed, purpose-built repository to store, share, and manage features for machine learning (ML) models. Features are inputs to ML models used during training and inference. For example, in an application that recommends a music playlist, features could include song ratings, listening duration, and listener demographics. Features are used repeatedly by multiple teams and feature quality is critical to ensure a highly accurate model. Also, when features used to train models offline in batch are made available for real-time inference, it’s hard to keep the two feature stores synchronized. SageMaker Feature Store provides a secured and unified store for feature use across the ML lifecycle. Store, share, and manage ML model features for training and inference to promote feature reuse across ML applications. Ingest features from any data source including streaming and batch such as application logs, service logs, clickstreams, sensors, etc.
  • 13
    Amazon SageMaker Data Wrangler
    Amazon SageMaker Data Wrangler reduces the time it takes to aggregate and prepare data for machine learning (ML) from weeks to minutes. With SageMaker Data Wrangler, you can simplify the process of data preparation and feature engineering, and complete each step of the data preparation workflow (including data selection, cleansing, exploration, visualization, and processing at scale) from a single visual interface. You can use SQL to select the data you want from a wide variety of data sources and import it quickly. Next, you can use the Data Quality and Insights report to automatically verify data quality and detect anomalies, such as duplicate rows and target leakage. SageMaker Data Wrangler contains over 300 built-in data transformations so you can quickly transform data without writing any code. Once you have completed your data preparation workflow, you can scale it to your full datasets using SageMaker data processing jobs; train, tune, and deploy models.
  • 14
    Amazon SageMaker Canvas
    Amazon SageMaker Canvas expands access to machine learning (ML) by providing business analysts with a visual interface that allows them to generate accurate ML predictions on their own, without requiring any ML experience or having to write a single line of code. Visual point-and-click interface to connect, prepare, analyze, and explore data for building ML models and generating accurate predictions. Automatically build ML models to run what-if analysis and generate single or bulk predictions with a few clicks. Boost collaboration between business analysts and data scientists by sharing, reviewing, and updating ML models across tools. Import ML models from anywhere and generate predictions directly in Amazon SageMaker Canvas. With Amazon SageMaker Canvas, you can import data from disparate sources, select values you want to predict, automatically prepare and explore data, and quickly and more easily build ML models. You can then analyze models and generate accurate predictions.
  • 15
    Amazon SageMaker Edge
    The SageMaker Edge Agent allows you to capture data and metadata based on triggers that you set so that you can retrain your existing models with real-world data or build new models. Additionally, this data can be used to conduct your own analysis, such as model drift analysis. We offer three options for deployment. GGv2 (~ size 100MB) is a fully integrated AWS IoT deployment mechanism. For those customers with a limited device capacity, we have a smaller built-in deployment mechanism within SageMaker Edge. For customers who have a preferred deployment mechanism, we support third party mechanisms that can be plugged into our user flow. Amazon SageMaker Edge Manager provides a dashboard so you can understand the performance of models running on each device across your fleet. The dashboard helps you visually understand overall fleet health and identify the problematic models through a dashboard in the console.
  • 16
    Amazon SageMaker Clarify
    Amazon SageMaker Clarify provides machine learning (ML) developers with purpose-built tools to gain greater insights into their ML training data and models. SageMaker Clarify detects and measures potential bias using a variety of metrics so that ML developers can address potential bias and explain model predictions. SageMaker Clarify can detect potential bias during data preparation, after model training, and in your deployed model. For instance, you can check for bias related to age in your dataset or in your trained model and receive a detailed report that quantifies different types of potential bias. SageMaker Clarify also includes feature importance scores that help you explain how your model makes predictions and produces explainability reports in bulk or real time through online explainability. You can use these reports to support customer or internal presentations or to identify potential issues with your model.
  • 17
    Amazon SageMaker JumpStart
    Amazon SageMaker JumpStart is a machine learning (ML) hub that can help you accelerate your ML journey. With SageMaker JumpStart, you can access built-in algorithms with pretrained models from model hubs, pretrained foundation models to help you perform tasks such as article summarization and image generation, and prebuilt solutions to solve common use cases. In addition, you can share ML artifacts, including ML models and notebooks, within your organization to accelerate ML model building and deployment. SageMaker JumpStart provides hundreds of built-in algorithms with pretrained models from model hubs, including TensorFlow Hub, PyTorch Hub, HuggingFace, and MxNet GluonCV. You can also access built-in algorithms using the SageMaker Python SDK. Built-in algorithms cover common ML tasks, such as data classifications (image, text, tabular) and sentiment analysis.
  • 18
    Amazon SageMaker Autopilot
    Amazon SageMaker Autopilot eliminates the heavy lifting of building ML models. You simply provide a tabular dataset and select the target column to predict, and SageMaker Autopilot will automatically explore different solutions to find the best model. You then can directly deploy the model to production with just one click or iterate on the recommended solutions to further improve the model quality. You can use Amazon SageMaker Autopilot even when you have missing data. SageMaker Autopilot automatically fills in the missing data, provides statistical insights about columns in your dataset, and automatically extracts information from non-numeric columns, such as date and time information from timestamps.
  • 19
    Amazon SageMaker Model Monitor
    With Amazon SageMaker Model Monitor, you can select the data you would like to monitor and analyze without the need to write any code. SageMaker Model Monitor lets you select data from a menu of options such as prediction output, and captures metadata such as timestamp, model name, and endpoint so you can analyze model predictions based on the metadata. You can specify the sampling rate of data capture as a percentage of overall traffic in the case of high volume real-time predictions, and the data is stored in your own Amazon S3 bucket. You can also encrypt this data, configure fine-grained security, define data retention policies, and implement access control mechanisms for secure access. Amazon SageMaker Model Monitor offers built-in analysis in the form of statistical rules, to detect drifts in data and model quality. You can also write custom rules and specify thresholds for each rule.
  • 20
    Amazon SageMaker Pipelines
    Using Amazon SageMaker Pipelines, you can create ML workflows with an easy-to-use Python SDK, and then visualize and manage your workflow using Amazon SageMaker Studio. You can be more efficient and scale faster by storing and reusing the workflow steps you create in SageMaker Pipelines. You can also get started quickly with built-in templates to build, test, register, and deploy models so you can get started with CI/CD in your ML environment quickly. Many customers have hundreds of workflows, each with a different version of the same model. With the SageMaker Pipelines model registry, you can track these versions in a central repository where it is easy to choose the right model for deployment based on your business requirements. You can use SageMaker Studio to browse and discover models, or you can access them through the SageMaker Python SDK.
  • 21
    Amazon SageMaker Model Deployment
    Amazon SageMaker makes it easy to deploy ML models to make predictions (also known as inference) at the best price-performance for any use case. It provides a broad selection of ML infrastructure and model deployment options to help meet all your ML inference needs. It is a fully managed service and integrates with MLOps tools, so you can scale your model deployment, reduce inference costs, manage models more effectively in production, and reduce operational burden. From low latency (a few milliseconds) and high throughput (hundreds of thousands of requests per second) to long-running inference for use cases such as natural language processing and computer vision, you can use Amazon SageMaker for all your inference needs.
  • 22
    Robust Intelligence

    Robust Intelligence

    Robust Intelligence

    The Robust Intelligence Platform integrates seamlessly into your ML lifecycle to eliminate model failures. The platform detects your model’s vulnerabilities, prevents aberrant data from entering your AI system, and detects statistical data issues like drift. At the core of our test-based approach is a single test. Each test measures your model’s robustness to a specific type of production model failure. Stress Testing runs hundreds of these tests to measure model production readiness. The results of these tests are used to auto-configure a custom AI Firewall that protects the model against the specific forms of failure to which a given model is susceptible. Finally, Continuous Testing runs these tests during production, providing automated root cause analysis informed by the underlying cause of any single test failure. Using all three elements of the Robust Intelligence platform together helps ensure ML Integrity.
  • 23
    Amazon Bedrock
    Amazon Bedrock is a fully managed service that simplifies building and scaling generative AI applications by providing access to a variety of high-performing foundation models (FMs) from leading AI companies such as AI21 Labs, Anthropic, Cohere, Meta, Mistral AI, Stability AI, and Amazon itself. Through a single API, developers can experiment with these models, customize them using techniques like fine-tuning and Retrieval Augmented Generation (RAG), and create agents that interact with enterprise systems and data sources. As a serverless platform, Amazon Bedrock eliminates the need for infrastructure management, allowing seamless integration of generative AI capabilities into applications with a focus on security, privacy, and responsible AI practices.
  • 24
    Rendered.ai

    Rendered.ai

    Rendered.ai

    Overcome challenges in acquiring data for machine learning and AI systems training. Rendered.ai is a PaaS designed for data scientists, engineers, and developers. Generate synthetic datasets for ML/AI training and validation. Experiment with sensor models, scene content, and post-processing effects. Characterize and catalog real and synthetic datasets. Download or move data to your own cloud repositories for processing and training. Power innovation and increase productivity with synthetic data as a capability. Build custom pipelines to model diverse sensors and computer vision inputs​. Start quickly with free, customizable Python sample code to model SAR, RGB satellite imagery, and more sensor types​. Experiment and iterate with flexible licensing that enables nearly unlimited content generation. Create labeled content rapidly in a hosted, high-performance computing environment​. Enable collaboration between data scientists and data engineers with a no-code configuration experience.
  • 25
    Acryl Data

    Acryl Data

    Acryl Data

    No more data catalog ghost towns. Acryl Cloud drives fast time-to-value via Shift Left practices for data producers and an intuitive UI for data consumers. Continuously detect data quality incidents in real-time, automate anomaly detection to prevent breakages, and drive fast resolution when they do occur. Acryl Cloud supports both push-based and pull-based metadata ingestion for easy maintenance, ensuring information is trustworthy, up-to-date, and definitive. Data should be operational. Go beyond simple visibility and use automated Metadata Tests to continuously expose data insights and surface new areas for improvement. Reduce confusion and accelerate resolution with clear asset ownership, automatic detection, streamlined alerts, and time-based lineage for tracing root causes.
  • 26
    AWS Neuron

    AWS Neuron

    Amazon Web Services

    It supports high-performance training on AWS Trainium-based Amazon Elastic Compute Cloud (Amazon EC2) Trn1 instances. For model deployment, it supports high-performance and low-latency inference on AWS Inferentia-based Amazon EC2 Inf1 instances and AWS Inferentia2-based Amazon EC2 Inf2 instances. With Neuron, you can use popular frameworks, such as TensorFlow and PyTorch, and optimally train and deploy machine learning (ML) models on Amazon EC2 Trn1, Inf1, and Inf2 instances with minimal code changes and without tie-in to vendor-specific solutions. AWS Neuron SDK, which supports Inferentia and Trainium accelerators, is natively integrated with PyTorch and TensorFlow. This integration ensures that you can continue using your existing workflows in these popular frameworks and get started with only a few lines of code changes. For distributed model training, the Neuron SDK supports libraries, such as Megatron-LM and PyTorch Fully Sharded Data Parallel (FSDP).
  • 27
    APERIO DataWise
    Data is used in every aspect of a processing plant or facility, it is underlying most operational processes, most business decisions, and most environmental events. Failures are often attributed to this same data, in terms of operator error, bad sensors, safety or environmental events, or poor analytics. This is where APERIO can alleviate these problems. Data integrity is a key element of Industry 4.0; the foundation upon which more advanced applications, such as predictive models, process optimization, and custom AI tools are developed. APERIO DataWise is the industry-leading provider of reliable, trusted data. Automate the quality of your PI data or digital twins continuously and at scale. Ensure validated data across the enterprise to improve asset reliability. Empower the operator to make better decisions. Detect threats made to operational data to ensure operational resilience. Accurately monitor & report sustainability metrics.
  • 28
    Cranium

    Cranium

    Cranium

    The AI revolution is here. Innovation is moving at light speed, and the regulation landscape is constantly evolving. How can you make sure that your AI systems — and those of your vendors — remain secure, trustworthy, and compliant? Cranium helps cybersecurity and data science teams understand everywhere that AI is impacting their systems, data or services. Secure your organization’s AI and machine learning systems to ensure they are compliant and trustworthy, without interrupting your workflow. Protect against adversarial threats without impacting how your team trains, tests and deploys AI models. Increase AI regulatory awareness and alignment within your organization. Showcase the security and trustworthiness of your AI systems.
  • 29
    Determined AI

    Determined AI

    Determined AI

    Distributed training without changing your model code, determined takes care of provisioning machines, networking, data loading, and fault tolerance. Our open source deep learning platform enables you to train models in hours and minutes, not days and weeks. Instead of arduous tasks like manual hyperparameter tuning, re-running faulty jobs, and worrying about hardware resources. Our distributed training implementation outperforms the industry standard, requires no code changes, and is fully integrated with our state-of-the-art training platform. With built-in experiment tracking and visualization, Determined records metrics automatically, makes your ML projects reproducible and allows your team to collaborate more easily. Your researchers will be able to build on the progress of their team and innovate in their domain, instead of fretting over errors and infrastructure.
  • 30
    WhyLabs

    WhyLabs

    WhyLabs

    Enable observability to detect data and ML issues faster, deliver continuous improvements, and avoid costly incidents. Start with reliable data. Continuously monitor any data-in-motion for data quality issues. Pinpoint data and model drift. Identify training-serving skew and proactively retrain. Detect model accuracy degradation by continuously monitoring key performance metrics. Identify risky behavior in generative AI applications and prevent data leakage. Protect your generative AI applications are safe from malicious actions. Improve AI applications through user feedback, monitoring, and cross-team collaboration. Integrate in minutes with purpose-built agents that analyze raw data without moving or duplicating it, ensuring privacy and security. Onboard the WhyLabs SaaS Platform for any use cases using the proprietary privacy-preserving integration. Security approved for healthcare and banks.