Compare the Top Machine Learning Software that integrates with Amazon S3 as of October 2025 - Page 2

This a list of Machine Learning software that integrates with Amazon S3. Use the filters on the left to add additional filters for products that have integrations with Amazon S3. View the products that work with Amazon S3 in the table below.

  • 1
    Feast

    Feast

    Tecton

    Make your offline data available for real-time predictions without having to build custom pipelines. Ensure data consistency between offline training and online inference, eliminating train-serve skew. Standardize data engineering workflows under one consistent framework. Teams use Feast as the foundation of their internal ML platforms. Feast doesn’t require the deployment and management of dedicated infrastructure. Instead, it reuses existing infrastructure and spins up new resources when needed. You are not looking for a managed solution and are willing to manage and maintain your own implementation. You have engineers that are able to support the implementation and management of Feast. You want to run pipelines that transform raw data into features in a separate system and integrate with it. You have unique requirements and want to build on top of an open source solution.
  • 2
    Zepl

    Zepl

    Zepl

    Sync, search and manage all the work across your data science team. Zepl’s powerful search lets you discover and reuse models and code. Use Zepl’s enterprise collaboration platform to query data from Snowflake, Athena or Redshift and build your models in Python. Use pivoting and dynamic forms for enhanced interactions with your data using heatmap, radar, and Sankey charts. Zepl creates a new container every time you run your notebook, providing you with the same image each time you run your models. Invite team members to join a shared space and work together in real time or simply leave their comments on a notebook. Use fine-grained access controls to share your work. Allow others have read, edit, and run access as well as enable collaboration and distribution. All notebooks are auto-saved and versioned. You can name, manage and roll back all versions through an easy-to-use interface, and export seamlessly into Github.
  • 3
    Amazon SageMaker Feature Store
    Amazon SageMaker Feature Store is a fully managed, purpose-built repository to store, share, and manage features for machine learning (ML) models. Features are inputs to ML models used during training and inference. For example, in an application that recommends a music playlist, features could include song ratings, listening duration, and listener demographics. Features are used repeatedly by multiple teams and feature quality is critical to ensure a highly accurate model. Also, when features used to train models offline in batch are made available for real-time inference, it’s hard to keep the two feature stores synchronized. SageMaker Feature Store provides a secured and unified store for feature use across the ML lifecycle. Store, share, and manage ML model features for training and inference to promote feature reuse across ML applications. Ingest features from any data source including streaming and batch such as application logs, service logs, clickstreams, sensors, etc.
  • 4
    Amazon SageMaker Data Wrangler
    Amazon SageMaker Data Wrangler reduces the time it takes to aggregate and prepare data for machine learning (ML) from weeks to minutes. With SageMaker Data Wrangler, you can simplify the process of data preparation and feature engineering, and complete each step of the data preparation workflow (including data selection, cleansing, exploration, visualization, and processing at scale) from a single visual interface. You can use SQL to select the data you want from a wide variety of data sources and import it quickly. Next, you can use the Data Quality and Insights report to automatically verify data quality and detect anomalies, such as duplicate rows and target leakage. SageMaker Data Wrangler contains over 300 built-in data transformations so you can quickly transform data without writing any code. Once you have completed your data preparation workflow, you can scale it to your full datasets using SageMaker data processing jobs; train, tune, and deploy models.
  • 5
    Outerbounds

    Outerbounds

    Outerbounds

    Design and develop data-intensive projects with human-friendly, open-source Metaflow. Run, scale, and deploy them reliably on the fully managed Outerbounds platform. One platform for all your ML and data science projects. Access data securely from your existing data warehouses. Compute with a cluster optimized for scale and cost. 24/7 managed orchestration for production workflows. Use results to power any application. Give your data scientists superpowers, approved by your engineers. Outerbounds Platform allows data scientists to develop rapidly, experiment at scale, and deploy to production confidently. All within the outer bounds of policies and processes defined by your engineers, running on your cloud account, fully managed by us. Security is in our DNA, not at the perimeter. The platform adapts to your policies and compliance requirements through multiple layers of security. Centralized auth, a strict permission boundary, and granular task execution roles.
  • 6
    Robust Intelligence

    Robust Intelligence

    Robust Intelligence

    The Robust Intelligence Platform integrates seamlessly into your ML lifecycle to eliminate model failures. The platform detects your model’s vulnerabilities, prevents aberrant data from entering your AI system, and detects statistical data issues like drift. At the core of our test-based approach is a single test. Each test measures your model’s robustness to a specific type of production model failure. Stress Testing runs hundreds of these tests to measure model production readiness. The results of these tests are used to auto-configure a custom AI Firewall that protects the model against the specific forms of failure to which a given model is susceptible. Finally, Continuous Testing runs these tests during production, providing automated root cause analysis informed by the underlying cause of any single test failure. Using all three elements of the Robust Intelligence platform together helps ensure ML Integrity.
  • 7
    Amazon EC2 UltraClusters
    Amazon EC2 UltraClusters enable you to scale to thousands of GPUs or purpose-built machine learning accelerators, such as AWS Trainium, providing on-demand access to supercomputing-class performance. They democratize supercomputing for ML, generative AI, and high-performance computing developers through a simple pay-as-you-go model without setup or maintenance costs. UltraClusters consist of thousands of accelerated EC2 instances co-located in a given AWS Availability Zone, interconnected using Elastic Fabric Adapter (EFA) networking in a petabit-scale nonblocking network. This architecture offers high-performance networking and access to Amazon FSx for Lustre, a fully managed shared storage built on a high-performance parallel file system, enabling rapid processing of massive datasets with sub-millisecond latencies. EC2 UltraClusters provide scale-out capabilities for distributed ML training and tightly coupled HPC workloads, reducing training times.