Best Artificial Intelligence Software for Amazon EC2 - Page 2

Compare the Top Artificial Intelligence Software that integrates with Amazon EC2 as of October 2025 - Page 2

This a list of Artificial Intelligence software that integrates with Amazon EC2. Use the filters on the left to add additional filters for products that have integrations with Amazon EC2. View the products that work with Amazon EC2 in the table below.

  • 1
    Amazon EC2 Trn2 Instances
    Amazon EC2 Trn2 instances, powered by AWS Trainium2 chips, are purpose-built for high-performance deep learning training of generative AI models, including large language models and diffusion models. They offer up to 50% cost-to-train savings over comparable Amazon EC2 instances. Trn2 instances support up to 16 Trainium2 accelerators, providing up to 3 petaflops of FP16/BF16 compute power and 512 GB of high-bandwidth memory. To facilitate efficient data and model parallelism, Trn2 instances feature NeuronLink, a high-speed, nonblocking interconnect, and support up to 1600 Gbps of second-generation Elastic Fabric Adapter (EFAv2) network bandwidth. They are deployed in EC2 UltraClusters, enabling scaling up to 30,000 Trainium2 chips interconnected with a nonblocking petabit-scale network, delivering 6 exaflops of compute performance. The AWS Neuron SDK integrates natively with popular machine learning frameworks like PyTorch and TensorFlow.
  • 2
    AWS Elastic Fabric Adapter (EFA)
    Elastic Fabric Adapter (EFA) is a network interface for Amazon EC2 instances that enables customers to run applications requiring high levels of inter-node communications at scale on AWS. Its custom-built operating system (OS) bypass hardware interface enhances the performance of inter-instance communications, which is critical to scaling these applications. With EFA, High-Performance Computing (HPC) applications using the Message Passing Interface (MPI) and Machine Learning (ML) applications using NVIDIA Collective Communications Library (NCCL) can scale to thousands of CPUs or GPUs. As a result, you get the application performance of on-premises HPC clusters with the on-demand elasticity and flexibility of the AWS cloud. EFA is available as an optional EC2 networking feature that you can enable on any supported EC2 instance at no additional cost. Plus, it works with the most commonly used interfaces, APIs, and libraries for inter-node communications.
  • 3
    MLlib

    MLlib

    Apache Software Foundation

    ​Apache Spark's MLlib is a scalable machine learning library that integrates seamlessly with Spark's APIs, supporting Java, Scala, Python, and R. It offers a comprehensive suite of algorithms and utilities, including classification, regression, clustering, collaborative filtering, and tools for constructing machine learning pipelines. MLlib's high-quality algorithms leverage Spark's iterative computation capabilities, delivering performance up to 100 times faster than traditional MapReduce implementations. It is designed to operate across diverse environments, running on Hadoop, Apache Mesos, Kubernetes, standalone clusters, or in the cloud, and accessing various data sources such as HDFS, HBase, and local files. This flexibility makes MLlib a robust solution for scalable and efficient machine learning tasks within the Apache Spark ecosystem. ​
  • 4
    Amazon EC2 G4 Instances
    Amazon EC2 G4 instances are optimized for machine learning inference and graphics-intensive applications. It offers a choice between NVIDIA T4 GPUs (G4dn) and AMD Radeon Pro V520 GPUs (G4ad). G4dn instances combine NVIDIA T4 GPUs with custom Intel Cascade Lake CPUs, providing a balance of compute, memory, and networking resources. These instances are ideal for deploying machine learning models, video transcoding, game streaming, and graphics rendering. G4ad instances, featuring AMD Radeon Pro V520 GPUs and 2nd-generation AMD EPYC processors, deliver cost-effective solutions for graphics workloads. Both G4dn and G4ad instances support Amazon Elastic Inference, allowing users to attach low-cost GPU-powered inference acceleration to Amazon EC2 and reduce deep learning inference costs. They are available in various sizes to accommodate different performance needs and are integrated with AWS services such as Amazon SageMaker, Amazon ECS, and Amazon EKS.
  • 5
    StackState

    StackState

    StackState

    StackState's Topology and Relationship-Based Observability platform lets you manage your dynamic IT environment more effectively by unifying performance data from your existing monitoring tools into a single topology. Enabling you to: 1. 80% Decreased MTTR: by identifying the root cause and alerting the right teams with the correct information. 2. 65% Fewer Outages: through real-time unified observability and more planful planning. 3. 3x Faster Releases: by giving time back to developers to increase implementations. Get started today with our free guided demo: https://www.stackstate.com/schedule-a-demo
  • 6
    Amazon CodeWhisperer
    Build apps faster with ML-powered coding companion. Accelerate application development with automatic code recommendations based on the code and comments in your IDE. Empower developers to use artificial intelligence (AI) responsibly to create syntactically correct and secure applications. Generate entire functions and logical code blocks without having to search and customize code snippets from the web. Stay focused and never leave the IDE, with real-time customized code recommendations for all your Java, Python, and JavaScript projects. Amazon CodeWhisperer is a machine learning (ML)–powered service that helps improve developer productivity by generating code recommendations based on their comments in natural language and code in the integrated development environment (IDE). Accelerate frontend and backend development by empowering developers with automatic code recommendations. Save time and effort by using CodeWhisperer to generate code to build and train your ML models.