Showing 133 open source projects for "training"

View related business solutions
  • $300 in Free Credit Towards Top Cloud Services Icon
    $300 in Free Credit Towards Top Cloud Services

    Build VMs, containers, AI, databases, storage—all in one place.

    Start your project in minutes. After credits run out, 20+ products include free monthly usage. Only pay when you're ready to scale.
    Get Started
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 1
    MMDeploy

    MMDeploy

    OpenMMLab Model Deployment Framework

    ...All kinds of modules in the SDK can be extended, such as Transform for image processing, Net for Neural Network inference, Module for postprocessing and so on. Install and build your target backend. ONNX Runtime is a cross-platform inference and training accelerator compatible with many popular ML/DNN frameworks. Please read getting_started for the basic usage of MMDeploy.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    Neural Tangents

    Neural Tangents

    Fast and Easy Infinite Neural Networks in Python

    ...It lets researchers define architectures from familiar building blocks—convolutions, pooling, residual connections, and nonlinearities—and obtain not only the finite network but also the corresponding Gaussian Process (GP) kernel of its infinite-width limit. With a single specification, you can compute NNGP and NTK kernels, perform exact GP inference, and study training dynamics analytically for infinitely wide networks. The library closely mirrors JAX’s stax API while extending it to return a kernel_fn alongside init_fn and apply_fn, enabling drop-in workflows for kernel computation. Kernel evaluation is highly optimized for speed and memory, and computations can be automatically distributed across accelerators with near-linear scaling.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    Evolutionary Algorithm

    Evolutionary Algorithm

    Evolutionary Algorithm using Python

    ...Users can explore basic genetic algorithm setups, match phrase examples, pathfinding challenges, and microbial GA variants, as well as evolution strategy approaches like NES. The project also links classical evolutionary approaches with neural networks, illustrating how evolution can be used for model training in reinforcement learning and supervised contexts.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    Neural Network Intelligence

    Neural Network Intelligence

    AutoML toolkit for automate machine learning lifecycle

    ...The tool manages automated machine learning (AutoML) experiments, dispatches and runs experiments' trial jobs generated by tuning algorithms to search the best neural architecture and/or hyper-parameters in different training environments like Local Machine, Remote Servers, OpenPAI, Kubeflow, FrameworkController on K8S (AKS etc.) DLWorkspace (aka. DLTS) AML (Azure Machine Learning) and other cloud options. NNI provides CommandLine Tool as well as an user friendly WebUI to manage training experiements.
    Downloads: 1 This Week
    Last Update:
    See Project
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • 5
    Horovod

    Horovod

    Distributed training framework for TensorFlow, Keras, PyTorch, etc.

    Horovod was originally developed by Uber to make distributed deep learning fast and easy to use, bringing model training time down from days and weeks to hours and minutes. With Horovod, an existing training script can be scaled up to run on hundreds of GPUs in just a few lines of Python code. Horovod can be installed on-premise or run out-of-the-box in cloud platforms, including AWS, Azure, and Databricks. Horovod can additionally run on top of Apache Spark, making it possible to unify data processing and model training into a single pipeline. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    fastMRI

    fastMRI

    A large open dataset + tools to speed up MRI scans using ML

    ...It includes reference implementations for key MRI reconstruction architectures such as U-Net and Variational Networks (VarNet), along with example scripts for model training and evaluation using the PyTorch Lightning framework. The project also releases several fully anonymized public MRI datasets, including knee, brain, and prostate scans.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 7
    iJEPA

    iJEPA

    Official codebase for I-JEPA

    ...The design scales naturally with Vision Transformer backbones and flexible masking strategies, and it trains stably at large batch sizes. i-JEPA’s predictions are made in embedding space, which is computationally efficient and better aligned with downstream discrimination tasks. The repository provides training recipes, data pipelines, and evaluation code that clarify which masking patterns and architectural choices matter most.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    learn2learn

    learn2learn

    A PyTorch Library for Meta-learning Research

    Learn2Learn is a PyTorch-based library focused on meta-learning and few-shot learning research. It provides reusable components and meta-learning algorithms, making it easier to build, train, and evaluate models that can quickly adapt to new tasks with minimal data. Learn2Learn is widely used in research for tasks such as few-shot classification, reinforcement learning, and optimization.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    ClassyVision

    ClassyVision

    An end-to-end PyTorch framework for image and video classification

    ...It offers high performance and scalability—capable of training models like ResNet-50 on ImageNet in just minutes—while remaining accessible to both researchers and production engineers. The library integrates seamlessly with PyTorch Hub for easy access to pretrained models and supports elastic training using PyTorch Elastic, making distributed training robust to changes in cluster resources or hardware failures.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Fully Managed MySQL, PostgreSQL, and SQL Server Icon
    Fully Managed MySQL, PostgreSQL, and SQL Server

    Automatic backups, patching, replication, and failover. Focus on your app, not your database.

    Cloud SQL handles your database ops end to end, so you can focus on your app.
    Try Free
  • 10
    Alpa

    Alpa

    Training and serving large-scale neural networks

    Alpa is a system for training and serving large-scale neural networks. Scaling neural networks to hundreds of billions of parameters has enabled dramatic breakthroughs such as GPT-3, but training and serving these large-scale neural networks require complicated distributed system techniques. Alpa aims to automate large-scale distributed training and serving with just a few lines of code.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    smclarify

    smclarify

    Fairness aware machine learning. Bias detection and mitigation

    ...A facet can have value(s) that designates that sample as "sensitive". Bias detection and mitigation for datasets and models. The label is a column or feature which is the target for training a machine learning model. The label can have value(s) that designates that sample as having a "positive" outcome. A bias measure is a function that returns a bias metric. A bias metric is a numerical value indicating the level of bias detected as determined by a particular bias measure. A collection of bias metrics for a given dataset or a combination of a dataset and model.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    FairScale

    FairScale

    PyTorch extensions for high performance and large scale training

    ...Its components are modular, so teams can adopt just the sharding optimizer or the pipeline engine without rewriting their training loop. FairScale puts emphasis on correctness and debuggability, offering hook points, logging, and reference examples for common trainer patterns. Although many ideas have since landed in core PyTorch, FairScale remains a valuable reference and a practical toolbox for squeezing more performance out of multi-GPU and multi-node jobs.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 13
    UnionML

    UnionML

    Build and deploy machine learning microservices

    ...Fit the rich ecosystem of tools and frameworks into a common protocol for machine learning. Using industry-standard machine learning methods, implement endpoints for fetching data, training models, serving predictions (and much more) to write a complete ML stack in one place. Data science, ML engineering, and MLOps practitioners can all gather around UnionML apps as a way of defining a single source of truth about your ML system’s behavior. This helps you maintain consistent code across your ML stack, from training to prediction logic.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    DeepCTR

    DeepCTR

    Package of deep-learning based CTR models

    ...You can use any complex model with model.fit(), and model.predict(). Provide tf.keras.Model like interface for quick experiment. Provide tensorflow estimator interface for large scale data and distributed training. It is compatible with both tf 1.x and tf 2.x. With the great success of deep learning,DNN-based techniques have been widely used in CTR prediction task. The data in CTR estimation task usually includes high sparse,high cardinality categorical features and some dense numerical features. Since DNN are good at handling dense numerical features,we usually map the sparse categorical features to dense numerical through embedding technique.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    Real-ESRGAN

    Real-ESRGAN

    Real-ESRGAN aims at developing Practical Algorithms

    Real-ESRGAN is a highly popular open-source project that provides practical algorithms for general image and video restoration using deep learning-based super-resolution techniques. It extends the original Enhanced Super-Resolution Generative Adversarial Network (ESRGAN) approach by training on synthetic degradations to make results more robust on real-world images, effectively enhancing resolution, reducing noise/artifacts, and reconstructing fine detail in low-quality imagery. The repository includes inference and training scripts, a model zoo with different pretrained models (including general and anime-oriented variants), and support for batch and arbitrary scaling, making it adaptable for diverse enhancement tasks. ...
    Downloads: 212 This Week
    Last Update:
    See Project
  • 16
    Jraph

    Jraph

    A Graph Neural Network Library in Jax

    Jraph (pronounced “giraffe”) is a lightweight JAX library developed by Google DeepMind for building and experimenting with graph neural networks (GNNs). It provides an efficient and flexible framework for representing, manipulating, and training models on graph-structured data. The core of Jraph is the GraphsTuple data structure, which enables users to define graphs with arbitrary node, edge, and global attributes, and to batch variable-sized graphs efficiently for JAX’s just-in-time compilation. The library includes a comprehensive set of utilities for batching, padding, masking, and partitioning graph data, making it ideal for distributed and large-scale GNN experiments. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    Fairseq

    Fairseq

    Facebook AI Research Sequence-to-Sequence Toolkit written in Python

    Fairseq(-py) is a sequence modeling toolkit that allows researchers and developers to train custom models for translation, summarization, language modeling and other text generation tasks. We provide reference implementations of various sequence modeling papers. Recent work by Microsoft and Google has shown that data parallel training can be made significantly more efficient by sharding the model parameters and optimizer state across data parallel workers. These ideas are encapsulated in the new FullyShardedDataParallel (FSDP) wrapper provided by fairscale. Fairseq can be extended through user-supplied plug-ins. Models define the neural network architecture and encapsulate all of the learnable parameters. ...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 18
    SageMaker MXNet Inference Toolkit

    SageMaker MXNet Inference Toolkit

    Toolkit for allowing inference and serving with MXNet in SageMaker

    ...This library provides default pre-processing, predict and postprocessing for certain MXNet model types and utilizes the SageMaker Inference Toolkit for starting up the model server, which is responsible for handling inference requests. AWS Deep Learning Containers (DLCs) are a set of Docker images for training and serving models in TensorFlow, TensorFlow 2, PyTorch, and MXNet. Deep Learning Containers provide optimized environments with TensorFlow and MXNet, Nvidia CUDA (for GPU instances), and Intel MKL (for CPU instances) libraries and are available in the Amazon Elastic Container Registry (Amazon ECR). The AWS DLCs are used in Amazon SageMaker as the default vehicles for your SageMaker jobs such as training, inference, transforms etc. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    Blankly

    Blankly

    Easily build, backtest and deploy your algo in just a few lines

    ...Models can be instantly backtested, paper traded, sandbox tested and run live by simply changing a single line. We built blankly for every type of quant including training & running ML models in the same environment, cross-exchange/cross-symbol arbitrage, and even long/short positions on stocks (all with built-in WebSockets). Blankly is the first framework to enable developers to backtest, paper trade, and go live across exchanges without modifying a single line of trading logic on stocks, crypto, and forex. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    AugLy

    AugLy

    A data augmentations library for audio, image, text, and video

    ...These sub-libraries include both function-based and class-based transforms, composition operators, and have the option to provide metadata about the transform applied, including its intensity. AugLy is a great library to utilize for augmenting your data in model training, or to evaluate the robustness gaps of your model! We designed AugLy to include many specific data augmentations that users perform in real life on internet platforms like Facebook's -- for example making an image into a meme, overlaying text/emojis on images/videos, reposting a screenshot from social media. While AugLy contains more generic data augmentations as well, it will be particularly useful to you if you're working on a problem like copy detection, hate speech detection, etc.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 21
    Deep Daze

    Deep Daze

    Simple command line tool for text to image generation

    ...Technique first devised and shared by Mario Klingemann, it allows you to prime the generator network with a starting image, before being steered towards the text. Simply specify the path to the image you wish to use, and optionally the number of initial training steps. We can also feed in an image as an optimization goal, instead of only priming the generator network. Deepdaze will then render its own interpretation of that image. The regular mode for texts only allows 77 tokens. If you want to visualize a full story/paragraph/song/poem, set create_story to True.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    Model Search

    Model Search

    Framework that implements AutoML algorithms

    ...Instead of hand-crafting models, you define a search space and objectives, then the system explores candidate architectures using controllers and population-based strategies. It supports multiple tasks (such as vision or text) by letting you express reusable building blocks—layers, cells, and topologies—that the search can recombine. Training, evaluation, and promotion of candidates are orchestrated automatically, with strong emphasis on reproducibility and fair comparisons. The framework logs trials, metrics, and artifacts so you can analyze what the search learned and why certain designs dominate. It’s intended as a platform for method development as much as for model discovery.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    Big Sleep

    Big Sleep

    A simple command line tool for text to image generation

    ...You can set the number of classes that you wish to restrict Big Sleep to use for the Big GAN with the --max-classes flag as follows (ex. 15 classes). This may lead to extra stability during training, at the cost of lost expressivity.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    SVoice (Speech Voice Separation)

    SVoice (Speech Voice Separation)

    We provide a PyTorch implementation of the paper Voice Separation

    ...Separate models are trained for different speaker counts, and the largest-capacity model dynamically determines the actual number of speakers in a mixture. The repository includes all necessary scripts for training, dataset preparation, distributed training, evaluation, and audio separation.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 25
    AWS Step Functions Data Science SDK

    AWS Step Functions Data Science SDK

    For building machine learning (ML) workflows and pipelines on AWS

    The AWS Step Functions Data Science SDK is an open-source library that allows data scientists to easily create workflows that process and publish machine learning models using Amazon SageMaker and AWS Step Functions. You can create machine learning workflows in Python that orchestrate AWS infrastructure at scale, without having to provision and integrate the AWS services separately. The best way to quickly review how the AWS Step Functions Data Science SDK works is to review the related...
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB