Showing 890 open source projects for "training"

View related business solutions
  • Custom VMs From 1 to 96 vCPUs With 99.95% Uptime Icon
    Custom VMs From 1 to 96 vCPUs With 99.95% Uptime

    General-purpose, compute-optimized, or GPU/TPU-accelerated. Built to your exact specs.

    Live migration and automatic failover keep workloads online through maintenance. One free e2-micro VM every month.
    Try Free
  • Earn up to 16% annual interest with Nexo. Icon
    Earn up to 16% annual interest with Nexo.

    More flexibility. More control.

    Generate interest, access liquidity without selling, and execute trades seamlessly. All in one platform. Geographic restrictions, eligibility, and terms apply.
    Get started with Nexo.
  • 1
    SageMaker Containers

    SageMaker Containers

    Create SageMaker-compatible Docker containers

    Amazon SageMaker is a fully managed service for data science and machine learning (ML) workflows. You can use Amazon SageMaker to simplify the process of building, training, and deploying ML models. To train a model, you can include your training script and dependencies in a Docker container that runs your training code. A container provides an effectively isolated environment, ensuring a consistent runtime and reliable training process. The SageMaker Training Toolkit can be easily added to any Docker container, making it compatible with SageMaker for training models. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    NLP-Models-Tensorflow

    NLP-Models-Tensorflow

    Gathers machine learning and Tensorflow deep learning models for NLP

    ...Each model implementation is designed to illustrate how common NLP architectures operate, such as recurrent neural networks, convolutional models for text processing, and transformer-style attention mechanisms. The project includes scripts for preparing datasets, training models, and evaluating performance on various text analysis tasks. Many implementations are designed for experimentation, allowing developers to adjust parameters, swap architectures, and test different preprocessing techniques.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    Delta ML

    Delta ML

    Deep learning based natural language and speech processing platform

    ...DELTA has been used for developing several state-of-the-art algorithms for publications and delivering real production to serve millions of users. It helps you to train, develop, and deploy NLP and/or speech models. Use configuration files to easily tune parameters and network structures. What you see in training is what you get in serving: all data processing and features extraction are integrated into a model graph. Text classification, named entity recognition, question and answering, text summarization, etc. Uniform I/O interfaces and no changes for new models.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    Linux-Intelligent-Ocr-Solution

    Linux-Intelligent-Ocr-Solution

    Easy-OCR solution and Tesseract trainer for GNU/Linux

    ...A Tesseract Trainer GUI is also shipped with this package. Forum : https://groups.google.com/forum/#!forum/lios Video Tutorial : https://www.youtube.com/playlist?list=PLn29o8rxtRe1zS1r2-yGm1DNMOZCgdU0i Tesseract Training Tutorial (beta) : https://www.youtube.com/watch?v=qLpCld4cdtk Source Code Github : https://github.com/Nalin-x-Linux/lios-3 Gitlab : https://gitlab.com/Nalin-x-Linux/lios-3 User guide is available in download page
    Downloads: 4 This Week
    Last Update:
    See Project
  • Streamline Azure Security with Palo Alto Networks VM-Series Icon
    Streamline Azure Security with Palo Alto Networks VM-Series

    Centrally manage physical and virtualized firewalls with Panorama

    Improve your security posture and reduce incident response time. Use the VM-Series to natively analyze Azure traffic and dynamically drive policy updates based on workload changes.
    Learn more
  • 5
    AdaNet

    AdaNet

    Fast and flexible AutoML with learning guarantees

    ...Adaptive neural architecture search and ensemble learning in a single train call. Regression, binary and multi-class classification, and multi-head task support. A tf.estimator.Estimator API for training, evaluation, prediction, and serving models.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    TFKit

    TFKit

    Handling multiple nlp task in one pipeline

    ...It leverages the use of transformers on many tasks with different models in this all-in-one framework. All you need is a little change of config. You can use tfkit for model training and evaluation with tfkit-train and tfkit-eval. The key to combine different task together is to make different task with same data format. All data will be in csv format - tfkit will use csv for all task, normally it will have two columns, first columns is the input of models, the second column is the output of models. Plane text with no tokenization - there is no need to tokenize text before training, or do re-calculating for tokenization, tfkit will handle it for you. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    Albedo

    Albedo

    A recommender system for discovering GitHub repos

    ...The project focuses on implicit feedback—stars, watches, and other engagement metrics—so it can build useful recommendations without explicit ratings. A reproducible setup and Makefile-driven workflow streamline tasks like spinning up services, loading datasets, training models, and generating candidate lists. Because it’s built around Spark’s scalable primitives, Albedo can experiment on substantial snapshots of GitHub metadata rather than toy corpora. The repo is also educational: it demonstrates a practical end-to-end pipeline from ingestion and feature preparation to training and ranking.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    PyText

    PyText

    A natural language modeling framework based on PyTorch

    ...It achieves this by providing simple and extensible interfaces and abstractions for model components, and by using PyTorch’s capabilities of exporting models for inference via the optimized Caffe2 execution engine. We use PyText at Facebook to iterate quickly on new modeling ideas and then seamlessly ship them at scale. Distributed-training support built on the new C10d backend in PyTorch 1.0. Mixed precision training support through APEX (trains faster with less GPU memory on NVIDIA Tensor Cores). Extensible components that allows easy creation of new models and tasks.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    pytorch-tutorial

    pytorch-tutorial

    PyTorch Tutorial for Deep Learning Researchers

    pytorch-tutorial is a highly popular educational repository that teaches deep learning with PyTorch through step-by-step examples and well-structured lessons. It is designed primarily for beginners and intermediate practitioners who want to understand PyTorch fundamentals and quickly move toward building real neural network models. The repository walks users through core concepts such as tensors, autograd, neural network modules, convolutional networks, recurrent networks, and transfer...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Compliant and Reliable File Transfers Backed by Top Security Certifications Icon
    Compliant and Reliable File Transfers Backed by Top Security Certifications

    Cerberus FTP Server delivers SOC 2 Type II certified security and FIPS 140-2 validated encryption.

    Stop relying on non-certified, legacy file transfer tools that creak under the weight of modern security demands. Get full audit trails, advanced access controls and more supported by an award-winning team of experts. Start your free 25-day trial today.
    Start Free Trial
  • 10
    ENAS in PyTorch

    ENAS in PyTorch

    PyTorch implementation of "Efficient Neural Architecture Search

    ...It is primarily intended as a research and educational codebase, helping practitioners understand how ENAS works in practice and how to reproduce results on benchmark datasets. The project includes training scripts, model definitions, and search procedures that show the full workflow from architecture sampling to evaluation. Because ENAS relies on shared weights among candidate models, the implementation emphasizes efficiency and experiment reproducibility.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    Tensor2Tensor

    Tensor2Tensor

    Library of deep learning models and datasets

    Deep Learning (DL) has enabled the rapid advancement of many useful technologies, such as machine translation, speech recognition and object detection. In the research community, one can find code open-sourced by the authors to help in replicating their results and further advancing deep learning. However, most of these DL systems use unique setups that require significant engineering effort and may only work for a specific problem or architecture, making it hard to run new experiments and...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 12
    DETR

    DETR

    End-to-end object detection with transformers

    PyTorch training code and pretrained models for DETR (DEtection TRansformer). We replace the full complex hand-crafted object detection pipeline with a Transformer, and match Faster R-CNN with a ResNet-50, obtaining 42 AP on COCO using half the computation power (FLOPs) and the same number of parameters. Inference in 50 lines of PyTorch. What it is.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    Super-résolution via CNN

    Super-résolution via CNN

    Super resolution using a CNN, based on the work of the DGtal team

    ...This program will generate "model_epoch_ .pth" files corresponding to the model at epoch n, in a folder saved_model_u t_bs bs_tbs tbs_lr lr, where corresponds to the scale factor, bsthe size of the training batch, tbsthe size of the test batch and lrto the learning rate. Low res images should be located in a "dataset/input" folder, and high res targets in a "dataset/target" folder, where each different quality image has the same name in both folders.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    Tensorflow and deep learning

    Tensorflow and deep learning

    A crash course in six episodes for software developers

    Tensorflow and deep learning repository is an educational deep learning crash course designed to help software developers quickly understand and apply machine learning concepts without requiring advanced academic background. It is structured as a series of guided lessons that combine theoretical explanations, practical examples, and runnable code, allowing learners to build intuition while actively experimenting with models. The repository covers core neural network concepts such as weights,...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    StarGAN

    StarGAN

    Official PyTorch Implementation

    ...Unlike earlier GAN approaches that required separate models for each domain pair, StarGAN enables flexible attribute transfer across multiple domains within one network, significantly improving efficiency and scalability. The repository includes full training and inference pipelines for tasks such as facial attribute manipulation and style transfer. It demonstrates adversarial training strategies, domain classification losses, and generator-discriminator coordination required for stable multi-domain translation. Researchers and practitioners often use the project as a reference when studying conditional GANs and advanced image synthesis techniques. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    COCO Annotator

    COCO Annotator

    Web-based image segmentation tool for object detection & localization

    COCO Annotator is a web-based image annotation tool designed for versatility and efficiently label images to create training data for image localization and object detection. It provides many distinct features including the ability to label an image segment (or part of a segment), track object instances, label objects with disconnected visible parts, and efficiently store and export annotations in the well-known COCO format. The annotation process is delivered through an intuitive and customizable interface and provides many tools for creating accurate datasets. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    Retro

    Retro

    Retro Games in Gym

    ...This design improves factual accuracy, reduces hallucinations, and enables smaller models to perform comparably to much larger ones by leveraging retrieval. The repository provides code and resources for training and evaluating RETRO models, along with infrastructure for integrating retrieval into the transformer pipeline. It includes example configurations, datasets, and utilities for building retrieval-augmented generation systems. RETRO represents an important step toward combining large-scale language modeling with information retrieval, offering researchers a foundation to study hybrid approaches for scaling AI responsibly.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 18
    PyTorch GAN Zoo

    PyTorch GAN Zoo

    A mix of GAN implementations including progressive growing

    ...It is built to support both researchers and developers who want to train, evaluate, and extend GANs efficiently across diverse datasets such as CelebA-HQ, FashionGen, DTD, and CIFAR-10. In addition to core GAN training, the repository includes tools for model evaluation, such as Inception Score and SWD metrics, as well as advanced features like GDPP for diverse generation and AC-GAN conditioning for class-specific synthesis. The framework also supports “inspirational generation,” enabling style or content transfer from reference images through pre-trained models.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    End-to-End Negotiator

    End-to-End Negotiator

    Deal or No Deal? End-to-End Learning for Negotiation Dialogues

    ...It enables agents to plan, reason, and communicate effectively to maximize outcomes in multi-turn negotiations over shared resources. The framework provides code for both supervised learning (training from human dialogue data) and reinforcement learning (via self-play and rollout-based planning). It introduces a hierarchical latent model, where high-level intents are first clustered and then translated into coherent language, improving dialogue diversity and goal consistency. The repository also includes the Negotiate dataset, comprising over 5,800 dialogues across 2,200 unique scenarios.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    MADDPG

    MADDPG

    Code for the MADDPG algorithm from a paper

    ...The repository implements a multi-agent reinforcement learning algorithm that extends DDPG to scenarios where multiple agents interact in shared environments. Each agent has its own policy, but training uses centralized critics conditioned on the observations and actions of all agents, enabling learning in cooperative, competitive, and mixed settings. The code is built on top of TensorFlow and integrates with the Multiagent Particle Environments (MPE) for benchmarking. Researchers can use it to reproduce the experiments presented in the paper, which demonstrate how agents learn behaviors such as coordination, competition, and communication. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 21
    RL Baselines Zoo

    RL Baselines Zoo

    A collection of 100+ pre-trained RL agents using Stable Baselines

    RL Baselines Zoo is a comprehensive training framework and collection of pre-trained RL agents using Stable-Baselines3. It offers tools for training, tuning, and evaluating RL algorithms across many standard environments, including MuJoCo, Atari, and robotics simulations. Designed for reproducible RL research and benchmarking, it includes scripts, hyperparameter presets, and best practices for training robust agents.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    Knock Knock

    Knock Knock

    Get notified when your training ends

    ...The goal of the project is to allow developers to monitor experiments remotely without needing to stay connected to the training environment. By adding only a few lines of code, the library can wrap around a training function and report execution status.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    SageMaker Chainer Containers

    SageMaker Chainer Containers

    Docker container for running Chainer scripts to train and host Chainer

    ...This repository also contains Dockerfiles which install this library, Chainer, and dependencies for building SageMaker Chainer images. Amazon SageMaker utilizes Docker containers to run all training jobs & inference endpoints. The Docker images are built from the Dockerfiles specified in Docker/. The Docker files are grouped based on Chainer version and separated based on Python version and processor type. The Docker images, used to run training & inference jobs, are built from both corresponding "base" and "final" Dockerfiles. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    micrograd

    micrograd

    A tiny scalar-valued autograd engine and a neural net library

    micrograd is a tiny, educational automatic differentiation engine focused on scalar values, built to show how backpropagation works end to end with minimal code. It constructs a dynamic computation graph as you perform math operations and then computes gradients by walking that graph backward, making it an approachable “from scratch” autograd reference. On top of the core autograd “Value” concept, the project includes a small neural network library that lets you define and train simple...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    BytePS

    BytePS

    A high performance and generic framework for distributed DNN training

    BytePS is a high-performance and generally distributed training framework. It supports TensorFlow, Keras, PyTorch, and MXNet, and can run on either TCP or RDMA networks. BytePS outperforms existing open-sourced distributed training frameworks by a large margin. For example, on BERT-large training, BytePS can achieve ~90% scaling efficiency with 256 GPUs (see below), which is much higher than Horovod+NCCL.
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB