Showing 230 open source projects for "training"

View related business solutions
  • Try Google Cloud Risk-Free With $300 in Credit Icon
    Try Google Cloud Risk-Free With $300 in Credit

    No hidden charges. No surprise bills. Cancel anytime.

    Use your credit across every product. Compute, storage, AI, analytics. When it runs out, 20+ products stay free. You only pay when you choose to.
    Start Free
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • 1
    picoGPT

    picoGPT

    An unnecessarily tiny implementation of GPT-2 in NumPy

    ...It allows users to understand how tokenization, transformer layers, attention mechanisms, and autoregressive text generation operate in modern large language models. The project uses a small amount of code to illustrate the essential mathematical operations involved in training and running a transformer-based neural network. Because the code is intentionally lightweight, it is often used as a teaching resource for students learning about natural language processing and deep learning architectures. Developers can explore the repository to understand how language models generate text and how transformer components interact within the architecture.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    minimalRL-pytorch

    minimalRL-pytorch

    Implementations of basic RL algorithms with minimal lines of codes

    minimalRL is a lightweight reinforcement learning repository that implements several classic algorithms using minimal PyTorch code. The project is designed primarily as an educational resource that demonstrates how reinforcement learning algorithms work internally without the complexity of large frameworks. Each algorithm implementation is contained within a single file and typically ranges from about 100 to 150 lines of code, making it easy for learners to inspect the entire implementation...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    sense2vec

    sense2vec

    Contextually-keyed word vectors

    sense2vec (Trask et. al, 2015) is a nice twist on word2vec that lets you learn more interesting and detailed word vectors. This library is a simple Python implementation for loading, querying and training sense2vec models. For more details, check out our blog post. To explore the semantic similarities across all Reddit comments of 2015 and 2019, see the interactive demo.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    FFCV

    FFCV

    Fast Forward Computer Vision (and other ML workloads!)

    ffcv is a drop-in data loading system that dramatically increases data throughput in model training. From gridding to benchmarking to fast research iteration, there are many reasons to want faster model training. Below we present premade codebases for training on ImageNet and CIFAR, including both (a) extensible codebases and (b) numerous premade training configurations.
    Downloads: 1 This Week
    Last Update:
    See Project
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 5
    Keras Attention Mechanism

    Keras Attention Mechanism

    Attention mechanism Implementation for Keras

    ...We consider two LSTM networks: one with this attention layer and the other one with a fully connected layer. Both have the same number of parameters for a fair comparison (250K). The attention is expected to be the highest after the delimiters. An overview of the training is shown below, where the top represents the attention map and the bottom the ground truth. As the training progresses, the model learns the task and the attention map converges to the ground truth. We consider many 1D sequences of the same length. The task is to find the maximum of each sequence. We give the full sequence processed by the RNN layer to the attention layer. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 6
    smclarify

    smclarify

    Fairness aware machine learning. Bias detection and mitigation

    ...A facet can have value(s) that designates that sample as "sensitive". Bias detection and mitigation for datasets and models. The label is a column or feature which is the target for training a machine learning model. The label can have value(s) that designates that sample as having a "positive" outcome. A bias measure is a function that returns a bias metric. A bias metric is a numerical value indicating the level of bias detected as determined by a particular bias measure. A collection of bias metrics for a given dataset or a combination of a dataset and model.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    NanoDet-Plus

    NanoDet-Plus

    Lightweight anchor-free object detection model

    ...In NanoDet-Plus, we propose a novel label assignment strategy with a simple assign guidance module (AGM) and a dynamic soft label assigner (DSLA) to solve the optimal label assignment problem in lightweight model training. We also introduce a light feature pyramid called Ghost-PAN to enhance multi-layer feature fusion. These improvements boost previous NanoDet's detection accuracy by 7 mAP on COCO dataset. NanoDet provide multi-backend C++ demo including ncnn, OpenVINO and MNN. There is also an Android demo based on ncnn library. Supports various backends including ncnn, MNN and OpenVINO. ...
    Downloads: 8 This Week
    Last Update:
    See Project
  • 8
    Sockeye

    Sockeye

    Sequence-to-sequence framework, focused on Neural Machine Translation

    Sockeye is an open-source sequence-to-sequence framework for Neural Machine Translation built on PyTorch. It implements distributed training and optimized inference for state-of-the-art models, powering Amazon Translate and other MT applications. For a quickstart guide to training a standard NMT model on any size of data, see the WMT 2014 English-German tutorial. If you are interested in collaborating or have any questions, please submit a pull request or issue. You can also send questions to sockeye-dev-at-amazon-dot-com. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    CodeContests

    CodeContests

    Large dataset of coding contests designed for AI and ML model training

    ...Each problem includes structured metadata, problem descriptions, paired input/output test cases, and multiple correct and incorrect solutions in various programming languages. The dataset is distributed in Riegeli format using Protocol Buffers, with separate training, validation, and test splits for reproducible machine learning experiments.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8 Monitoring Tools in One APM. Install in 5 Minutes. Icon
    8 Monitoring Tools in One APM. Install in 5 Minutes.

    Errors, performance, logs, uptime, hosts, anomalies, dashboards, and check-ins. One interface.

    AppSignal works out of the box for Ruby, Elixir, Node.js, Python, and more. 30-day free trial, no credit card required.
    Start Free
  • 10
    Merlion

    Merlion

    A Machine Learning Framework for Time Series Intelligence

    Merlion is a Python library for time series intelligence. It provides an end-to-end machine learning framework that includes loading and transforming data, building and training models, post-processing model outputs, and evaluating model performance. It supports various time series learning tasks, including forecasting, anomaly detection, and change point detection for both univariate and multivariate time series. This library aims to provide engineers and researchers a one-stop solution to rapidly develop models for their specific time series needs, and benchmark them across multiple time series datasets.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    UnionML

    UnionML

    Build and deploy machine learning microservices

    ...Fit the rich ecosystem of tools and frameworks into a common protocol for machine learning. Using industry-standard machine learning methods, implement endpoints for fetching data, training models, serving predictions (and much more) to write a complete ML stack in one place. Data science, ML engineering, and MLOps practitioners can all gather around UnionML apps as a way of defining a single source of truth about your ML system’s behavior. This helps you maintain consistent code across your ML stack, from training to prediction logic.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    LSTMs for Human Activity Recognition

    LSTMs for Human Activity Recognition

    Human Activity Recognition example using TensorFlow on smartphone

    ...Through the use of sequential neural network architectures, the system learns patterns in motion data that correspond to activities such as walking, sitting, standing, or climbing stairs. The repository includes data preprocessing scripts, neural network architecture definitions, and training pipelines that allow researchers to reproduce and modify the experiments. It serves as an educational example of how deep learning models can process temporal sensor signals for pattern recognition tasks.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    DeepCTR

    DeepCTR

    Package of deep-learning based CTR models

    ...You can use any complex model with model.fit(), and model.predict(). Provide tf.keras.Model like interface for quick experiment. Provide tensorflow estimator interface for large scale data and distributed training. It is compatible with both tf 1.x and tf 2.x. With the great success of deep learning,DNN-based techniques have been widely used in CTR prediction task. The data in CTR estimation task usually includes high sparse,high cardinality categorical features and some dense numerical features. Since DNN are good at handling dense numerical features,we usually map the sparse categorical features to dense numerical through embedding technique.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    DialoGPT

    DialoGPT

    Large-scale pretraining for dialogue

    ...The model was trained on a massive dataset of approximately 147 million conversational exchanges extracted from Reddit discussion threads, allowing it to learn patterns of natural human conversation. DialoGPT provides multiple pretrained model sizes and includes code for training, fine-tuning, and evaluating dialogue generation models. The repository also contains scripts for preparing conversation datasets and reproducing experimental benchmarks related to conversational AI research.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    Elephas

    Elephas

    Distributed Deep learning with Keras & Spark

    Elephas is an extension of Keras, which allows you to run distributed deep learning models at scale with Spark. Elephas currently supports a number of applications. Elephas brings deep learning with Keras to Spark. Elephas intends to keep the simplicity and high usability of Keras, thereby allowing for fast prototyping of distributed models, which can be run on massive data sets. Elephas implements a class of data-parallel algorithms on top of Keras, using Spark's RDDs and data frames. Keras...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    KAIR

    KAIR

    Image Restoration Toolbox (PyTorch). Training and testing codes

    Image restoration toolbox (PyTorch). Training and testing codes for DPIR, USRNet, DnCNN, FFDNet, SRMD, DPSR, BSR/GAN, SwinIR.
    Downloads: 12 This Week
    Last Update:
    See Project
  • 17
    Auto-PyTorch

    Auto-PyTorch

    Automatic architecture search and hyperparameter optimization

    While early AutoML frameworks focused on optimizing traditional ML pipelines and their hyperparameters, another trend in AutoML is to focus on neural architecture search. To bring the best of these two worlds together, we developed Auto-PyTorch, which jointly and robustly optimizes the network architecture and the training hyperparameters to enable fully automated deep learning (AutoDL). Auto-PyTorch is mainly developed to support tabular data (classification, regression) and time series data (forecasting). The newest features in Auto-PyTorch for tabular data are described in the paper "Auto-PyTorch Tabular: Multi-Fidelity MetaLearning for Efficient and Robust AutoDL" (see below for bibtex ref). ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    StudioGAN

    StudioGAN

    StudioGAN is a Pytorch library providing implementations of networks

    StudioGAN is a Pytorch library providing implementations of representative Generative Adversarial Networks (GANs) for conditional/unconditional image generation. StudioGAN aims to offer an identical playground for modern GANs so that machine learning researchers can readily compare and analyze a new idea. Moreover, StudioGAN provides an unprecedented-scale benchmark for generative models. The benchmark includes results from GANs (BigGAN-Deep, StyleGAN-XL), auto-regressive models (MaskGIT,...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    Alphafold2

    Alphafold2

    Unofficial Pytorch implementation / replication of Alphafold2

    ...Deepmind has open sourced the official code in Jax, along with the weights! This repository will now be geared towards a straight pytorch translation with some improvements on positional encoding. lhatsk has reported training a modified trunk of this repository, using the same setup as trRosetta, with competitive results. The underlying assumption is that the trunk works on the residue level, and then constitutes to atomic level for the structure module, whether it be SE3 Transformers, E(n)-Transformer, or EGNN doing the refinement.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    MuZero General

    MuZero General

    A commented and documented implementation of MuZero

    ...The framework is modular so that users can easily add new environments by defining the game logic and associated hyperparameters. It also includes support for distributed training, GPU acceleration, and monitoring tools for tracking learning progress.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    YOLOX

    YOLOX

    YOLOX is a high-performance anchor-free YOLO, exceeding yolov3~v5

    ...Except special cases, we always recommend using our COCO pre-trained weights for initializing the model. As YOLOX is an anchor-free detector with only several hyper-parameters, most of the time good results can be obtained with no changes to the models or training settings.
    Downloads: 16 This Week
    Last Update:
    See Project
  • 22
    Guild AI

    Guild AI

    Experiment tracking, ML developer tools

    Guild AI is an open-source experiment tracking toolkit designed to bring systematic control to machine learning workflows, enabling users to build better models faster. It automatically captures every detail of training runs as unique experiments, facilitating comprehensive tracking and analysis. Users can compare and analyze runs to deepen their understanding and incrementally improve models. Guild AI simplifies hyperparameter tuning by applying state-of-the-art algorithms through straightforward commands, eliminating the need for complex trial setups. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    Catalyst

    Catalyst

    Accelerated deep learning R&D

    Catalyst is a PyTorch framework for accelerated Deep Learning research and development. It allows you to write compact but full-featured Deep Learning pipelines with just a few lines of code. With Catalyst you get a full set of features including a training loop with metrics, model checkpointing and more, all without the boilerplate. Catalyst is focused on reproducibility, rapid experimentation, and codebase reuse so you can break the cycle of writing another regular train loop and make something totally new. Catalyst is compatible with Python 3.6+. PyTorch 1.1+, and has been tested on Ubuntu 16.04/18.04/20.04, macOS 10.15, Windows 10 and Windows Subsystem for Linux. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    EasyNLP

    EasyNLP

    EasyNLP: A Comprehensive and Easy-to-use NLP Toolkit

    ...It has powered more than 10 BUs and more than 20 business scenarios within the Alibaba group. It is seamlessly integrated to Platform of AI (PAI) products, including PAI-DSW for development, PAI-DLC for cloud-native training, PAI-EAS for serving, and PAI-Designer for zero-code model training.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    TensorFlowOnSpark

    TensorFlowOnSpark

    TensorFlowOnSpark brings TensorFlow programs to Apache Spark clusters

    By combining salient features from the TensorFlow deep learning framework with Apache Spark and Apache Hadoop, TensorFlowOnSpark enables distributed deep learning on a cluster of GPU and CPU servers. It enables both distributed TensorFlow training and inferencing on Spark clusters, with a goal to minimize the amount of code changes required to run existing TensorFlow programs on a shared grid.
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB