19 projects for "training" with 2 filters applied:

  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • Earn up to 16% annual interest with Nexo. Icon
    Earn up to 16% annual interest with Nexo.

    Let your crypto work for you

    Put idle assets to work with competitive interest rates, borrow without selling, and trade with precision. All in one platform. Geographic restrictions, eligibility, and terms apply.
    Get started with Nexo.
  • 1
    Megatron

    Megatron

    Ongoing research training transformer models at scale

    Megatron is a large, powerful transformer developed by the Applied Deep Learning Research team at NVIDIA. This repository is for ongoing research on training large transformer language models at scale. We developed efficient, model-parallel (tensor, sequence, and pipeline), and multi-node pre-training of transformer based models such as GPT, BERT, and T5 using mixed precision. Megatron is also used in NeMo Megatron, a framework to help enterprises overcome the challenges of building and training sophisticated natural language processing models with billions and trillions of parameters. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 2
    DeepSpeed

    DeepSpeed

    Deep learning optimization library: makes distributed training easy

    ...Achieve extreme compression for an unparalleled inference latency and model size reduction with low costs DeepSpeed offers a confluence of system innovations, that has made large scale DL training effective, and efficient, greatly improved ease of use, and redefined the DL training landscape in terms of scale that is possible. These innovations such as ZeRO, 3D-Parallelism, DeepSpeed-MoE, ZeRO-Infinity, etc. fall under the training pillar.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 3
    AudioCraft

    AudioCraft

    Audiocraft is a library for audio processing and generation

    AudioCraft is a PyTorch library for text-to-audio and text-to-music generation, packaging research models and tooling for training and inference. It includes MusicGen for music generation conditioned on text (and optionally melody) and AudioGen for text-conditioned sound effects and environmental audio. Both models operate over discrete audio tokens produced by a neural codec (EnCodec), which acts like a tokenizer for waveforms and enables efficient sequence modeling.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 4
    DLRM

    DLRM

    An implementation of a deep learning recommendation model (DLRM)

    ...The implementation is optimized for performance at scale, supporting multi-GPU and multi-node execution, quantization, embedding partitioning, and pipelined I/O to feed huge embeddings efficiently. It includes data loaders for standard benchmarks (like Criteo), training scripts, evaluation tools, and capabilities like mixed precision, gradient compression, and memory fusion to maximize throughput.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Our Free Plans just got better! | Auth0 Icon
    Our Free Plans just got better! | Auth0

    With up to 25k MAUs and unlimited Okta connections, our Free Plan lets you focus on what you do best—building great apps.

    You asked, we delivered! Auth0 is excited to expand our Free and Paid plans to include more options so you can focus on building, deploying, and scaling applications without having to worry about your security. Auth0 now, thank yourself later.
    Try free now
  • 5
    vJEPA-2

    vJEPA-2

    PyTorch code and models for VJEPA2 self-supervised learning from video

    ...Trained representations transfer well to downstream tasks such as action recognition, temporal localization, and video retrieval, often with simple linear probes or light fine-tuning. The repository typically includes end-to-end recipes—data pipelines, augmentation policies, training scripts, and evaluation harnesses.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    Deep Learning Is Nothing

    Deep Learning Is Nothing

    Deep learning concepts in an approachable style

    ...The goal is to replace buzzwords with intuition so learners can reason about architectures and training dynamics with confidence.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    Make-A-Video - Pytorch (wip)

    Make-A-Video - Pytorch (wip)

    Implementation of Make-A-Video, new SOTA text to video generator

    ...Passing in images (if one were to pretrain on images first), both temporal convolution and attention will be automatically skipped. In other words, you can use this straightforwardly in your 2d Unet and then port it over to a 3d Unet once that phase of the training is done.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 8
    JEPA

    JEPA

    PyTorch code and models for V-JEPA self-supervised learning from video

    ...This makes learning focus on semantics and structure, yielding features that transfer well with simple linear probes and minimal fine-tuning. The repository provides training recipes, data pipelines, and evaluation utilities for image JEPA variants and often includes ablations that illuminate which masking and architectural choices matter. Because the objective is non-autoregressive and operates in embedding space, JEPA tends to be compute-efficient and stable at scale. The approach has become a strong alternative to contrastive or pixel-reconstruction methods for representation learning.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    Deep-Learning-Interview-Book

    Deep-Learning-Interview-Book

    Interview guide for machine learning, mathematics, and deep learning

    ...It spans the core math (linear algebra, probability, optimization) and the practitioner topics candidates actually face, like CNNs, RNNs/Transformers, attention, regularization, and training tricks. Explanations emphasize intuition first, then key formulas and common pitfalls, so you can reason through unseen questions rather than memorize trivia. Many entries connect theory to implementation details, including how choices in activation, initialization, or normalization affect convergence and stability. The content is organized for fast review before an interview loop but is also deep enough for systematic study over weeks. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • $300 in Free Credit Towards Top Cloud Services Icon
    $300 in Free Credit Towards Top Cloud Services

    Build VMs, containers, AI, databases, storage—all in one place.

    Start your project in minutes. After credits run out, 20+ products include free monthly usage. Only pay when you're ready to scale.
    Get Started
  • 10
    The Hypersim Dataset

    The Hypersim Dataset

    Photorealistic Synthetic Dataset for Holistic Indoor Scene

    ...It provides richly annotated renderings—RGB, depth, surface normals, instance and semantic segmentations, and material/lighting metadata—produced from high-fidelity virtual environments. The dataset spans diverse furniture layouts, room types, and camera trajectories, enabling robust training for geometry, segmentation, and SLAM-adjacent tasks. Rendering pipelines and utilities allow researchers to reproduce sequences, generate novel views, or extract task-specific supervision. Because the data are perfectly labeled and controllable, Hypersim is well suited for pretraining and for studying domain transfer to real imagery. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    Deep Learning Models

    Deep Learning Models

    A collection of various deep learning architectures, models, and tips

    This repository collects clear, well-documented implementations of deep learning models and training utilities written by Sebastian Raschka. The code favors readability and pedagogy: components are organized so you can trace data flow through layers, losses, optimizers, and evaluation. Examples span fundamental architectures—MLPs, CNNs, RNN/Transformers—and practical tasks like image classification or text modeling. Reproducible training scripts and configuration files make it straightforward to rerun experiments or adapt them to your own datasets. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    Exposure Correction

    Exposure Correction

    Learning multi-scale deep model correcting over- and under- exposed

    ...The method employs a multi-scale framework that learns to enhance images by adjusting exposure levels across different spatial resolutions. This allows the model to preserve fine details while correcting global lighting inconsistencies. The repository includes pre-trained models, datasets, and training/testing code to enable reproducibility and experimentation. By leveraging this framework, researchers and developers can apply exposure correction to a wide range of natural images, improving visual quality without manual editing. The project serves both as a research reference and a practical tool for computational photography and image enhancement.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 13
    GPT-NeoX

    GPT-NeoX

    Implementation of model parallel autoregressive transformers on GPUs

    This repository records EleutherAI's library for training large-scale language models on GPUs. Our current framework is based on NVIDIA's Megatron Language Model and has been augmented with techniques from DeepSpeed as well as some novel optimizations. We aim to make this repo a centralized and accessible place to gather techniques for training large-scale autoregressive language models, and accelerate research into large-scale training.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 14
    Deeplearning.ai

    Deeplearning.ai

    Study notes, summaries, and auxiliary materials for deep learning

    Deeplearning.ai collects study notes, summaries, and auxiliary materials aligned with the popular deep learning course series many learners take early in their AI journey. It distills core ideas such as optimization, regularization, convolutional networks, sequence models, and practical training tricks. The explanations aim to bridge theory and practice, often connecting mathematical intuition to code-level implications. By organizing the content as “books” or structured notes, it gives students a consistent reference to revisit as models and tooling evolve. Many learners use it to supplement course videos, reinforcing concepts before implementing assignments or projects. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 15
    deep-q-learning

    deep-q-learning

    Minimal Deep Q Learning (DQN & DDQN) implementations in Keras

    The deep-q-learning repository authored by keon provides a Python-based implementation of the Deep Q-Learning algorithm — a cornerstone method in reinforcement learning. It implements the core logic needed to train an agent using Q-learning with neural networks (i.e. approximating Q-values via deep nets), setting up environment interaction loops, experience replay, network updates, and policy behavior. For learners and researchers interested in reinforcement learning, this repo offers a...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    TensorFlow Course

    TensorFlow Course

    Simple and ready-to-use tutorials for TensorFlow

    This repository houses a highly popular (~16k stars) set of TensorFlow tutorials and example code aimed at beginners and intermediate users. It includes Jupyter notebooks and scripts that cover neural network fundamentals, model training, deployment, and more, with support for Google Colab.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    Learn_Deep_Learning_in_6_Weeks

    Learn_Deep_Learning_in_6_Weeks

    This is the Curriculum for "Learn Deep Learning in 6 Weeks"

    ...It begins with neural network fundamentals and moves through convolutional and recurrent architectures, optimization strategies, regularization, and transfer learning. The materials emphasize code-first understanding: building small models, training them on accessible datasets, and analyzing their behavior. Each week culminates in a tangible outcome—such as a working classifier or sequence model—so progress is visible and motivating. The plan also introduces practical considerations like GPU usage, checkpoints, and debugging training dynamics. It aims to give you enough breadth to recognize common patterns and enough depth to implement them on your own problems.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    LearningToCompare_FSL

    LearningToCompare_FSL

    Learning to Compare: Relation Network for Few-Shot Learning

    ...The core idea implemented here is the relation network, which learns to compare pairs of feature embeddings and output relation scores that indicate whether two images belong to the same class, enabling classification from only a handful of labeled examples. The repository provides training and evaluation code for standard few-shot benchmarks such as miniImageNet and Omniglot, making it possible to reproduce the experimental results reported in the paper. It includes model definitions, data loading logic, episodic training loops, and scripts that implement the N-way K-shot evaluation protocol common in few-shot research. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    Deep Reinforcement Learning TensorFlow

    Deep Reinforcement Learning TensorFlow

    TensorFlow implementation of Deep Reinforcement Learning papers

    ...It includes implementations of well-known algorithms such as Deep Q-Networks (DQN), policy gradients, and related variants, demonstrating how neural networks can be trained through interaction with simulated environments. The project is commonly used by learners who want to move beyond theory and understand the practical mechanics of training RL agents. Visualization utilities and training scripts help users monitor learning progress and debug experiments.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB