Showing 113 open source projects for "training"

View related business solutions
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 1
    Opacus

    Opacus

    Training PyTorch models with differential privacy

    Opacus is a library that enables training PyTorch models with differential privacy. It supports training with minimal code changes required on the client, has little impact on training performance, and allows the client to online track the privacy budget expended at any given moment. Vectorized per-sample gradient computation that is 10x faster than micro batching. Supports most types of PyTorch models and can be used with minimal modification to the original neural network. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 2
    LLM Datasets

    LLM Datasets

    Curated list of datasets and tools for post-training

    ...The repository aims to make datasets easy to inspect and transform, with scripts for downloading, deduping, cleaning, and converting to formats like JSONL that slot into training pipelines. It highlights instruction-tuning and conversation-style corpora while also pointing to code, math, or domain-specific sets for targeted capabilities. Quality is a recurring theme: examples and utilities help filter low-value samples, enforce length limits, and split train/validation consistently so results are comparable. ...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 3
    CoreNet

    CoreNet

    CoreNet: A library for training deep neural networks

    CoreNet is Apple’s internal deep learning framework for distributed neural network training, designed for high scalability, low-latency communication, and strong hardware efficiency. It focuses on enabling large-scale model training across clusters of GPUs and accelerators by optimizing data flow and parallelism strategies. CoreNet provides abstractions for data, tensor, and pipeline parallelism, allowing models to scale without code duplication or heavy manual configuration. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    MuJoCo Playground

    MuJoCo Playground

    An open source library for GPU-accelerated robot learning

    ...The project includes classic control benchmarks from dm_control, advanced quadruped and bipedal locomotion systems, and dexterous as well as non-prehensile manipulation setups. It also offers optional vision-based training capabilities through integration with Madrona-MJX, allowing researchers to train policies directly from image input on GPUs. MuJoCo Playground supports both the MJX JAX implementation and the Warp physics engine, enabling flexible use across research pipelines. The environments are designed for fast training, compatibility with reinforcement learning libraries, and real-time trajectory visualization using rscope.
    Downloads: 8 This Week
    Last Update:
    See Project
  • Try Google Cloud Risk-Free With $300 in Credit Icon
    Try Google Cloud Risk-Free With $300 in Credit

    No hidden charges. No surprise bills. Cancel anytime.

    Use your credit across every product. Compute, storage, AI, analytics. When it runs out, 20+ products stay free. You only pay when you choose to.
    Start Free
  • 5
    AudioCraft

    AudioCraft

    Audiocraft is a library for audio processing and generation

    AudioCraft is a PyTorch library for text-to-audio and text-to-music generation, packaging research models and tooling for training and inference. It includes MusicGen for music generation conditioned on text (and optionally melody) and AudioGen for text-conditioned sound effects and environmental audio. Both models operate over discrete audio tokens produced by a neural codec (EnCodec), which acts like a tokenizer for waveforms and enables efficient sequence modeling.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 6
    Flax

    Flax

    Flax is a neural network library for JAX

    ...Modules define parameterized computations, but initialization and application remain side-effect free, which pairs naturally with JAX’s staging and compilation model. Flax emphasizes composability: optimizers, training loops, and checkpointing are provided as examples or utilities rather than monolithic frameworks, encouraging research-friendly customization. The library is widely used in vision, language, and reinforcement learning, often serving as a thin layer atop NumPy-like JAX primitives. Tutorials and examples show patterns for multi-host training, mixed precision, and advanced input pipelines that scale from laptops to TPUs.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 7
    Axon

    Axon

    Nx-powered Neural Networks

    ...Model Creation API – A high-level model creation API which manages model initialization and application. Optimization API – An API for creating and using first-order optimization techniques based on the Optax library. Training API – An API for quickly training models, inspired by PyTorch Ignite. Axon provides abstractions that enable easy integration while maintaining a level of separation between each component. You should be able to use any of the APIs without dependencies on others. By decoupling the APIs, Axon gives you full control over each aspect of creating and training a neural network. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    Petastorm

    Petastorm

    Petastorm library enables single machine or distributed training

    Petastorm library enables single machine or distributed training and evaluation of deep learning models from datasets in Apache Parquet format. It supports ML frameworks such as Tensorflow, Pytorch, and PySpark and can be used from pure Python code. Petastorm is an open-source data access library developed at Uber ATG. This library enables single machine or distributed training and evaluation of deep learning models directly from datasets in Apache Parquet format.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    Lightweight' GAN

    Lightweight' GAN

    Implementation of 'lightweight' GAN, proposed in ICLR 2021

    ...You can turn on automatic mixed precision with one flag --amp. You should expect it to be 33% faster and save up to 40% memory. Aim is an open-source experiment tracker that logs your training runs, and enables a beautiful UI to compare them.
    Downloads: 0 This Week
    Last Update:
    See Project
  • AI-generated apps that pass security review Icon
    AI-generated apps that pass security review

    Stop waiting on engineering. Build production-ready internal tools with AI—on your company data, in your cloud.

    Retool lets you generate dashboards, admin panels, and workflows directly on your data. Type something like “Build me a revenue dashboard on my Stripe data” and get a working app with security, permissions, and compliance built in from day one. Whether on our cloud or self-hosted, create the internal software your team needs without compromising enterprise standards or control.
    Try Retool free
  • 10
    Anomalib

    Anomalib

    An anomaly detection library comprising state-of-the-art algorithms

    Anomalib is an open-source deep learning library focused on anomaly detection and localization tasks, collecting state-of-the-art algorithms and tools under one modular framework. It provides implementations of leading anomaly detection methods drawn from current research, as well as a full set of utilities for training, evaluating, benchmarking, and deploying these models on both public and private datasets. Anomalib emphasizes flexibility and reproducibility: you can use its simple APIs to plug in custom models, track experiments, tune hyperparameters, and generate visualizations that highlight anomalous regions. Its design supports unsupervised or semi-supervised paradigms, making it especially powerful for scenarios where only “normal” data is readily available and defects must be detected without exhaustive labeling. ...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 11
    TensorRT

    TensorRT

    C++ library for high performance inference on NVIDIA GPUs

    NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference. It includes a deep learning inference optimizer and runtime that delivers low latency and high throughput for deep learning inference applications. TensorRT-based applications perform up to 40X faster than CPU-only platforms during inference. With TensorRT, you can optimize neural network models trained in all major frameworks, calibrate for lower precision with high accuracy, and deploy to hyperscale data centers,...
    Downloads: 10 This Week
    Last Update:
    See Project
  • 12
    Lightly

    Lightly

    A python library for self-supervised learning on images

    ...Our solution can be applied before any data annotation step and the learned representations can be used to visualize and analyze datasets. This allows selecting the best core set of samples for model training through advanced filtering. We provide PyTorch, PyTorch Lightning and PyTorch Lightning distributed examples for each of the models to kickstart your project. Lightly requires Python 3.6+ but we recommend using Python 3.7+. We recommend installing Lightly in a Linux or OSX environment. With lightly, you can use the latest self-supervised learning methods in a modular way using the full power of PyTorch. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    jstraining

    jstraining

    Training materials designed to teach JavaScript

    jstraining is a curated set of training materials designed to teach JavaScript with a practical, hands-on approach. It organizes slides, exercises, and example projects into a progressive curriculum that moves from language fundamentals to tooling and best practices. Learners write code early and often, reinforcing concepts like functions, objects, modules, and asynchronous control flow.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    SYSADMIN Test Questions

    SYSADMIN Test Questions

    A collection of Linux Sysadmin Test Questions and Answers

    ...Many questions highlight real-world troubleshooting patterns rather than purely theoretical trivia, reflecting what admins face daily. As an interview prep resource, it helps candidates assess breadth and identify weak spots before deeper study. For teams, it doubles as a training checklist to standardize baseline knowledge across junior and mid-level engineers.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    Tunix

    Tunix

    A JAX-native LLM Post-Training Library

    Tunix is a JAX-native library for post-training large language models, bringing supervised fine-tuning, reinforcement learning–based alignment, and knowledge distillation into one coherent toolkit. It embraces JAX’s strengths—functional programming, jit compilation, and effortless multi-device execution—so experiments scale from a single GPU to pods of TPUs with minimal code changes.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    DeepEP

    DeepEP

    DeepEP: an efficient expert-parallel communication library

    ...The library also supports low-precision operations (such as FP8) to reduce memory and bandwidth usage during communication. DeepEP is aimed at large-scale model inference or training systems where expert parallelism is used to scale model capacity without replicating entire networks.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    Avalanche

    Avalanche

    End-to-End Library for Continual Learning based on PyTorch

    ...Provides all the necessary utilities concerning model training. This includes simple and efficient ways of implementing new continual learning strategies as well as a set of pre-implemented CL baselines and state-of-the-art algorithms you will be able to use for comparison! Avalanche the first experiment of an End-to-end Library for reproducible continual learning research & development where you can find benchmarks, algorithms, etc.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    CO3D (Common Objects in 3D)

    CO3D (Common Objects in 3D)

    Tooling for the Common Objects In 3D dataset

    ...Each of the thousands of sequences in CO3Dv2 captures a common object (from categories like cars, chairs, or plants) from multiple real-world viewpoints. The dataset includes RGB images, depth maps, masks, and camera poses for each frame, along with pre-defined training, validation, and testing splits for both few-view and many-view reconstruction tasks.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    DGL

    DGL

    Python package built to ease deep learning on graph

    Build your models with PyTorch, TensorFlow or Apache MXNet. Fast and memory-efficient message passing primitives for training Graph Neural Networks. Scale to giant graphs via multi-GPU acceleration and distributed training infrastructure. DGL empowers a variety of domain-specific projects including DGL-KE for learning large-scale knowledge graph embeddings, DGL-LifeSci for bioinformatics and cheminformatics, and many others. We are keen to bringing graphs closer to deep learning researchers. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    XGBoost

    XGBoost

    Scalable and Flexible Gradient Boosting

    XGBoost is an optimized distributed gradient boosting library, designed to be scalable, flexible, portable and highly efficient. It supports regression, classification, ranking and user defined objectives, and runs on all major operating systems and cloud platforms. XGBoost works by implementing machine learning algorithms under the Gradient Boosting framework. It also offers parallel tree boosting (GBDT, GBRT or GBM) that can quickly and accurately solve many data science problems....
    Downloads: 6 This Week
    Last Update:
    See Project
  • 21
    MongoDB JVM Driver

    MongoDB JVM Driver

    The MongoDB drivers for Java, Kotlin, and Scala

    Welcome to the documentation site for the Java Driver, the MongoDB driver for synchronous Java applications. Download it using Maven or Gradle, or set up a runnable project by following our Quick Start guide. For tutorials on how to use the MongoDB JVM Drivers, please reference MongoDB University. Additional tutorials, videos, and code examples using both the Java Driver and the Kotlin Driver can also be found in the MongoDB Developer Center.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    TensorFlow.js

    TensorFlow.js

    TensorFlow.js is a library for machine learning in JavaScript

    TensorFlow.js is a library for machine learning in JavaScript. Develop ML models in JavaScript, and use ML directly in the browser or in Node.js. Use off-the-shelf JavaScript models or convert Python TensorFlow models to run in the browser or under Node.js. Retrain pre-existing ML models using your own data. Build and train models directly in JavaScript using flexible and intuitive APIs. Tensors are the core datastructure of TensorFlow.js They are a generalization of vectors and matrices to...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 23
    Karpathy-Inspired Claude Code Guidelines

    Karpathy-Inspired Claude Code Guidelines

    A single CLAUDE.md file to improve Claude Code behavior

    ...The project organizes a progressive path through exercises, notebooks, code examples, and practical mini-projects that echo Karpathy’s approach to “learning by doing,” where students build core concepts from first principles rather than consuming superficial abstractions. It covers topics like implementing backpropagation from scratch, understanding convolutional and recurrent networks, building simple training loops, and exploring real datasets with hands-on code. This collection makes abstract theoretical ideas concrete by walking learners through real code and tangible outcomes, helping demystify parts of machine learning that often feel opaque in purely textbook settings.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 24
    Introduction to Programming Principles

    Introduction to Programming Principles

    Categorized overview of Programming Principles & Patterns

    Introduction to Programming Principles is a living catalog of fundamental software development principles collected in one place so developers can reference and teach them easily. It documents well-known concepts like SOLID, DRY, KISS, YAGNI, and others, and explains when and why to apply them. The goal is to move principles from being fuzzy tribal knowledge to explicit, written, teachable guidelines. Because it’s on GitHub, it can evolve as the community suggests new principles,...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    Deep Learning Essay Reading

    Deep Learning Essay Reading

    Read classic and new deep learning papers paragraph by paragraph

    ...Instead of reading entire dense academic papers, contributors provide structured breakdowns and insights into the most influential research from the past decade, often including explanation highlights and key takeaways. The content spans foundational models, architectures, and training methodologies across computer vision, natural language processing, generative models, and other machine learning domains. These summaries help students, researchers, and engineers stay up to date with breakthroughs in the field without needing to sift through full academic documents. With thousands of stars and forks, this repository has become a widely referenced learning resource for anyone interested in understanding the technical ideas behind major advancements.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • 3
  • 4
  • 5
  • Next
MongoDB Logo MongoDB