Showing 890 open source projects for "training"

View related business solutions
  • Earn up to 16% annual interest with Nexo. Icon
    Earn up to 16% annual interest with Nexo.

    More flexibility. More control.

    Generate interest, access liquidity without selling, and execute trades seamlessly. All in one platform. Geographic restrictions, eligibility, and terms apply.
    Get started with Nexo.
  • Enterprise-grade ITSM, for every business Icon
    Enterprise-grade ITSM, for every business

    Give your IT, operations, and business teams the ability to deliver exceptional services—without the complexity.

    Freshservice is an intuitive, AI-powered platform that helps IT, operations, and business teams deliver exceptional service without the usual complexity. Automate repetitive tasks, resolve issues faster, and provide seamless support across the organization. From managing incidents and assets to driving smarter decisions, Freshservice makes it easy to stay efficient and scale with confidence.
    Try it Free
  • 1
    VITS

    VITS

    Conditional Variational Autoencoder with Adversarial Learning

    ...Unlike traditional two-stage systems that separately train an acoustic model and a vocoder, VITS trains an end-to-end model that maps text directly to waveform using a conditional variational autoencoder combined with normalizing flows and adversarial training. This architecture enables parallel generation (fast inference) while achieving speech quality that rivals or surpasses many two-stage systems. The repository provides training and inference pipelines for common datasets such as LJ Speech (single-speaker) and VCTK (multi-speaker), including filelists, configs, and preprocessing scripts. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    Chatette

    Chatette

    A powerful dataset generator for Rasa NLU, inspired by Chatito

    Chatette is a Python-based tool for generating training datasets for Natural Language Understanding (NLU) models, particularly those used with Rasa NLU. It employs a domain-specific language to define templates, enabling the creation of diverse and extensive training examples for intent classification and entity recognition.​
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    ML workspace

    ML workspace

    All-in-one web-based IDE specialized for machine learning

    All-in-one web-based development environment for machine learning. The ML workspace is an all-in-one web-based IDE specialized for machine learning and data science. It is simple to deploy and gets you started within minutes to productively built ML solutions on your own machines. This workspace is the ultimate tool for developers preloaded with a variety of popular data science libraries (e.g., Tensorflow, PyTorch, Keras, Sklearn) and dev tools (e.g., Jupyter, VS Code, Tensorboard)...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 4
    Transformer TTS

    Transformer TTS

    Implementation of a Transformer based neural network

    ...This design addresses common autoregressive issues such as repetition, skipped words, and unstable attention, and results in robust, fast synthesis where all frames are predicted in parallel. The repository ships with tooling to build datasets (especially LJSpeech) and create training data, plus scripts to train both the aligner and the TTS model, monitor training with TensorBoard, and resume or reset training runs.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Our Free Plans just got better! | Auth0 Icon
    Our Free Plans just got better! | Auth0

    With up to 25k MAUs and unlimited Okta connections, our Free Plan lets you focus on what you do best—building great apps.

    You asked, we delivered! Auth0 is excited to expand our Free and Paid plans to include more options so you can focus on building, deploying, and scaling applications without having to worry about your security. Auth0 now, thank yourself later.
    Try free now
  • 5
    U-Net Fusion RFI

    U-Net Fusion RFI

    U-Net for RFI Detection based on @jakeret's implementation

    See original code here: https://github.com/jakeret/tf_unet Currently this project is based on Tensorflow 1.13 code base and there are no plans to transfer to TF version 2. The primary improvements to this code base include a training and evaluation framework, along with a fusion based approach to detection, combining a number of models (currently hard coded to two trained models) along with Sum Threshold as an additional "expert." Additional work is being done to add custom layers to this model for further experimentation, including Squeeze/Excitation layers (unimplemented.) ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    YOLOv4-large

    YOLOv4-large

    Scaled-YOLOv4: Scaling Cross Stage Partial Network

    YOLOv4-large is an open-source implementation of the Scaled-YOLOv4 object detection architecture, designed to improve both the accuracy and scalability of real-time computer vision models. The project provides a PyTorch implementation of the Scaled-YOLOv4 framework, which extends the original YOLOv4 architecture using Cross Stage Partial (CSP) networks and new scaling techniques. Unlike earlier object detection systems that only scale depth or width, this architecture scales multiple aspects...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    SageMaker TensorFlow Serving Container

    SageMaker TensorFlow Serving Container

    A TensorFlow Serving solution for use in SageMaker

    ...Some of the build and tests scripts interact with resources in your AWS account. Be sure to set your default AWS credentials and region using aws configure before using these scripts. Amazon SageMaker uses Docker containers to run all training jobs and inference endpoints. The Docker images are built from the Dockerfiles in docker/. The Dockerfiles are grouped based on the version of TensorFlow Serving they support. Each supported processor type (e.g. "cpu", "gpu", "ei") has a different Dockerfile in each group. If your are testing locally, building the image is enough. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    PyCls

    PyCls

    Codebase for Image Classification Research, written in PyTorch

    pycls is a focused PyTorch codebase for image classification research that emphasizes reproducibility and strong, transparent baselines. It popularized families like RegNet and supports classic architectures (ResNet, ResNeXt) with clean implementations and consistent training recipes. The repository includes highly tuned schedules, augmentations, and regularization settings that make it straightforward to match reported accuracy without guesswork. Distributed training and mixed precision are first-class, enabling fast experiments on multi-GPU setups with simple, declarative configs. Model definitions are concise and modular, making it easy to prototype new blocks or swap backbones while keeping the rest of the pipeline unchanged. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    Nerfies

    Nerfies

    This is the code for Deformable Neural Radiance Fields

    ...Instead of assuming a static world, the method learns a canonical space plus a deformation field that maps changing poses or expressions back to that space during training. This lets the system generate photorealistic novel views of nonrigid subjects—faces, bodies, cloth—while preserving fine detail and consistent lighting. The training pipeline handles imperfect captures by modeling camera poses, exposure variations, and background segmentation, producing stable geometry and appearance. A set of utilities manages dataset preparation, pose estimation, and checkpoints so researchers can reproduce results on their own footage. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Go From AI Idea to AI App Fast Icon
    Go From AI Idea to AI App Fast

    One platform to build, fine-tune, and deploy ML models. No MLOps team required.

    Access Gemini 3 and 200+ models. Build chatbots, agents, or custom models with built-in monitoring and scaling.
    Try Free
  • 10
    tensorflow_template_application

    tensorflow_template_application

    TensorFlow template application for deep learning

    tensorflow_template_application is a template project that demonstrates how to structure scalable applications built with TensorFlow. The repository provides a standardized architecture that helps developers organize machine learning code into clear components such as data processing, model training, evaluation, and deployment. Instead of focusing on a specific algorithm, the project emphasizes software engineering practices that make machine learning systems easier to maintain and extend. The template includes configuration files, scripts, and project structures that help teams build reproducible experiments and production-ready pipelines. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    SimCSE

    SimCSE

    SimCSE: Simple Contrastive Learning of Sentence Embeddings

    SimCSE (Simple Contrastive Learning of Sentence Embeddings) is a machine learning framework for training sentence embeddings using contrastive learning. It improves representation learning for NLP tasks.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 12
    SRU

    SRU

    Training RNNs as Fast as CNNs

    ...In this work, we propose the Simple Recurrent Unit (SRU), a light recurrent unit that balances model capacity and scalability. SRU is designed to provide expressive recurrence, enable highly parallelized implementation, and comes with careful initialization to facilitate the training of deep models. We demonstrate the effectiveness of SRU on multiple NLP tasks. SRU achieves 5--9x speed-up over cuDNN-optimized LSTM on classification and question answering datasets, and delivers stronger results than LSTM and convolutional models. We also obtain an average of 0.7 BLEU improvement over the Transformer model on the translation by incorporating SRU into the architecture. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    AliceMind

    AliceMind

    ALIbaba's Collection of Encoder-decoders from MinD

    This repository provides pre-trained encoder-decoder models and its related optimization techniques developed by Alibaba's MinD (Machine IntelligeNce of Damo) Lab. Pre-trained models for natural language understanding (NLU). We extend BERT to a new model, StructBERT, by incorporating language structures into pre-training. Specifically, we pre-train StructBERT with two auxiliary tasks to make the most of the sequential order of words and sentences, which leverage language structures at the word and sentence levels, respectively. Pre-trained models for natural language generation (NLG). We propose a novel scheme that jointly pre-trains an autoencoding and autoregressive language model on a large unlabeled corpus, specifically designed for generating new text conditioned on context. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 14
    TimeSformer

    TimeSformer

    The official pytorch implementation of our paper

    ...Because the attention is global across frames, TimeSformer can reason about dependencies across long time spans, not just local neighborhoods. The official implementation in PyTorch provides configurations, pretrained models, and training scripts that make it straightforward to evaluate or fine-tune on video datasets. TimeSformer was influential in showing that pure transformer architectures—without convolutional backbones—can perform strongly on video classification tasks. Its flexible attention design allows experimenting with different factoring (spatial-then-temporal, joint, etc.) to trade off compute, memory, and accuracy.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    Minkowski Engine

    Minkowski Engine

    Auto-diff neural network library for high-dimensional sparse tensors

    The Minkowski Engine is an auto-differentiation library for sparse tensors. It supports all standard neural network layers such as convolution, pooling, unspooling, and broadcasting operations for sparse tensors. The Minkowski Engine supports various functions that can be built on a sparse tensor. We list a few popular network architectures and applications here. To run the examples, please install the package and run the command in the package root directory. Compressing a neural network to...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    DrQA

    DrQA

    Reading Wikipedia to Answer Open-Domain Questions

    DrQA is an open-domain question answering system that reads large text corpora—famously Wikipedia—to answer natural language questions with extractive spans. It follows a two-stage pipeline: a fast document retriever first narrows down candidate articles, and a neural machine reader then predicts the exact answer span from those passages. The retriever relies on classic IR features (like TF-IDF and n-gram statistics) to remain lightweight and scalable to millions of documents. The reader is...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    UniVL

    UniVL

    Official implementation for UniVL video and language training models

    UniVL is a video-language pretrain model. It is designed with four modules and five objectives for both video language understanding and generation tasks. It is also a flexible model for most of the multimodal downstream tasks considering both efficiency and effectiveness.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    Denoiser

    Denoiser

    Real Time Speech Enhancement in the Waveform Domain (Interspeech 2020)

    ...The project supports both offline denoising (batch inference) and live audio processing (e.g. via loopback audio interfaces), making it practical for real-time use in calls or recording. The codebase includes training and evaluation scripts, configuration management via Hydra, and pretrained models on standard noise datasets.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 19
    XLM (Cross-lingual Language Model)

    XLM (Cross-lingual Language Model)

    PyTorch original implementation of Cross-lingual Language Model

    ...Using a shared subword vocabulary, XLM learns language-agnostic features that work well for classification and sequence labeling tasks such as XNLI, NER, and POS without target-language supervision. The repository provides preprocessing pipelines, training code, and fine-tuning scripts so you can reproduce benchmark results or adapt models to your own multilingual corpora. Pretrained checkpoints cover dozens of languages and multiple model sizes, balancing quality and compute needs.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    DensePose

    DensePose

    A real-time approach for mapping all human pixels of 2D RGB images

    ...This enables detailed understanding of human shape, motion, and surface appearance directly from images or videos. The repository includes the DensePose network architecture, training code, pretrained models, and dataset tools for annotation and visualization. DensePose is widely used in augmented reality, motion capture, virtual try-on, and visual effects applications because it enables real-time 3D human mapping from 2D inputs. The model architecture builds on Mask R-CNN, using additional regression heads to predict UV coordinates that map image pixels to 3D surfaces.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 21
    Semantic Segmentation in PyTorch

    Semantic Segmentation in PyTorch

    Semantic segmentation models, datasets & losses implemented in PyTorch

    Semantic segmentation models, datasets and losses implemented in PyTorch. PyTorch and Torchvision needs to be installed before running the scripts, together with PIL and opencv for data-preprocessing and tqdm for showing the training progress. PyTorch v1.1 is supported (using the new supported tensoboard); can work with earlier versions, but instead of using tensoboard, use tensoboardX. Poly learning rate, where the learning rate is scaled down linearly from the starting value down to zero during training. Considered as the go-to scheduler for semantic segmentation. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    CapsGNN

    CapsGNN

    A PyTorch implementation of "Capsule Graph Neural Network"

    A PyTorch implementation of "Capsule Graph Neural Network" (ICLR 2019). The high-quality node embeddings learned from the Graph Neural Networks (GNNs) have been applied to a wide range of node-based applications and some of them have achieved state-of-the-art (SOTA) performance. However, when applying node embeddings learned from GNNs to generate graph embeddings, the scalar node representation may not suffice to preserve the node/graph properties efficiently, resulting in sub-optimal graph...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    MLOps Course

    MLOps Course

    Learn how to design, develop, deploy and iterate on ML apps

    ...The repository itself contains configuration, code examples, and links to accompanying lessons hosted on the Made With ML site, which provide detailed narrative explanations and diagrams. Instead of focusing only on model training, the course emphasizes best practices like modular code design, CI/CD, containerization, reproducibility, and responsible ML (including monitoring and feedback loops). This makes it particularly valuable for engineers transitioning from “notebooks and prototypes” to real systems that must be robust, maintainable, and observable in production.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 24
    CRSLab

    CRSLab

    CRSLab is an open-source toolkit

    ...CRSLab has the following highlights. Comprehensive benchmark models and datasets: We have integrated commonly-used 6 datasets and 18 models, including graph neural network and pre-training models such as R-GCN, BERT and GPT-2. We have preprocessed these datasets to support these models, and release for downloading. Extensive and standard evaluation protocols: We support a series of widely-adopted evaluation protocols for testing and comparing different CRS. General and extensible structure: We design a general and extensible structure to unify various conversational recommendation datasets and models, in which we integrate various built-in interfaces and functions for quickly development. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    Pipeline for training Language Models

    Pipeline for training Language Models

    Pipeline for training Language Models using PyTorch.

    Pipeline for training Language Models using PyTorch. Inspired by Yandex Data School NLP Course (week 03: Language Modeling) Prepared text file with space-separated words on each line.
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB