Showing 28 open source projects for "training"

View related business solutions
  • Try Google Cloud Risk-Free With $300 in Credit Icon
    Try Google Cloud Risk-Free With $300 in Credit

    No hidden charges. No surprise bills. Cancel anytime.

    Use your credit across every product. Compute, storage, AI, analytics. When it runs out, 20+ products stay free. You only pay when you choose to.
    Start Free
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • 1
    RefineNet

    RefineNet

    RefineNet: Multi-Path Refinement Networks

    ...It provides trained models for datasets such as PASCAL VOC 2012, Cityscapes, NYUDv2, Person_Parts, PASCAL_Context, SUNRGBD, and ADE20k, with versions based on ResNet-101 and ResNet-152 backbones. The repository supports both single-scale and multi-scale prediction, with scripts for training, testing, and evaluating segmentation performance. While this codebase is specific to MATLAB and MatConvNet, a PyTorch implementation and lighter-weight variants are also available from the community.
    Downloads: 8 This Week
    Last Update:
    See Project
  • 2
    Gin Config

    Gin Config

    Gin provides a lightweight configuration framework for Python

    ...Gin is particularly popular in TensorFlow and PyTorch projects, where researchers and developers need to tune numerous interdependent parameters across models, datasets, optimizers, and training pipelines.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 3
    seq2seq

    seq2seq

    A general-purpose encoder-decoder framework for Tensorflow

    seq2seq is an early, influential TensorFlow reference implementation for sequence-to-sequence learning with attention, covering tasks like neural machine translation, summarization, and dialogue. It packaged encoders, decoders, attention mechanisms, and beam search into a modular training and inference framework. The codebase showcased best practices for batching, bucketing by sequence length, and handling variable-length sequences efficiently on GPUs. Researchers used it as a baseline to reproduce classic results and to prototype new attention variants and training tricks. It also offered scripts for data preprocessing, evaluation, and exporting models for serving. ...
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB