Showing 650 open source projects for "train"

View related business solutions
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • Build Securely on Azure with Proven Frameworks Icon
    Build Securely on Azure with Proven Frameworks

    Lay a foundation for success with Tested Reference Architectures developed by Fortinet’s experts. Learn more in this white paper.

    Moving to the cloud brings new challenges. How can you manage a larger attack surface while ensuring great network performance? Turn to Fortinet’s Tested Reference Architectures, blueprints for designing and securing cloud environments built by cybersecurity experts. Learn more and explore use cases in this white paper.
    Download Now
  • 1
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    deepfakes_faceswap

    deepfakes_faceswap

    Deepfakes Software For All

    Faceswap is the leading free and open source multi-platform deepfakes software. When faceswapping was first developed and published, the technology was groundbreaking, it was a huge step in AI development. It was also completely ignored outside of academia because the code was confusing and fragmentary. It required a thorough understanding of complicated AI techniques and took a lot of effort to figure it out. Until one individual brought it together into a single, cohesive collection.
    Downloads: 18 This Week
    Last Update:
    See Project
  • 3
    Colossal-AI

    Colossal-AI

    Making large AI models cheaper, faster and more accessible

    ...Together with better performance come larger model sizes. This imposes challenges to the memory wall of the current accelerator hardware such as GPU. It is never ideal to train large models such as Vision Transformer, BERT, and GPT on a single GPU or a single machine. There is an urgent demand to train models in a distributed environment. However, distributed training, especially model parallelism, often requires domain expertise in computer systems and architecture. It remains a challenge for AI researchers to implement complex distributed training solutions for their models. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 4
    Tokenizers

    Tokenizers

    Fast State-of-the-Art Tokenizers optimized for Research and Production

    ...Tokenizers provides an implementation of today’s most used tokenizers, with a focus on performance and versatility. These tokenizers are also used in Transformers. Train new vocabularies and tokenize, using today’s most used tokenizers. Extremely fast (both training and tokenization), thanks to the Rust implementation. Takes less than 20 seconds to tokenize a GB of text on a server’s CPU. Easy to use, but also extremely versatile. Designed for both research and production. Full alignment tracking. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Forever Free Full-Stack Observability | Grafana Cloud Icon
    Forever Free Full-Stack Observability | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • 5
    DeepSeed

    DeepSeed

    Deep learning optimization library making distributed training easy

    ...DeepSpeed delivers extreme-scale model training for everyone, from data scientists training on massive supercomputers to those training on low-end clusters or even on a single GPU. Using current generation of GPU clusters with hundreds of devices, 3D parallelism of DeepSpeed can efficiently train deep learning models with trillions of parameters. With just a single GPU, ZeRO-Offload of DeepSpeed can train models with over 10B parameters, 10x bigger than the state of arts, democratizing multi-billion-parameter model training such that many deep learning scientists can explore bigger and better models. Sparse attention of DeepSpeed powers an order-of-magnitude longer input sequence and obtains up to 6x faster execution comparing with dense transformers.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    DeepSqueak

    DeepSqueak

    DeepSqueak Using Machine Vision to Accelerate Bioacoustics Research

    Using Machine Vision to Accelerate Bioacoustics Research.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 7
    LLaMA-Factory

    LLaMA-Factory

    Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)

    LLaMA-Factory is a fine-tuning and training framework for Meta's LLaMA language models. It enables researchers and developers to train and customize LLaMA models efficiently using advanced optimization techniques.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 8
    AutoTrain Advanced

    AutoTrain Advanced

    Faster and easier training and deployments

    AutoTrain Advanced is an open-source machine learning training framework developed by Hugging Face that simplifies the process of training and fine-tuning state-of-the-art AI models. The project provides a no-code and low-code interface that allows users to train models using custom datasets without needing extensive expertise in machine learning engineering. It supports a wide range of tasks including text classification, sequence-to-sequence modeling, token classification, sentence embedding training, and large language model fine-tuning. The system integrates closely with the Hugging Face ecosystem and allows developers to train models using datasets hosted on the Hugging Face Hub. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    verl-agent

    verl-agent

    Designed for training LLM/VLM agents via RL

    ...This step-wise interaction model makes it possible to train agents to operate in long-horizon scenarios where decisions depend on cumulative context and previous outcomes. Developers can configure memory modules that determine how historical information is stored and incorporated into each step of the reasoning process.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Build Securely on AWS with Proven Frameworks Icon
    Build Securely on AWS with Proven Frameworks

    Lay a foundation for success with Tested Reference Architectures developed by Fortinet’s experts. Learn more in this white paper.

    Moving to the cloud brings new challenges. How can you manage a larger attack surface while ensuring great network performance? Turn to Fortinet’s Tested Reference Architectures, blueprints for designing and securing cloud environments built by cybersecurity experts. Learn more and explore use cases in this white paper.
    Download Now
  • 10
    ReCall

    ReCall

    Learning to Reason with Search for LLMs via Reinforcement Learning

    ...Instead of relying purely on static knowledge stored inside the model, ReCall allows the language model to dynamically decide when it should retrieve information or invoke external capabilities during the reasoning process. The framework uses reinforcement learning to train models to perform these tool calls effectively while solving multi-step reasoning tasks.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    MaxText

    MaxText

    A simple, performant and scalable Jax LLM

    MaxText is a high-performance, highly scalable open-source framework designed to train and fine-tune large language models using the JAX ecosystem. The project acts as both a reference implementation and a practical training library that demonstrates best practices for building and scaling transformer-based language models on modern accelerator hardware. It is optimized to run efficiently on Google Cloud TPUs and GPUs, enabling researchers and engineers to train models ranging from small experiments to extremely large distributed workloads. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    Agent Reinforcement Trainer

    Agent Reinforcement Trainer

    Train multi-step agents for real-world tasks using GRPO

    ...Instead of just manually crafting prompts or relying on supervised fine-tuning, ART uses techniques like Group Relative Policy Optimization (GRPO) to let agents learn from environmental feedback and reward signals. The framework is designed to integrate easily with Python applications, abstracting much of the RL infrastructure so developers can train agents without deep RL expertise or heavy infrastructure overhead. ART also supports scalable training patterns, observability tools, and integration with hosted platforms like Weights & Biases, and it provides notebooks that demonstrate training on standard benchmarks and tasks.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    PyG

    PyG

    Graph Neural Network Library for PyTorch

    PyG (PyTorch Geometric) is a library built upon PyTorch to easily write and train Graph Neural Networks (GNNs) for a wide range of applications related to structured data. It consists of various methods for deep learning on graphs and other irregular structures, also known as geometric deep learning, from a variety of published papers. In addition, it consists of easy-to-use mini-batch loaders for operating on many small and single giant graphs, multi GPU-support, DataPipe support, distributed graph learning via Quiver, a large number of common benchmark datasets (based on simple interfaces to create your own), the GraphGym experiment manager, and helpful transforms, both for learning on arbitrary graphs as well as on 3D meshes or point clouds. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    Train Traffic Simulator

    Train Traffic Simulator

    A game about station masters job , managing trains which are arriving

    This program is a simulator for managing arrival and departure for different train traffic for different station , here the player acts just like a station master . Trains are created at different intervals of time with some metadata , the player is supposed to modify the signals and the switches to stop trains at the station or let them pass without any interruption. The goal will be to prevent delays , achieve maximum throughput and most importantly not cause collisions .
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    i.am.ai

    i.am.ai

    Roadmap to becoming an Artificial Intelligence Expert in 2022

    ...The project presents visual charts that outline multiple career paths such as data scientist, machine learning engineer, and AI specialist, helping learners understand what to study and in what order. It was originally created to train internal employees but was released publicly to support the broader community. The roadmap emphasizes foundational skills like mathematics, programming, and data handling before progressing into deep learning and specialized domains. Rather than prescribing a single path, it helps users navigate the AI landscape and understand which tools fit different scenarios. ...
    Downloads: 14 This Week
    Last Update:
    See Project
  • 16
    SuperDuperDB

    SuperDuperDB

    Integrate, train and manage any AI models and APIs with your database

    ...SuperDuperDB enables vector search in your existing database. Integrate and combine models from Sklearn, PyTorch, HuggingFace with AI APIs such as OpenAI to build even the most complex AI applications and workflows. Train models on your data in your datastore simply by querying without additional ingestion and pre-processing.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    lightning AI

    lightning AI

    The most intuitive, flexible, way for researchers to build models

    ...Download the code and type 'lightning run app'. Feel free to ssh into any machine and run from there as well. In research, we often have multiple separate scripts to train models, finetune them, collect results and more.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    AI4U

    AI4U

    Multi-engine plugin to specify agents with reinforcement learning

    ...Reinforcement learning promises to overcome traditional navigation mesh mechanisms in games and to provide more autonomous characters. AI4U can be integrated into Imitation Learning through Behavioral Cloning or Generative Adversarial Imitation Learning present on stable-baslines. Train using multiple concurrent Unity/Godot environment instances. Unity/Godot environment partial control from Python. Wrap Unity/Godot learning environments as a gym.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    tslearn

    tslearn

    The machine learning toolkit for time series analysis in Python

    The machine learning toolkit for time series analysis in Python. tslearn expects a time series dataset to be formatted as a 3D numpy array. The three dimensions correspond to the number of time series, the number of measurements per time series and the number of dimensions respectively (n_ts, max_sz, d). In order to get the data in the right format.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    Imagen - Pytorch

    Imagen - Pytorch

    Implementation of Imagen, Google's Text-to-Image Neural Network

    Implementation of Imagen, Google's Text-to-Image Neural Network that beats DALL-E2, in Pytorch. It is the new SOTA for text-to-image synthesis. Architecturally, it is actually much simpler than DALL-E2. It consists of a cascading DDPM conditioned on text embeddings from a large pre-trained T5 model (attention network). It also contains dynamic clipping for improved classifier-free guidance, noise level conditioning, and a memory-efficient unit design. It appears neither CLIP nor prior...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    DataDreamer

    DataDreamer

    DataDreamer: Prompt. Generate Synthetic Data. Train & Align Models

    DataDreamer is a tool designed to assist in the generation and manipulation of synthetic data for various applications, including testing and machine learning.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    Rhino

    Rhino

    On-device Speech-to-Intent engine powered by deep learning

    ...Develop voice features with a few lines of code using intuitive and cross-platform SDKs. Deliver voice AI everywhere: on-device, mobile, web browsers, on-premise, or cloud. Measure adoption, learn, and iterate. Continuously re-design and re-train to optimize engagement. Building accurate, responsive, and private voice technology is difficult. We learned the hard way, so you don’t have to. Picovoice heavily invests in R&D to offer superior voice AI that surpasses even Big Tech in accuracy and efficiency. Picovoice researchers do not follow recent frameworks and techniques but build them.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    AWS IoT FleetWise Edge

    AWS IoT FleetWise Edge

    AWS IoT FleetWise Edge Agent

    ...AWS IoT FleetWise makes it easy and cost-effective for automakers to collect, transform, and transfer vehicle data to the cloud in near-real-time and use it to build applications with analytics and machine learning that improve vehicle quality, safety, and autonomy. Train autonomous vehicles (AVs) and advanced driver assistance systems (ADAS) with camera data collected from a fleet of production vehicles. Improve electric vehicle (EV) battery range estimates with crowdsourced environmental data, such as weather and driving conditions, from nearby vehicles. Collect select data from nearby vehicles and use it to notify drivers of changing road conditions, such as lane closures or construction. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    Kubeflow Training Operator

    Kubeflow Training Operator

    Distributed ML Training and Fine-Tuning on Kubernetes

    Kubeflow Training Operator is a Kubernetes-native project for fine-tuning and scalable distributed training of machine learning (ML) models created with various ML frameworks such as PyTorch, TensorFlow, XGBoost, MPI, Paddle, and others.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    LLM101n

    LLM101n

    LLM101n: Let's build a Storyteller

    ...The materials favor compact, readable code and incremental steps, so learners can verify each concept before moving on. You’ll see how data pipelines, batching, masking, and positional encodings fit together to train a small GPT-style model end to end. The repo often complements explanations with runnable notebooks or scripts, encouraging experimentation and modification. By the end, the focus is less on polishing a production system and more on internalizing how LLM components interact to produce coherent text.
    Downloads: 2 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB