Open Source Linux Reinforcement Learning Frameworks - Page 2

Reinforcement Learning Frameworks for Linux

View 1 business solution
  • Full-stack observability with actually useful AI | Grafana Cloud Icon
    Full-stack observability with actually useful AI | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • 1
    Transformer Reinforcement Learning X

    Transformer Reinforcement Learning X

    A repo for distributed training of language models with Reinforcement

    trlX is a distributed training framework designed from the ground up to focus on fine-tuning large language models with reinforcement learning using either a provided reward function or a reward-labeled dataset. Training support for Hugging Face models is provided by Accelerate-backed trainers, allowing users to fine-tune causal and T5-based language models of up to 20B parameters, such as facebook/opt-6.7b, EleutherAI/gpt-neox-20b, and google/flan-t5-xxl. For models beyond 20B parameters, trlX provides NVIDIA NeMo-backed trainers that leverage efficient parallelism techniques to scale effectively.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 2
    Weights and Biases

    Weights and Biases

    Tool for visualizing and tracking your machine learning experiments

    Use W&B to build better models faster. Track and visualize all the pieces of your machine learning pipeline, from datasets to production models. Quickly identify model regressions. Use W&B to visualize results in real time, all in a central dashboard. Focus on the interesting ML. Spend less time manually tracking results in spreadsheets and text files. Capture dataset versions with W&B Artifacts to identify how changing data affects your resulting models. Reproduce any model, with saved code, hyperparameters, launch commands, input data, and resulting model weights. Set wandb.config once at the beginning of your script to save your hyperparameters, input settings (like dataset name or model type), and any other independent variables for your experiments. This is useful for analyzing your experiments and reproducing your work in the future. Setting configs also allows you to visualize the relationships between features of your model architecture or data pipeline and model performance.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 3
    verl

    verl

    Volcano Engine Reinforcement Learning for LLMs

    VERL is a reinforcement-learning–oriented toolkit designed to train and align modern AI systems, from language models to decision-making agents. It brings together supervised fine-tuning, preference modeling, and online RL into one coherent training stack so teams can move from raw data to aligned policies with minimal glue code. The library focuses on scalability and efficiency, offering distributed training loops, mixed precision, and replay/buffering utilities that keep accelerators busy. It ships with reference implementations of popular alignment algorithms and clear examples that make it straightforward to reproduce baselines before customizing. Data pipelines treat human feedback, simulated environments, and synthetic preferences as interchangeable sources, which helps with rapid experimentation. VERL is meant for both research and production hardening: logging, checkpointing, and evaluation suites are built in so you can track learning dynamics and regressions over time.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 4
    SkyAI
    Highly modularized Reinforcement Learning library for real/simulation robots to learn behaviors. Our ultimate goal is to develop an artificial intelligence (AI) program with which the robots can learn to behave as their users wish.
    Downloads: 2 This Week
    Last Update:
    See Project
  • AI-generated apps that pass security review Icon
    AI-generated apps that pass security review

    Stop waiting on engineering. Build production-ready internal tools with AI—on your company data, in your cloud.

    Retool lets you generate dashboards, admin panels, and workflows directly on your data. Type something like “Build me a revenue dashboard on my Stripe data” and get a working app with security, permissions, and compliance built in from day one. Whether on our cloud or self-hosted, create the internal software your team needs without compromising enterprise standards or control.
    Try Retool free
  • 5
    festival3os

    festival3os

    mods to the Festival sokoban solver to run on OSX + Win + linux

    Mods to the Festival sokoban solver that allow building on OSX, Linux, & Windows
    Downloads: 2 This Week
    Last Update:
    See Project
  • 6
    tic tac toe AI

    tic tac toe AI

    simplest AI programme of tic-tac-toe game

    This is a program of tic tac toe game it currently is the 1.0 version of this this is my program - an AI program which plays tic-tac-toe, it is an AI program which is given knowledge on the basis of my previous analysis and knowledge about playing tic-tac-toe. I have made it to be playable with players right now but I can make it for AI vs AI, AI vs player, player vs player as well. Using a settings option. I think this program has enough IQ to defeat a normal person. This is the update 1.1 of this game. My future visions about this program is: v 1.0.1 --> bug fixes v 1.1 --> (added) click interaction _______________________________________________________________________________________________________________________________________________ v 1.2 --> addition of reinforcement learning (cache data different for each computer unlike v1.3). v 1.3 --> addition of cloud reinforcement learning (optional; chosen from settings). ... & more
    Downloads: 2 This Week
    Last Update:
    See Project
  • 7
    AI4U

    AI4U

    Multi-engine plugin to specify agents with reinforcement learning

    AI4U is a multi-engine plugin (Godot and Unity) that allows you to design Non-Player Characters (NPCs) of games using an agent abstraction. In addition, AI4U has a low-level API that allows you to connect the agent to any algorithm made available in Python by the reinforcement learning community specifically and by the Artificial Intelligence community in general. Reinforcement learning promises to overcome traditional navigation mesh mechanisms in games and to provide more autonomous characters. AI4U can be integrated into Imitation Learning through Behavioral Cloning or Generative Adversarial Imitation Learning present on stable-baslines. Train using multiple concurrent Unity/Godot environment instances. Unity/Godot environment partial control from Python. Wrap Unity/Godot learning environments as a gym.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    Acme

    Acme

    A library of reinforcement learning components and agents

    Acme is a framework from DeepMind for building scalable and reproducible reinforcement learning agents. It emphasizes modular components, distributed training, and ease of experimentation.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    Alibi Explain

    Alibi Explain

    Algorithms for explaining machine learning models

    Alibi is a Python library aimed at machine learning model inspection and interpretation. The focus of the library is to provide high-quality implementations of black-box, white-box, local and global explanation methods for classification and regression models.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Go from Code to Production URL in Seconds Icon
    Go from Code to Production URL in Seconds

    Cloud Run deploys apps in any language instantly. Scales to zero. Pay only when code runs.

    Skip the Kubernetes configs. Cloud Run handles HTTPS, scaling, and infrastructure automatically. Two million requests free per month.
    Try it free
  • 10
    AndroidEnv

    AndroidEnv

    RL research on Android devices

    android_env is a reinforcement learning (RL) environment developed by Google DeepMind that enables agents to interact with Android applications directly as a learning environment. It provides a standardized API for training agents to perform tasks on Android apps, supporting tasks ranging from games to productivity apps, making it suitable for research in real-world RL settings.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    AnyTrading

    AnyTrading

    The most simple, flexible, and comprehensive OpenAI Gym trading

    gym-anytrading is an OpenAI Gym-compatible environment designed for developing and testing reinforcement learning algorithms on trading strategies. It simulates trading environments for financial markets, including stocks and forex.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12

    Astrape

    Optical-packet node transceiver frequency allocation

    In an optical network scenario which consists of multiple nodes (whiteboxes) at its edges and ROADMs in-between, the coherent transceiver average laser configuration time is improved. The process is evaluated according to a testbed setup. This is facilitated in the appropriate lab equipment (or via simulation when required). For that purpose, a software agent (Netconf server) residing at the whiteboxes, is developed receiving input from the Software-Defined Networking (SDN) packet controller (PacketCTL - a Netconf client). Then, configuration of the local transceiver laser frequencies of the controlled pluggable devices takes place, for facilitating the connectivity in-between the ROADM network. Also, the agent records and reports back telemetry data (feedback) which is used by the PacketCTL's resource-allocating mechanism to improve efficiency within the network topology.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    Atropos

    Atropos

    Language Model Reinforcement Learning Environments frameworks

    Atropos is a comprehensive open-source framework for reinforcement learning (RL) environments tailored specifically to work with large language models (LLMs). Designed as a scalable ecosystem of environment microservices, Atropos allows researchers and developers to collect, evaluate, and manage trajectories (sequences of actions and outcomes) generated by LLMs across a variety of tasks—from static dataset benchmarks to dynamic interactive games and real-world scenario environments. It provides foundational tooling for asynchronous RL loops where environment services communicate with trainers and inference engines, enabling complex workflow orchestration in distributed and parallel setups. This framework facilitates experimentation with RLHF (Reinforcement Learning from Human Feedback), RLAIF, or multi-turn training approaches by abstracting environment logic, scoring, and logging into reusable components.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    Best-of Machine Learning with Python

    Best-of Machine Learning with Python

    A ranked list of awesome machine learning Python libraries

    This curated list contains 900 awesome open-source projects with a total of 3.3M stars grouped into 34 categories. All projects are ranked by a project-quality score, which is calculated based on various metrics automatically collected from GitHub and different package managers. If you like to add or update projects, feel free to open an issue, submit a pull request, or directly edit the projects.yaml. Contributions are very welcome! General-purpose machine learning and deep learning frameworks.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    BindsNET

    BindsNET

    Simulation of spiking neural networks (SNNs) using PyTorch

    A Python package used for simulating spiking neural networks (SNNs) on CPUs or GPUs using PyTorch Tensor functionality. BindsNET is a spiking neural network simulation library geared towards the development of biologically inspired algorithms for machine learning. This package is used as part of ongoing research on applying SNNs to machine learning (ML) and reinforcement learning (RL) problems in the Biologically Inspired Neural & Dynamical Systems (BINDS) lab.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16

    CLSquare

    Closed Loop Simulation System

    Closed Loop Simulation System (CLSquare) is an integrated architecture to train, test and compare reinforcement learning controllers on different plants. CLSquare provides simulated plants as well as interfaces to real plants.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    CORL

    CORL

    High-quality single-file implementations of SOTA Offline

    CORL (Collection of Reinforcement Learning Environments for Control Tasks) is a modular and extensible set of high-quality reinforcement learning environments focused on continuous control and robotics. It aims to offer standardized environments suitable for benchmarking state-of-the-art RL algorithms in control tasks, including physics-based simulations and custom-designed scenarios.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    Catalyst

    Catalyst

    Accelerated deep learning R&D

    Catalyst is a PyTorch framework for accelerated Deep Learning research and development. It allows you to write compact but full-featured Deep Learning pipelines with just a few lines of code. With Catalyst you get a full set of features including a training loop with metrics, model checkpointing and more, all without the boilerplate. Catalyst is focused on reproducibility, rapid experimentation, and codebase reuse so you can break the cycle of writing another regular train loop and make something totally new. Catalyst is compatible with Python 3.6+. PyTorch 1.1+, and has been tested on Ubuntu 16.04/18.04/20.04, macOS 10.15, Windows 10 and Windows Subsystem for Linux. It's part of the PyTorch Ecosystem, as well as the Catalyst Ecosystem which includes Alchemy (experiments logging & visualization) and Reaction (convenient deep learning models serving).
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    ChainerRL

    ChainerRL

    ChainerRL is a deep reinforcement learning library

    ChainerRL (this repository) is a deep reinforcement learning library that implements various state-of-the-art deep reinforcement algorithms in Python using Chainer, a flexible deep learning framework. PFRL is the PyTorch analog of ChainerRL. ChainerRL has a set of accompanying visualization tools in order to aid developers' ability to understand and debug their RL agents. With this visualization tool, the behavior of ChainerRL agents can be easily inspected from a browser UI. Environments that support the subset of OpenAI Gym's interface (reset and step methods) can be used.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    CleanRL

    CleanRL

    High-quality single file implementation of Deep Reinforcement Learning

    CleanRL is a Deep Reinforcement Learning library that provides high-quality single-file implementation with research-friendly features. The implementation is clean and simple, yet we can scale it to run thousands of experiments using AWS Batch. CleanRL is not a modular library and therefore it is not meant to be imported. At the cost of duplicate code, we make all implementation details of a DRL algorithm variant easy to understand, so CleanRL comes with its own pros and cons. You should consider using CleanRL if you want to 1) understand all implementation details of an algorithm's variant or 2) prototype advanced features that other modular DRL libraries do not support (CleanRL has minimal lines of code so it gives you great debugging experience and you don't have to do a lot of subclassing like sometimes in modular DRL libraries).
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    Coach

    Coach

    Enables easy experimentation with state of the art algorithms

    Coach is a python framework that models the interaction between an agent and an environment in a modular way. With Coach, it is possible to model an agent by combining various building blocks, and training the agent on multiple environments. The available environments allow testing the agent in different fields such as robotics, autonomous driving, games and more. It exposes a set of easy-to-use APIs for experimenting with new RL algorithms and allows simple integration of new environments to solve. Coach collects statistics from the training process and supports advanced visualization techniques for debugging the agent being trained. Coach supports many state-of-the-art reinforcement learning algorithms, which are separated into three main classes - value optimization, policy optimization, and imitation learning. Coach supports a large number of environments which can be solved using reinforcement learning.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    ConvNetJS

    ConvNetJS

    Deep learning in Javascript to train convolutional neural networks

    ConvNetJS is a Javascript library for training Deep Learning models (Neural Networks) entirely in your browser. Open a tab and you're training. No software requirements, no compilers, no installations, no GPUs, no sweat. ConvNetJS is an implementation of Neural networks, together with nice browser-based demos. It currently supports common Neural Network modules (fully connected layers, non-linearities), classification (SVM/Softmax) and Regression (L2) cost functions, ability to specify and train Convolutional Networks that process images, and experimental Reinforcement Learning modules, based on Deep Q Learning. The library allows you to formulate and solve Neural Networks in Javascript. If you would like to add features to the library, you will have to change the code in src/ and then compile the library into the build/ directory. The compilation script simply concatenates files in src/ and then minifies the result.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    Cosmos-RL

    Cosmos-RL

    Cosmos-RL is a flexible and scalable Reinforcement Learning framework

    Cosmos-RL is a scalable reinforcement learning framework designed specifically for physical AI systems such as robotics, autonomous agents, and multimodal models. It provides a distributed training architecture that separates policy learning and environment rollout processes, enabling efficient and asynchronous reinforcement learning at scale. The framework supports multiple parallelism strategies, including tensor, pipeline, and data parallelism, allowing it to leverage large GPU clusters effectively. It is built with compatibility in mind, supporting popular model families such as LLaMA, Qwen, and diffusion-based world models, as well as integration with Hugging Face ecosystems. cosmos-rl also includes support for advanced RL algorithms, low-precision training, and fault-tolerant execution, making it suitable for large-scale production workloads.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    DI-engine

    DI-engine

    OpenDILab Decision AI Engine

    DI-engine is a unified reinforcement learning (RL) platform for reproducible and scalable RL research. It offers modular pipelines for various RL algorithms, with an emphasis on production-level training and evaluation.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    Deep Reinforcement Learning for Keras

    Deep Reinforcement Learning for Keras

    Deep Reinforcement Learning for Keras.

    keras-rl implements some state-of-the-art deep reinforcement learning algorithms in Python and seamlessly integrates with the deep learning library Keras. Furthermore, keras-rl works with OpenAI Gym out of the box. This means that evaluating and playing around with different algorithms is easy. Of course, you can extend keras-rl according to your own needs. You can use built-in Keras callbacks and metrics or define your own. Even more so, it is easy to implement your own environments and even algorithms by simply extending some simple abstract classes. Documentation is available online.
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB