TRFL, developed by Google DeepMind, is a TensorFlow-based library that provides a collection of essential building blocks for reinforcement learning (RL) algorithms. Pronounced “truffle,” it simplifies the implementation of RL agents by offering reusable components such as loss functions, value estimation tools, and temporal difference (TD) learning operators. The library is designed to integrate seamlessly with TensorFlow, allowing users to define differentiable RL objectives and train models using standard optimization routines. TRFL supports both CPU and GPU TensorFlow environments, though TensorFlow itself must be installed separately. It exposes clean, modular APIs for various RL methods including Q-learning, policy gradient, and actor-critic algorithms, among others. Each function returns not only the computed loss tensor but also a detailed structure containing auxiliary information like TD errors and targets.

Features

  • Provides modular TensorFlow operations for reinforcement learning algorithms
  • Includes Q-learning, actor-critic, policy gradient, and value-based losses
  • Returns structured outputs with loss and diagnostic information
  • Fully differentiable for use in end-to-end RL training pipelines
  • Works with both CPU and GPU versions of TensorFlow
  • Lightweight design for easy integration into custom RL research projects

Project Samples

Project Activity

See All Activity >

Categories

Libraries

License

Apache License V2.0

Follow TRFL

TRFL Web Site

Other Useful Business Software
Fully Managed MySQL, PostgreSQL, and SQL Server Icon
Fully Managed MySQL, PostgreSQL, and SQL Server

Automatic backups, patching, replication, and failover. Focus on your app, not your database.

Cloud SQL handles your database ops end to end, so you can focus on your app.
Try Free
Rate This Project
Login To Rate This Project

User Reviews

Be the first to post a review of TRFL!

Additional Project Details

Operating Systems

Linux, Mac

Programming Language

Python

Related Categories

Python Libraries

Registered

2025-10-09