There are 4 repositories under offline-reinforcement-learning topic.
High-quality single-file implementations of SOTA Offline and Offline-to-Online RL algorithms: AWAC, BC, CQL, DT, EDAC, IQL, SAC-N, TD3+BC, LB-SAC, SPOT, Cal-QL, ReBRAC
JAX (Flax) implementation of algorithms for Deep Reinforcement Learning with continuous action spaces.
An elegant PyTorch offline reinforcement learning library for researchers.
📚 List of Top-tier Conference Papers on Reinforcement Learning (RL),including: NeurIPS, ICML, AAAI, IJCAI, AAMAS, ICLR, ICRA, etc.
Minimal implementation of Decision Transformer: Reinforcement Learning via Sequence Modeling in PyTorch for mujoco control tasks in OpenAI gym
Datasets with baselines for Offline MARL.
A simple and easy-to-use autonomous driving environment for reinforcement learning, based on the CARLA simulator.
PyTorch implementation of the Offline Reinforcement Learning algorithm CQL. Includes the versions DQN-CQL and SAC-CQL for discrete and continuous action spaces.
Python interface for accessing the near real-world offline reinforcement learning (NeoRL) benchmark datasets
Unified Implementations of Offline Reinforcement Learning Algorithms
Code release for Efficient Planning in a Compact Latent Action Space (ICLR2023) https://arxiv.org/abs/2208.10291.
[ICLR 2024] The official implementation of "Safe Offline Reinforcement Learning with Feasibility-Guided Diffusion Model"
Official PyTorch implementation of "Uncertainty-Based Offline Reinforcement Learning with Diversified Q-Ensemble" (NeurIPS'21)
Pytorch Implementation of Stochastic MuZero for gym environment. This algorithm is capable of supporting a wide range of action and observation spaces, including both discrete and continuous variations.
[ICLR 2025] The offical implementation of "PSEC: Skill Expansion and Composition in Parameter Space", a new framework designed to facilitate efficient and flexible skill expansion and composition, iteratively evolve the agents' capabilities and efficiently address new challenges
[NeurIPS 2022 Oral] The official implementation of POR in "A Policy-Guided Imitation Approach for Offline Reinforcement Learning"
Author's implementation of ReBRAC, a minimalist improvement upon TD3+BC
Official implementation for "Anti-Exploration by Random Network Distillation", ICML 2023
Single-file SAC-N implementation on jax with flax and equinox. 10x faster than pytorch
Code for FOCAL Paper Published at ICLR 2021
An open-source Python library for Reinforcement Learning (RL), designed to model, optimize, and control dynamic systems.
Official implementation of "Direct Preference-based Policy Optimization without Reward Modeling" (NeurIPS 2023)
[NeurIPS 2023] The official implementation of "Offline Multi-Agent Reinforcement Learning with Implicit Global-to-Local Value Regularization"
[ICML 2025 Poster] Official PyTorch Implementation of "Habitizing Diffusion Planning for Efficient and Effective Decision Making"
[ICML 2022] The official implementation of DWBC in "Discriminator-Weighted Offline Imitation Learning from Suboptimal Demonstrations"
[ICML 2024] The offical implementation of A2PR, a simple way to achieve SOTA in offline reinforcement learning with an adaptive advantage-guided policy regularization method, in Pytorch
Code for NeurIPS 2022 paper Exploiting Reward Shifting in Value-Based Deep RL
A Production Tool for Embodied AI
Code for ICLR 2022 paper Rethinking Goal-Conditioned Supervised Learning and Its Connection to Offline RL.
[NeurIPS 2024] Code for Federated Ensemble-Directed Offline Reinforcement Learning
Experiment for Understanding the Effects of Dataset Characteristics on Offline Reinforcement Learning
Official code repo for paper: Hybrid RL: Using both offline and online data can make RL efficient.