OpenDILab Decision AI Engine. The Most Comprehensive Reinforcement Learning Framework B.P.
-
Updated
May 29, 2025 - Python
OpenDILab Decision AI Engine. The Most Comprehensive Reinforcement Learning Framework B.P.
Classic papers and resources on recommendation
Python implementations of contextual bandits algorithms
Code to reproduce the experiments in Sample Efficient Reinforcement Learning via Model-Ensemble Exploration and Exploitation (MEEE).
This is the pytorch implementation of ICML 2018 paper - Self-Imitation Learning.
Code for NeurIPS 2022 paper Exploiting Reward Shifting in Value-Based Deep RL
Repository Containing Comparison of two methods for dealing with Exploration-Exploitation dilemma for MultiArmed Bandits
The official code release for "Langevin Soft Actor-Critic: Efficient Exploration through Uncertainty-Driven Critic Learning", ICLR 2025
Official implementation of LECO (NeurIPS'22)
The official code release for Provable and Practical: Efficient Exploration in Reinforcement Learning via Langevin Monte Carlo, ICLR 2024.
Deep Intrinsically Motivated Exploration in Continuous Control
A short implementation of bandit algorithms - ETC, UCB, MOSS and KL-UCB
The GitHub repository for "Accelerating Approximate Thompson Sampling with Underdamped Langevin Monte Carlo", AISTATS 2024.
The official code release for "More Efficient Randomized Exploration for Reinforcement Learning via Approximate Sampling", Reinforcement Learning Conference (RLC) 2024
This project focuses on comparing different Reinforcement Learning Algorithms, including monte-carlo, q-learning, lambda q-learning epsilon-greedy variations, etc.
Research Thesis - Reinforcement Learning
An Optimistic Approach to the Q-Network Error in Actor-Critic Methods
A reinforcement learning project where a snake learns to navigate and survive in a dynamic environment through Q-learning.
over-parameterization = exploration ?
Add a description, image, and links to the exploration-exploitation topic page so that developers can more easily learn about it.
To associate your repository with the exploration-exploitation topic, visit your repo's landing page and select "manage topics."