Skip to content
This repository was archived by the owner on Aug 29, 2024. It is now read-only.

Commit 2ce885d

Browse files
Rishabh Agarwalagarwl
authored andcommitted
Update README with dataset visualization colab.
PiperOrigin-RevId: 320332055
1 parent 3e42af5 commit 2ce885d

File tree

1 file changed

+5
-2
lines changed

1 file changed

+5
-2
lines changed

README.md

Lines changed: 5 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -40,7 +40,6 @@ run the command:
4040
gsutil -m cp -R gs://atari-replay-datasets/dqn/[GAME_NAME]
4141
```
4242

43-
4443
This data can be generated by running the online agents using
4544
[`batch_rl/baselines/train.py`](https://github.com/google-research/batch_rl/blob/master/batch_rl/baselines/train.py) for 200 million frames
4645
(standard protocol). Note that the dataset consists of approximately 50 million
@@ -58,6 +57,10 @@ agent's previous action again, instead of the agent's new action.
5857
[gcp_bucket]: https://console.cloud.google.com/storage/browser/atari-replay-datasets
5958
[project_page]: https://offline-rl.github.io
6059

60+
### Dataset Visualization Colab
61+
A colab showing how to load the dataset using dopamine replay buffers and
62+
visualize the observations can be found [here](https://colab.research.google.com/drive/17JQzRpIrguMjmZN5bL-RYzDJD5GSV8MJ?usp=sharing).
63+
6164
## Asymptotic Performance of offline agents on Atari-replay dataset
6265

6366
<div>
@@ -190,7 +193,7 @@ If you find this open source release useful, please reference in your paper:
190193
> An Optimistic Perspective on Offline Reinforcement Learning
191194
> *International Conference on Machine Learning (ICML)*.
192195
193-
@article{agarwal2019optimistic,
196+
@article{agarwal2020optimistic,
194197
title={An Optimistic Perspective on Offline Reinforcement Learning},
195198
author={Agarwal, Rishabh and Schuurmans, Dale and Norouzi, Mohammad},
196199
journal={International Conference on Machine Learning},

0 commit comments

Comments
 (0)