Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

WIP: [Impala] Scalable Distributed Deep-RL with ImportanceWeighted Actor-Learner Architectures #2147

Closed
wants to merge 46 commits into from

Conversation

joneswong
Copy link
Contributor

@joneswong joneswong commented May 28, 2018

This is a work-in-progress PR for implementing Impala, a scalable distributed reinforcement learning algorithm.
The impala optimizer has been implemented.
To do

  • Impala agent (v-trace)
  • support usage of LSTM and deep residual network
  • multi-gpu support (allreduce)

and after all the above items have been done, we must carefully test impala via reproducing experimental results reported by original paper.

Related issue number

1924

@AmplabJenkins
Copy link

Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/Ray-PRB/5654/
Test PASSed.

@@ -0,0 +1,3 @@
from ray.rllib.impala.impala import ImpalaAgent, DEFAULT_CONFIG
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Let's add

from __future__ import absolute_import
from __future__ import division
from __future__ import print_function

to the top of the file

@@ -1,3 +1,4 @@
from ray.rllib.optimizers.impala_optimizer import ImpalaOptimizer
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Looks like we forgot before, but let's also add

from __future__ import absolute_import
from __future__ import division
from __future__ import print_function

to the top of this file

@AmplabJenkins
Copy link

Test FAILed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/Ray-PRB/5779/
Test FAILed.

@AmplabJenkins
Copy link

Test PASSed.
Refer to this link for build results (access rights to CI server needed):
https://amplab.cs.berkeley.edu/jenkins//job/Ray-PRB/5850/
Test PASSed.

@ericl
Copy link
Contributor

ericl commented Jun 25, 2018

FYI, #2299

will add efficient batched LSTM support similar to that described in the IMPALA paper.

@robertnishihara
Copy link
Collaborator

@ericl @joneswong is this PR still relevant or should it be closed?

@ericl ericl closed this Oct 27, 2018
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

4 participants