-
Notifications
You must be signed in to change notification settings - Fork 6.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
WIP: [Impala] Scalable Distributed Deep-RL with ImportanceWeighted Actor-Learner Architectures #2147
Conversation
contribute DDPG and related test configurations to Ray RLlib
Test PASSed. |
@@ -0,0 +1,3 @@ | |||
from ray.rllib.impala.impala import ImpalaAgent, DEFAULT_CONFIG |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Let's add
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
to the top of the file
@@ -1,3 +1,4 @@ | |||
from ray.rllib.optimizers.impala_optimizer import ImpalaOptimizer |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Looks like we forgot before, but let's also add
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
to the top of this file
Test FAILed. |
Test PASSed. |
FYI, #2299 will add efficient batched LSTM support similar to that described in the IMPALA paper. |
@ericl @joneswong is this PR still relevant or should it be closed? |
This is a work-in-progress PR for implementing Impala, a scalable distributed reinforcement learning algorithm.
The impala optimizer has been implemented.
To do
and after all the above items have been done, we must carefully test impala via reproducing experimental results reported by original paper.
Related issue number
1924