We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 13c0d3f commit 4be74beCopy full SHA for 4be74be
contents/12_Proximal_Policy_Optimization/DPPO.py
@@ -125,7 +125,7 @@ def work(self):
125
s = s_
126
ep_r += r
127
128
- GLOBAL_UPDATE_COUNTER += 1 # count to minimum batch size
+ GLOBAL_UPDATE_COUNTER += 1 # count to minimum batch size, no need to wait other workers
129
if t == EP_LEN - 1 or GLOBAL_UPDATE_COUNTER >= MIN_BATCH_SIZE:
130
v_s_ = self.ppo.get_v(s_)
131
discounted_r = [] # compute discounted reward
0 commit comments