From 85a65d8f92188c5047bfae8b98166e392fee955f Mon Sep 17 00:00:00 2001 From: Noa Nabeshima <31455280+noanabeshima@users.noreply.github.com> Date: Thu, 12 Dec 2019 18:51:39 -0800 Subject: [PATCH] Update PPO.py --- PPO.py | 3 +-- 1 file changed, 1 insertion(+), 2 deletions(-) diff --git a/PPO.py b/PPO.py index 5383418..3b36d33 100644 --- a/PPO.py +++ b/PPO.py @@ -23,8 +23,7 @@ def clear_memory(self): class ActorCritic(nn.Module): def __init__(self, state_dim, action_dim, n_latent_var): super(ActorCritic, self).__init__() - self.affine = nn.Linear(state_dim, n_latent_var) - + # actor self.action_layer = nn.Sequential( nn.Linear(state_dim, n_latent_var),