Skip to content

Commit

Permalink
Fixes formatting
Browse files Browse the repository at this point in the history
  • Loading branch information
lcswillems committed Aug 10, 2022
1 parent 2397934 commit 18f5e17
Show file tree
Hide file tree
Showing 2 changed files with 11 additions and 20 deletions.
13 changes: 5 additions & 8 deletions scripts/evaluate.py
Original file line number Diff line number Diff line change
Expand Up @@ -41,6 +41,7 @@
print(f"Device: {device}\n")

# Load environments

envs = []
for i in range(args.procs):
env = utils.make_env(args.env, args.seed + 10000 * i)
Expand Down Expand Up @@ -75,16 +76,14 @@
obss, rewards, dones, _ = env.step(actions)
agent.analyze_feedbacks(rewards, dones)

log_episode_return += torch.tensor(rewards,
device=device, dtype=torch.float)
log_episode_return += torch.tensor(rewards, device=device, dtype=torch.float)
log_episode_num_frames += torch.ones(args.procs, device=device)

for i, done in enumerate(dones):
if done:
log_done_counter += 1
logs["return_per_episode"].append(log_episode_return[i].item())
logs["num_frames_per_episode"].append(
log_episode_num_frames[i].item())
logs["num_frames_per_episode"].append(log_episode_num_frames[i].item())

mask = 1 - torch.tensor(dones, device=device, dtype=torch.float)
log_episode_return *= mask
Expand All @@ -111,8 +110,6 @@
if n > 0:
print("\n{} worst episodes:".format(n))

indexes = sorted(range(
len(logs["return_per_episode"])), key=lambda k: logs["return_per_episode"][k])
indexes = sorted(range(len(logs["return_per_episode"])), key=lambda k: logs["return_per_episode"][k])
for i in indexes[:n]:
print("- episode {}: R={}, F={}".format(i,
logs["return_per_episode"][i], logs["num_frames_per_episode"][i]))
print("- episode {}: R={}, F={}".format(i, logs["return_per_episode"][i], logs["num_frames_per_episode"][i]))
18 changes: 6 additions & 12 deletions scripts/train.py
Original file line number Diff line number Diff line change
Expand Up @@ -111,8 +111,7 @@

# Load observations preprocessor

obs_space, preprocess_obss = utils.get_obss_preprocessor(
envs[0].observation_space)
obs_space, preprocess_obss = utils.get_obss_preprocessor(envs[0].observation_space)
if "vocab" in status:
preprocess_obss.vocab.load_vocab(status["vocab"])
txt_logger.info("Observations preprocessor loaded")
Expand Down Expand Up @@ -167,22 +166,17 @@
fps = logs["num_frames"]/(update_end_time - update_start_time)
duration = int(time.time() - start_time)
return_per_episode = utils.synthesize(logs["return_per_episode"])
rreturn_per_episode = utils.synthesize(
logs["reshaped_return_per_episode"])
num_frames_per_episode = utils.synthesize(
logs["num_frames_per_episode"])
rreturn_per_episode = utils.synthesize(logs["reshaped_return_per_episode"])
num_frames_per_episode = utils.synthesize(logs["num_frames_per_episode"])

header = ["update", "frames", "FPS", "duration"]
data = [update, num_frames, fps, duration]
header += ["rreturn_" + key for key in rreturn_per_episode.keys()]
data += rreturn_per_episode.values()
header += ["num_frames_" +
key for key in num_frames_per_episode.keys()]
header += ["num_frames_" + key for key in num_frames_per_episode.keys()]
data += num_frames_per_episode.values()
header += ["entropy", "value",
"policy_loss", "value_loss", "grad_norm"]
data += [logs["entropy"], logs["value"], logs["policy_loss"],
logs["value_loss"], logs["grad_norm"]]
header += ["entropy", "value", "policy_loss", "value_loss", "grad_norm"]
data += [logs["entropy"], logs["value"], logs["policy_loss"], logs["value_loss"], logs["grad_norm"]]

txt_logger.info(
"U {} | F {:06} | FPS {:04.0f} | D {} | rR:μσmM {:.2f} {:.2f} {:.2f} {:.2f} | F:μσmM {:.1f} {:.1f} {} {} | H {:.3f} | V {:.3f} | pL {:.3f} | vL {:.3f} | ∇ {:.3f}"
Expand Down

0 comments on commit 18f5e17

Please sign in to comment.