Skip to content

Commit

Permalink
Use full dataset for eval (SequentialSampler in Distributed setting)
Browse files Browse the repository at this point in the history
  • Loading branch information
VictorSanh committed Dec 3, 2019
1 parent f434bfc commit 48cbf26
Show file tree
Hide file tree
Showing 4 changed files with 4 additions and 4 deletions.
2 changes: 1 addition & 1 deletion examples/run_glue.py
Original file line number Diff line number Diff line change
Expand Up @@ -231,7 +231,7 @@ def evaluate(args, model, tokenizer, prefix=""):

args.eval_batch_size = args.per_gpu_eval_batch_size * max(1, args.n_gpu)
# Note that DistributedSampler samples randomly
eval_sampler = SequentialSampler(eval_dataset) if args.local_rank == -1 else DistributedSampler(eval_dataset)
eval_sampler = SequentialSampler(eval_dataset)
eval_dataloader = DataLoader(eval_dataset, sampler=eval_sampler, batch_size=args.eval_batch_size)

# multi-gpu eval
Expand Down
2 changes: 1 addition & 1 deletion examples/run_lm_finetuning.py
Original file line number Diff line number Diff line change
Expand Up @@ -300,7 +300,7 @@ def evaluate(args, model, tokenizer, prefix=""):

args.eval_batch_size = args.per_gpu_eval_batch_size * max(1, args.n_gpu)
# Note that DistributedSampler samples randomly
eval_sampler = SequentialSampler(eval_dataset) if args.local_rank == -1 else DistributedSampler(eval_dataset)
eval_sampler = SequentialSampler(eval_dataset)
eval_dataloader = DataLoader(eval_dataset, sampler=eval_sampler, batch_size=args.eval_batch_size)

# multi-gpu evaluate
Expand Down
2 changes: 1 addition & 1 deletion examples/run_multiple_choice.py
Original file line number Diff line number Diff line change
Expand Up @@ -226,7 +226,7 @@ def evaluate(args, model, tokenizer, prefix="", test=False):

args.eval_batch_size = args.per_gpu_eval_batch_size * max(1, args.n_gpu)
# Note that DistributedSampler samples randomly
eval_sampler = SequentialSampler(eval_dataset) if args.local_rank == -1 else DistributedSampler(eval_dataset)
eval_sampler = SequentialSampler(eval_dataset)
eval_dataloader = DataLoader(eval_dataset, sampler=eval_sampler, batch_size=args.eval_batch_size)

# multi-gpu evaluate
Expand Down
2 changes: 1 addition & 1 deletion examples/run_xnli.py
Original file line number Diff line number Diff line change
Expand Up @@ -206,7 +206,7 @@ def evaluate(args, model, tokenizer, prefix=""):

args.eval_batch_size = args.per_gpu_eval_batch_size * max(1, args.n_gpu)
# Note that DistributedSampler samples randomly
eval_sampler = SequentialSampler(eval_dataset) if args.local_rank == -1 else DistributedSampler(eval_dataset)
eval_sampler = SequentialSampler(eval_dataset)
eval_dataloader = DataLoader(eval_dataset, sampler=eval_sampler, batch_size=args.eval_batch_size)

# multi-gpu eval
Expand Down

1 comment on commit 48cbf26

@VictorSanh
Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks for pointing that out @ethanjperez (cf 96e8350)!

Please sign in to comment.