-
Notifications
You must be signed in to change notification settings - Fork 110
Low CPU and GPU utilization when running in the gpu mode. #15
Comments
EDIT: ignore the first part of this, I misread TLDR: Based on what's in the README, + some of the closed issues and parts of cfvpy/selfplay.py, I think the intended use case for non-zero (And, if you only have one GPU, that's what the CPU-based data generation, i.e. setting selfplay.cpu_gen_threads=60 in the example, is for) Having said all that, some comments by the authors do seem to suggest that it's probably ok to modify the relevant selfplay.py code to also use GPU 0 for data generation -> https://github.com/facebookresearch/rebel/blob/master/cfvpy/selfplay.py#L193. So might be worth trying that out perhaps? Longer explanation: I'm not 100% certain about any of this (I'm a complete newbie to CUDA / cuDNN), but after reading through some of the closed issues + the code, I think:
On a side note, for anyone else poking around this stuff in 2024, here's a couple quick things I've noticed:
|
Here is my script:
python run.py --adhoc --cfg conf/c02_selfplay/liars_sp.yaml env.num_dice=1 env.num_faces=4 env.subgame_params.use_cfr=true selfplay.cpu_gen_threads=0 selfplay.threads_per_gpu=16
My computer configuration is 22080ti + 2cpu with 24 thread per cpu.
The log seems quite normal, just collecting experience and training. However, I find that the utilization of cpu is only 5% and the gpu is 6%. Is this normal? And could you tell me how can I fasten the process? Thanks a lot!
The text was updated successfully, but these errors were encountered: