@@ -15,11 +15,57 @@ export OMP_NUM_THREADS=8
15
15
pip install -r ${BASE} /requirements.txt
16
16
17
17
# train dummy
18
- for strategy in ddp colossalai_gemini colossalai_zero2; do
19
- torchrun --standalone --nproc_per_node=2 ${BASE} /train_dummy.py --strategy ${strategy} --num_episodes 2 --max_timesteps 3 --update_timesteps 3 --max_epochs 3 --experience_batch_size 4 --train_batch_size 4
20
- done
18
+ python ${BASE} /train_dummy.py --strategy naive --num_episodes 1 \
19
+ --max_timesteps 2 --update_timesteps 2 \
20
+ --max_epochs 1 --train_batch_size 2 --lora_rank 4
21
+
22
+ torchrun --standalone --nproc_per_node=2 ${BASE} /train_dummy.py \
23
+ --strategy colossalai_gemini --num_episodes 1 --max_timesteps 2 \
24
+ --update_timesteps 2 --max_epochs 1 --train_batch_size 2\
25
+ --pretrain ' facebook/opt-350m' --model opt --lora_rank 4\
26
+ --save_path ${BASE} /actor_checkpoint_dummy.pt
27
+ python ${BASE} /inference.py --model_path ${BASE} /actor_checkpoint_dummy.pt --pretrain ' facebook/opt-350m' --model opt
28
+
29
+ torchrun --standalone --nproc_per_node=2 ${BASE} /train_dummy.py \
30
+ --strategy ddp --num_episodes 1 --max_timesteps 2 \
31
+ --update_timesteps 2 --max_epochs 1 --train_batch_size 2\
32
+ --pretrain ' facebook/opt-350m' --model opt --lora_rank 4\
33
+ --save_path ${BASE} /actor_checkpoint_dummy.pt
34
+ python ${BASE} /inference.py --model_path ${BASE} /actor_checkpoint_dummy.pt --pretrain ' facebook/opt-350m' --model opt
35
+
36
+ torchrun --standalone --nproc_per_node=2 ${BASE} /train_dummy.py \
37
+ --strategy colossalai_zero2 --num_episodes 1 --max_timesteps 2 \
38
+ --update_timesteps 2 --max_epochs 1 --train_batch_size 2\
39
+ --pretrain ' gpt2' --model gpt2 --lora_rank 4\
40
+ --save_path ${BASE} /actor_checkpoint_dummy.pt
41
+ python ${BASE} /inference.py --model_path ${BASE} /actor_checkpoint_dummy.pt --pretrain ' gpt2' --model gpt2
42
+
43
+ rm -rf ${BASE} /actor_checkpoint_dummy.pt
21
44
22
45
# train prompts
23
- for strategy in ddp colossalai_gemini colossalai_zero2; do
24
- torchrun --standalone --nproc_per_node=2 ${BASE} /train_prompts.py $PROMPT_PATH --strategy ${strategy} --num_episodes 2 --max_timesteps 3 --update_timesteps 3 --max_epochs 3
25
- done
46
+ python ${BASE} /train_prompts.py $PROMPT_PATH --strategy naive --num_episodes 1 \
47
+ --max_timesteps 2 --update_timesteps 2 \
48
+ --max_epochs 1 --train_batch_size 2 --lora_rank 4
49
+
50
+ torchrun --standalone --nproc_per_node=2 ${BASE} /train_prompts.py $PROMPT_PATH \
51
+ --strategy colossalai_zero2 --num_episodes 1 --max_timesteps 2 \
52
+ --update_timesteps 2 --max_epochs 1 --train_batch_size 2\
53
+ --pretrain ' facebook/opt-350m' --model opt --lora_rank 4\
54
+ --save_path ${BASE} /actor_checkpoint_prompts.pt
55
+ python ${BASE} /inference.py --model_path ${BASE} /actor_checkpoint_prompts.pt --pretrain ' facebook/opt-350m' --model opt
56
+
57
+ torchrun --standalone --nproc_per_node=2 ${BASE} /train_prompts.py $PROMPT_PATH \
58
+ --strategy ddp --num_episodes 1 --max_timesteps 2 \
59
+ --update_timesteps 2 --max_epochs 1 --train_batch_size 2\
60
+ --pretrain ' gpt2' --model gpt2 --lora_rank 4\
61
+ --save_path ${BASE} /actor_checkpoint_prompts.pt
62
+ python ${BASE} /inference.py --model_path ${BASE} /actor_checkpoint_prompts.pt --pretrain ' gpt2' --model gpt2
63
+
64
+ torchrun --standalone --nproc_per_node=2 ${BASE} /train_prompts.py $PROMPT_PATH \
65
+ --strategy colossalai_gemini --num_episodes 1 --max_timesteps 2 \
66
+ --update_timesteps 2 --max_epochs 1 --train_batch_size 2\
67
+ --pretrain ' gpt2' --model gpt2 --lora_rank 4\
68
+ --save_path ${BASE} /actor_checkpoint_prompts.pt
69
+ python ${BASE} /inference.py --model_path ${BASE} /actor_checkpoint_prompts.pt --pretrain ' gpt2' --model gpt2
70
+
71
+ rm -rf ${BASE} /actor_checkpoint_prompts.pt
0 commit comments