Skip to content

Commit

Permalink
Put LR back, need to save larger checkpoints to weka to prevent timeouts
Browse files Browse the repository at this point in the history
  • Loading branch information
jakep-allenai committed Oct 17, 2024
1 parent e141c91 commit 529d51d
Show file tree
Hide file tree
Showing 2 changed files with 2 additions and 2 deletions.
2 changes: 1 addition & 1 deletion pdelfin/train/config/qwen2vl-7b-lora.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -46,7 +46,7 @@ hparams:
gradient_accumulation_steps: 4
gradient_checkpointing: true
clip_grad_norm: 1.0
learning_rate: 3e-4
learning_rate: 1e-4
max_steps: 10000
pad_multiple_of: 16
log_every_steps: 10
Expand Down
2 changes: 1 addition & 1 deletion scripts/qwen2vl-7b-gantry.sh
Original file line number Diff line number Diff line change
Expand Up @@ -10,7 +10,7 @@ then
fi


EXTRA_ARGS="-c pdelfin/train/config/qwen2vl-7b.yaml --num_proc 64 --save.path \"s3://ai2-oe-data/jakep/experiments/qwen2vl-pdf/v1/models/\${BEAKER_USER_ID}\""
EXTRA_ARGS="-c pdelfin/train/config/qwen2vl-7b.yaml --num_proc 64 --save.path \"/data/jakep/experiments/qwen2vl-pdf/v1/models/\${BEAKER_USER_ID}\""

run_name=$(basename "$0" .sh)

Expand Down

0 comments on commit 529d51d

Please sign in to comment.