-
Notifications
You must be signed in to change notification settings - Fork 8
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
Merge pull request #7 from valence-labs/slurm
Slurm
- Loading branch information
Showing
4 changed files
with
24 additions
and
6 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,10 @@ | ||
accelerate launch --config_file config/accelerate.yaml \ | ||
scripts/model_trainer.py --tokenizer "tokenizer/tokenizer-custom.json" \ | ||
--dataset data/ --text_column "input" \ | ||
--is_tokenized False --streaming True \ | ||
--num_labels 1 --include_descriptors False \ | ||
--gradient_accumulation_steps 2 --wandb_watch 'gradients' \ | ||
--per_device_train_batch_size 32 --num_train_epochs 5 --save_steps 2000 --save_total_limit 10 \ | ||
--eval_accumulation_steps 100 --logging_steps 200 --logging_first_step True \ | ||
--save_safetensors True --do_train True --output_dir output/test/ \ | ||
--learning_rate 5e-4 --warmup_steps 1000 --gradient_checkpointing True --max_steps 15_000 |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -1,10 +1,10 @@ | ||
accelerate launch --config_file config/accelerate.yaml \ | ||
scripts/model_trainer.py --tokenizer "tokenizer/tokenizer-custom.json" \ | ||
--dataset data/ --text_column "input" \ | ||
--dataset ~/data/ --text_column "input" \ | ||
--is_tokenized False --streaming True \ | ||
--num_labels 1 --include_descriptors False \ | ||
--gradient_accumulation_steps 2 --wandb_watch 'gradients' \ | ||
--per_device_train_batch_size 32 --num_train_epochs 5 --save_steps 2000 --save_total_limit 10 \ | ||
--eval_accumulation_steps 100 --logging_steps 200 --logging_first_step True \ | ||
--save_safetensors True --do_train True --output_dir output/test/ \ | ||
--learning_rate 5e-4 --warmup_steps 500 --gradient_checkpointing True --max_steps 15000 | ||
--per_device_train_batch_size 64 --num_train_epochs 2 --save_steps 5000 --save_total_limit 10 \ | ||
--eval_accumulation_steps 100 --logging_steps 500 --logging_first_step True \ | ||
--save_safetensors True --do_train True --output_dir output/safe/ \ | ||
--learning_rate 5e-5 --warmup_steps 2500 --gradient_checkpointing True --max_steps 30_000_000 |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters