Guest User

Llama 7B RM

a guest
Oct 16th, 2023
614
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
Bash 1.43 KB | None | 0 0
  1. #!/bin/bash
  2. # Please run this script under ${project_id} in project directory of
  3. #   https://github.com/shizhediao/llm-ft
  4. #     COMMIT: d5fecf30ba8011067b10cf51fede53a5ab6574e4
  5.  
  6. deepspeed_args="--master_port=11000"      # Default argument
  7. if [ $# -ge 1 ]; then
  8.   deepspeed_args="$1"
  9. fi
  10.  
  11. exp_id=llama-7b-rm-not-formatted
  12. project_dir=$(cd "$(dirname $0)"/..; pwd)
  13. output_dir=${project_dir}/output_models/${exp_id}
  14. log_dir=${project_dir}/log/${exp_id}
  15.  
  16. dataset_path=${project_dir}/data/hh_rlhf/rm-not-formatted/hh_rlhf_rm_data.json
  17. if [ ! -d data/hh_rlhf ]; then
  18.   cd data && ./download.sh hh_rlhf && cd -
  19. fi
  20.  
  21. mkdir -p ${output_dir} ${log_dir}
  22.  
  23. deepspeed ${deepspeed_args} \
  24.   examples/reward_modeling.py \
  25.     --model_name_or_path output_models/llama-7b-sft-not-formatted \
  26.     --dataset_path ${dataset_path} \
  27.     --output_dir ${output_dir} --overwrite_output_dir \
  28.     --num_train_epochs 1 \
  29.     --learning_rate 5e-6 \
  30.     --block_size 512 \
  31.     --per_device_train_batch_size 2 \
  32.     --per_device_eval_batch_size 1 \
  33.     --deepspeed configs/ds_config_zero3.json \
  34.     --fp16 \
  35.     --run_name llama-7b-rm \
  36.     --validation_split_percentage 10 \
  37.     --logging_steps 10 \
  38.     --do_train \
  39.     --ddp_timeout 72000 \
  40.     --save_steps 999999 \
  41.     --evaluation_strategy steps \
  42.     --eval_steps 1000 \
  43.     --weight_decay 0.001 \
  44.     --dataloader_num_workers 1 \
  45.     | tee ${log_dir}/train.log \
  46.     2> ${log_dir}/train.err
Advertisement
Add Comment
Please, Sign In to add comment