Skip to content

Commit 77bbf65

Browse files
authored
disable valset by default (hiyouga#6690)
1 parent 4d0f662 commit 77bbf65

30 files changed

+142
-114
lines changed

examples/extras/adam_mini/qwen2_full_sft.yaml

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -34,7 +34,7 @@ bf16: true
3434
ddp_timeout: 180000000
3535

3636
### eval
37-
val_size: 0.1
38-
per_device_eval_batch_size: 1
39-
eval_strategy: steps
40-
eval_steps: 500
37+
# val_size: 0.1
38+
# per_device_eval_batch_size: 1
39+
# eval_strategy: steps
40+
# eval_steps: 500

examples/extras/apollo/llama3_full_sft.yaml

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -39,7 +39,7 @@ pure_bf16: true
3939
ddp_timeout: 180000000
4040

4141
### eval
42-
val_size: 0.1
43-
per_device_eval_batch_size: 1
44-
eval_strategy: steps
45-
eval_steps: 500
42+
# val_size: 0.1
43+
# per_device_eval_batch_size: 1
44+
# eval_strategy: steps
45+
# eval_steps: 500

examples/extras/badam/llama3_full_sft.yaml

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -37,7 +37,7 @@ lr_scheduler_type: cosine
3737
warmup_ratio: 0.1
3838

3939
### eval
40-
val_size: 0.1
41-
per_device_eval_batch_size: 1
42-
eval_strategy: steps
43-
eval_steps: 500
40+
# val_size: 0.1
41+
# per_device_eval_batch_size: 1
42+
# eval_strategy: steps
43+
# eval_steps: 500

examples/extras/fsdp_qlora/llama3_lora_sft.yaml

Lines changed: 5 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -7,6 +7,7 @@ trust_remote_code: true
77
stage: sft
88
do_train: true
99
finetuning_type: lora
10+
lora_rank: 8
1011
lora_target: all
1112

1213
### dataset
@@ -35,7 +36,7 @@ bf16: true
3536
ddp_timeout: 180000000
3637

3738
### eval
38-
val_size: 0.1
39-
per_device_eval_batch_size: 1
40-
eval_strategy: steps
41-
eval_steps: 500
39+
# val_size: 0.1
40+
# per_device_eval_batch_size: 1
41+
# eval_strategy: steps
42+
# eval_steps: 500

examples/extras/galore/llama3_full_sft.yaml

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -38,7 +38,7 @@ pure_bf16: true
3838
ddp_timeout: 180000000
3939

4040
### eval
41-
val_size: 0.1
42-
per_device_eval_batch_size: 1
43-
eval_strategy: steps
44-
eval_steps: 500
41+
# val_size: 0.1
42+
# per_device_eval_batch_size: 1
43+
# eval_strategy: steps
44+
# eval_steps: 500

examples/extras/llama_pro/llama3_freeze_sft.yaml

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -36,7 +36,7 @@ bf16: true
3636
ddp_timeout: 180000000
3737

3838
### eval
39-
val_size: 0.1
40-
per_device_eval_batch_size: 1
41-
eval_strategy: steps
42-
eval_steps: 500
39+
# val_size: 0.1
40+
# per_device_eval_batch_size: 1
41+
# eval_strategy: steps
42+
# eval_steps: 500

examples/extras/loraplus/llama3_lora_sft.yaml

Lines changed: 5 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -6,6 +6,7 @@ trust_remote_code: true
66
stage: sft
77
do_train: true
88
finetuning_type: lora
9+
lora_rank: 8
910
lora_target: all
1011
loraplus_lr_ratio: 16.0
1112

@@ -35,7 +36,7 @@ bf16: true
3536
ddp_timeout: 180000000
3637

3738
### eval
38-
val_size: 0.1
39-
per_device_eval_batch_size: 1
40-
eval_strategy: steps
41-
eval_steps: 500
39+
# val_size: 0.1
40+
# per_device_eval_batch_size: 1
41+
# eval_strategy: steps
42+
# eval_steps: 500

examples/extras/mod/llama3_full_sft.yaml

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -35,7 +35,7 @@ pure_bf16: true
3535
ddp_timeout: 180000000
3636

3737
### eval
38-
val_size: 0.1
39-
per_device_eval_batch_size: 1
40-
eval_strategy: steps
41-
eval_steps: 500
38+
# val_size: 0.1
39+
# per_device_eval_batch_size: 1
40+
# eval_strategy: steps
41+
# eval_steps: 500

examples/extras/pissa/llama3_lora_sft.yaml

Lines changed: 5 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -6,6 +6,7 @@ trust_remote_code: true
66
stage: sft
77
do_train: true
88
finetuning_type: lora
9+
lora_rank: 8
910
lora_target: all
1011
pissa_init: true
1112
pissa_iter: 16
@@ -37,7 +38,7 @@ bf16: true
3738
ddp_timeout: 180000000
3839

3940
### eval
40-
val_size: 0.1
41-
per_device_eval_batch_size: 1
42-
eval_strategy: steps
43-
eval_steps: 500
41+
# val_size: 0.1
42+
# per_device_eval_batch_size: 1
43+
# eval_strategy: steps
44+
# eval_steps: 500

examples/train_full/llama3_full_sft.yaml

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -34,7 +34,7 @@ bf16: true
3434
ddp_timeout: 180000000
3535

3636
### eval
37-
val_size: 0.1
38-
per_device_eval_batch_size: 1
39-
eval_strategy: steps
40-
eval_steps: 500
37+
# val_size: 0.1
38+
# per_device_eval_batch_size: 1
39+
# eval_strategy: steps
40+
# eval_steps: 500

0 commit comments

Comments
 (0)