-
Notifications
You must be signed in to change notification settings - Fork 15
Expand file tree
/
Copy pathdefaults.yaml
More file actions
84 lines (76 loc) · 1.47 KB
/
defaults.yaml
File metadata and controls
84 lines (76 loc) · 1.47 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
constants:
embed_dim: 1024
img_size: 224
data:
data_set: imagenet
data_path:
nb_classes: 1000
eval_crop_ratio: 0.875
generation_kwargs:
tokenizer:
validation_tokenizer:
train:
shuffle: true
dataset:
dataloader:
_target_: torch.utils.data.DataLoader
_partial_: true
batch_size: 64
num_workers: 10
pin_memory: true
drop_last: true
collator:
_target_: torch.utils.data.default_collate
_partial_: true
validation:
optim:
grad_clip:
lr_scalers:
gradient_accumulation_steps: 1
lr_decay:
freeze_list:
unfreeze_list:
probe_hparam_search: false
scheduler_type: fvcore
scheduler:
model:
modality:
checkpoint: # Add path for finetuning
checkpoint_filters:
checkpoint_renamer:
meta_model:
loss:
val_loss:
metrics:
_target_: l3m.metrics.MetricsComputer # Empty list of metrics
wandb:
use_wandb: false
wandb_config: ./.wandb.yaml
watch_freq:
project: l3m
experiment:
test_frequency: 2000
torch_compile: false
dtype: bfloat16
output_dir:
shared_dir:
device: cuda
ignore_resume_iteration: false
find_unused_parameters: false
seed: 0
ckpt_save_freq: 1000
start_iteration: 0
dist_eval: true
distributed: false
world_size: 1
dist_url: env://
nccl_timeout_mins: 30
eval: false
resume:
no_sync_gradient_accumulation: false
amp_enabled: true
fsdp:
sharding_strategy: FULL_SHARD
param_dtype: bf16
reduce_dtype: bf16
buffer_dtype: fp32