forked from facebookresearch/omnilingual-asr
-
Notifications
You must be signed in to change notification settings - Fork 0
Expand file tree
/
Copy pathctc-finetune-recommendation.yaml
More file actions
50 lines (44 loc) · 1.25 KB
/
ctc-finetune-recommendation.yaml
File metadata and controls
50 lines (44 loc) · 1.25 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
# Copyright (c) Meta Platforms, Inc. and affiliates.
# All rights reserved.
#
# This source code is licensed under the BSD-style license found in the
# LICENSE file in the root directory of this source tree.
model:
name: "omniASR_CTC_300M" # omniASR_CTC_1B, omniASR_CTC_3B
dataset:
name: "example_dataset"
train_split: "train"
valid_split: "dev"
storage_mode: "MIXTURE_PARQUET"
task_mode: "ASR"
mixture_parquet_storage_config:
dataset_summary_path: "/path/to/your/dataset/language_distribution_0.tsv"
beta_corpus: 0.5
beta_language: 0.5
fragment_loading:
cache: True
asr_task_config:
max_audio_len: 960_000
max_num_elements: 7_680_000
batch_shuffle_window: 1
normalize_audio: true
example_shuffle_window: 0 # full-batch shuffling by default
tokenizer:
name: "omniASR_tokenizer"
optimizer:
config:
lr: 1e-05
trainer:
freeze_encoder_for_n_steps: 0
mixed_precision:
dtype: "torch.bfloat16"
grad_accumulation:
num_batches: 1 # tune this parameter if you are getting OOM
regime:
num_steps: 5_000
validate_every_n_steps: 500
validate_after_n_steps: 500
checkpoint_every_n_steps: 500
checkpoint_after_n_steps: 500
publish_metrics_every_n_steps: 500
publish_metrics_after_n_steps: 500