File size: 2,388 Bytes
b760891
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
accum_grad: 2
cmvn: global_cmvn
cmvn_conf:
  cmvn_file: global_cmvn
  is_json_cmvn: true
context_module: cppn
context_module_conf:
  attention_heads: 4
  dropout_rate: 0.1
  embedding_size: 768
  encoder_layers: 2
ctc: ctc
ctc_conf:
  ctc_blank_id: 0
dataset: asr
dataset_conf:
  batch_conf:
    batch_size: 128
    batch_type: static
  context_conf:
    batch_num_context: 600
    len_max: 5
    len_min: 2
    mode: train
    utt_num_context: 5
  ctc_label: true
  cycle: 100
  fbank_conf:
    dither: 0.1
    frame_length: 25
    frame_shift: 10
    num_mel_bins: 80
  filter_conf:
    max_length: 3000
    min_length: 0
    token_max_length: 200
    token_min_length: 1
  no_time_idx: 3
  remove_punctuation: true
  remove_timestamp: true
  resample_conf:
    resample_rate: 16000
  shuffle: true
  shuffle_conf:
    shuffle_size: 512
  sort: true
  sort_conf:
    sort_size: 512
  spec_aug: true
  spec_aug_conf:
    max_f: 10
    max_t: 50
    num_f_mask: 2
    num_t_mask: 2
  speed_perturb: true
  time_apply_prob: 0.0
decoder: transformer
decoder_conf:
  attention_heads: 12
  dropout_rate: 0.1
  linear_units: 3072
  num_blocks: 12
  positional_dropout_rate: 0.1
  self_attention_dropout_rate: 0.1
  src_attention_dropout_rate: 0.1
  use_sdpa: true
dtype: fp32
encoder: e_branchformer
encoder_conf:
  activation_type: swish
  attention_dropout_rate: 0.1
  attention_heads: 12
  causal: false
  cgmlp_conv_kernel: 31
  cgmlp_linear_units: 3072
  dropout_rate: 0.1
  gate_activation: identity
  input_layer: conv2d
  linear_units: 3072
  merge_conv_kernel: 31
  num_blocks: 12
  output_size: 768
  pos_enc_layer_type: rel_pos
  positional_dropout_rate: 0.1
  selfattention_layer_type: rel_selfattn
  use_linear_after_conv: false
  use_sdpa: true
grad_clip: 5
input_dim: 80
log_interval: 10
max_epoch: 100
model: asr_model
model_conf:
  ctc_weight: 0.3
  length_normalized_loss: false
  lsm_weight: 0.1
model_dir: exp/nonstream_base_avg_40_step_353000.pt_sampleone
optim: adam
optim_conf:
  lr: 0.0005
output_dim: 18173
save_interval: 200
save_states: model_only
scheduler: warmuplr
scheduler_conf:
  warmup_steps: 2048
stats_dialect: true
tokenizer: char
tokenizer_conf:
  special_tokens:
    <asr>: 4
    <blank>: 0
    <eos>: 3
    <sos>: 2
    <unk>: 1
  split_with_space: false
  symbol_table_path: data/dict/units.txt
train_engine: torch_ddp
use_amp: false
vocab_size: 18173