| model: CTTransformer |
| model_conf: |
| ignore_id: 0 |
| embed_unit: 516 |
| att_unit: 516 |
| dropout_rate: 0.1 |
| punc_list: |
| - <unk> |
| - _ |
| - , |
| - 。 |
| - ? |
| - 、 |
| punc_weight: |
| - 1.0 |
| - 1.0 |
| - 1.0 |
| - 1.0 |
| - 1.0 |
| - 1.0 |
| sentence_end_id: 3 |
|
|
| encoder: SANMEncoder |
| encoder_conf: |
| input_size: 516 |
| output_size: 516 |
| attention_heads: 12 |
| linear_units: 2048 |
| num_blocks: 12 |
| dropout_rate: 0.1 |
| positional_dropout_rate: 0.1 |
| attention_dropout_rate: 0.0 |
| input_layer: pe |
| pos_enc_class: SinusoidalPositionEncoder |
| normalize_before: true |
| kernel_size: 11 |
| sanm_shfit: 0 |
| selfattention_layer_type: sanm |
| padding_idx: 0 |
|
|
| tokenizer: CharTokenizer |
| tokenizer_conf: |
| unk_symbol: <unk> |
|
|
|
|
|
|
|
|