File size: 1,651 Bytes
e26f1fa | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 | {
"model": {
"extractor_mode": "layer_norm",
"extractor_conv_bias": false,
"extractor_conv_layer_config": [
[
512,
10,
5
],
[
512,
3,
2
],
[
512,
3,
2
],
[
512,
3,
2
],
[
512,
3,
2
],
[
512,
2,
2
],
[
512,
2,
2
]
],
"encoder_embed_dim": 768,
"encoder_projection_dropout": 0,
"encoder_pos_conv_kernel": 95,
"encoder_pos_conv_groups": 16,
"encoder_pos_conv_depth": 5,
"encoder_num_layers": 12,
"encoder_num_heads": 12,
"encoder_attention_dropout": 0.1,
"encoder_ff_interm_features": 3072,
"encoder_ff_interm_dropout": 0.0,
"encoder_dropout": 0.1,
"encoder_layer_norm_first": false,
"encoder_layer_drop": 0.0,
"encoder_qkv_bias": false,
"codebook_size": 256,
"codebook_decay": 0.9,
"num_codebooks": 8,
"ema_start_decay": 0.999,
"ema_final_decay": 0.9999,
"ema_final_step": 30000,
"ema_exclude_layers": [
"pos_conv_embed"
],
"freeze_step": 200000,
"ema_timescale": 20000,
"ema_threshold": 1e-07
}
}
|