File size: 1,651 Bytes
4693bac
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
{
    "model": {
        "extractor_mode": "layer_norm",
        "extractor_conv_bias": false,
        "extractor_conv_layer_config": [
            [
                512,
                10,
                5
            ],
            [
                512,
                3,
                2
            ],
            [
                512,
                3,
                2
            ],
            [
                512,
                3,
                2
            ],
            [
                512,
                3,
                2
            ],
            [
                512,
                2,
                2
            ],
            [
                512,
                2,
                2
            ]
        ],
        "encoder_embed_dim": 768,
        "encoder_projection_dropout": 0,
        "encoder_pos_conv_kernel": 95,
        "encoder_pos_conv_groups": 16,
        "encoder_pos_conv_depth": 5,
        "encoder_num_layers": 12,
        "encoder_num_heads": 12,
        "encoder_attention_dropout": 0.1,
        "encoder_ff_interm_features": 3072,
        "encoder_ff_interm_dropout": 0.0,
        "encoder_dropout": 0.1,
        "encoder_layer_norm_first": false,
        "encoder_layer_drop": 0.0,
        "encoder_qkv_bias": false,
        "codebook_size": 256,
        "codebook_decay": 0.9,
        "num_codebooks": 8,
        "ema_start_decay": 0.999,
        "ema_final_decay": 0.9999,
        "ema_final_step": 30000,
        "ema_exclude_layers": [
            "pos_conv_embed"
        ],
        "freeze_step": 200000,
        "ema_timescale": 20000,
        "ema_threshold": 1e-07
    }
}