add checkpoints
Browse files- pretrained_models/.DS_Store +0 -0
- pretrained_models/autoenc/autoencoder_config.yaml +45 -0
- pretrained_models/autoenc/checkpoint.pth +3 -0
- pretrained_models/checkerboard-L-2x/checkpoint-last.pth +3 -0
- pretrained_models/checkerboard-L-2x/config.yaml +37 -0
- pretrained_models/checkerboard-L-4x/checkpoint-last.pth +3 -0
- pretrained_models/checkerboard-L-4x/config.yaml +37 -0
pretrained_models/.DS_Store
ADDED
|
Binary file (6.15 kB). View file
|
|
|
pretrained_models/autoenc/autoencoder_config.yaml
ADDED
|
@@ -0,0 +1,45 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
model:
|
| 2 |
+
base_learning_rate: 0.0001
|
| 3 |
+
scale_lr: false
|
| 4 |
+
target: cbgen.autoencoder.VQModel
|
| 5 |
+
init:
|
| 6 |
+
checkpoint: pretrained_models/llamagen/vq_ds16_c2i.pt
|
| 7 |
+
ignore_keys: ["quantize."]
|
| 8 |
+
freeze_encoder: true
|
| 9 |
+
freeze_decoder: true
|
| 10 |
+
freeze_decoder_last_layer: true
|
| 11 |
+
freeze_quant_convs: true
|
| 12 |
+
params:
|
| 13 |
+
monitor: "val/qloss"
|
| 14 |
+
embed_dim: 8
|
| 15 |
+
n_embed: 4096
|
| 16 |
+
#n_embed: 16384
|
| 17 |
+
quantize_prob: 1.0
|
| 18 |
+
p_train_random_scale: 0.9
|
| 19 |
+
normalize: true
|
| 20 |
+
quantize_iter_start: 1000
|
| 21 |
+
requant_pyramid_levels: 1 # -1 means all levels, 1 means only the top level
|
| 22 |
+
resolution: 256
|
| 23 |
+
# uses defaults for llg_vq_model class defs
|
| 24 |
+
ddconfig: {}
|
| 25 |
+
|
| 26 |
+
data:
|
| 27 |
+
target: main_train_ae.DataModuleFromConfig
|
| 28 |
+
params:
|
| 29 |
+
num_workers: 32
|
| 30 |
+
batch_size: 32
|
| 31 |
+
wrap: true
|
| 32 |
+
train:
|
| 33 |
+
target: arexper.ldm_dataset.DatasetTrain
|
| 34 |
+
params:
|
| 35 |
+
train_dir: /datasets/imagenet/ILSVRC/Data/CLS-LOC/train
|
| 36 |
+
dataset_name: imagenet
|
| 37 |
+
size: 256
|
| 38 |
+
degradation: pil_nearest
|
| 39 |
+
validation:
|
| 40 |
+
target: arexper.ldm_dataset.DatasetVal
|
| 41 |
+
params:
|
| 42 |
+
val_dir: /datasets/imagenet/ILSVRC/Data/CLS-LOC/val
|
| 43 |
+
dataset_name: imagenet
|
| 44 |
+
size: 256
|
| 45 |
+
degradation: pil_nearest
|
pretrained_models/autoenc/checkpoint.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7503e2afd945546a0ba77cbb863227e8e5b1e1498190725c9358c6fed5720afa
|
| 3 |
+
size 287552963
|
pretrained_models/checkerboard-L-2x/checkpoint-last.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d0b1c226a3b6bc84827855d2228bc21642774d99145d5978636f2a1e3553f297
|
| 3 |
+
size 4698506735
|
pretrained_models/checkerboard-L-2x/config.yaml
ADDED
|
@@ -0,0 +1,37 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
model:
|
| 2 |
+
img_size: 256
|
| 3 |
+
scale_ratio: 2
|
| 4 |
+
class_cond: true
|
| 5 |
+
num_classes: 1000
|
| 6 |
+
random_block_size: true
|
| 7 |
+
random_replace: 0.1
|
| 8 |
+
random_replace_sampled: false # "sampled" gt or just upscaled for random replace
|
| 9 |
+
use_head_mlp: false
|
| 10 |
+
num_mixing_layers: 2
|
| 11 |
+
autoenc:
|
| 12 |
+
config_path: pretrained_models/autoenc/autoencoder_config.yaml
|
| 13 |
+
checkpoint_path: pretrained_models/autoenc/checkpoint.pth
|
| 14 |
+
params:
|
| 15 |
+
requant_pyramid_levels: -1 # needed for loading pretrained from config
|
| 16 |
+
|
| 17 |
+
llama_config:
|
| 18 |
+
hidden_size: 1024
|
| 19 |
+
intermediate_size: 3072
|
| 20 |
+
num_attention_heads: 16
|
| 21 |
+
num_hidden_layers: 20
|
| 22 |
+
|
| 23 |
+
args:
|
| 24 |
+
batch_size: 64
|
| 25 |
+
lr: 5.e-5
|
| 26 |
+
lr_step_factor: 0.2
|
| 27 |
+
warmup_epochs: 1
|
| 28 |
+
optimizer: adamw
|
| 29 |
+
weight_decay: 0.01
|
| 30 |
+
num_workers: 8
|
| 31 |
+
data_path: data/imagenet
|
| 32 |
+
lr_schedule: step
|
| 33 |
+
epochs: 205
|
| 34 |
+
lr_step_epochs: [200]
|
| 35 |
+
eval_freq: 5
|
| 36 |
+
cfg: 2.0
|
| 37 |
+
cfg_start_step: 5
|
pretrained_models/checkerboard-L-4x/checkpoint-last.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e359552cb2b1d7ace912d0aaac2450e80680e2635b9cb537442aba3823692fa9
|
| 3 |
+
size 4697565524
|
pretrained_models/checkerboard-L-4x/config.yaml
ADDED
|
@@ -0,0 +1,37 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
model:
|
| 2 |
+
img_size: 256
|
| 3 |
+
scale_ratio: 4
|
| 4 |
+
class_cond: true
|
| 5 |
+
num_classes: 1000
|
| 6 |
+
random_block_size: true
|
| 7 |
+
random_replace: 0.1
|
| 8 |
+
random_replace_sampled: false # "sampled" gt or just upscaled for random replace
|
| 9 |
+
use_head_mlp: false
|
| 10 |
+
num_mixing_layers: 2
|
| 11 |
+
autoenc:
|
| 12 |
+
config_path: pretrained_models/autoenc/autoencoder_config.yaml
|
| 13 |
+
checkpoint_path: pretrained_models/autoenc/checkpoint.pth
|
| 14 |
+
params:
|
| 15 |
+
requant_pyramid_levels: -1 # needed for loading pretrained from config
|
| 16 |
+
|
| 17 |
+
llama_config:
|
| 18 |
+
hidden_size: 1024
|
| 19 |
+
intermediate_size: 3072
|
| 20 |
+
num_attention_heads: 16
|
| 21 |
+
num_hidden_layers: 20
|
| 22 |
+
|
| 23 |
+
args:
|
| 24 |
+
batch_size: 64
|
| 25 |
+
lr: 5.e-5
|
| 26 |
+
lr_step_factor: 0.2
|
| 27 |
+
warmup_epochs: 1
|
| 28 |
+
optimizer: adamw
|
| 29 |
+
weight_decay: 0.01
|
| 30 |
+
num_workers: 8
|
| 31 |
+
data_path: data/imagenet
|
| 32 |
+
lr_schedule: step
|
| 33 |
+
epochs: 205
|
| 34 |
+
lr_step_epochs: [200]
|
| 35 |
+
eval_freq: 5
|
| 36 |
+
cfg: 2.0
|
| 37 |
+
cfg_start_step: 5
|