File size: 4,509 Bytes
d1303b6 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 | {
"datasets": {
"assemble_data": {
"data_features": "src_pc,tgt_pc,imgs,lang,asset,category,pre_pose",
"data_mix": "official_twobytwo_pose_qa+official_twobytwo_pose_instruction_qa",
"dataset_py": "hdf5datasets",
"default_point_end_token": "<point_end>",
"default_point_patch_token": "<point>",
"default_point_start_token": "<point_start>",
"drop_last": true,
"eval_with_val": true,
"import_rotation": true,
"import_translation": false,
"label_pad_token_id": -100,
"num_samples": 100,
"num_workers": 0,
"per_device_batch_size": 4,
"point_token_len": 256,
"shuffle": true,
"split_train_val": true,
"test_mix": "twobytwo_final_hdf5",
"train_mix": "biassembly_final_hdf5+partnet_final_hdf5+partnext_final_hdf5+twobytwo_final_hdf5"
},
"point_token_len": 256
},
"framework": {
"action_model": {
"action_dim": 7,
"action_hidden_dim": 1024,
"action_horizon": 1,
"add_pos_embed": true,
"diffusion_model_cfg": {
"cross_attention_dim": 2048,
"dropout": 0.2,
"final_dropout": true,
"interleave_self_attention": true,
"norm_type": "ada_norm",
"num_layers": 16,
"output_dim": 1024,
"positional_embeddings": null
},
"expected_len": 20,
"fast_tokenizer_num_samples": 100000,
"future_action_window_size": 7,
"hidden_size": 1024,
"max_seq_len": 1024,
"model_path": "/gemini/space/users/jingzhi/project/AssemPVLM-main/starVLA/results/train_vndgcnn_jingzhi15/fast_tokenizer_trained",
"noise_beta_alpha": 1.5,
"noise_beta_beta": 1.0,
"noise_s": 0.999,
"num_inference_timesteps": 4,
"num_target_vision_tokens": 32,
"num_timestep_buckets": 1000,
"pad_token_id": 2047,
"past_action_window_size": 0,
"repeated_diffusion_steps": 8,
"scale": 900,
"state_dim": 7,
"tokenizer_vocab_size": 201,
"type": "linear",
"use_expected_len": false,
"vocab_size": 2048
},
"assemlm_py": "assemlm",
"framework_py": "PVLA_Fast",
"name": "AssemLM",
"point_encoder": {
"model_path": "/gemini/space/users/jingzhi/project/AssemPVLM-main/assempvlm/models_local/pretrain_vndgcnn/1000-network_A.pth",
"pc_feat_dim": 512,
"resume_from_checkpoint": false,
"type": "vn_dgcnn"
},
"point_projector": {
"backbone_output_dim": 12,
"project_output_dim": 2048,
"projection_hidden_dim": [
1024,
2048
],
"projection_hidden_layer": 2
},
"use_category_in_instruction": true,
"vlm": {
"base_vlm": "vlm",
"build_version": 0,
"local_files_only": true,
"saved_version": 1
}
},
"is_debug": false,
"model_type": "assemlm_hf",
"run_id": "assemlm21",
"run_root_dir": "./results",
"seed": 42,
"trackers": [
"jsonl",
"wandb"
],
"trainer": {
"enable_gradient_checkpointing": true,
"enable_mixed_precision_training": true,
"epochs": 1,
"eval_end_step": 1000000,
"eval_interval": 200,
"eval_start_step": 0,
"freeze_modules": null,
"gen_do_sample": false,
"gen_temperature": 1.0,
"gradient_accumulation_steps": 1,
"gradient_clipping": 1.0,
"is_resume": true,
"learning_rate": {
"action_model": 0.0001,
"base": 1e-05,
"pvlm_interface": 5e-06
},
"logging_frequency": 10,
"lr_scheduler_type": "cosine_with_min_lr",
"max_gen_tokens": 50,
"max_grad_norm": 1.0,
"max_train_steps": 1000000,
"num_warmup_steps": 50000,
"optimizer": {
"betas": [
0.9,
0.95
],
"eps": 1e-08,
"name": "AdamW",
"weight_decay": 1e-08
},
"pretrained_checkpoint": "/gemini/space/users/jingzhi/project/AssemPVLM-main/starVLA/results/train_vndgcnn_jingzhi19/checkpoints/latest/pytorch_model/mp_rank_00_model_states.pt",
"resume_epoch": null,
"resume_from_checkpoint": "/gemini/space/users/jingzhi/project/AssemPVLM-main/starVLA/results/train_vndgcnn_jingzhi19/checkpoints/latest",
"resume_step": null,
"save_end_step": 1000000,
"save_interval": 200,
"save_start_step": 0,
"scheduler_specific_kwargs": {
"min_lr": 5e-07
},
"train_fast_tokenizer": false,
"vlm_max_length": 1800,
"vlm_padding": true,
"warmup_ratio": 0.1,
"weight_decay": 0.0
},
"transformers_version": "4.57.0"
}
|