usr256864 commited on
Commit
c2b6eb2
·
verified ·
1 Parent(s): cd1fcf7

Upload 5 files

Browse files
config_kwargs.json ADDED
@@ -0,0 +1,79 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "monitor_metric": "eval_MRR@NM_0,eval_MRR@NM_1",
3
+ "trainer_kwargs": {
4
+ "class_name": "CROSS_MODAL_IR"
5
+ },
6
+ "trainee_kwargs": {
7
+ "class_name": "CLIP_Encoder",
8
+ "freeze_prefixes": [],
9
+ "use_attention": false,
10
+ "symmetric_CL": false,
11
+ "weighted_loss": false,
12
+ "image_type": false,
13
+ "mlm_type": false,
14
+ "tie_weights": false,
15
+ "loss": {
16
+ "class_name": "NLLLoss",
17
+ "align_uniform": true
18
+ }
19
+ },
20
+ "data_module_kwargs": {
21
+ "class_name": "cross_modal_kb_DataModule_article",
22
+ "data_processor": {
23
+ "class_name": "kb_viquae_data_processor",
24
+ "dataset_path": "../../all_data/kb_dataset_new/",
25
+ "kb_path": "../../all_data/passages/",
26
+ "entity_kb_path": "../../all_data/kb/",
27
+ "validation_dataset_path": "../../all_data/"
28
+ },
29
+ "input_key": "input",
30
+ "passage_key": "passage",
31
+ "relevant_indices_key": "BM25_provenance_indices",
32
+ "irrelevant_indices_key": "BM25_irrelevant_indices",
33
+ "use_image": true,
34
+ "add_positives": false,
35
+ "cross_modal_viquae_valid": true,
36
+ "use_mep": false,
37
+ "use_mlm": false,
38
+ "use_CL": true,
39
+ "augmented": false,
40
+ "random_mask": false,
41
+ "dataloader_kwargs": {
42
+ "num_workers": 6,
43
+ "prefetch_factor": 2
44
+ },
45
+ "tokenizer_kwargs": {
46
+ "class_name": "CLIPTokenizer",
47
+ "pretrained_model_name_or_path": "clip-vit-base-patch32_tokenizer"
48
+ },
49
+ "tokenization_kwargs": {
50
+ "max_length": 77,
51
+ "padding": "longest"
52
+ },
53
+ "image_processor_kwargs": {
54
+ "class_name": "ImageFormatter",
55
+ "feature_extractor_kwargs": {
56
+ "class_name": "CLIPFeatureExtractor",
57
+ "pretrained_model_name_or_path": "clip-vit-base-patch32_FE"
58
+ }
59
+ }
60
+ },
61
+ "text_encoder_kwargs": {
62
+ "class_name": "CLIP_Text_Encoder",
63
+ "checkpoint_name": "text_encoder",
64
+ "base_encoder_kwargs": {
65
+ "class_name": "CLIPModel",
66
+ "pretrained_model_name_or_path": "/gpfswork/rech/der/ufc61ee/my_transformers_cache/clip-vit-base-patch32"
67
+ },
68
+ "inference_path": "saved_models/CLIP_article/text_encoder"
69
+ },
70
+ "image_encoder_kwargs": {
71
+ "class_name": "CLIP_Image_Encoder",
72
+ "checkpoint_name": "image_encoder",
73
+ "base_encoder_kwargs": {
74
+ "class_name": "CLIPModel",
75
+ "pretrained_model_name_or_path": "/gpfswork/rech/der/ufc61ee/my_transformers_cache/clip-vit-base-patch32"
76
+ },
77
+ "inference_path": "saved_models/CLIP_article/image_encoder"
78
+ }
79
+ }
experiment_params.json ADDED
@@ -0,0 +1,153 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "subparser": null,
3
+ "max_seq_length": 256,
4
+ "question_max_seq_length": 256,
5
+ "RC": false,
6
+ "coco": false,
7
+ "flickr": false,
8
+ "aokvqa": false,
9
+ "grad_check": true,
10
+ "MJL": false,
11
+ "debug_run_time": false,
12
+ "only_neighbors": false,
13
+ "full_kb": false,
14
+ "image_eval": false,
15
+ "embedding_column": "my_DPR_few_shot",
16
+ "output_ds": null,
17
+ "output_kb": null,
18
+ "embed": false,
19
+ "data_path": null,
20
+ "train_batch_size": 1000,
21
+ "eval_batch_size": 1000,
22
+ "viquae_dev_batch_size": 1000,
23
+ "transformer_model_name": "bert-base-uncased",
24
+ "gradient_accumulation_steps": 1,
25
+ "warmup_proportion": 0.1,
26
+ "weight_decay": 0.01,
27
+ "adam_beta1": 0.9,
28
+ "adam_beta2": 0.999,
29
+ "warmup_steps": 4,
30
+ "adam_epsilon": 1e-08,
31
+ "num_train_epochs": 10,
32
+ "learning_rate": 2e-06,
33
+ "dropout": 0.5,
34
+ "bert_hidden_size": 768,
35
+ "MM": null,
36
+ "use_graph_P": null,
37
+ "use_graph_Q": null,
38
+ "syntactic_P": null,
39
+ "syntactic_Q": null,
40
+ "NS": null,
41
+ "IC_Q": false,
42
+ "IC_P": false,
43
+ "gcn_lr": 2e-05,
44
+ "head_lr": 2e-06,
45
+ "func": "cat",
46
+ "graph_pooling": "mean",
47
+ "max_nbr_nodes": 2,
48
+ "num_neib": 1,
49
+ "neib_depth": 1,
50
+ "undirect": false,
51
+ "node_hidden_size": 128,
52
+ "max_num_relations": null,
53
+ "max_num_syntactic_relations": null,
54
+ "map_wikidataRelCode_to_idx": null,
55
+ "graph_layout": "fr",
56
+ "residual": false,
57
+ "num_gcn": null,
58
+ "node_alias_max_len": 32,
59
+ "attention": false,
60
+ "num_paths": 1,
61
+ "kg_embed": false,
62
+ "graph_only": false,
63
+ "freeze": false,
64
+ "sanity_run": false,
65
+ "enhanced_question": false,
66
+ "use_entity_type": false,
67
+ "filter_edges": false,
68
+ "draw": false,
69
+ "layer_norm": false,
70
+ "filters_3": 768,
71
+ "tensor_neurons": 16,
72
+ "output_dir": "saved_models/CLIP_article",
73
+ "main_dir": "/gpfsdswork/projects/rech/der/ufc61ee/MEERQAT/ViQuAE/meerqat/GP-VQA",
74
+ "experiment_dir": "None",
75
+ "experiment_name": "CLIP_article",
76
+ "search": null,
77
+ "xlnet": false,
78
+ "tune_loss": false,
79
+ "tune_dev": true,
80
+ "tune_valid_loss": false,
81
+ "sanity_val_steps": 0,
82
+ "cpu": false,
83
+ "keep_in_memory": false,
84
+ "N": 0,
85
+ "nbr_workers": 1,
86
+ "num_proc": 1,
87
+ "local_cache": "/gpfswork/rech/der/ufc61ee/my_transformers_cache",
88
+ "IMAGE_PATH": null,
89
+ "transformer_path": null,
90
+ "resume_from": null,
91
+ "checkpoint": null,
92
+ "linear_MEP_checkpoint": null,
93
+ "linear_MLM_checkpoint": null,
94
+ "config": "experiments/ir/viquae_kb_full/clip/config_clip_article.json",
95
+ "context_checkpoint": null,
96
+ "grad_accum": 1,
97
+ "cls_token": "[CLS]",
98
+ "sep_token": "[SEP]",
99
+ "pad_token": "[PAD]",
100
+ "cls_token_at_end": false,
101
+ "mask_padding_with_zero": true,
102
+ "pad_on_left": false,
103
+ "sequence_a_segment_id": 0,
104
+ "pad_token_segment_id": 0,
105
+ "cls_token_segment_id": 0,
106
+ "pad_token_label_id": 0,
107
+ "pad_token_id": 0,
108
+ "past_index": -1,
109
+ "world_size": 1,
110
+ "use_lstm": false,
111
+ "use_entity": false,
112
+ "use_question_graph": false,
113
+ "use_question_objects": false,
114
+ "EFeat": null,
115
+ "NFeat": null,
116
+ "ltn": false,
117
+ "use_kelm": false,
118
+ "unshared": false,
119
+ "mlm": false,
120
+ "test": false,
121
+ "split_dpr": false,
122
+ "rename_model": false,
123
+ "mask_rate": 0.15,
124
+ "pretrained_triviaq": false,
125
+ "error_analysis": false,
126
+ "debug": false,
127
+ "stop_debug": false,
128
+ "entity_linking": false,
129
+ "entity_linking_split": false,
130
+ "build_entity_paths": null,
131
+ "object_detection": null,
132
+ "update_wikidata_ids": null,
133
+ "update_None_ids": null,
134
+ "dependency_parsing": null,
135
+ "image_captionning": null,
136
+ "build_graph": false,
137
+ "from_bert": false,
138
+ "fast": false,
139
+ "get_examples": false,
140
+ "get_statistics": null,
141
+ "word_vocab_size": 0,
142
+ "char_vocab_size": 0,
143
+ "max_word_len": 30,
144
+ "device": null,
145
+ "embedding_matrix": null,
146
+ "w2v_file": "word_vector_200d.vec",
147
+ "word_emb_dim": 200,
148
+ "char_lstm": false,
149
+ "char_cnn": false,
150
+ "seed": 32,
151
+ "tune_batch_size": false,
152
+ "model_names": null
153
+ }
model-epoch=05-eval_MRR@NM_0=0.4627_sd=32.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c404255c2617efb880beb7494483787beae15d16948ff7e3c89746a7ca77eac4
3
+ size 1815837916
results.txt ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ -------- dev---------
2
+ [0.46274155378341675]
validation_metrics.txt ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ [{'MRR@NM': 0.3684380584589463, 'hits@1': 0.27678830164285234, 'avg_L_align': tensor(1.4444), 'avg_L_uniform': tensor(-2.1290)}, {'MRR@NM': 0.27367666416643144, 'hits@1': 0.18372319688109162}]
2
+ [{'MRR@NM': 0.4243174365696836, 'hits@1': 0.3220614787900403, 'avg_L_align': tensor(1.4181), 'avg_L_uniform': tensor(-1.9166)}, {'MRR@NM': 0.32139479322781445, 'hits@1': 0.2309941520467836}]
3
+ [{'MRR@NM': 0.4421935825064815, 'hits@1': 0.3378881433762065, 'avg_L_align': tensor(1.4131), 'avg_L_uniform': tensor(-1.9543)}, {'MRR@NM': 0.32120630824863855, 'hits@1': 0.22514619883040934}]
4
+ [{'MRR@NM': 0.4545746362170726, 'hits@1': 0.3494795033548071, 'avg_L_align': tensor(1.4129), 'avg_L_uniform': tensor(-2.0277)}, {'MRR@NM': 0.3198178910308339, 'hits@1': 0.22660818713450293}]
5
+ [{'MRR@NM': 0.46088537715427935, 'hits@1': 0.35416062950001115, 'avg_L_align': tensor(1.4164), 'avg_L_uniform': tensor(-2.1373)}, {'MRR@NM': 0.30702571062989, 'hits@1': 0.2149122807017544}]
6
+ [{'MRR@NM': 0.46274156319615783, 'hits@1': 0.35592162457368315, 'avg_L_align': tensor(1.4235), 'avg_L_uniform': tensor(-2.2792)}, {'MRR@NM': 0.3126044940448983, 'hits@1': 0.2202729044834308}]
7
+ [{'MRR@NM': 0.46134979593148656, 'hits@1': 0.3537593901161365, 'avg_L_align': tensor(1.4337), 'avg_L_uniform': tensor(-2.4167)}, {'MRR@NM': 0.3118104677421376, 'hits@1': 0.21978557504873295}]
8
+ [{'MRR@NM': 0.45893284627342656, 'hits@1': 0.3517086110429995, 'avg_L_align': tensor(1.4450), 'avg_L_uniform': tensor(-2.5418)}, {'MRR@NM': 0.3164403163511111, 'hits@1': 0.22319688109161792}]
9
+ [{'MRR@NM': 0.456976254547464, 'hits@1': 0.3501705267381467, 'avg_L_align': tensor(1.4546), 'avg_L_uniform': tensor(-2.6405)}, {'MRR@NM': 0.3106033856166459, 'hits@1': 0.22076023391812866}]
10
+ [{'MRR@NM': 0.45557986387468385, 'hits@1': 0.3493234658166336, 'avg_L_align': tensor(1.4613), 'avg_L_uniform': tensor(-2.7086)}, {'MRR@NM': 0.31200438202869274, 'hits@1': 0.2168615984405458}]