usr256864 commited on
Commit
6df4d35
·
verified ·
1 Parent(s): 0952acb

Upload folder using huggingface_hub

Browse files
args.txt ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ --train_batch_size 1
2
+ --eval_batch_size 1
3
+ --num_train_epochs 20
4
+ --nbr_gpus 1
5
+ --learning_rate 5e-6
6
+ --grad_check
7
+ --config experiments/event_extraction/config_code_generator.json
8
+ --non_strict_load
9
+ --ds ../datasets/textee/wikievents
10
+ --answer_generator HiTZ/GoLLIE-7B
11
+ --guidelines new_Llama-31-8B-Instruct_annotation_guidelines
12
+ --k_schemes 10
13
+ --num_passages 5
config_kwargs.json ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "monitor_metric": "mean_f1",
3
+ "trainer_kwargs": {
4
+ "class_name": "Event_Extractor_Code"
5
+ },
6
+ "trainee_kwargs": {
7
+ "class_name": "EE",
8
+ "prompt_cls": "event_extraction_code",
9
+ "loss": {
10
+ "class_name": "NLLLoss"
11
+ },
12
+ "freeze_prefixes": []
13
+ },
14
+ "data_module_kwargs": {
15
+ "class_name": "event_extraction_DataModule",
16
+ "data_processor": {
17
+ "class_name": "viquae_data_processor",
18
+ "train_path": "viquae/train/",
19
+ "val_path": "viquae/val/",
20
+ "test_path": "viquae/test/"
21
+ },
22
+ "dataloader_kwargs": {
23
+ "num_workers": 6,
24
+ "prefetch_factor": 2
25
+ }
26
+ },
27
+ "answer_generator_kwargs": {
28
+ "dividable": true,
29
+ "checkpoint_name": "answer_generator",
30
+ "rag_training": false,
31
+ "rag": false,
32
+ "wo_kb": true,
33
+ "generator_only": true,
34
+ "entity_prompt": false,
35
+ "beam_search": true,
36
+ "best_answer": false,
37
+ "best_entity": false,
38
+ "inference_path": "saved_models/wikievents/GoLLIE-7B/code_generator/train/2025-09-26/19-34-03/answer_generator"
39
+ },
40
+ "all_datasets_path": "../datasets"
41
+ }
experiment_params.json ADDED
@@ -0,0 +1,248 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "max_seq_length": 256,
3
+ "question_max_seq_length": 256,
4
+ "RC": false,
5
+ "coco": false,
6
+ "flickr": false,
7
+ "aokvqa": false,
8
+ "grad_check": true,
9
+ "MJL": false,
10
+ "debug_run_time": false,
11
+ "debug_output": false,
12
+ "only_neighbors": false,
13
+ "full_kb": false,
14
+ "image_eval": false,
15
+ "embedding_column": "my_DPR_few_shot",
16
+ "output_ds": null,
17
+ "output_kb": null,
18
+ "embed": false,
19
+ "data_path": null,
20
+ "train_batch_size": 1,
21
+ "eval_batch_size": 1,
22
+ "viquae_dev_batch_size": 1000,
23
+ "transformer_model_name": "bert-base-uncased",
24
+ "gradient_accumulation_steps": 1,
25
+ "warmup_proportion": 0.1,
26
+ "weight_decay": 0.01,
27
+ "lr_decay": false,
28
+ "adam_beta1": 0.9,
29
+ "adam_beta2": 0.999,
30
+ "warmup_steps": 4,
31
+ "adam_epsilon": 1e-08,
32
+ "num_train_epochs": 20,
33
+ "learning_rate": 5e-06,
34
+ "dropout": 0.5,
35
+ "bert_hidden_size": 768,
36
+ "MM": null,
37
+ "use_graph_P": null,
38
+ "use_graph_Q": null,
39
+ "syntactic_P": null,
40
+ "syntactic_Q": null,
41
+ "NS": false,
42
+ "IC_Q": false,
43
+ "IC_P": false,
44
+ "gcn_lr": 2e-05,
45
+ "clip_lr": null,
46
+ "dpr_lr": null,
47
+ "head_lr": 2e-06,
48
+ "func": "cat",
49
+ "graph_pooling": "mean",
50
+ "max_nbr_nodes": 2,
51
+ "num_neib": 1,
52
+ "neib_depth": 1,
53
+ "undirect": false,
54
+ "node_hidden_size": 128,
55
+ "max_num_relations": null,
56
+ "max_num_syntactic_relations": null,
57
+ "map_wikidataRelCode_to_idx": null,
58
+ "graph_layout": "fr",
59
+ "residual": false,
60
+ "num_gcn": null,
61
+ "node_alias_max_len": 32,
62
+ "attention": false,
63
+ "num_paths": 1,
64
+ "kg_embed": false,
65
+ "graph_only": false,
66
+ "freeze": false,
67
+ "sanity_run": false,
68
+ "scale_losses": false,
69
+ "check_answers": false,
70
+ "load_4bit": false,
71
+ "load_8bit": false,
72
+ "device_id": null,
73
+ "drop_last": false,
74
+ "disable_checkpointing": false,
75
+ "samples": null,
76
+ "sample_id": null,
77
+ "datasets_root_dir": "../datasets",
78
+ "ds": "../datasets/textee/wikievents",
79
+ "supp_train_ds": null,
80
+ "guidelines": "new_Llama-31-8B-Instruct_annotation_guidelines",
81
+ "num_trainable_params": null,
82
+ "num_all_param": null,
83
+ "trainable_ratio": null,
84
+ "train_sanity_run": false,
85
+ "enhanced_question": false,
86
+ "use_entity_type": false,
87
+ "filter_edges": false,
88
+ "draw": false,
89
+ "layer_norm": false,
90
+ "filters_3": 768,
91
+ "tensor_neurons": 16,
92
+ "output_dir": "saved_models/wikievents/GoLLIE-7B/code_generator/train/2025-09-26/19-34-03",
93
+ "main_dir": "/fscratch/oadjali/EEE",
94
+ "experiment_dir": "/fscratch/oadjali/EEE/saved_models/wikievents/GoLLIE-7B/code_generator/train/2025-09-26/19-34-03",
95
+ "experiment_name": null,
96
+ "search": null,
97
+ "input_key": null,
98
+ "train_input_key": null,
99
+ "train_search_run": null,
100
+ "val_search_run": null,
101
+ "test_search_run": null,
102
+ "no_resize": false,
103
+ "xlnet": false,
104
+ "tune_loss": false,
105
+ "tune_dev": true,
106
+ "tune_valid_loss": false,
107
+ "sanity_val_steps": 0,
108
+ "cpu": false,
109
+ "keep_in_memory": false,
110
+ "nbr_gpus": 1,
111
+ "N": 0,
112
+ "nbr_workers": 1,
113
+ "num_proc": 1,
114
+ "acc": "gpu",
115
+ "fast_dev_run": null,
116
+ "local_cache": "../my_transformers_cache",
117
+ "IMAGE_PATH": null,
118
+ "transformer_path": null,
119
+ "resume_from": null,
120
+ "answer_generator": "HiTZ/GoLLIE-7B",
121
+ "query_model": null,
122
+ "context_model": null,
123
+ "answer_generator_2": null,
124
+ "checkpoint": null,
125
+ "query_checkpoint": null,
126
+ "checkpoint_2": null,
127
+ "non_strict_load": true,
128
+ "trainer_checkpoint": null,
129
+ "linear_MEP_checkpoint": null,
130
+ "linear_MLM_checkpoint": null,
131
+ "config": "experiments/event_extraction/config_code_generator.json",
132
+ "context_checkpoint": null,
133
+ "grad_accum": 1,
134
+ "cls_token": "[CLS]",
135
+ "sep_token": "[SEP]",
136
+ "pad_token": "[PAD]",
137
+ "cls_token_at_end": false,
138
+ "mask_padding_with_zero": true,
139
+ "pad_on_left": false,
140
+ "sequence_a_segment_id": 0,
141
+ "pad_token_segment_id": 0,
142
+ "cls_token_segment_id": 0,
143
+ "pad_token_label_id": 0,
144
+ "pad_token_id": 0,
145
+ "past_index": -1,
146
+ "world_size": 1,
147
+ "use_lstm": false,
148
+ "use_entity": false,
149
+ "use_question_graph": false,
150
+ "use_question_objects": false,
151
+ "EFeat": null,
152
+ "NFeat": null,
153
+ "ltn": false,
154
+ "use_kelm": false,
155
+ "unshared": false,
156
+ "mlm": false,
157
+ "test": false,
158
+ "validate": false,
159
+ "eval_on_start": false,
160
+ "predict": null,
161
+ "split_dpr": false,
162
+ "rename_model": false,
163
+ "mask_rate": 0.15,
164
+ "pretrained_triviaq": false,
165
+ "error_analysis": false,
166
+ "debug": false,
167
+ "stop_debug": false,
168
+ "entity_linking": false,
169
+ "entity_linking_split": false,
170
+ "build_entity_paths": null,
171
+ "object_detection": null,
172
+ "update_wikidata_ids": null,
173
+ "update_None_ids": null,
174
+ "dependency_parsing": null,
175
+ "image_captionning": null,
176
+ "build_graph": false,
177
+ "from_bert": false,
178
+ "fast": false,
179
+ "get_examples": false,
180
+ "get_statistics": null,
181
+ "word_vocab_size": 0,
182
+ "char_vocab_size": 0,
183
+ "max_word_len": 30,
184
+ "device": "cuda",
185
+ "embedding_matrix": null,
186
+ "new_vocab_embeddings": null,
187
+ "w2v_file": "word_vector_200d.vec",
188
+ "word_emb_dim": 200,
189
+ "char_lstm": false,
190
+ "char_cnn": false,
191
+ "seed": 42,
192
+ "tune_batch_size": false,
193
+ "val_check_interval": null,
194
+ "model_names": null,
195
+ "build_hf_local_cache": true,
196
+ "strategy": null,
197
+ "monitor_metric": "mean_f1",
198
+ "precision": 32,
199
+ "no_lora": false,
200
+ "early_stopping": true,
201
+ "reduce_on_plateau": true,
202
+ "num_generations": 2,
203
+ "num_passages": 5,
204
+ "k_schemes": 10,
205
+ "use_liger_loss": false,
206
+ "kl_beta": 0.0,
207
+ "TI_alpha": 1.0,
208
+ "TC_alpha": 1.0,
209
+ "AI_alpha": 1.0,
210
+ "AC_alpha": 1.0,
211
+ "schema_retrieval_val": false,
212
+ "schema_retrieval_train": false,
213
+ "EE_2steps": false,
214
+ "EE_SL": false,
215
+ "save_ckpt": false,
216
+ "lora_config": {
217
+ "peft_type": "LORA",
218
+ "auto_mapping": null,
219
+ "base_model_name_or_path": null,
220
+ "revision": null,
221
+ "task_type": "CAUSAL_LM",
222
+ "inference_mode": false,
223
+ "r": 8,
224
+ "target_modules": [
225
+ "q_proj",
226
+ "v_proj"
227
+ ],
228
+ "lora_alpha": 16,
229
+ "lora_dropout": 0.05,
230
+ "fan_in_fan_out": false,
231
+ "bias": "none",
232
+ "use_rslora": false,
233
+ "modules_to_save": null,
234
+ "init_lora_weights": true,
235
+ "layers_to_transform": null,
236
+ "layers_pattern": null,
237
+ "rank_pattern": {},
238
+ "alpha_pattern": {},
239
+ "megatron_config": null,
240
+ "megatron_core": "megatron.core",
241
+ "loftq_config": {},
242
+ "use_dora": false
243
+ },
244
+ "quantization_config": null,
245
+ "hf": false,
246
+ "slurm_scipt_path": null,
247
+ "enabled_args": " --train_batch_size 1\n --eval_batch_size 1\n --num_train_epochs 20\n --nbr_gpus 1\n --learning_rate 5e-6\n --grad_check\n --config experiments/event_extraction/config_code_generator.json\n --non_strict_load\n --ds ../datasets/textee/wikievents\n --answer_generator HiTZ/GoLLIE-7B\n --guidelines new_Llama-31-8B-Instruct_annotation_guidelines\n --k_schemes 10\n --num_passages 5"
248
+ }
log.txt ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ trainable params: 4194304 || all params: 6742740992 || trainable%: 0.06220473254091146
2
+
3
+ epoch:0 {'TC_prec': 0.3167701863354037, 'TC_rec': 0.12085308056872038, 'TI_prec': 0.45962732919254656, 'TI_rec': 0.17535545023696683, 'AC_prec': 0.10723860589812333, 'AC_rec': 0.06825938566552901, 'AI_prec': 0.13404825737265416, 'AI_rec': 0.08532423208191127, 'TI': 0.2538593481989709, 'TC': 0.17495711835334476, 'AI': 0.1042752867570386, 'AC': 0.08342022940563087, 'mean_f1': 0.15412799567874627}
4
+ avg_train_loss tensor(1.6375)
5
+ lr 5e-06
6
+
7
+ epoch:1 {'TC_prec': 0.3235294117647059, 'TC_rec': 0.13033175355450238, 'TI_prec': 0.4764705882352941, 'TI_rec': 0.19194312796208532, 'AC_prec': 0.1174934725848564, 'AC_rec': 0.07679180887372014, 'AI_prec': 0.1409921671018277, 'AI_rec': 0.09215017064846416, 'TI': 0.27364864864864863, 'TC': 0.18581081081081083, 'AI': 0.11145510835913311, 'AC': 0.09287925696594428, 'mean_f1': 0.1659484561961342}
8
+ avg_train_loss tensor(1.1741)
9
+ lr 5e-06
10
+
11
+ epoch:2 {'TC_prec': 0.3176470588235294, 'TC_rec': 0.12796208530805686, 'TI_prec': 0.47058823529411764, 'TI_rec': 0.1895734597156398, 'AC_prec': 0.11345646437994723, 'AC_rec': 0.07337883959044368, 'AI_prec': 0.13192612137203166, 'AI_rec': 0.08532423208191127, 'TI': 0.2702702702702703, 'TC': 0.1824324324324324, 'AI': 0.10362694300518135, 'AC': 0.08911917098445596, 'mean_f1': 0.161362204173085}
12
+ avg_train_loss tensor(1.0730)
13
+ lr 5e-06
14
+
15
+ epoch:3 {'TC_prec': 0.3275862068965517, 'TC_rec': 0.13507109004739337, 'TI_prec': 0.4827586206896552, 'TI_rec': 0.1990521327014218, 'AC_prec': 0.11286089238845144, 'AC_rec': 0.07337883959044368, 'AI_prec': 0.13385826771653545, 'AI_rec': 0.08703071672354949, 'TI': 0.2818791946308725, 'TC': 0.1912751677852349, 'AI': 0.10548086866597725, 'AC': 0.08893485005170632, 'mean_f1': 0.16689252028344775}
16
+ avg_train_loss tensor(1.0316)
17
+ lr 5e-06
18
+
19
+ epoch:4 {'TC_prec': 0.3176470588235294, 'TC_rec': 0.12796208530805686, 'TI_prec': 0.4764705882352941, 'TI_rec': 0.19194312796208532, 'AC_prec': 0.11466666666666667, 'AC_rec': 0.07337883959044368, 'AI_prec': 0.13333333333333333, 'AI_rec': 0.08532423208191127, 'TI': 0.27364864864864863, 'TC': 0.1824324324324324, 'AI': 0.10405827263267431, 'AC': 0.08949011446409989, 'mean_f1': 0.16240736704446382}
20
+ avg_train_loss tensor(1.0073)
21
+ lr 5e-06
22
+
23
+ epoch:5 {'TC_prec': 0.313953488372093, 'TC_rec': 0.12796208530805686, 'TI_prec': 0.48255813953488375, 'TI_rec': 0.1966824644549763, 'AC_prec': 0.11052631578947368, 'AC_rec': 0.07167235494880546, 'AI_prec': 0.13421052631578947, 'AI_rec': 0.08703071672354949, 'TI': 0.27946127946127947, 'TC': 0.18181818181818182, 'AI': 0.10559006211180125, 'AC': 0.08695652173913043, 'mean_f1': 0.16345651128259822}
24
+ avg_train_loss tensor(0.9908)
25
+ lr 5e-06
26
+
27
+ epoch:6 {'TC_prec': 0.32335329341317365, 'TC_rec': 0.12796208530805686, 'TI_prec': 0.48502994011976047, 'TI_rec': 0.19194312796208532, 'AC_prec': 0.12054794520547946, 'AC_rec': 0.07508532423208192, 'AI_prec': 0.14246575342465753, 'AI_rec': 0.08873720136518772, 'TI': 0.2750424448217318, 'TC': 0.1833616298811545, 'AI': 0.10935856992639327, 'AC': 0.092534174553102, 'mean_f1': 0.1650742047955954}
28
+ avg_train_loss tensor(0.9788)
29
+ lr 5e-06
30
+
31
+ epoch:7 {'TC_prec': 0.33532934131736525, 'TC_rec': 0.13270142180094788, 'TI_prec': 0.47904191616766467, 'TI_rec': 0.1895734597156398, 'AC_prec': 0.11653116531165311, 'AC_rec': 0.07337883959044368, 'AI_prec': 0.13821138211382114, 'AI_rec': 0.08703071672354949, 'TI': 0.27164685908319186, 'TC': 0.19015280135823429, 'AI': 0.1068062827225131, 'AC': 0.0900523560209424, 'mean_f1': 0.16466457479622043}
32
+ avg_train_loss tensor(0.9733)
33
+ lr 2.5e-06
34
+
35
+ [TEST]
36
+ {'TC_prec': 0.2796610169491525, 'TC_rec': 0.1746031746031746, 'TI_prec': 0.3728813559322034, 'TI_rec': 0.2328042328042328, 'AC_prec': 0.0947136563876652, 'AC_rec': 0.083984375, 'AI_prec': 0.11674008810572688, 'AI_rec': 0.103515625, 'TI': 0.2866449511400651, 'TC': 0.21498371335504887, 'AI': 0.10973084886128363, 'AC': 0.08902691511387163, 'mean_f1': 0.1750966071175673}
lora_config.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "peft_type": "LORA",
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": null,
5
+ "revision": null,
6
+ "task_type": "CAUSAL_LM",
7
+ "inference_mode": false,
8
+ "r": 8,
9
+ "target_modules": [
10
+ "q_proj",
11
+ "v_proj"
12
+ ],
13
+ "lora_alpha": 16,
14
+ "lora_dropout": 0.05,
15
+ "fan_in_fan_out": false,
16
+ "bias": "none",
17
+ "use_rslora": false,
18
+ "modules_to_save": null,
19
+ "init_lora_weights": true,
20
+ "layers_to_transform": null,
21
+ "layers_pattern": null,
22
+ "rank_pattern": {},
23
+ "alpha_pattern": {},
24
+ "megatron_config": null,
25
+ "megatron_core": "megatron.core",
26
+ "loftq_config": {},
27
+ "use_dora": false
28
+ }
metrics.tex ADDED
@@ -0,0 +1,112 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ epoch:0
3
+ =========================================
4
+ \begin{table}
5
+ \caption{Evaluation Metrics}
6
+ \label{tab:metrics}
7
+ \begin{tabular}{rrrr}
8
+ \toprule
9
+ TI & TC & AI & AC \\
10
+ \midrule
11
+ 25.39 & 17.50 & 10.43 & 8.34 \\
12
+ \bottomrule
13
+ \end{tabular}
14
+ \end{table}
15
+
16
+ epoch:1
17
+ =========================================
18
+ \begin{table}
19
+ \caption{Evaluation Metrics}
20
+ \label{tab:metrics}
21
+ \begin{tabular}{rrrr}
22
+ \toprule
23
+ TI & TC & AI & AC \\
24
+ \midrule
25
+ 27.36 & 18.58 & 11.15 & 9.29 \\
26
+ \bottomrule
27
+ \end{tabular}
28
+ \end{table}
29
+
30
+ epoch:2
31
+ =========================================
32
+ \begin{table}
33
+ \caption{Evaluation Metrics}
34
+ \label{tab:metrics}
35
+ \begin{tabular}{rrrr}
36
+ \toprule
37
+ TI & TC & AI & AC \\
38
+ \midrule
39
+ 27.03 & 18.24 & 10.36 & 8.91 \\
40
+ \bottomrule
41
+ \end{tabular}
42
+ \end{table}
43
+
44
+ epoch:3
45
+ =========================================
46
+ \begin{table}
47
+ \caption{Evaluation Metrics}
48
+ \label{tab:metrics}
49
+ \begin{tabular}{rrrr}
50
+ \toprule
51
+ TI & TC & AI & AC \\
52
+ \midrule
53
+ 28.19 & 19.13 & 10.55 & 8.89 \\
54
+ \bottomrule
55
+ \end{tabular}
56
+ \end{table}
57
+
58
+ epoch:4
59
+ =========================================
60
+ \begin{table}
61
+ \caption{Evaluation Metrics}
62
+ \label{tab:metrics}
63
+ \begin{tabular}{rrrr}
64
+ \toprule
65
+ TI & TC & AI & AC \\
66
+ \midrule
67
+ 27.36 & 18.24 & 10.41 & 8.95 \\
68
+ \bottomrule
69
+ \end{tabular}
70
+ \end{table}
71
+
72
+ epoch:5
73
+ =========================================
74
+ \begin{table}
75
+ \caption{Evaluation Metrics}
76
+ \label{tab:metrics}
77
+ \begin{tabular}{rrrr}
78
+ \toprule
79
+ TI & TC & AI & AC \\
80
+ \midrule
81
+ 27.95 & 18.18 & 10.56 & 8.70 \\
82
+ \bottomrule
83
+ \end{tabular}
84
+ \end{table}
85
+
86
+ epoch:6
87
+ =========================================
88
+ \begin{table}
89
+ \caption{Evaluation Metrics}
90
+ \label{tab:metrics}
91
+ \begin{tabular}{rrrr}
92
+ \toprule
93
+ TI & TC & AI & AC \\
94
+ \midrule
95
+ 27.50 & 18.34 & 10.94 & 9.25 \\
96
+ \bottomrule
97
+ \end{tabular}
98
+ \end{table}
99
+
100
+ epoch:7
101
+ =========================================
102
+ \begin{table}
103
+ \caption{Evaluation Metrics}
104
+ \label{tab:metrics}
105
+ \begin{tabular}{rrrr}
106
+ \toprule
107
+ TI & TC & AI & AC \\
108
+ \midrule
109
+ 27.16 & 19.02 & 10.68 & 9.01 \\
110
+ \bottomrule
111
+ \end{tabular}
112
+ \end{table}
model-epoch=03-mean_f1=0.1669_sd=42.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:812d283bbcb77631a3a04b23066e0509a69b67edd8931898c49de27f72b0670f
3
+ size 13502640837
predictions/predictions_0.json ADDED
@@ -0,0 +1 @@
 
 
1
+ null
predictions/predictions_1.json ADDED
@@ -0,0 +1 @@
 
 
1
+ null
predictions/predictions_2.json ADDED
@@ -0,0 +1 @@
 
 
1
+ null
predictions/predictions_3.json ADDED
@@ -0,0 +1 @@
 
 
1
+ null
predictions/predictions_4.json ADDED
@@ -0,0 +1 @@
 
 
1
+ null
predictions/predictions_5.json ADDED
@@ -0,0 +1 @@
 
 
1
+ null
predictions/predictions_6.json ADDED
@@ -0,0 +1 @@
 
 
1
+ null
predictions/predictions_7.json ADDED
@@ -0,0 +1 @@
 
 
1
+ null
qualitative_analysis.txt ADDED
The diff for this file is too large to render. See raw diff
 
quantization_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ null
test_metrics.tex ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ =========================================
2
+ \begin{table}
3
+ \caption{Evaluation Metrics}
4
+ \label{tab:metrics}
5
+ \begin{tabular}{rrrr}
6
+ \toprule
7
+ TI & TC & AI & AC \\
8
+ \midrule
9
+ 28.66 & 21.50 & 10.97 & 8.90 \\
10
+ \bottomrule
11
+ \end{tabular}
12
+ \end{table}