ringorsolya commited on
Commit
e7bc0f1
·
verified ·
1 Parent(s): d791a5c

Upload eval_results.json with huggingface_hub

Browse files
Files changed (1) hide show
  1. eval_results.json +27 -63
eval_results.json CHANGED
@@ -1,84 +1,48 @@
1
  {
2
  "overall": {
3
- "accuracy": 0.4108175318619832,
4
- "f1_macro": 0.1941274108021563,
5
- "f1_weighted": 0.23925283131749744
6
  },
7
  "per_language": {
8
- "cz": {
9
- "n": 4602,
10
- "accuracy": 0.41090830073880924,
11
- "f1_macro": 0.19415781097592277,
12
- "f1_weighted": 0.23934316854985005
13
- },
14
- "en": {
15
- "n": 4596,
16
- "accuracy": 0.41079199303742386,
17
- "f1_macro": 0.19411885667283568,
18
- "f1_weighted": 0.23922741605634057
19
- },
20
- "fr": {
21
- "n": 4569,
22
- "accuracy": 0.4108119938717444,
23
- "f1_macro": 0.19412555590029992,
24
- "f1_weighted": 0.23924732004258895
25
- },
26
- "ger": {
27
- "n": 4599,
28
- "accuracy": 0.41074146553598606,
29
- "f1_macro": 0.19410193177147553,
30
- "f1_weighted": 0.23917713575754554
31
- },
32
  "hun": {
33
  "n": 4603,
34
- "accuracy": 0.41081903106669565,
35
- "f1_macro": 0.19412791294528284,
36
- "f1_weighted": 0.2392543232975428
37
- },
38
- "pl": {
39
- "n": 4603,
40
- "accuracy": 0.41081903106669565,
41
- "f1_macro": 0.19412791294528284,
42
- "f1_weighted": 0.2392543232975428
43
- },
44
- "sk": {
45
- "n": 4598,
46
- "accuracy": 0.4108307959982601,
47
- "f1_macro": 0.19413185345049075,
48
- "f1_weighted": 0.23926603164504806
49
  }
50
  },
51
  "classification_report": {
52
  "negative": {
53
- "precision": 0.4108175318619832,
54
- "recall": 1.0,
55
- "f1-score": 0.5823822324064689,
56
- "support": 13216.0
57
  },
58
  "neutral": {
59
- "precision": 0.0,
60
- "recall": 0.0,
61
- "f1-score": 0.0,
62
- "support": 9559.0
63
  },
64
  "positive": {
65
- "precision": 0.0,
66
- "recall": 0.0,
67
- "f1-score": 0.0,
68
- "support": 9395.0
69
  },
70
- "accuracy": 0.4108175318619832,
71
  "macro avg": {
72
- "precision": 0.13693917728732774,
73
- "recall": 0.3333333333333333,
74
- "f1-score": 0.1941274108021563,
75
- "support": 32170.0
76
  },
77
  "weighted avg": {
78
- "precision": 0.16877104448517158,
79
- "recall": 0.4108175318619832,
80
- "f1-score": 0.23925283131749744,
81
- "support": 32170.0
82
  }
83
  }
84
  }
 
1
  {
2
  "overall": {
3
+ "accuracy": 0.8442320225939605,
4
+ "f1_macro": 0.8387464047460437,
5
+ "f1_weighted": 0.8435908941071462
6
  },
7
  "per_language": {
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
8
  "hun": {
9
  "n": 4603,
10
+ "accuracy": 0.8442320225939605,
11
+ "f1_macro": 0.8387464047460437,
12
+ "f1_weighted": 0.8435908941071462
 
 
 
 
 
 
 
 
 
 
 
 
13
  }
14
  },
15
  "classification_report": {
16
  "negative": {
17
+ "precision": 0.8884718498659517,
18
+ "recall": 0.8762559492332099,
19
+ "f1-score": 0.882321618743344,
20
+ "support": 1891.0
21
  },
22
  "neutral": {
23
+ "precision": 0.7978311386522076,
24
+ "recall": 0.7529239766081871,
25
+ "f1-score": 0.7747273411056789,
26
+ "support": 1368.0
27
  },
28
  "positive": {
29
+ "precision": 0.8286109191430546,
30
+ "recall": 0.8921130952380952,
31
+ "f1-score": 0.8591902543891079,
32
+ "support": 1344.0
33
  },
34
+ "accuracy": 0.8442320225939605,
35
  "macro avg": {
36
+ "precision": 0.8383046358870713,
37
+ "recall": 0.8404310070264974,
38
+ "f1-score": 0.8387464047460437,
39
+ "support": 4603.0
40
  },
41
  "weighted avg": {
42
+ "precision": 0.8440552555074952,
43
+ "recall": 0.8442320225939605,
44
+ "f1-score": 0.8435908941071462,
45
+ "support": 4603.0
46
  }
47
  }
48
  }