Add evaluation results for model lightonai/Reason-ModernColBERT revision a1f75a3d01084caf9d4cea3500a498533352f001

#4
by Samoed - opened
.eval_results/BrightAopsRetrieval.yaml ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ - dataset:
2
+ id: mteb/BRIGHT
3
+ task_id: BrightAopsRetrieval_default_standard
4
+ revision: c26703e6600d97c579ee2985f16cf307db13ed85
5
+ value: 8.251
6
+ date: '2026-03-09'
7
+ notes: Obtained using MTEB v2.10.8
8
+ source:
9
+ url: https://github.com/embeddings-benchmark/mteb/
10
+ name: Obtained using MTEB v2.10.8
11
+ user: mteb
12
+ - dataset:
13
+ id: mteb/BRIGHT
14
+ task_id: BrightAopsRetrieval
15
+ revision: c26703e6600d97c579ee2985f16cf307db13ed85
16
+ value: 8.251
17
+ date: '2026-03-09'
18
+ notes: Obtained using MTEB v2.10.8
19
+ source:
20
+ url: https://github.com/embeddings-benchmark/mteb/
21
+ name: Obtained using MTEB v2.10.8
22
+ user: mteb
.eval_results/BrightBiologyRetrieval.yaml ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ - dataset:
2
+ id: mteb/BRIGHT
3
+ task_id: BrightBiologyRetrieval_default_standard
4
+ revision: c26703e6600d97c579ee2985f16cf307db13ed85
5
+ value: 33.869
6
+ date: '2026-03-03'
7
+ notes: Obtained using MTEB v2.10.3
8
+ source:
9
+ url: https://github.com/embeddings-benchmark/mteb/
10
+ name: Obtained using MTEB v2.10.3
11
+ user: mteb
12
+ - dataset:
13
+ id: mteb/BRIGHT
14
+ task_id: BrightBiologyRetrieval
15
+ revision: c26703e6600d97c579ee2985f16cf307db13ed85
16
+ value: 33.869
17
+ date: '2026-03-03'
18
+ notes: Obtained using MTEB v2.10.3
19
+ source:
20
+ url: https://github.com/embeddings-benchmark/mteb/
21
+ name: Obtained using MTEB v2.10.3
22
+ user: mteb
.eval_results/BrightEarthScienceRetrieval.yaml ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ - dataset:
2
+ id: mteb/BRIGHT
3
+ task_id: BrightEarthScienceRetrieval_default_standard
4
+ revision: c26703e6600d97c579ee2985f16cf307db13ed85
5
+ value: 41.705
6
+ date: '2026-03-03'
7
+ notes: Obtained using MTEB v2.10.3
8
+ source:
9
+ url: https://github.com/embeddings-benchmark/mteb/
10
+ name: Obtained using MTEB v2.10.3
11
+ user: mteb
12
+ - dataset:
13
+ id: mteb/BRIGHT
14
+ task_id: BrightEarthScienceRetrieval
15
+ revision: c26703e6600d97c579ee2985f16cf307db13ed85
16
+ value: 41.705
17
+ date: '2026-03-03'
18
+ notes: Obtained using MTEB v2.10.3
19
+ source:
20
+ url: https://github.com/embeddings-benchmark/mteb/
21
+ name: Obtained using MTEB v2.10.3
22
+ user: mteb
.eval_results/BrightEconomicsRetrieval.yaml ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ - dataset:
2
+ id: mteb/BRIGHT
3
+ task_id: BrightEconomicsRetrieval_default_standard
4
+ revision: c26703e6600d97c579ee2985f16cf307db13ed85
5
+ value: 24.552
6
+ date: '2026-03-03'
7
+ notes: Obtained using MTEB v2.10.3
8
+ source:
9
+ url: https://github.com/embeddings-benchmark/mteb/
10
+ name: Obtained using MTEB v2.10.3
11
+ user: mteb
12
+ - dataset:
13
+ id: mteb/BRIGHT
14
+ task_id: BrightEconomicsRetrieval
15
+ revision: c26703e6600d97c579ee2985f16cf307db13ed85
16
+ value: 24.552
17
+ date: '2026-03-03'
18
+ notes: Obtained using MTEB v2.10.3
19
+ source:
20
+ url: https://github.com/embeddings-benchmark/mteb/
21
+ name: Obtained using MTEB v2.10.3
22
+ user: mteb
.eval_results/BrightLeetcodeRetrieval.yaml ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ - dataset:
2
+ id: mteb/BRIGHT
3
+ task_id: BrightLeetcodeRetrieval_default_standard
4
+ revision: c26703e6600d97c579ee2985f16cf307db13ed85
5
+ value: 30.857
6
+ date: '2026-03-09'
7
+ notes: Obtained using MTEB v2.10.8
8
+ source:
9
+ url: https://github.com/embeddings-benchmark/mteb/
10
+ name: Obtained using MTEB v2.10.8
11
+ user: mteb
12
+ - dataset:
13
+ id: mteb/BRIGHT
14
+ task_id: BrightLeetcodeRetrieval
15
+ revision: c26703e6600d97c579ee2985f16cf307db13ed85
16
+ value: 30.857
17
+ date: '2026-03-09'
18
+ notes: Obtained using MTEB v2.10.8
19
+ source:
20
+ url: https://github.com/embeddings-benchmark/mteb/
21
+ name: Obtained using MTEB v2.10.8
22
+ user: mteb
.eval_results/BrightPonyRetrieval.yaml ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ - dataset:
2
+ id: mteb/BRIGHT
3
+ task_id: BrightPonyRetrieval_default_standard
4
+ revision: c26703e6600d97c579ee2985f16cf307db13ed85
5
+ value: 8.735
6
+ date: '2026-03-03'
7
+ notes: Obtained using MTEB v2.10.3
8
+ source:
9
+ url: https://github.com/embeddings-benchmark/mteb/
10
+ name: Obtained using MTEB v2.10.3
11
+ user: mteb
12
+ - dataset:
13
+ id: mteb/BRIGHT
14
+ task_id: BrightPonyRetrieval
15
+ revision: c26703e6600d97c579ee2985f16cf307db13ed85
16
+ value: 8.735
17
+ date: '2026-03-03'
18
+ notes: Obtained using MTEB v2.10.3
19
+ source:
20
+ url: https://github.com/embeddings-benchmark/mteb/
21
+ name: Obtained using MTEB v2.10.3
22
+ user: mteb
.eval_results/BrightPsychologyRetrieval.yaml ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ - dataset:
2
+ id: mteb/BRIGHT
3
+ task_id: BrightPsychologyRetrieval_default_standard
4
+ revision: c26703e6600d97c579ee2985f16cf307db13ed85
5
+ value: 31.037
6
+ date: '2026-03-03'
7
+ notes: Obtained using MTEB v2.10.3
8
+ source:
9
+ url: https://github.com/embeddings-benchmark/mteb/
10
+ name: Obtained using MTEB v2.10.3
11
+ user: mteb
12
+ - dataset:
13
+ id: mteb/BRIGHT
14
+ task_id: BrightPsychologyRetrieval
15
+ revision: c26703e6600d97c579ee2985f16cf307db13ed85
16
+ value: 31.037
17
+ date: '2026-03-03'
18
+ notes: Obtained using MTEB v2.10.3
19
+ source:
20
+ url: https://github.com/embeddings-benchmark/mteb/
21
+ name: Obtained using MTEB v2.10.3
22
+ user: mteb
.eval_results/BrightRoboticsRetrieval.yaml ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ - dataset:
2
+ id: mteb/BRIGHT
3
+ task_id: BrightRoboticsRetrieval_default_standard
4
+ revision: c26703e6600d97c579ee2985f16cf307db13ed85
5
+ value: 21.811
6
+ date: '2026-03-03'
7
+ notes: Obtained using MTEB v2.10.3
8
+ source:
9
+ url: https://github.com/embeddings-benchmark/mteb/
10
+ name: Obtained using MTEB v2.10.3
11
+ user: mteb
12
+ - dataset:
13
+ id: mteb/BRIGHT
14
+ task_id: BrightRoboticsRetrieval
15
+ revision: c26703e6600d97c579ee2985f16cf307db13ed85
16
+ value: 21.811
17
+ date: '2026-03-03'
18
+ notes: Obtained using MTEB v2.10.3
19
+ source:
20
+ url: https://github.com/embeddings-benchmark/mteb/
21
+ name: Obtained using MTEB v2.10.3
22
+ user: mteb
.eval_results/BrightStackoverflowRetrieval.yaml ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ - dataset:
2
+ id: mteb/BRIGHT
3
+ task_id: BrightStackoverflowRetrieval_default_standard
4
+ revision: c26703e6600d97c579ee2985f16cf307db13ed85
5
+ value: 24.255
6
+ date: '2026-03-03'
7
+ notes: Obtained using MTEB v2.10.3
8
+ source:
9
+ url: https://github.com/embeddings-benchmark/mteb/
10
+ name: Obtained using MTEB v2.10.3
11
+ user: mteb
12
+ - dataset:
13
+ id: mteb/BRIGHT
14
+ task_id: BrightStackoverflowRetrieval
15
+ revision: c26703e6600d97c579ee2985f16cf307db13ed85
16
+ value: 24.255
17
+ date: '2026-03-03'
18
+ notes: Obtained using MTEB v2.10.3
19
+ source:
20
+ url: https://github.com/embeddings-benchmark/mteb/
21
+ name: Obtained using MTEB v2.10.3
22
+ user: mteb
.eval_results/BrightSustainableLivingRetrieval.yaml ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ - dataset:
2
+ id: mteb/BRIGHT
3
+ task_id: BrightSustainableLivingRetrieval_default_standard
4
+ revision: c26703e6600d97c579ee2985f16cf307db13ed85
5
+ value: 20.215
6
+ date: '2026-03-03'
7
+ notes: Obtained using MTEB v2.10.3
8
+ source:
9
+ url: https://github.com/embeddings-benchmark/mteb/
10
+ name: Obtained using MTEB v2.10.3
11
+ user: mteb
12
+ - dataset:
13
+ id: mteb/BRIGHT
14
+ task_id: BrightSustainableLivingRetrieval
15
+ revision: c26703e6600d97c579ee2985f16cf307db13ed85
16
+ value: 20.215
17
+ date: '2026-03-03'
18
+ notes: Obtained using MTEB v2.10.3
19
+ source:
20
+ url: https://github.com/embeddings-benchmark/mteb/
21
+ name: Obtained using MTEB v2.10.3
22
+ user: mteb
.eval_results/BrightTheoremQAQuestionsRetrieval.yaml ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ - dataset:
2
+ id: mteb/BRIGHT
3
+ task_id: BrightTheoremQAQuestionsRetrieval_default_standard
4
+ revision: c26703e6600d97c579ee2985f16cf307db13ed85
5
+ value: 18.327
6
+ date: '2026-03-09'
7
+ notes: Obtained using MTEB v2.10.8
8
+ source:
9
+ url: https://github.com/embeddings-benchmark/mteb/
10
+ name: Obtained using MTEB v2.10.8
11
+ user: mteb
12
+ - dataset:
13
+ id: mteb/BRIGHT
14
+ task_id: BrightTheoremQAQuestionsRetrieval
15
+ revision: c26703e6600d97c579ee2985f16cf307db13ed85
16
+ value: 18.327
17
+ date: '2026-03-09'
18
+ notes: Obtained using MTEB v2.10.8
19
+ source:
20
+ url: https://github.com/embeddings-benchmark/mteb/
21
+ name: Obtained using MTEB v2.10.8
22
+ user: mteb
.eval_results/BrightTheoremQATheoremsRetrieval.yaml ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ - dataset:
2
+ id: mteb/BRIGHT
3
+ task_id: BrightTheoremQATheoremsRetrieval_default_standard
4
+ revision: c26703e6600d97c579ee2985f16cf307db13ed85
5
+ value: 9.29
6
+ date: '2026-03-09'
7
+ notes: Obtained using MTEB v2.10.8
8
+ source:
9
+ url: https://github.com/embeddings-benchmark/mteb/
10
+ name: Obtained using MTEB v2.10.8
11
+ user: mteb
12
+ - dataset:
13
+ id: mteb/BRIGHT
14
+ task_id: BrightTheoremQATheoremsRetrieval
15
+ revision: c26703e6600d97c579ee2985f16cf307db13ed85
16
+ value: 9.29
17
+ date: '2026-03-09'
18
+ notes: Obtained using MTEB v2.10.8
19
+ source:
20
+ url: https://github.com/embeddings-benchmark/mteb/
21
+ name: Obtained using MTEB v2.10.8
22
+ user: mteb