dataset
stringlengths
5
115
config
stringlengths
1
162
split
stringlengths
1
228
num_examples
int64
3
341M
column_name
stringlengths
0
77.9k
null_count
int64
0
62.9M
null_proportion
float64
0
1
min
int64
0
9.25M
max
int64
0
1.07B
mean
float64
0
90.4M
median
float64
0
80.1M
std
float64
0
130M
histogram
dict
partial
bool
2 classes
1024m/mMGTD-Corpus
version v0.3
English
634,012
Original text
0
0
9
234,745
1,933.32825
1,314
2,574.12075
{ "bin_edges": [ 9, 23483, 46957, 70431, 93905, 117379, 140853, 164327, 187801, 211275, 234745 ], "hist": [ 633202, 645, 108, 40, 9, 4, 1, 1, 1, 1 ] }
false
open-llm-leaderboard/Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3-details
Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3__leaderboard_bbh_tracking_shuffled_objects_seven_objects
2024_07_22T10_01_37.284764
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3-details
Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3__leaderboard_bbh_tracking_shuffled_objects_seven_objects
2024_07_22T10_01_37.284764
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3-details
Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3__leaderboard_bbh_tracking_shuffled_objects_seven_objects
latest
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3-details
Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3__leaderboard_bbh_tracking_shuffled_objects_seven_objects
latest
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
1024m/mMGTD-Corpus
version v0.3
Indonesian
310,361
Modified text
1
0
10
65,058
778.21434
357
1,438.56405
{ "bin_edges": [ 10, 6515, 13020, 19525, 26030, 32535, 39040, 45545, 52050, 58555, 65058 ], "hist": [ 305650, 4204, 433, 44, 20, 7, 1, 0, 0, 1 ] }
false
1024m/mMGTD-Corpus
version v0.3
Indonesian
310,361
Original text
0
0
10
65,058
776.00043
354
1,447.58364
{ "bin_edges": [ 10, 6515, 13020, 19525, 26030, 32535, 39040, 45545, 52050, 58555, 65058 ], "hist": [ 305572, 4276, 441, 43, 20, 7, 1, 0, 0, 1 ] }
false
open-llm-leaderboard/arcee-ai__Arcee-Spark-details
arcee-ai__Arcee-Spark__leaderboard_bbh_snarks
2024_07_22T09_52_59.637349
178
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 178 ] }
false
open-llm-leaderboard/arcee-ai__Arcee-Spark-details
arcee-ai__Arcee-Spark__leaderboard_bbh_snarks
2024_07_22T09_52_59.637349
178
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 178 ] }
false
open-llm-leaderboard/arcee-ai__Arcee-Spark-details
arcee-ai__Arcee-Spark__leaderboard_bbh_snarks
latest
178
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 178 ] }
false
open-llm-leaderboard/arcee-ai__Arcee-Spark-details
arcee-ai__Arcee-Spark__leaderboard_bbh_snarks
latest
178
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 178 ] }
false
1024m/mMGTD-Corpus
version v0.3
Italian
85,520
Modified text
4
0.00005
8
16,701
740.20323
649
429.46011
{ "bin_edges": [ 8, 1678, 3348, 5018, 6688, 8358, 10028, 11698, 13368, 15038, 16701 ], "hist": [ 82392, 3025, 83, 9, 3, 1, 1, 0, 1, 1 ] }
false
1024m/mMGTD-Corpus
version v0.3
Italian
85,520
Original text
0
0
8
16,701
688.12067
608
391.50787
{ "bin_edges": [ 8, 1678, 3348, 5018, 6688, 8358, 10028, 11698, 13368, 15038, 16701 ], "hist": [ 83566, 1866, 74, 9, 2, 1, 0, 0, 1, 1 ] }
false
1024m/mMGTD-Corpus
version v0.3
Japanese
32,764
Modified text
7
0.00021
21
41,233
578.99621
285
779.10456
{ "bin_edges": [ 21, 4143, 8265, 12387, 16509, 20631, 24753, 28875, 32997, 37119, 41233 ], "hist": [ 32605, 133, 12, 4, 0, 1, 1, 0, 0, 1 ] }
false
1024m/mMGTD-Corpus
version v0.3
Japanese
32,764
Original text
0
0
180
27,504
612.94705
222
839.40899
{ "bin_edges": [ 180, 2913, 5646, 8379, 11112, 13845, 16578, 19311, 22044, 24777, 27504 ], "hist": [ 32049, 642, 53, 12, 6, 1, 0, 0, 0, 1 ] }
false
1024m/mMGTD-Corpus
version v0.3
Hindi
103,822
Modified text
9
0.00009
14
48,453
2,329.49287
1,685
2,157.53369
{ "bin_edges": [ 14, 4858, 9702, 14546, 19390, 24234, 29078, 33922, 38766, 43610, 48453 ], "hist": [ 92815, 9713, 1053, 169, 42, 13, 4, 2, 0, 2 ] }
false
1024m/mMGTD-Corpus
version v0.3
Hindi
103,822
Original text
0
0
14
48,453
2,377.19411
1,729
2,188.61529
{ "bin_edges": [ 14, 4858, 9702, 14546, 19390, 24234, 29078, 33922, 38766, 43610, 48453 ], "hist": [ 92378, 10088, 1110, 181, 43, 13, 4, 2, 1, 2 ] }
false
open-llm-leaderboard/Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3-details
Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3__leaderboard_bbh_logical_deduction_three_objects
2024_07_22T10_01_37.284764
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3-details
Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3__leaderboard_bbh_logical_deduction_three_objects
2024_07_22T10_01_37.284764
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3-details
Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3__leaderboard_bbh_logical_deduction_three_objects
latest
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3-details
Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3__leaderboard_bbh_logical_deduction_three_objects
latest
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/arcee-ai__Arcee-Spark-details
arcee-ai__Arcee-Spark__leaderboard_bbh_reasoning_about_colored_objects
2024_07_22T09_52_59.637349
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/arcee-ai__Arcee-Spark-details
arcee-ai__Arcee-Spark__leaderboard_bbh_reasoning_about_colored_objects
2024_07_22T09_52_59.637349
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/arcee-ai__Arcee-Spark-details
arcee-ai__Arcee-Spark__leaderboard_bbh_reasoning_about_colored_objects
latest
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/arcee-ai__Arcee-Spark-details
arcee-ai__Arcee-Spark__leaderboard_bbh_reasoning_about_colored_objects
latest
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
1024m/mMGTD-Corpus
version v0.3
Korean
199,279
Modified text
5
0.00003
7
35,327
257.8861
200
326.69878
{ "bin_edges": [ 7, 3540, 7073, 10606, 14139, 17672, 21205, 24738, 28271, 31804, 35327 ], "hist": [ 198948, 310, 14, 1, 0, 0, 0, 0, 0, 1 ] }
false
1024m/mMGTD-Corpus
version v0.3
Korean
199,279
Original text
0
0
7
35,327
245.43323
191
326.14518
{ "bin_edges": [ 7, 3540, 7073, 10606, 14139, 17672, 21205, 24738, 28271, 31804, 35327 ], "hist": [ 198944, 321, 12, 1, 0, 0, 0, 0, 0, 1 ] }
false
1024m/mMGTD-Corpus
version v0.3
Polish
185,252
Modified text
3
0.00002
9
28,736
392.47616
342
252.44775
{ "bin_edges": [ 9, 2882, 5755, 8628, 11501, 14374, 17247, 20120, 22993, 25866, 28736 ], "hist": [ 185204, 37, 2, 3, 1, 1, 0, 0, 0, 1 ] }
false
1024m/mMGTD-Corpus
version v0.3
Polish
185,252
Original text
0
0
9
10,735
378.81015
335
212.68605
{ "bin_edges": [ 9, 1082, 2155, 3228, 4301, 5374, 6447, 7520, 8593, 9666, 10735 ], "hist": [ 183054, 2062, 119, 10, 3, 1, 1, 0, 1, 1 ] }
false
open-llm-leaderboard/Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3-details
Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3__leaderboard_bbh_date_understanding
2024_07_22T10_01_37.284764
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3-details
Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3__leaderboard_bbh_date_understanding
2024_07_22T10_01_37.284764
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3-details
Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3__leaderboard_bbh_date_understanding
latest
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3-details
Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3__leaderboard_bbh_date_understanding
latest
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
1024m/mMGTD-Corpus
version v0.3
Romanian
76,890
Modified text
0
0
13
20,674
531.97071
453
365.94804
{ "bin_edges": [ 13, 2080, 4147, 6214, 8281, 10348, 12415, 14482, 16549, 18616, 20674 ], "hist": [ 76400, 473, 7, 1, 3, 1, 2, 2, 0, 1 ] }
false
1024m/mMGTD-Corpus
version v0.3
Romanian
76,890
Original text
0
0
13
9,280
494.0922
431
296.56951
{ "bin_edges": [ 13, 940, 1867, 2794, 3721, 4648, 5575, 6502, 7429, 8356, 9280 ], "hist": [ 71847, 4660, 329, 37, 10, 3, 3, 0, 0, 1 ] }
false
1024m/mMGTD-Corpus
version v0.3
Persian
102,594
Modified text
0
0
15
122,274
1,994.73529
1,260
2,571.42612
{ "bin_edges": [ 15, 12241, 24467, 36693, 48919, 61145, 73371, 85597, 97823, 110049, 122274 ], "hist": [ 101720, 751, 82, 25, 11, 2, 1, 1, 0, 1 ] }
false
1024m/mMGTD-Corpus
version v0.3
Persian
102,594
Original text
0
0
13
122,274
2,033.73457
1,297
2,619.96276
{ "bin_edges": [ 13, 12240, 24467, 36694, 48921, 61148, 73375, 85602, 97829, 110056, 122274 ], "hist": [ 101681, 780, 89, 28, 11, 2, 1, 1, 0, 1 ] }
false
1024m/mMGTD-Corpus
version v0.3
Portuguese
214,694
Modified text
4
0.00002
18
158,104
1,478.25673
526
2,443.97973
{ "bin_edges": [ 18, 15827, 31636, 47445, 63254, 79063, 94872, 110681, 126490, 142299, 158104 ], "hist": [ 213978, 667, 33, 5, 5, 0, 1, 0, 0, 1 ] }
false
1024m/mMGTD-Corpus
version v0.3
Portuguese
214,694
Original text
0
0
18
158,104
1,489.21973
504
2,487.66123
{ "bin_edges": [ 18, 15827, 31636, 47445, 63254, 79063, 94872, 110681, 126490, 142299, 158104 ], "hist": [ 213946, 702, 34, 5, 5, 0, 1, 0, 0, 1 ] }
false
open-llm-leaderboard/arcee-ai__Arcee-Spark-details
arcee-ai__Arcee-Spark__leaderboard_bbh_tracking_shuffled_objects_five_objects
2024_07_22T09_52_59.637349
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/arcee-ai__Arcee-Spark-details
arcee-ai__Arcee-Spark__leaderboard_bbh_tracking_shuffled_objects_five_objects
2024_07_22T09_52_59.637349
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/arcee-ai__Arcee-Spark-details
arcee-ai__Arcee-Spark__leaderboard_bbh_tracking_shuffled_objects_five_objects
latest
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/arcee-ai__Arcee-Spark-details
arcee-ai__Arcee-Spark__leaderboard_bbh_tracking_shuffled_objects_five_objects
latest
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
1024m/mMGTD-Corpus
version v0.3
Turkish
124,472
Modified text
0
0
11
39,334
1,170.6495
543
1,688.26365
{ "bin_edges": [ 11, 3944, 7877, 11810, 15743, 19676, 23609, 27542, 31475, 35408, 39334 ], "hist": [ 116764, 6208, 1163, 239, 64, 22, 3, 4, 3, 2 ] }
false
1024m/mMGTD-Corpus
version v0.3
Turkish
124,472
Original text
0
0
11
39,334
1,174.08829
532
1,715.83833
{ "bin_edges": [ 11, 3944, 7877, 11810, 15743, 19676, 23609, 27542, 31475, 35408, 39334 ], "hist": [ 116558, 6363, 1206, 242, 65, 24, 3, 4, 4, 3 ] }
false
1024m/mMGTD-Corpus
version v0.3
Spanish
202,096
Modified text
1
0
6
57,912
1,541.26238
563
2,477.20809
{ "bin_edges": [ 6, 5797, 11588, 17379, 23170, 28961, 34752, 40543, 46334, 52125, 57912 ], "hist": [ 185897, 14443, 1438, 245, 47, 16, 2, 1, 5, 1 ] }
false
1024m/mMGTD-Corpus
version v0.3
Spanish
202,096
Original text
0
0
6
57,912
1,553.02924
544
2,526.30424
{ "bin_edges": [ 6, 5797, 11588, 17379, 23170, 28961, 34752, 40543, 46334, 52125, 57912 ], "hist": [ 185206, 15042, 1511, 264, 47, 17, 2, 1, 5, 1 ] }
false
1024m/mMGTD-Corpus
version v0.3
Russian
207,875
Modified text
2
0.00001
14
119,088
1,749.63001
604
2,788.2701
{ "bin_edges": [ 14, 11922, 23830, 35738, 47646, 59554, 71462, 83370, 95278, 107186, 119088 ], "hist": [ 205325, 2356, 150, 28, 7, 2, 0, 1, 3, 1 ] }
false
1024m/mMGTD-Corpus
version v0.3
Russian
207,875
Original text
0
0
14
119,088
1,758.9954
596
2,811.35248
{ "bin_edges": [ 14, 11922, 23830, 35738, 47646, 59554, 71462, 83370, 95278, 107186, 119088 ], "hist": [ 205289, 2391, 151, 29, 7, 2, 1, 1, 3, 1 ] }
false
1024m/mMGTD-Corpus
version v0.3
Ukrainian
158,118
Modified text
2
0.00001
9
127,676
1,298.1319
530
2,221.67497
{ "bin_edges": [ 9, 12776, 25543, 38310, 51077, 63844, 76611, 89378, 102145, 114912, 127676 ], "hist": [ 157426, 604, 55, 22, 3, 1, 3, 0, 1, 1 ] }
false
1024m/mMGTD-Corpus
version v0.3
Ukrainian
158,118
Original text
0
0
9
127,676
1,304.37919
522
2,243.81306
{ "bin_edges": [ 9, 12776, 25543, 38310, 51077, 63844, 76611, 89378, 102145, 114912, 127676 ], "hist": [ 157409, 620, 58, 22, 3, 1, 3, 0, 1, 1 ] }
false
1024m/mMGTD-Corpus
version v0.3
Vietnamese
129,747
Modified text
7
0.00005
12
164,135
1,603.37008
513
3,342.86149
{ "bin_edges": [ 12, 16425, 32838, 49251, 65664, 82077, 98490, 114903, 131316, 147729, 164135 ], "hist": [ 129080, 443, 124, 52, 20, 15, 5, 0, 0, 1 ] }
false
1024m/mMGTD-Corpus
version v0.3
Vietnamese
129,747
Original text
0
0
12
164,135
1,620.5668
503
3,413.86623
{ "bin_edges": [ 12, 16425, 32838, 49251, 65664, 82077, 98490, 114903, 131316, 147729, 164135 ], "hist": [ 129062, 453, 133, 55, 22, 16, 5, 0, 0, 1 ] }
false
open-llm-leaderboard/Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3-details
Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3__leaderboard_math_intermediate_algebra_hard
2024_07_22T10_01_37.284764
280
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 280 ] }
false
open-llm-leaderboard/Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3-details
Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3__leaderboard_math_intermediate_algebra_hard
2024_07_22T10_01_37.284764
280
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 280 ] }
false
open-llm-leaderboard/Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3-details
Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3__leaderboard_math_intermediate_algebra_hard
2024_07_22T10_01_37.284764
280
target
0
0
1
41
7.19643
4
6.56323
{ "bin_edges": [ 1, 6, 11, 16, 21, 26, 31, 36, 41, 41 ], "hist": [ 155, 39, 59, 17, 7, 1, 1, 0, 1 ] }
false
open-llm-leaderboard/Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3-details
Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3__leaderboard_math_intermediate_algebra_hard
2024_07_22T10_01_37.284764
280
target_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 280 ] }
false
open-llm-leaderboard/Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3-details
Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3__leaderboard_math_intermediate_algebra_hard
latest
280
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 280 ] }
false
open-llm-leaderboard/Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3-details
Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3__leaderboard_math_intermediate_algebra_hard
latest
280
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 280 ] }
false
open-llm-leaderboard/Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3-details
Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3__leaderboard_math_intermediate_algebra_hard
latest
280
target
0
0
1
41
7.19643
4
6.56323
{ "bin_edges": [ 1, 6, 11, 16, 21, 26, 31, 36, 41, 41 ], "hist": [ 155, 39, 59, 17, 7, 1, 1, 0, 1 ] }
false
open-llm-leaderboard/Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3-details
Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3__leaderboard_math_intermediate_algebra_hard
latest
280
target_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 280 ] }
false
open-llm-leaderboard/arcee-ai__Arcee-Spark-details
arcee-ai__Arcee-Spark__leaderboard_bbh_tracking_shuffled_objects_three_objects
2024_07_22T09_52_59.637349
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/arcee-ai__Arcee-Spark-details
arcee-ai__Arcee-Spark__leaderboard_bbh_tracking_shuffled_objects_three_objects
2024_07_22T09_52_59.637349
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/arcee-ai__Arcee-Spark-details
arcee-ai__Arcee-Spark__leaderboard_bbh_tracking_shuffled_objects_three_objects
latest
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/arcee-ai__Arcee-Spark-details
arcee-ai__Arcee-Spark__leaderboard_bbh_tracking_shuffled_objects_three_objects
latest
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3-details
Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3__leaderboard_math_num_theory_hard
2024_07_22T10_01_37.284764
154
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 154 ] }
false
open-llm-leaderboard/Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3-details
Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3__leaderboard_math_num_theory_hard
2024_07_22T10_01_37.284764
154
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 154 ] }
false
open-llm-leaderboard/Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3-details
Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3__leaderboard_math_num_theory_hard
2024_07_22T10_01_37.284764
154
target
0
0
1
15
2.68182
2
2.25703
{ "bin_edges": [ 1, 3, 5, 7, 9, 11, 13, 15, 15 ], "hist": [ 99, 40, 7, 1, 2, 4, 0, 1 ] }
false
open-llm-leaderboard/Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3-details
Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3__leaderboard_math_num_theory_hard
2024_07_22T10_01_37.284764
154
target_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 154 ] }
false
open-llm-leaderboard/Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3-details
Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3__leaderboard_math_num_theory_hard
latest
154
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 154 ] }
false
open-llm-leaderboard/Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3-details
Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3__leaderboard_math_num_theory_hard
latest
154
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 154 ] }
false
open-llm-leaderboard/Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3-details
Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3__leaderboard_math_num_theory_hard
latest
154
target
0
0
1
15
2.68182
2
2.25703
{ "bin_edges": [ 1, 3, 5, 7, 9, 11, 13, 15, 15 ], "hist": [ 99, 40, 7, 1, 2, 4, 0, 1 ] }
false
open-llm-leaderboard/Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3-details
Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3__leaderboard_math_num_theory_hard
latest
154
target_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 154 ] }
false
open-llm-leaderboard/arcee-ai__Arcee-Spark-details
arcee-ai__Arcee-Spark__leaderboard_bbh_tracking_shuffled_objects_seven_objects
2024_07_22T09_52_59.637349
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/arcee-ai__Arcee-Spark-details
arcee-ai__Arcee-Spark__leaderboard_bbh_tracking_shuffled_objects_seven_objects
2024_07_22T09_52_59.637349
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/arcee-ai__Arcee-Spark-details
arcee-ai__Arcee-Spark__leaderboard_bbh_tracking_shuffled_objects_seven_objects
latest
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/arcee-ai__Arcee-Spark-details
arcee-ai__Arcee-Spark__leaderboard_bbh_tracking_shuffled_objects_seven_objects
latest
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3-details
Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3__leaderboard_math_algebra_hard
2024_07_22T10_01_37.284764
307
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 307 ] }
false
open-llm-leaderboard/Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3-details
Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3__leaderboard_math_algebra_hard
2024_07_22T10_01_37.284764
307
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 307 ] }
false
open-llm-leaderboard/Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3-details
Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3__leaderboard_math_algebra_hard
2024_07_22T10_01_37.284764
307
target
0
0
1
34
5.35505
3
6.10965
{ "bin_edges": [ 1, 5, 9, 13, 17, 21, 25, 29, 33, 34 ], "hist": [ 202, 33, 38, 19, 1, 9, 0, 3, 2 ] }
false
open-llm-leaderboard/Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3-details
Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3__leaderboard_math_algebra_hard
2024_07_22T10_01_37.284764
307
target_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 307 ] }
false
open-llm-leaderboard/Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3-details
Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3__leaderboard_math_algebra_hard
latest
307
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 307 ] }
false
open-llm-leaderboard/Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3-details
Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3__leaderboard_math_algebra_hard
latest
307
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 307 ] }
false
open-llm-leaderboard/Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3-details
Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3__leaderboard_math_algebra_hard
latest
307
target
0
0
1
34
5.35505
3
6.10965
{ "bin_edges": [ 1, 5, 9, 13, 17, 21, 25, 29, 33, 34 ], "hist": [ 202, 33, 38, 19, 1, 9, 0, 3, 2 ] }
false
open-llm-leaderboard/Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3-details
Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3__leaderboard_math_algebra_hard
latest
307
target_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 307 ] }
false
open-llm-leaderboard/arcee-ai__Arcee-Spark-details
arcee-ai__Arcee-Spark__leaderboard_math_counting_and_prob_hard
2024_07_22T09_52_59.637349
123
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 123 ] }
false
open-llm-leaderboard/arcee-ai__Arcee-Spark-details
arcee-ai__Arcee-Spark__leaderboard_math_counting_and_prob_hard
2024_07_22T09_52_59.637349
123
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 123 ] }
false
open-llm-leaderboard/arcee-ai__Arcee-Spark-details
arcee-ai__Arcee-Spark__leaderboard_math_counting_and_prob_hard
2024_07_22T09_52_59.637349
123
target
0
0
1
18
8.09756
11
5.31224
{ "bin_edges": [ 1, 3, 5, 7, 9, 11, 13, 15, 17, 18 ], "hist": [ 31, 20, 7, 1, 0, 28, 25, 10, 1 ] }
false
open-llm-leaderboard/arcee-ai__Arcee-Spark-details
arcee-ai__Arcee-Spark__leaderboard_math_counting_and_prob_hard
2024_07_22T09_52_59.637349
123
target_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 123 ] }
false
open-llm-leaderboard/arcee-ai__Arcee-Spark-details
arcee-ai__Arcee-Spark__leaderboard_math_counting_and_prob_hard
latest
123
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 123 ] }
false
open-llm-leaderboard/arcee-ai__Arcee-Spark-details
arcee-ai__Arcee-Spark__leaderboard_math_counting_and_prob_hard
latest
123
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 123 ] }
false
open-llm-leaderboard/arcee-ai__Arcee-Spark-details
arcee-ai__Arcee-Spark__leaderboard_math_counting_and_prob_hard
latest
123
target
0
0
1
18
8.09756
11
5.31224
{ "bin_edges": [ 1, 3, 5, 7, 9, 11, 13, 15, 17, 18 ], "hist": [ 31, 20, 7, 1, 0, 28, 25, 10, 1 ] }
false
open-llm-leaderboard/arcee-ai__Arcee-Spark-details
arcee-ai__Arcee-Spark__leaderboard_math_counting_and_prob_hard
latest
123
target_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 123 ] }
false
open-llm-leaderboard/Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3-details
Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3__leaderboard_math_prealgebra_hard
2024_07_22T10_01_37.284764
193
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 193 ] }
false
open-llm-leaderboard/Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3-details
Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3__leaderboard_math_prealgebra_hard
2024_07_22T10_01_37.284764
193
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 193 ] }
false
open-llm-leaderboard/Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3-details
Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3__leaderboard_math_prealgebra_hard
2024_07_22T10_01_37.284764
193
target
0
0
1
20
4.68394
3
4.33957
{ "bin_edges": [ 1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 20 ], "hist": [ 87, 52, 10, 3, 6, 16, 13, 4, 1, 1 ] }
false
open-llm-leaderboard/Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3-details
Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3__leaderboard_math_prealgebra_hard
2024_07_22T10_01_37.284764
193
target_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 193 ] }
false
open-llm-leaderboard/Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3-details
Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3__leaderboard_math_prealgebra_hard
latest
193
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 193 ] }
false
open-llm-leaderboard/Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3-details
Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3__leaderboard_math_prealgebra_hard
latest
193
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 193 ] }
false
open-llm-leaderboard/Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3-details
Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3__leaderboard_math_prealgebra_hard
latest
193
target
0
0
1
20
4.68394
3
4.33957
{ "bin_edges": [ 1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 20 ], "hist": [ 87, 52, 10, 3, 6, 16, 13, 4, 1, 1 ] }
false