picocreator commited on
Commit
f030095
·
1 Parent(s): 726c775

all RWKV based results

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. lm-eval-output/RWKV/rwkv-4-world-1b5/ai2_arc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +3 -0
  2. lm-eval-output/RWKV/rwkv-4-world-1b5/ai2_arc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +132 -0
  3. lm-eval-output/RWKV/rwkv-4-world-1b5/ai2_arc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +3 -0
  4. lm-eval-output/RWKV/rwkv-4-world-1b5/anli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +3 -0
  5. lm-eval-output/RWKV/rwkv-4-world-1b5/anli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +161 -0
  6. lm-eval-output/RWKV/rwkv-4-world-1b5/anli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +3 -0
  7. lm-eval-output/RWKV/rwkv-4-world-1b5/arc_challenge/dtype=bfloat16,trust_remote_code=True-num_fewshot=1-nvidia-gpu/result-jsonl.tar.gz +3 -0
  8. lm-eval-output/RWKV/rwkv-4-world-1b5/arc_challenge/dtype=bfloat16,trust_remote_code=True-num_fewshot=1-nvidia-gpu/results.json +70 -0
  9. lm-eval-output/RWKV/rwkv-4-world-1b5/arc_challenge/dtype=bfloat16,trust_remote_code=True-num_fewshot=1-nvidia-gpu/taskrun.log +3 -0
  10. lm-eval-output/RWKV/rwkv-4-world-1b5/arc_challenge/dtype=bfloat16,trust_remote_code=True-num_fewshot=10-nvidia-gpu/result-jsonl.tar.gz +3 -0
  11. lm-eval-output/RWKV/rwkv-4-world-1b5/arc_challenge/dtype=bfloat16,trust_remote_code=True-num_fewshot=10-nvidia-gpu/results.json +70 -0
  12. lm-eval-output/RWKV/rwkv-4-world-1b5/arc_challenge/dtype=bfloat16,trust_remote_code=True-num_fewshot=10-nvidia-gpu/taskrun.log +3 -0
  13. lm-eval-output/RWKV/rwkv-4-world-1b5/arc_challenge/dtype=bfloat16,trust_remote_code=True-num_fewshot=2-nvidia-gpu/result-jsonl.tar.gz +3 -0
  14. lm-eval-output/RWKV/rwkv-4-world-1b5/arc_challenge/dtype=bfloat16,trust_remote_code=True-num_fewshot=2-nvidia-gpu/results.json +70 -0
  15. lm-eval-output/RWKV/rwkv-4-world-1b5/arc_challenge/dtype=bfloat16,trust_remote_code=True-num_fewshot=2-nvidia-gpu/taskrun.log +3 -0
  16. lm-eval-output/RWKV/rwkv-4-world-1b5/arc_challenge/dtype=bfloat16,trust_remote_code=True-num_fewshot=25-nvidia-gpu/result-jsonl.tar.gz +3 -0
  17. lm-eval-output/RWKV/rwkv-4-world-1b5/arc_challenge/dtype=bfloat16,trust_remote_code=True-num_fewshot=25-nvidia-gpu/results.json +70 -0
  18. lm-eval-output/RWKV/rwkv-4-world-1b5/arc_challenge/dtype=bfloat16,trust_remote_code=True-num_fewshot=25-nvidia-gpu/taskrun.log +3 -0
  19. lm-eval-output/RWKV/rwkv-4-world-1b5/arc_challenge/dtype=bfloat16,trust_remote_code=True-num_fewshot=5-nvidia-gpu/result-jsonl.tar.gz +3 -0
  20. lm-eval-output/RWKV/rwkv-4-world-1b5/arc_challenge/dtype=bfloat16,trust_remote_code=True-num_fewshot=5-nvidia-gpu/results.json +70 -0
  21. lm-eval-output/RWKV/rwkv-4-world-1b5/arc_challenge/dtype=bfloat16,trust_remote_code=True-num_fewshot=5-nvidia-gpu/taskrun.log +3 -0
  22. lm-eval-output/RWKV/rwkv-4-world-1b5/arithmetic/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +3 -0
  23. lm-eval-output/RWKV/rwkv-4-world-1b5/arithmetic/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +378 -0
  24. lm-eval-output/RWKV/rwkv-4-world-1b5/arithmetic/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +3 -0
  25. lm-eval-output/RWKV/rwkv-4-world-1b5/arithmetic__/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +3 -0
  26. lm-eval-output/RWKV/rwkv-4-world-1b5/arithmetic__/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +364 -0
  27. lm-eval-output/RWKV/rwkv-4-world-1b5/arithmetic__/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +3 -0
  28. lm-eval-output/RWKV/rwkv-4-world-1b5/asdiv/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +3 -0
  29. lm-eval-output/RWKV/rwkv-4-world-1b5/asdiv/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +55 -0
  30. lm-eval-output/RWKV/rwkv-4-world-1b5/asdiv/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +3 -0
  31. lm-eval-output/RWKV/rwkv-4-world-1b5/blimp/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +3 -0
  32. lm-eval-output/RWKV/rwkv-4-world-1b5/blimp/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +2249 -0
  33. lm-eval-output/RWKV/rwkv-4-world-1b5/blimp/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +3 -0
  34. lm-eval-output/RWKV/rwkv-4-world-1b5/boolq/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +3 -0
  35. lm-eval-output/RWKV/rwkv-4-world-1b5/boolq/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +62 -0
  36. lm-eval-output/RWKV/rwkv-4-world-1b5/boolq/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +3 -0
  37. lm-eval-output/RWKV/rwkv-4-world-1b5/cb/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +3 -0
  38. lm-eval-output/RWKV/rwkv-4-world-1b5/cb/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +68 -0
  39. lm-eval-output/RWKV/rwkv-4-world-1b5/cb/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +3 -0
  40. lm-eval-output/RWKV/rwkv-4-world-1b5/ceval-valid/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +3 -0
  41. lm-eval-output/RWKV/rwkv-4-world-1b5/ceval-valid/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +2590 -0
  42. lm-eval-output/RWKV/rwkv-4-world-1b5/ceval-valid/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +3 -0
  43. lm-eval-output/RWKV/rwkv-4-world-1b5/cmmlu/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +3 -0
  44. lm-eval-output/RWKV/rwkv-4-world-1b5/cmmlu/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +0 -0
  45. lm-eval-output/RWKV/rwkv-4-world-1b5/cmmlu/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +3 -0
  46. lm-eval-output/RWKV/rwkv-4-world-1b5/cola/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +3 -0
  47. lm-eval-output/RWKV/rwkv-4-world-1b5/cola/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +60 -0
  48. lm-eval-output/RWKV/rwkv-4-world-1b5/cola/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +3 -0
  49. lm-eval-output/RWKV/rwkv-4-world-1b5/copa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +3 -0
  50. lm-eval-output/RWKV/rwkv-4-world-1b5/copa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +58 -0
lm-eval-output/RWKV/rwkv-4-world-1b5/ai2_arc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:59b2a7caaff669b6eb9132471d728fe321c89e6ad3234be9d3092602be9405ca
3
+ size 681678
lm-eval-output/RWKV/rwkv-4-world-1b5/ai2_arc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json ADDED
@@ -0,0 +1,132 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "ai2_arc": {
4
+ "acc,none": 0.4811161217587373,
5
+ "acc_stderr,none": 0.05031958627263597,
6
+ "acc_norm,none": 0.44701240135287484,
7
+ "acc_norm_stderr,none": 0.033658934383932894,
8
+ "alias": "ai2_arc"
9
+ },
10
+ "arc_challenge": {
11
+ "acc,none": 0.2696245733788396,
12
+ "acc_stderr,none": 0.012968040686869154,
13
+ "acc_norm,none": 0.310580204778157,
14
+ "acc_norm_stderr,none": 0.013522292098053059,
15
+ "alias": " - arc_challenge"
16
+ },
17
+ "arc_easy": {
18
+ "acc,none": 0.5854377104377104,
19
+ "acc_stderr,none": 0.01010888921244777,
20
+ "acc_norm,none": 0.5143097643097643,
21
+ "acc_norm_stderr,none": 0.010255580881603625,
22
+ "alias": " - arc_easy"
23
+ }
24
+ },
25
+ "groups": {
26
+ "ai2_arc": {
27
+ "acc,none": 0.4811161217587373,
28
+ "acc_stderr,none": 0.05031958627263597,
29
+ "acc_norm,none": 0.44701240135287484,
30
+ "acc_norm_stderr,none": 0.033658934383932894,
31
+ "alias": "ai2_arc"
32
+ }
33
+ },
34
+ "configs": {
35
+ "arc_challenge": {
36
+ "task": "arc_challenge",
37
+ "group": [
38
+ "ai2_arc"
39
+ ],
40
+ "dataset_path": "allenai/ai2_arc",
41
+ "dataset_name": "ARC-Challenge",
42
+ "training_split": "train",
43
+ "validation_split": "validation",
44
+ "test_split": "test",
45
+ "doc_to_text": "Question: {{question}}\nAnswer:",
46
+ "doc_to_target": "{{choices.label.index(answerKey)}}",
47
+ "doc_to_choice": "{{choices.text}}",
48
+ "description": "",
49
+ "target_delimiter": " ",
50
+ "fewshot_delimiter": "\n\n",
51
+ "metric_list": [
52
+ {
53
+ "metric": "acc",
54
+ "aggregation": "mean",
55
+ "higher_is_better": true
56
+ },
57
+ {
58
+ "metric": "acc_norm",
59
+ "aggregation": "mean",
60
+ "higher_is_better": true
61
+ }
62
+ ],
63
+ "output_type": "multiple_choice",
64
+ "repeats": 1,
65
+ "should_decontaminate": true,
66
+ "doc_to_decontamination_query": "Question: {{question}}\nAnswer:",
67
+ "metadata": {
68
+ "version": 1.0
69
+ }
70
+ },
71
+ "arc_easy": {
72
+ "task": "arc_easy",
73
+ "group": [
74
+ "ai2_arc"
75
+ ],
76
+ "dataset_path": "allenai/ai2_arc",
77
+ "dataset_name": "ARC-Easy",
78
+ "training_split": "train",
79
+ "validation_split": "validation",
80
+ "test_split": "test",
81
+ "doc_to_text": "Question: {{question}}\nAnswer:",
82
+ "doc_to_target": "{{choices.label.index(answerKey)}}",
83
+ "doc_to_choice": "{{choices.text}}",
84
+ "description": "",
85
+ "target_delimiter": " ",
86
+ "fewshot_delimiter": "\n\n",
87
+ "metric_list": [
88
+ {
89
+ "metric": "acc",
90
+ "aggregation": "mean",
91
+ "higher_is_better": true
92
+ },
93
+ {
94
+ "metric": "acc_norm",
95
+ "aggregation": "mean",
96
+ "higher_is_better": true
97
+ }
98
+ ],
99
+ "output_type": "multiple_choice",
100
+ "repeats": 1,
101
+ "should_decontaminate": true,
102
+ "doc_to_decontamination_query": "Question: {{question}}\nAnswer:",
103
+ "metadata": {
104
+ "version": 1.0
105
+ }
106
+ }
107
+ },
108
+ "versions": {
109
+ "ai2_arc": "N/A",
110
+ "arc_challenge": 1.0,
111
+ "arc_easy": 1.0
112
+ },
113
+ "n-shot": {
114
+ "ai2_arc": 0,
115
+ "arc_challenge": 0,
116
+ "arc_easy": 0
117
+ },
118
+ "config": {
119
+ "model": "hf",
120
+ "model_args": "pretrained=RWKV/rwkv-4-world-1b5,dtype=bfloat16,trust_remote_code=True",
121
+ "batch_size": "auto",
122
+ "batch_sizes": [
123
+ 64
124
+ ],
125
+ "device": null,
126
+ "use_cache": null,
127
+ "limit": null,
128
+ "bootstrap_iters": 100000,
129
+ "gen_kwargs": null
130
+ },
131
+ "git_hash": "99f5004"
132
+ }
lm-eval-output/RWKV/rwkv-4-world-1b5/ai2_arc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9657e6cf4615943a3cc413a40181c2c048c716c0dbe55b910e03c461daeb87b7
3
+ size 13597
lm-eval-output/RWKV/rwkv-4-world-1b5/anli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1f50a9d069f9fd13c6a21d8c78516d724526f39fadf9fb949523417d5a0ec111
3
+ size 1078126
lm-eval-output/RWKV/rwkv-4-world-1b5/anli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json ADDED
@@ -0,0 +1,161 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "anli": {
4
+ "acc,none": 0.334375,
5
+ "acc_stderr,none": 0.02042049510113634,
6
+ "alias": "anli"
7
+ },
8
+ "anli_r1": {
9
+ "acc,none": 0.307,
10
+ "acc_stderr,none": 0.014593284892852621,
11
+ "alias": " - anli_r1"
12
+ },
13
+ "anli_r2": {
14
+ "acc,none": 0.33,
15
+ "acc_stderr,none": 0.014876872027456732,
16
+ "alias": " - anli_r2"
17
+ },
18
+ "anli_r3": {
19
+ "acc,none": 0.36083333333333334,
20
+ "acc_stderr,none": 0.013869180252444865,
21
+ "alias": " - anli_r3"
22
+ }
23
+ },
24
+ "groups": {
25
+ "anli": {
26
+ "acc,none": 0.334375,
27
+ "acc_stderr,none": 0.02042049510113634,
28
+ "alias": "anli"
29
+ }
30
+ },
31
+ "configs": {
32
+ "anli_r1": {
33
+ "task": "anli_r1",
34
+ "group": [
35
+ "anli"
36
+ ],
37
+ "dataset_path": "anli",
38
+ "training_split": "train_r1",
39
+ "validation_split": "dev_r1",
40
+ "test_split": "test_r1",
41
+ "doc_to_text": "{{premise}}\nQuestion: {{hypothesis}} True, False, or Neither?\nAnswer:",
42
+ "doc_to_target": "{{['True', 'Neither', 'False'][label]}}",
43
+ "doc_to_choice": [
44
+ "True",
45
+ "Neither",
46
+ "False"
47
+ ],
48
+ "description": "",
49
+ "target_delimiter": " ",
50
+ "fewshot_delimiter": "\n\n",
51
+ "metric_list": [
52
+ {
53
+ "metric": "acc",
54
+ "aggregation": "mean",
55
+ "higher_is_better": true
56
+ }
57
+ ],
58
+ "output_type": "multiple_choice",
59
+ "repeats": 1,
60
+ "should_decontaminate": true,
61
+ "doc_to_decontamination_query": "premise",
62
+ "metadata": {
63
+ "version": 1.0
64
+ }
65
+ },
66
+ "anli_r2": {
67
+ "task": "anli_r2",
68
+ "group": [
69
+ "anli"
70
+ ],
71
+ "dataset_path": "anli",
72
+ "training_split": "train_r2",
73
+ "validation_split": "dev_r2",
74
+ "test_split": "test_r2",
75
+ "doc_to_text": "{{premise}}\nQuestion: {{hypothesis}} True, False, or Neither?\nAnswer:",
76
+ "doc_to_target": "{{['True', 'Neither', 'False'][label]}}",
77
+ "doc_to_choice": [
78
+ "True",
79
+ "Neither",
80
+ "False"
81
+ ],
82
+ "description": "",
83
+ "target_delimiter": " ",
84
+ "fewshot_delimiter": "\n\n",
85
+ "metric_list": [
86
+ {
87
+ "metric": "acc",
88
+ "aggregation": "mean",
89
+ "higher_is_better": true
90
+ }
91
+ ],
92
+ "output_type": "multiple_choice",
93
+ "repeats": 1,
94
+ "should_decontaminate": true,
95
+ "doc_to_decontamination_query": "premise",
96
+ "metadata": {
97
+ "version": 1.0
98
+ }
99
+ },
100
+ "anli_r3": {
101
+ "task": "anli_r3",
102
+ "group": [
103
+ "anli"
104
+ ],
105
+ "dataset_path": "anli",
106
+ "training_split": "train_r3",
107
+ "validation_split": "dev_r3",
108
+ "test_split": "test_r3",
109
+ "doc_to_text": "{{premise}}\nQuestion: {{hypothesis}} True, False, or Neither?\nAnswer:",
110
+ "doc_to_target": "{{['True', 'Neither', 'False'][label]}}",
111
+ "doc_to_choice": [
112
+ "True",
113
+ "Neither",
114
+ "False"
115
+ ],
116
+ "description": "",
117
+ "target_delimiter": " ",
118
+ "fewshot_delimiter": "\n\n",
119
+ "metric_list": [
120
+ {
121
+ "metric": "acc",
122
+ "aggregation": "mean",
123
+ "higher_is_better": true
124
+ }
125
+ ],
126
+ "output_type": "multiple_choice",
127
+ "repeats": 1,
128
+ "should_decontaminate": true,
129
+ "doc_to_decontamination_query": "premise",
130
+ "metadata": {
131
+ "version": 1.0
132
+ }
133
+ }
134
+ },
135
+ "versions": {
136
+ "anli": "N/A",
137
+ "anli_r1": 1.0,
138
+ "anli_r2": 1.0,
139
+ "anli_r3": 1.0
140
+ },
141
+ "n-shot": {
142
+ "anli": 0,
143
+ "anli_r1": 0,
144
+ "anli_r2": 0,
145
+ "anli_r3": 0
146
+ },
147
+ "config": {
148
+ "model": "hf",
149
+ "model_args": "pretrained=RWKV/rwkv-4-world-1b5,dtype=bfloat16,trust_remote_code=True",
150
+ "batch_size": "auto",
151
+ "batch_sizes": [
152
+ 64
153
+ ],
154
+ "device": null,
155
+ "use_cache": null,
156
+ "limit": null,
157
+ "bootstrap_iters": 100000,
158
+ "gen_kwargs": null
159
+ },
160
+ "git_hash": "99f5004"
161
+ }
lm-eval-output/RWKV/rwkv-4-world-1b5/anli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7b03db730e4f56e503d20df5a8ba10f894f65720fd2173692c9528a43d3d39e4
3
+ size 14840
lm-eval-output/RWKV/rwkv-4-world-1b5/arc_challenge/dtype=bfloat16,trust_remote_code=True-num_fewshot=1-nvidia-gpu/result-jsonl.tar.gz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f8f3cf7e19994a9732e9d2d8d12710cf682e85f1c6361722a6dd49f6eea4f1da
3
+ size 328530
lm-eval-output/RWKV/rwkv-4-world-1b5/arc_challenge/dtype=bfloat16,trust_remote_code=True-num_fewshot=1-nvidia-gpu/results.json ADDED
@@ -0,0 +1,70 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "arc_challenge": {
4
+ "acc,none": 0.2713310580204778,
5
+ "acc_stderr,none": 0.012993807727545789,
6
+ "acc_norm,none": 0.30802047781569963,
7
+ "acc_norm_stderr,none": 0.01349142951729204,
8
+ "alias": "arc_challenge"
9
+ }
10
+ },
11
+ "configs": {
12
+ "arc_challenge": {
13
+ "task": "arc_challenge",
14
+ "group": [
15
+ "ai2_arc"
16
+ ],
17
+ "dataset_path": "allenai/ai2_arc",
18
+ "dataset_name": "ARC-Challenge",
19
+ "training_split": "train",
20
+ "validation_split": "validation",
21
+ "test_split": "test",
22
+ "doc_to_text": "Question: {{question}}\nAnswer:",
23
+ "doc_to_target": "{{choices.label.index(answerKey)}}",
24
+ "doc_to_choice": "{{choices.text}}",
25
+ "description": "",
26
+ "target_delimiter": " ",
27
+ "fewshot_delimiter": "\n\n",
28
+ "num_fewshot": 1,
29
+ "metric_list": [
30
+ {
31
+ "metric": "acc",
32
+ "aggregation": "mean",
33
+ "higher_is_better": true
34
+ },
35
+ {
36
+ "metric": "acc_norm",
37
+ "aggregation": "mean",
38
+ "higher_is_better": true
39
+ }
40
+ ],
41
+ "output_type": "multiple_choice",
42
+ "repeats": 1,
43
+ "should_decontaminate": true,
44
+ "doc_to_decontamination_query": "Question: {{question}}\nAnswer:",
45
+ "metadata": {
46
+ "version": 1.0
47
+ }
48
+ }
49
+ },
50
+ "versions": {
51
+ "arc_challenge": 1.0
52
+ },
53
+ "n-shot": {
54
+ "arc_challenge": 1
55
+ },
56
+ "config": {
57
+ "model": "hf",
58
+ "model_args": "pretrained=RWKV/rwkv-4-world-1b5,dtype=bfloat16,trust_remote_code=True",
59
+ "batch_size": "auto",
60
+ "batch_sizes": [
61
+ 64
62
+ ],
63
+ "device": null,
64
+ "use_cache": null,
65
+ "limit": null,
66
+ "bootstrap_iters": 100000,
67
+ "gen_kwargs": null
68
+ },
69
+ "git_hash": "99f5004"
70
+ }
lm-eval-output/RWKV/rwkv-4-world-1b5/arc_challenge/dtype=bfloat16,trust_remote_code=True-num_fewshot=1-nvidia-gpu/taskrun.log ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2aa98c66aea8f3313d74a45e0dc3493cf25a0e644fae01eec267d119835dbfb6
3
+ size 12197
lm-eval-output/RWKV/rwkv-4-world-1b5/arc_challenge/dtype=bfloat16,trust_remote_code=True-num_fewshot=10-nvidia-gpu/result-jsonl.tar.gz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d814d17319946128ce333da1c1b59ecc6dbfe82c35aa6eed6a41ea1a2c6d8317
3
+ size 1076385
lm-eval-output/RWKV/rwkv-4-world-1b5/arc_challenge/dtype=bfloat16,trust_remote_code=True-num_fewshot=10-nvidia-gpu/results.json ADDED
@@ -0,0 +1,70 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "arc_challenge": {
4
+ "acc,none": 0.28668941979522183,
5
+ "acc_stderr,none": 0.013214986329274774,
6
+ "acc_norm,none": 0.31399317406143346,
7
+ "acc_norm_stderr,none": 0.013562691224726291,
8
+ "alias": "arc_challenge"
9
+ }
10
+ },
11
+ "configs": {
12
+ "arc_challenge": {
13
+ "task": "arc_challenge",
14
+ "group": [
15
+ "ai2_arc"
16
+ ],
17
+ "dataset_path": "allenai/ai2_arc",
18
+ "dataset_name": "ARC-Challenge",
19
+ "training_split": "train",
20
+ "validation_split": "validation",
21
+ "test_split": "test",
22
+ "doc_to_text": "Question: {{question}}\nAnswer:",
23
+ "doc_to_target": "{{choices.label.index(answerKey)}}",
24
+ "doc_to_choice": "{{choices.text}}",
25
+ "description": "",
26
+ "target_delimiter": " ",
27
+ "fewshot_delimiter": "\n\n",
28
+ "num_fewshot": 10,
29
+ "metric_list": [
30
+ {
31
+ "metric": "acc",
32
+ "aggregation": "mean",
33
+ "higher_is_better": true
34
+ },
35
+ {
36
+ "metric": "acc_norm",
37
+ "aggregation": "mean",
38
+ "higher_is_better": true
39
+ }
40
+ ],
41
+ "output_type": "multiple_choice",
42
+ "repeats": 1,
43
+ "should_decontaminate": true,
44
+ "doc_to_decontamination_query": "Question: {{question}}\nAnswer:",
45
+ "metadata": {
46
+ "version": 1.0
47
+ }
48
+ }
49
+ },
50
+ "versions": {
51
+ "arc_challenge": 1.0
52
+ },
53
+ "n-shot": {
54
+ "arc_challenge": 10
55
+ },
56
+ "config": {
57
+ "model": "hf",
58
+ "model_args": "pretrained=RWKV/rwkv-4-world-1b5,dtype=bfloat16,trust_remote_code=True",
59
+ "batch_size": "auto",
60
+ "batch_sizes": [
61
+ 64
62
+ ],
63
+ "device": null,
64
+ "use_cache": null,
65
+ "limit": null,
66
+ "bootstrap_iters": 100000,
67
+ "gen_kwargs": null
68
+ },
69
+ "git_hash": "99f5004"
70
+ }
lm-eval-output/RWKV/rwkv-4-world-1b5/arc_challenge/dtype=bfloat16,trust_remote_code=True-num_fewshot=10-nvidia-gpu/taskrun.log ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c37204e7dd84a816c39badb7f327ef522c1cd464e8efe064a53c145a1f253580
3
+ size 12275
lm-eval-output/RWKV/rwkv-4-world-1b5/arc_challenge/dtype=bfloat16,trust_remote_code=True-num_fewshot=2-nvidia-gpu/result-jsonl.tar.gz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:08fb1cac17f6bdabfff60b01c509cb3e8ee4633dd44f082831a4ae802e5552d2
3
+ size 423894
lm-eval-output/RWKV/rwkv-4-world-1b5/arc_challenge/dtype=bfloat16,trust_remote_code=True-num_fewshot=2-nvidia-gpu/results.json ADDED
@@ -0,0 +1,70 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "arc_challenge": {
4
+ "acc,none": 0.2901023890784983,
5
+ "acc_stderr,none": 0.013261573677520773,
6
+ "acc_norm,none": 0.3122866894197952,
7
+ "acc_norm_stderr,none": 0.013542598541688065,
8
+ "alias": "arc_challenge"
9
+ }
10
+ },
11
+ "configs": {
12
+ "arc_challenge": {
13
+ "task": "arc_challenge",
14
+ "group": [
15
+ "ai2_arc"
16
+ ],
17
+ "dataset_path": "allenai/ai2_arc",
18
+ "dataset_name": "ARC-Challenge",
19
+ "training_split": "train",
20
+ "validation_split": "validation",
21
+ "test_split": "test",
22
+ "doc_to_text": "Question: {{question}}\nAnswer:",
23
+ "doc_to_target": "{{choices.label.index(answerKey)}}",
24
+ "doc_to_choice": "{{choices.text}}",
25
+ "description": "",
26
+ "target_delimiter": " ",
27
+ "fewshot_delimiter": "\n\n",
28
+ "num_fewshot": 2,
29
+ "metric_list": [
30
+ {
31
+ "metric": "acc",
32
+ "aggregation": "mean",
33
+ "higher_is_better": true
34
+ },
35
+ {
36
+ "metric": "acc_norm",
37
+ "aggregation": "mean",
38
+ "higher_is_better": true
39
+ }
40
+ ],
41
+ "output_type": "multiple_choice",
42
+ "repeats": 1,
43
+ "should_decontaminate": true,
44
+ "doc_to_decontamination_query": "Question: {{question}}\nAnswer:",
45
+ "metadata": {
46
+ "version": 1.0
47
+ }
48
+ }
49
+ },
50
+ "versions": {
51
+ "arc_challenge": 1.0
52
+ },
53
+ "n-shot": {
54
+ "arc_challenge": 2
55
+ },
56
+ "config": {
57
+ "model": "hf",
58
+ "model_args": "pretrained=RWKV/rwkv-4-world-1b5,dtype=bfloat16,trust_remote_code=True",
59
+ "batch_size": "auto",
60
+ "batch_sizes": [
61
+ 64
62
+ ],
63
+ "device": null,
64
+ "use_cache": null,
65
+ "limit": null,
66
+ "bootstrap_iters": 100000,
67
+ "gen_kwargs": null
68
+ },
69
+ "git_hash": "99f5004"
70
+ }
lm-eval-output/RWKV/rwkv-4-world-1b5/arc_challenge/dtype=bfloat16,trust_remote_code=True-num_fewshot=2-nvidia-gpu/taskrun.log ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f61e035d6b81efded8f05c154684c4d777e982041f7e7af4a572a713862c915b
3
+ size 12197
lm-eval-output/RWKV/rwkv-4-world-1b5/arc_challenge/dtype=bfloat16,trust_remote_code=True-num_fewshot=25-nvidia-gpu/result-jsonl.tar.gz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4dd90017878f2ee98990abda13ef42a816cca246cd6c82ce8ef2303339f70376
3
+ size 2211160
lm-eval-output/RWKV/rwkv-4-world-1b5/arc_challenge/dtype=bfloat16,trust_remote_code=True-num_fewshot=25-nvidia-gpu/results.json ADDED
@@ -0,0 +1,70 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "arc_challenge": {
4
+ "acc,none": 0.28242320819112626,
5
+ "acc_stderr,none": 0.013155456884097224,
6
+ "acc_norm,none": 0.3122866894197952,
7
+ "acc_norm_stderr,none": 0.013542598541688065,
8
+ "alias": "arc_challenge"
9
+ }
10
+ },
11
+ "configs": {
12
+ "arc_challenge": {
13
+ "task": "arc_challenge",
14
+ "group": [
15
+ "ai2_arc"
16
+ ],
17
+ "dataset_path": "allenai/ai2_arc",
18
+ "dataset_name": "ARC-Challenge",
19
+ "training_split": "train",
20
+ "validation_split": "validation",
21
+ "test_split": "test",
22
+ "doc_to_text": "Question: {{question}}\nAnswer:",
23
+ "doc_to_target": "{{choices.label.index(answerKey)}}",
24
+ "doc_to_choice": "{{choices.text}}",
25
+ "description": "",
26
+ "target_delimiter": " ",
27
+ "fewshot_delimiter": "\n\n",
28
+ "num_fewshot": 25,
29
+ "metric_list": [
30
+ {
31
+ "metric": "acc",
32
+ "aggregation": "mean",
33
+ "higher_is_better": true
34
+ },
35
+ {
36
+ "metric": "acc_norm",
37
+ "aggregation": "mean",
38
+ "higher_is_better": true
39
+ }
40
+ ],
41
+ "output_type": "multiple_choice",
42
+ "repeats": 1,
43
+ "should_decontaminate": true,
44
+ "doc_to_decontamination_query": "Question: {{question}}\nAnswer:",
45
+ "metadata": {
46
+ "version": 1.0
47
+ }
48
+ }
49
+ },
50
+ "versions": {
51
+ "arc_challenge": 1.0
52
+ },
53
+ "n-shot": {
54
+ "arc_challenge": 25
55
+ },
56
+ "config": {
57
+ "model": "hf",
58
+ "model_args": "pretrained=RWKV/rwkv-4-world-1b5,dtype=bfloat16,trust_remote_code=True",
59
+ "batch_size": "auto",
60
+ "batch_sizes": [
61
+ 32
62
+ ],
63
+ "device": null,
64
+ "use_cache": null,
65
+ "limit": null,
66
+ "bootstrap_iters": 100000,
67
+ "gen_kwargs": null
68
+ },
69
+ "git_hash": "99f5004"
70
+ }
lm-eval-output/RWKV/rwkv-4-world-1b5/arc_challenge/dtype=bfloat16,trust_remote_code=True-num_fewshot=25-nvidia-gpu/taskrun.log ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e6e11c31f5c0a643a5754cdfe52c5e67119a76ffd9aa30de5b27ac152c6a08ed
3
+ size 14143
lm-eval-output/RWKV/rwkv-4-world-1b5/arc_challenge/dtype=bfloat16,trust_remote_code=True-num_fewshot=5-nvidia-gpu/result-jsonl.tar.gz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9607c8977920848df7cf1280e91787b5ae465dae532c5d48d9b9b9a52feabb8f
3
+ size 680876
lm-eval-output/RWKV/rwkv-4-world-1b5/arc_challenge/dtype=bfloat16,trust_remote_code=True-num_fewshot=5-nvidia-gpu/results.json ADDED
@@ -0,0 +1,70 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "arc_challenge": {
4
+ "acc,none": 0.2815699658703072,
5
+ "acc_stderr,none": 0.013143376735009019,
6
+ "acc_norm,none": 0.310580204778157,
7
+ "acc_norm_stderr,none": 0.013522292098053055,
8
+ "alias": "arc_challenge"
9
+ }
10
+ },
11
+ "configs": {
12
+ "arc_challenge": {
13
+ "task": "arc_challenge",
14
+ "group": [
15
+ "ai2_arc"
16
+ ],
17
+ "dataset_path": "allenai/ai2_arc",
18
+ "dataset_name": "ARC-Challenge",
19
+ "training_split": "train",
20
+ "validation_split": "validation",
21
+ "test_split": "test",
22
+ "doc_to_text": "Question: {{question}}\nAnswer:",
23
+ "doc_to_target": "{{choices.label.index(answerKey)}}",
24
+ "doc_to_choice": "{{choices.text}}",
25
+ "description": "",
26
+ "target_delimiter": " ",
27
+ "fewshot_delimiter": "\n\n",
28
+ "num_fewshot": 5,
29
+ "metric_list": [
30
+ {
31
+ "metric": "acc",
32
+ "aggregation": "mean",
33
+ "higher_is_better": true
34
+ },
35
+ {
36
+ "metric": "acc_norm",
37
+ "aggregation": "mean",
38
+ "higher_is_better": true
39
+ }
40
+ ],
41
+ "output_type": "multiple_choice",
42
+ "repeats": 1,
43
+ "should_decontaminate": true,
44
+ "doc_to_decontamination_query": "Question: {{question}}\nAnswer:",
45
+ "metadata": {
46
+ "version": 1.0
47
+ }
48
+ }
49
+ },
50
+ "versions": {
51
+ "arc_challenge": 1.0
52
+ },
53
+ "n-shot": {
54
+ "arc_challenge": 5
55
+ },
56
+ "config": {
57
+ "model": "hf",
58
+ "model_args": "pretrained=RWKV/rwkv-4-world-1b5,dtype=bfloat16,trust_remote_code=True",
59
+ "batch_size": "auto",
60
+ "batch_sizes": [
61
+ 64
62
+ ],
63
+ "device": null,
64
+ "use_cache": null,
65
+ "limit": null,
66
+ "bootstrap_iters": 100000,
67
+ "gen_kwargs": null
68
+ },
69
+ "git_hash": "99f5004"
70
+ }
lm-eval-output/RWKV/rwkv-4-world-1b5/arc_challenge/dtype=bfloat16,trust_remote_code=True-num_fewshot=5-nvidia-gpu/taskrun.log ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a006a61e9498a2037d56c09083e26b7f7242aced406c258e52faa062d5ceae38
3
+ size 12194
lm-eval-output/RWKV/rwkv-4-world-1b5/arithmetic/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:beefa415acc7816b8018aa1012ae17c2e9f7cb8c4c41551f91df9e2be73f130a
3
+ size 579405
lm-eval-output/RWKV/rwkv-4-world-1b5/arithmetic/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json ADDED
@@ -0,0 +1,378 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "arithmetic": {
4
+ "acc,none": 0.038,
5
+ "acc_stderr,none": 0.038029216765005085,
6
+ "alias": "arithmetic"
7
+ },
8
+ "arithmetic_1dc": {
9
+ "acc,none": 0.0125,
10
+ "acc_stderr,none": 0.0024849471787626713,
11
+ "alias": " - arithmetic_1dc"
12
+ },
13
+ "arithmetic_2da": {
14
+ "acc,none": 0.0695,
15
+ "acc_stderr,none": 0.005687798389997829,
16
+ "alias": " - arithmetic_2da"
17
+ },
18
+ "arithmetic_2dm": {
19
+ "acc,none": 0.1855,
20
+ "acc_stderr,none": 0.008693829210029837,
21
+ "alias": " - arithmetic_2dm"
22
+ },
23
+ "arithmetic_2ds": {
24
+ "acc,none": 0.081,
25
+ "acc_stderr,none": 0.006102304405675846,
26
+ "alias": " - arithmetic_2ds"
27
+ },
28
+ "arithmetic_3da": {
29
+ "acc,none": 0.011,
30
+ "acc_stderr,none": 0.0023328568559933755,
31
+ "alias": " - arithmetic_3da"
32
+ },
33
+ "arithmetic_3ds": {
34
+ "acc,none": 0.0175,
35
+ "acc_stderr,none": 0.00293277608892907,
36
+ "alias": " - arithmetic_3ds"
37
+ },
38
+ "arithmetic_4da": {
39
+ "acc,none": 0.0025,
40
+ "acc_stderr,none": 0.0011169148353275286,
41
+ "alias": " - arithmetic_4da"
42
+ },
43
+ "arithmetic_4ds": {
44
+ "acc,none": 0.0005,
45
+ "acc_stderr,none": 0.0005000000000000148,
46
+ "alias": " - arithmetic_4ds"
47
+ },
48
+ "arithmetic_5da": {
49
+ "acc,none": 0.0,
50
+ "acc_stderr,none": 0.0,
51
+ "alias": " - arithmetic_5da"
52
+ },
53
+ "arithmetic_5ds": {
54
+ "acc,none": 0.0,
55
+ "acc_stderr,none": 0.0,
56
+ "alias": " - arithmetic_5ds"
57
+ }
58
+ },
59
+ "groups": {
60
+ "arithmetic": {
61
+ "acc,none": 0.038,
62
+ "acc_stderr,none": 0.038029216765005085,
63
+ "alias": "arithmetic"
64
+ }
65
+ },
66
+ "configs": {
67
+ "arithmetic_1dc": {
68
+ "task": "arithmetic_1dc",
69
+ "group": [
70
+ "arithmetic"
71
+ ],
72
+ "dataset_path": "EleutherAI/arithmetic",
73
+ "dataset_name": "arithmetic_1dc",
74
+ "validation_split": "validation",
75
+ "doc_to_text": "{{context}}",
76
+ "doc_to_target": "{{completion}}",
77
+ "description": "",
78
+ "target_delimiter": " ",
79
+ "fewshot_delimiter": "\n\n",
80
+ "metric_list": [
81
+ {
82
+ "metric": "acc",
83
+ "aggregation": "mean",
84
+ "higher_is_better": true
85
+ }
86
+ ],
87
+ "output_type": "loglikelihood",
88
+ "repeats": 1,
89
+ "should_decontaminate": false,
90
+ "metadata": {
91
+ "version": 1.0
92
+ }
93
+ },
94
+ "arithmetic_2da": {
95
+ "task": "arithmetic_2da",
96
+ "group": [
97
+ "arithmetic"
98
+ ],
99
+ "dataset_path": "EleutherAI/arithmetic",
100
+ "dataset_name": "arithmetic_2da",
101
+ "validation_split": "validation",
102
+ "doc_to_text": "{{context}}",
103
+ "doc_to_target": "{{completion}}",
104
+ "description": "",
105
+ "target_delimiter": " ",
106
+ "fewshot_delimiter": "\n\n",
107
+ "metric_list": [
108
+ {
109
+ "metric": "acc",
110
+ "aggregation": "mean",
111
+ "higher_is_better": true
112
+ }
113
+ ],
114
+ "output_type": "loglikelihood",
115
+ "repeats": 1,
116
+ "should_decontaminate": false,
117
+ "metadata": {
118
+ "version": 1.0
119
+ }
120
+ },
121
+ "arithmetic_2dm": {
122
+ "task": "arithmetic_2dm",
123
+ "group": [
124
+ "arithmetic"
125
+ ],
126
+ "dataset_path": "EleutherAI/arithmetic",
127
+ "dataset_name": "arithmetic_2dm",
128
+ "validation_split": "validation",
129
+ "doc_to_text": "{{context}}",
130
+ "doc_to_target": "{{completion}}",
131
+ "description": "",
132
+ "target_delimiter": " ",
133
+ "fewshot_delimiter": "\n\n",
134
+ "metric_list": [
135
+ {
136
+ "metric": "acc",
137
+ "aggregation": "mean",
138
+ "higher_is_better": true
139
+ }
140
+ ],
141
+ "output_type": "loglikelihood",
142
+ "repeats": 1,
143
+ "should_decontaminate": false,
144
+ "metadata": {
145
+ "version": 1.0
146
+ }
147
+ },
148
+ "arithmetic_2ds": {
149
+ "task": "arithmetic_2ds",
150
+ "group": [
151
+ "arithmetic"
152
+ ],
153
+ "dataset_path": "EleutherAI/arithmetic",
154
+ "dataset_name": "arithmetic_2ds",
155
+ "validation_split": "validation",
156
+ "doc_to_text": "{{context}}",
157
+ "doc_to_target": "{{completion}}",
158
+ "description": "",
159
+ "target_delimiter": " ",
160
+ "fewshot_delimiter": "\n\n",
161
+ "metric_list": [
162
+ {
163
+ "metric": "acc",
164
+ "aggregation": "mean",
165
+ "higher_is_better": true
166
+ }
167
+ ],
168
+ "output_type": "loglikelihood",
169
+ "repeats": 1,
170
+ "should_decontaminate": false,
171
+ "metadata": {
172
+ "version": 1.0
173
+ }
174
+ },
175
+ "arithmetic_3da": {
176
+ "task": "arithmetic_3da",
177
+ "group": [
178
+ "arithmetic"
179
+ ],
180
+ "dataset_path": "EleutherAI/arithmetic",
181
+ "dataset_name": "arithmetic_3da",
182
+ "validation_split": "validation",
183
+ "doc_to_text": "{{context}}",
184
+ "doc_to_target": "{{completion}}",
185
+ "description": "",
186
+ "target_delimiter": " ",
187
+ "fewshot_delimiter": "\n\n",
188
+ "metric_list": [
189
+ {
190
+ "metric": "acc",
191
+ "aggregation": "mean",
192
+ "higher_is_better": true
193
+ }
194
+ ],
195
+ "output_type": "loglikelihood",
196
+ "repeats": 1,
197
+ "should_decontaminate": false,
198
+ "metadata": {
199
+ "version": 1.0
200
+ }
201
+ },
202
+ "arithmetic_3ds": {
203
+ "task": "arithmetic_3ds",
204
+ "group": [
205
+ "arithmetic"
206
+ ],
207
+ "dataset_path": "EleutherAI/arithmetic",
208
+ "dataset_name": "arithmetic_3ds",
209
+ "validation_split": "validation",
210
+ "doc_to_text": "{{context}}",
211
+ "doc_to_target": "{{completion}}",
212
+ "description": "",
213
+ "target_delimiter": " ",
214
+ "fewshot_delimiter": "\n\n",
215
+ "metric_list": [
216
+ {
217
+ "metric": "acc",
218
+ "aggregation": "mean",
219
+ "higher_is_better": true
220
+ }
221
+ ],
222
+ "output_type": "loglikelihood",
223
+ "repeats": 1,
224
+ "should_decontaminate": false,
225
+ "metadata": {
226
+ "version": 1.0
227
+ }
228
+ },
229
+ "arithmetic_4da": {
230
+ "task": "arithmetic_4da",
231
+ "group": [
232
+ "arithmetic"
233
+ ],
234
+ "dataset_path": "EleutherAI/arithmetic",
235
+ "dataset_name": "arithmetic_4da",
236
+ "validation_split": "validation",
237
+ "doc_to_text": "{{context}}",
238
+ "doc_to_target": "{{completion}}",
239
+ "description": "",
240
+ "target_delimiter": " ",
241
+ "fewshot_delimiter": "\n\n",
242
+ "metric_list": [
243
+ {
244
+ "metric": "acc",
245
+ "aggregation": "mean",
246
+ "higher_is_better": true
247
+ }
248
+ ],
249
+ "output_type": "loglikelihood",
250
+ "repeats": 1,
251
+ "should_decontaminate": false,
252
+ "metadata": {
253
+ "version": 1.0
254
+ }
255
+ },
256
+ "arithmetic_4ds": {
257
+ "task": "arithmetic_4ds",
258
+ "group": [
259
+ "arithmetic"
260
+ ],
261
+ "dataset_path": "EleutherAI/arithmetic",
262
+ "dataset_name": "arithmetic_4ds",
263
+ "validation_split": "validation",
264
+ "doc_to_text": "{{context}}",
265
+ "doc_to_target": "{{completion}}",
266
+ "description": "",
267
+ "target_delimiter": " ",
268
+ "fewshot_delimiter": "\n\n",
269
+ "metric_list": [
270
+ {
271
+ "metric": "acc",
272
+ "aggregation": "mean",
273
+ "higher_is_better": true
274
+ }
275
+ ],
276
+ "output_type": "loglikelihood",
277
+ "repeats": 1,
278
+ "should_decontaminate": false,
279
+ "metadata": {
280
+ "version": 1.0
281
+ }
282
+ },
283
+ "arithmetic_5da": {
284
+ "task": "arithmetic_5da",
285
+ "group": [
286
+ "arithmetic"
287
+ ],
288
+ "dataset_path": "EleutherAI/arithmetic",
289
+ "dataset_name": "arithmetic_5da",
290
+ "validation_split": "validation",
291
+ "doc_to_text": "{{context}}",
292
+ "doc_to_target": "{{completion}}",
293
+ "description": "",
294
+ "target_delimiter": " ",
295
+ "fewshot_delimiter": "\n\n",
296
+ "metric_list": [
297
+ {
298
+ "metric": "acc",
299
+ "aggregation": "mean",
300
+ "higher_is_better": true
301
+ }
302
+ ],
303
+ "output_type": "loglikelihood",
304
+ "repeats": 1,
305
+ "should_decontaminate": false,
306
+ "metadata": {
307
+ "version": 1.0
308
+ }
309
+ },
310
+ "arithmetic_5ds": {
311
+ "task": "arithmetic_5ds",
312
+ "group": [
313
+ "arithmetic"
314
+ ],
315
+ "dataset_path": "EleutherAI/arithmetic",
316
+ "dataset_name": "arithmetic_5ds",
317
+ "validation_split": "validation",
318
+ "doc_to_text": "{{context}}",
319
+ "doc_to_target": "{{completion}}",
320
+ "description": "",
321
+ "target_delimiter": " ",
322
+ "fewshot_delimiter": "\n\n",
323
+ "metric_list": [
324
+ {
325
+ "metric": "acc",
326
+ "aggregation": "mean",
327
+ "higher_is_better": true
328
+ }
329
+ ],
330
+ "output_type": "loglikelihood",
331
+ "repeats": 1,
332
+ "should_decontaminate": false,
333
+ "metadata": {
334
+ "version": 1.0
335
+ }
336
+ }
337
+ },
338
+ "versions": {
339
+ "arithmetic": "N/A",
340
+ "arithmetic_1dc": 1.0,
341
+ "arithmetic_2da": 1.0,
342
+ "arithmetic_2dm": 1.0,
343
+ "arithmetic_2ds": 1.0,
344
+ "arithmetic_3da": 1.0,
345
+ "arithmetic_3ds": 1.0,
346
+ "arithmetic_4da": 1.0,
347
+ "arithmetic_4ds": 1.0,
348
+ "arithmetic_5da": 1.0,
349
+ "arithmetic_5ds": 1.0
350
+ },
351
+ "n-shot": {
352
+ "arithmetic": 0,
353
+ "arithmetic_1dc": 0,
354
+ "arithmetic_2da": 0,
355
+ "arithmetic_2dm": 0,
356
+ "arithmetic_2ds": 0,
357
+ "arithmetic_3da": 0,
358
+ "arithmetic_3ds": 0,
359
+ "arithmetic_4da": 0,
360
+ "arithmetic_4ds": 0,
361
+ "arithmetic_5da": 0,
362
+ "arithmetic_5ds": 0
363
+ },
364
+ "config": {
365
+ "model": "hf",
366
+ "model_args": "pretrained=RWKV/rwkv-4-world-1b5,dtype=bfloat16,trust_remote_code=True",
367
+ "batch_size": "auto",
368
+ "batch_sizes": [
369
+ 64
370
+ ],
371
+ "device": null,
372
+ "use_cache": null,
373
+ "limit": null,
374
+ "bootstrap_iters": 100000,
375
+ "gen_kwargs": null
376
+ },
377
+ "git_hash": "99f5004"
378
+ }
lm-eval-output/RWKV/rwkv-4-world-1b5/arithmetic/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:762502092e838f4958628f42e3f26568dc20304ef540f88cb74087fcb0107097
3
+ size 24836
lm-eval-output/RWKV/rwkv-4-world-1b5/arithmetic__/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:beefa415acc7816b8018aa1012ae17c2e9f7cb8c4c41551f91df9e2be73f130a
3
+ size 579405
lm-eval-output/RWKV/rwkv-4-world-1b5/arithmetic__/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json ADDED
@@ -0,0 +1,364 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "arithmetic_5ds": {
4
+ "acc,none": 0.0,
5
+ "acc_stderr,none": 0.0,
6
+ "alias": "arithmetic_5ds"
7
+ },
8
+ "arithmetic_5da": {
9
+ "acc,none": 0.0,
10
+ "acc_stderr,none": 0.0,
11
+ "alias": "arithmetic_5da"
12
+ },
13
+ "arithmetic_4ds": {
14
+ "acc,none": 0.0005,
15
+ "acc_stderr,none": 0.0005000000000000148,
16
+ "alias": "arithmetic_4ds"
17
+ },
18
+ "arithmetic_4da": {
19
+ "acc,none": 0.0025,
20
+ "acc_stderr,none": 0.0011169148353275286,
21
+ "alias": "arithmetic_4da"
22
+ },
23
+ "arithmetic_3ds": {
24
+ "acc,none": 0.0175,
25
+ "acc_stderr,none": 0.00293277608892907,
26
+ "alias": "arithmetic_3ds"
27
+ },
28
+ "arithmetic_3da": {
29
+ "acc,none": 0.011,
30
+ "acc_stderr,none": 0.0023328568559933755,
31
+ "alias": "arithmetic_3da"
32
+ },
33
+ "arithmetic_2ds": {
34
+ "acc,none": 0.081,
35
+ "acc_stderr,none": 0.006102304405675846,
36
+ "alias": "arithmetic_2ds"
37
+ },
38
+ "arithmetic_2dm": {
39
+ "acc,none": 0.1855,
40
+ "acc_stderr,none": 0.008693829210029837,
41
+ "alias": "arithmetic_2dm"
42
+ },
43
+ "arithmetic_2da": {
44
+ "acc,none": 0.0695,
45
+ "acc_stderr,none": 0.005687798389997829,
46
+ "alias": "arithmetic_2da"
47
+ },
48
+ "arithmetic_1dc": {
49
+ "acc,none": 0.0125,
50
+ "acc_stderr,none": 0.0024849471787626713,
51
+ "alias": "arithmetic_1dc"
52
+ }
53
+ },
54
+ "configs": {
55
+ "arithmetic_1dc": {
56
+ "task": "arithmetic_1dc",
57
+ "group": [
58
+ "arithmetic"
59
+ ],
60
+ "dataset_path": "EleutherAI/arithmetic",
61
+ "dataset_name": "arithmetic_1dc",
62
+ "validation_split": "validation",
63
+ "doc_to_text": "{{context}}",
64
+ "doc_to_target": "{{completion}}",
65
+ "description": "",
66
+ "target_delimiter": " ",
67
+ "fewshot_delimiter": "\n\n",
68
+ "metric_list": [
69
+ {
70
+ "metric": "acc",
71
+ "aggregation": "mean",
72
+ "higher_is_better": true
73
+ }
74
+ ],
75
+ "output_type": "loglikelihood",
76
+ "repeats": 1,
77
+ "should_decontaminate": false,
78
+ "metadata": {
79
+ "version": 1.0
80
+ }
81
+ },
82
+ "arithmetic_2da": {
83
+ "task": "arithmetic_2da",
84
+ "group": [
85
+ "arithmetic"
86
+ ],
87
+ "dataset_path": "EleutherAI/arithmetic",
88
+ "dataset_name": "arithmetic_2da",
89
+ "validation_split": "validation",
90
+ "doc_to_text": "{{context}}",
91
+ "doc_to_target": "{{completion}}",
92
+ "description": "",
93
+ "target_delimiter": " ",
94
+ "fewshot_delimiter": "\n\n",
95
+ "metric_list": [
96
+ {
97
+ "metric": "acc",
98
+ "aggregation": "mean",
99
+ "higher_is_better": true
100
+ }
101
+ ],
102
+ "output_type": "loglikelihood",
103
+ "repeats": 1,
104
+ "should_decontaminate": false,
105
+ "metadata": {
106
+ "version": 1.0
107
+ }
108
+ },
109
+ "arithmetic_2dm": {
110
+ "task": "arithmetic_2dm",
111
+ "group": [
112
+ "arithmetic"
113
+ ],
114
+ "dataset_path": "EleutherAI/arithmetic",
115
+ "dataset_name": "arithmetic_2dm",
116
+ "validation_split": "validation",
117
+ "doc_to_text": "{{context}}",
118
+ "doc_to_target": "{{completion}}",
119
+ "description": "",
120
+ "target_delimiter": " ",
121
+ "fewshot_delimiter": "\n\n",
122
+ "metric_list": [
123
+ {
124
+ "metric": "acc",
125
+ "aggregation": "mean",
126
+ "higher_is_better": true
127
+ }
128
+ ],
129
+ "output_type": "loglikelihood",
130
+ "repeats": 1,
131
+ "should_decontaminate": false,
132
+ "metadata": {
133
+ "version": 1.0
134
+ }
135
+ },
136
+ "arithmetic_2ds": {
137
+ "task": "arithmetic_2ds",
138
+ "group": [
139
+ "arithmetic"
140
+ ],
141
+ "dataset_path": "EleutherAI/arithmetic",
142
+ "dataset_name": "arithmetic_2ds",
143
+ "validation_split": "validation",
144
+ "doc_to_text": "{{context}}",
145
+ "doc_to_target": "{{completion}}",
146
+ "description": "",
147
+ "target_delimiter": " ",
148
+ "fewshot_delimiter": "\n\n",
149
+ "metric_list": [
150
+ {
151
+ "metric": "acc",
152
+ "aggregation": "mean",
153
+ "higher_is_better": true
154
+ }
155
+ ],
156
+ "output_type": "loglikelihood",
157
+ "repeats": 1,
158
+ "should_decontaminate": false,
159
+ "metadata": {
160
+ "version": 1.0
161
+ }
162
+ },
163
+ "arithmetic_3da": {
164
+ "task": "arithmetic_3da",
165
+ "group": [
166
+ "arithmetic"
167
+ ],
168
+ "dataset_path": "EleutherAI/arithmetic",
169
+ "dataset_name": "arithmetic_3da",
170
+ "validation_split": "validation",
171
+ "doc_to_text": "{{context}}",
172
+ "doc_to_target": "{{completion}}",
173
+ "description": "",
174
+ "target_delimiter": " ",
175
+ "fewshot_delimiter": "\n\n",
176
+ "metric_list": [
177
+ {
178
+ "metric": "acc",
179
+ "aggregation": "mean",
180
+ "higher_is_better": true
181
+ }
182
+ ],
183
+ "output_type": "loglikelihood",
184
+ "repeats": 1,
185
+ "should_decontaminate": false,
186
+ "metadata": {
187
+ "version": 1.0
188
+ }
189
+ },
190
+ "arithmetic_3ds": {
191
+ "task": "arithmetic_3ds",
192
+ "group": [
193
+ "arithmetic"
194
+ ],
195
+ "dataset_path": "EleutherAI/arithmetic",
196
+ "dataset_name": "arithmetic_3ds",
197
+ "validation_split": "validation",
198
+ "doc_to_text": "{{context}}",
199
+ "doc_to_target": "{{completion}}",
200
+ "description": "",
201
+ "target_delimiter": " ",
202
+ "fewshot_delimiter": "\n\n",
203
+ "metric_list": [
204
+ {
205
+ "metric": "acc",
206
+ "aggregation": "mean",
207
+ "higher_is_better": true
208
+ }
209
+ ],
210
+ "output_type": "loglikelihood",
211
+ "repeats": 1,
212
+ "should_decontaminate": false,
213
+ "metadata": {
214
+ "version": 1.0
215
+ }
216
+ },
217
+ "arithmetic_4da": {
218
+ "task": "arithmetic_4da",
219
+ "group": [
220
+ "arithmetic"
221
+ ],
222
+ "dataset_path": "EleutherAI/arithmetic",
223
+ "dataset_name": "arithmetic_4da",
224
+ "validation_split": "validation",
225
+ "doc_to_text": "{{context}}",
226
+ "doc_to_target": "{{completion}}",
227
+ "description": "",
228
+ "target_delimiter": " ",
229
+ "fewshot_delimiter": "\n\n",
230
+ "metric_list": [
231
+ {
232
+ "metric": "acc",
233
+ "aggregation": "mean",
234
+ "higher_is_better": true
235
+ }
236
+ ],
237
+ "output_type": "loglikelihood",
238
+ "repeats": 1,
239
+ "should_decontaminate": false,
240
+ "metadata": {
241
+ "version": 1.0
242
+ }
243
+ },
244
+ "arithmetic_4ds": {
245
+ "task": "arithmetic_4ds",
246
+ "group": [
247
+ "arithmetic"
248
+ ],
249
+ "dataset_path": "EleutherAI/arithmetic",
250
+ "dataset_name": "arithmetic_4ds",
251
+ "validation_split": "validation",
252
+ "doc_to_text": "{{context}}",
253
+ "doc_to_target": "{{completion}}",
254
+ "description": "",
255
+ "target_delimiter": " ",
256
+ "fewshot_delimiter": "\n\n",
257
+ "metric_list": [
258
+ {
259
+ "metric": "acc",
260
+ "aggregation": "mean",
261
+ "higher_is_better": true
262
+ }
263
+ ],
264
+ "output_type": "loglikelihood",
265
+ "repeats": 1,
266
+ "should_decontaminate": false,
267
+ "metadata": {
268
+ "version": 1.0
269
+ }
270
+ },
271
+ "arithmetic_5da": {
272
+ "task": "arithmetic_5da",
273
+ "group": [
274
+ "arithmetic"
275
+ ],
276
+ "dataset_path": "EleutherAI/arithmetic",
277
+ "dataset_name": "arithmetic_5da",
278
+ "validation_split": "validation",
279
+ "doc_to_text": "{{context}}",
280
+ "doc_to_target": "{{completion}}",
281
+ "description": "",
282
+ "target_delimiter": " ",
283
+ "fewshot_delimiter": "\n\n",
284
+ "metric_list": [
285
+ {
286
+ "metric": "acc",
287
+ "aggregation": "mean",
288
+ "higher_is_better": true
289
+ }
290
+ ],
291
+ "output_type": "loglikelihood",
292
+ "repeats": 1,
293
+ "should_decontaminate": false,
294
+ "metadata": {
295
+ "version": 1.0
296
+ }
297
+ },
298
+ "arithmetic_5ds": {
299
+ "task": "arithmetic_5ds",
300
+ "group": [
301
+ "arithmetic"
302
+ ],
303
+ "dataset_path": "EleutherAI/arithmetic",
304
+ "dataset_name": "arithmetic_5ds",
305
+ "validation_split": "validation",
306
+ "doc_to_text": "{{context}}",
307
+ "doc_to_target": "{{completion}}",
308
+ "description": "",
309
+ "target_delimiter": " ",
310
+ "fewshot_delimiter": "\n\n",
311
+ "metric_list": [
312
+ {
313
+ "metric": "acc",
314
+ "aggregation": "mean",
315
+ "higher_is_better": true
316
+ }
317
+ ],
318
+ "output_type": "loglikelihood",
319
+ "repeats": 1,
320
+ "should_decontaminate": false,
321
+ "metadata": {
322
+ "version": 1.0
323
+ }
324
+ }
325
+ },
326
+ "versions": {
327
+ "arithmetic_1dc": 1.0,
328
+ "arithmetic_2da": 1.0,
329
+ "arithmetic_2dm": 1.0,
330
+ "arithmetic_2ds": 1.0,
331
+ "arithmetic_3da": 1.0,
332
+ "arithmetic_3ds": 1.0,
333
+ "arithmetic_4da": 1.0,
334
+ "arithmetic_4ds": 1.0,
335
+ "arithmetic_5da": 1.0,
336
+ "arithmetic_5ds": 1.0
337
+ },
338
+ "n-shot": {
339
+ "arithmetic_1dc": 0,
340
+ "arithmetic_2da": 0,
341
+ "arithmetic_2dm": 0,
342
+ "arithmetic_2ds": 0,
343
+ "arithmetic_3da": 0,
344
+ "arithmetic_3ds": 0,
345
+ "arithmetic_4da": 0,
346
+ "arithmetic_4ds": 0,
347
+ "arithmetic_5da": 0,
348
+ "arithmetic_5ds": 0
349
+ },
350
+ "config": {
351
+ "model": "hf",
352
+ "model_args": "pretrained=RWKV/rwkv-4-world-1b5,dtype=bfloat16,trust_remote_code=True",
353
+ "batch_size": "auto",
354
+ "batch_sizes": [
355
+ 64
356
+ ],
357
+ "device": null,
358
+ "use_cache": null,
359
+ "limit": null,
360
+ "bootstrap_iters": 100000,
361
+ "gen_kwargs": null
362
+ },
363
+ "git_hash": "99f5004"
364
+ }
lm-eval-output/RWKV/rwkv-4-world-1b5/arithmetic__/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eae8a027d0a328b29ea7be287929a54a71fb608d6f0cba3758fc88e062fc2a55
3
+ size 21471
lm-eval-output/RWKV/rwkv-4-world-1b5/asdiv/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b0853fc15a8dc1bd7daefe7190d2bb59cf17df268c5d65de0193775a8af1dcb2
3
+ size 263936
lm-eval-output/RWKV/rwkv-4-world-1b5/asdiv/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json ADDED
@@ -0,0 +1,55 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "asdiv": {
4
+ "acc,none": 0.0,
5
+ "acc_stderr,none": 0.0,
6
+ "alias": "asdiv"
7
+ }
8
+ },
9
+ "configs": {
10
+ "asdiv": {
11
+ "task": "asdiv",
12
+ "dataset_path": "EleutherAI/asdiv",
13
+ "validation_split": "validation",
14
+ "doc_to_text": "{{body}}\nQuestion:{{question}}\nAnswer:",
15
+ "doc_to_target": "{{answer.split(' (')[0]}}",
16
+ "description": "",
17
+ "target_delimiter": " ",
18
+ "fewshot_delimiter": "\n\n",
19
+ "metric_list": [
20
+ {
21
+ "metric": "acc",
22
+ "aggregation": "mean",
23
+ "higher_is_better": true
24
+ }
25
+ ],
26
+ "output_type": "loglikelihood",
27
+ "repeats": 1,
28
+ "should_decontaminate": true,
29
+ "doc_to_decontamination_query": "{{body}} {{question}}",
30
+ "metadata": {
31
+ "version": 1.0
32
+ }
33
+ }
34
+ },
35
+ "versions": {
36
+ "asdiv": 1.0
37
+ },
38
+ "n-shot": {
39
+ "asdiv": 0
40
+ },
41
+ "config": {
42
+ "model": "hf",
43
+ "model_args": "pretrained=RWKV/rwkv-4-world-1b5,dtype=bfloat16,trust_remote_code=True",
44
+ "batch_size": "auto",
45
+ "batch_sizes": [
46
+ 64
47
+ ],
48
+ "device": null,
49
+ "use_cache": null,
50
+ "limit": null,
51
+ "bootstrap_iters": 100000,
52
+ "gen_kwargs": null
53
+ },
54
+ "git_hash": "99f5004"
55
+ }
lm-eval-output/RWKV/rwkv-4-world-1b5/asdiv/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a05e21dbf179b04307139fe1d86cc8c5401ab577112ddf7043d636e06d798fb1
3
+ size 14999
lm-eval-output/RWKV/rwkv-4-world-1b5/blimp/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:debb30c695eed5af59514b9967be7f0d053946cb82a719ba2a6a614e5a9fc8b3
3
+ size 4241567
lm-eval-output/RWKV/rwkv-4-world-1b5/blimp/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json ADDED
@@ -0,0 +1,2249 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "blimp": {
4
+ "acc,none": 0.8387313432835821,
5
+ "acc_stderr,none": 0.14773171752507513,
6
+ "alias": "blimp"
7
+ },
8
+ "blimp_adjunct_island": {
9
+ "acc,none": 0.886,
10
+ "acc_stderr,none": 0.01005510343582333,
11
+ "alias": " - blimp_adjunct_island"
12
+ },
13
+ "blimp_anaphor_gender_agreement": {
14
+ "acc,none": 0.995,
15
+ "acc_stderr,none": 0.0022315868748448825,
16
+ "alias": " - blimp_anaphor_gender_agreement"
17
+ },
18
+ "blimp_anaphor_number_agreement": {
19
+ "acc,none": 0.998,
20
+ "acc_stderr,none": 0.001413505570557809,
21
+ "alias": " - blimp_anaphor_number_agreement"
22
+ },
23
+ "blimp_animate_subject_passive": {
24
+ "acc,none": 0.816,
25
+ "acc_stderr,none": 0.012259457340938588,
26
+ "alias": " - blimp_animate_subject_passive"
27
+ },
28
+ "blimp_animate_subject_trans": {
29
+ "acc,none": 0.901,
30
+ "acc_stderr,none": 0.009449248027662744,
31
+ "alias": " - blimp_animate_subject_trans"
32
+ },
33
+ "blimp_causative": {
34
+ "acc,none": 0.786,
35
+ "acc_stderr,none": 0.012975838021968783,
36
+ "alias": " - blimp_causative"
37
+ },
38
+ "blimp_complex_NP_island": {
39
+ "acc,none": 0.619,
40
+ "acc_stderr,none": 0.015364734787007436,
41
+ "alias": " - blimp_complex_NP_island"
42
+ },
43
+ "blimp_coordinate_structure_constraint_complex_left_branch": {
44
+ "acc,none": 0.792,
45
+ "acc_stderr,none": 0.012841374572096928,
46
+ "alias": " - blimp_coordinate_structure_constraint_complex_left_branch"
47
+ },
48
+ "blimp_coordinate_structure_constraint_object_extraction": {
49
+ "acc,none": 0.879,
50
+ "acc_stderr,none": 0.010318210380946092,
51
+ "alias": " - blimp_coordinate_structure_constraint_object_extraction"
52
+ },
53
+ "blimp_determiner_noun_agreement_1": {
54
+ "acc,none": 0.994,
55
+ "acc_stderr,none": 0.002443352199329814,
56
+ "alias": " - blimp_determiner_noun_agreement_1"
57
+ },
58
+ "blimp_determiner_noun_agreement_2": {
59
+ "acc,none": 0.99,
60
+ "acc_stderr,none": 0.003148000938676759,
61
+ "alias": " - blimp_determiner_noun_agreement_2"
62
+ },
63
+ "blimp_determiner_noun_agreement_irregular_1": {
64
+ "acc,none": 0.96,
65
+ "acc_stderr,none": 0.0061998740663370515,
66
+ "alias": " - blimp_determiner_noun_agreement_irregular_1"
67
+ },
68
+ "blimp_determiner_noun_agreement_irregular_2": {
69
+ "acc,none": 0.968,
70
+ "acc_stderr,none": 0.005568393575081374,
71
+ "alias": " - blimp_determiner_noun_agreement_irregular_2"
72
+ },
73
+ "blimp_determiner_noun_agreement_with_adj_2": {
74
+ "acc,none": 0.962,
75
+ "acc_stderr,none": 0.006049181150584942,
76
+ "alias": " - blimp_determiner_noun_agreement_with_adj_2"
77
+ },
78
+ "blimp_determiner_noun_agreement_with_adj_irregular_1": {
79
+ "acc,none": 0.935,
80
+ "acc_stderr,none": 0.007799733061832022,
81
+ "alias": " - blimp_determiner_noun_agreement_with_adj_irregular_1"
82
+ },
83
+ "blimp_determiner_noun_agreement_with_adj_irregular_2": {
84
+ "acc,none": 0.939,
85
+ "acc_stderr,none": 0.007572076091557425,
86
+ "alias": " - blimp_determiner_noun_agreement_with_adj_irregular_2"
87
+ },
88
+ "blimp_determiner_noun_agreement_with_adjective_1": {
89
+ "acc,none": 0.991,
90
+ "acc_stderr,none": 0.0029879638431426544,
91
+ "alias": " - blimp_determiner_noun_agreement_with_adjective_1"
92
+ },
93
+ "blimp_distractor_agreement_relational_noun": {
94
+ "acc,none": 0.902,
95
+ "acc_stderr,none": 0.00940661918462124,
96
+ "alias": " - blimp_distractor_agreement_relational_noun"
97
+ },
98
+ "blimp_distractor_agreement_relative_clause": {
99
+ "acc,none": 0.78,
100
+ "acc_stderr,none": 0.013106173040661754,
101
+ "alias": " - blimp_distractor_agreement_relative_clause"
102
+ },
103
+ "blimp_drop_argument": {
104
+ "acc,none": 0.826,
105
+ "acc_stderr,none": 0.011994493230973409,
106
+ "alias": " - blimp_drop_argument"
107
+ },
108
+ "blimp_ellipsis_n_bar_1": {
109
+ "acc,none": 0.872,
110
+ "acc_stderr,none": 0.010570133761108658,
111
+ "alias": " - blimp_ellipsis_n_bar_1"
112
+ },
113
+ "blimp_ellipsis_n_bar_2": {
114
+ "acc,none": 0.894,
115
+ "acc_stderr,none": 0.009739551265785129,
116
+ "alias": " - blimp_ellipsis_n_bar_2"
117
+ },
118
+ "blimp_existential_there_object_raising": {
119
+ "acc,none": 0.885,
120
+ "acc_stderr,none": 0.010093407594904612,
121
+ "alias": " - blimp_existential_there_object_raising"
122
+ },
123
+ "blimp_existential_there_quantifiers_1": {
124
+ "acc,none": 0.994,
125
+ "acc_stderr,none": 0.002443352199329814,
126
+ "alias": " - blimp_existential_there_quantifiers_1"
127
+ },
128
+ "blimp_existential_there_quantifiers_2": {
129
+ "acc,none": 0.346,
130
+ "acc_stderr,none": 0.015050266127564441,
131
+ "alias": " - blimp_existential_there_quantifiers_2"
132
+ },
133
+ "blimp_existential_there_subject_raising": {
134
+ "acc,none": 0.925,
135
+ "acc_stderr,none": 0.00833333333333334,
136
+ "alias": " - blimp_existential_there_subject_raising"
137
+ },
138
+ "blimp_expletive_it_object_raising": {
139
+ "acc,none": 0.804,
140
+ "acc_stderr,none": 0.012559527926707361,
141
+ "alias": " - blimp_expletive_it_object_raising"
142
+ },
143
+ "blimp_inchoative": {
144
+ "acc,none": 0.696,
145
+ "acc_stderr,none": 0.014553205687950422,
146
+ "alias": " - blimp_inchoative"
147
+ },
148
+ "blimp_intransitive": {
149
+ "acc,none": 0.846,
150
+ "acc_stderr,none": 0.011419913065098689,
151
+ "alias": " - blimp_intransitive"
152
+ },
153
+ "blimp_irregular_past_participle_adjectives": {
154
+ "acc,none": 0.96,
155
+ "acc_stderr,none": 0.006199874066337051,
156
+ "alias": " - blimp_irregular_past_participle_adjectives"
157
+ },
158
+ "blimp_irregular_past_participle_verbs": {
159
+ "acc,none": 0.909,
160
+ "acc_stderr,none": 0.009099549538400234,
161
+ "alias": " - blimp_irregular_past_participle_verbs"
162
+ },
163
+ "blimp_irregular_plural_subject_verb_agreement_1": {
164
+ "acc,none": 0.949,
165
+ "acc_stderr,none": 0.006960420062571416,
166
+ "alias": " - blimp_irregular_plural_subject_verb_agreement_1"
167
+ },
168
+ "blimp_irregular_plural_subject_verb_agreement_2": {
169
+ "acc,none": 0.926,
170
+ "acc_stderr,none": 0.008282064512704163,
171
+ "alias": " - blimp_irregular_plural_subject_verb_agreement_2"
172
+ },
173
+ "blimp_left_branch_island_echo_question": {
174
+ "acc,none": 0.452,
175
+ "acc_stderr,none": 0.015746235865880677,
176
+ "alias": " - blimp_left_branch_island_echo_question"
177
+ },
178
+ "blimp_left_branch_island_simple_question": {
179
+ "acc,none": 0.89,
180
+ "acc_stderr,none": 0.009899393819724447,
181
+ "alias": " - blimp_left_branch_island_simple_question"
182
+ },
183
+ "blimp_matrix_question_npi_licensor_present": {
184
+ "acc,none": 0.539,
185
+ "acc_stderr,none": 0.01577110420128319,
186
+ "alias": " - blimp_matrix_question_npi_licensor_present"
187
+ },
188
+ "blimp_npi_present_1": {
189
+ "acc,none": 0.588,
190
+ "acc_stderr,none": 0.015572363292015088,
191
+ "alias": " - blimp_npi_present_1"
192
+ },
193
+ "blimp_npi_present_2": {
194
+ "acc,none": 0.64,
195
+ "acc_stderr,none": 0.015186527932040124,
196
+ "alias": " - blimp_npi_present_2"
197
+ },
198
+ "blimp_only_npi_licensor_present": {
199
+ "acc,none": 0.923,
200
+ "acc_stderr,none": 0.008434580140240627,
201
+ "alias": " - blimp_only_npi_licensor_present"
202
+ },
203
+ "blimp_only_npi_scope": {
204
+ "acc,none": 0.75,
205
+ "acc_stderr,none": 0.013699915608779773,
206
+ "alias": " - blimp_only_npi_scope"
207
+ },
208
+ "blimp_passive_1": {
209
+ "acc,none": 0.904,
210
+ "acc_stderr,none": 0.009320454434783213,
211
+ "alias": " - blimp_passive_1"
212
+ },
213
+ "blimp_passive_2": {
214
+ "acc,none": 0.914,
215
+ "acc_stderr,none": 0.008870325962594766,
216
+ "alias": " - blimp_passive_2"
217
+ },
218
+ "blimp_principle_A_c_command": {
219
+ "acc,none": 0.839,
220
+ "acc_stderr,none": 0.011628164696727197,
221
+ "alias": " - blimp_principle_A_c_command"
222
+ },
223
+ "blimp_principle_A_case_1": {
224
+ "acc,none": 1.0,
225
+ "acc_stderr,none": 0.0,
226
+ "alias": " - blimp_principle_A_case_1"
227
+ },
228
+ "blimp_principle_A_case_2": {
229
+ "acc,none": 0.975,
230
+ "acc_stderr,none": 0.004939574819698448,
231
+ "alias": " - blimp_principle_A_case_2"
232
+ },
233
+ "blimp_principle_A_domain_1": {
234
+ "acc,none": 0.995,
235
+ "acc_stderr,none": 0.0022315868748448847,
236
+ "alias": " - blimp_principle_A_domain_1"
237
+ },
238
+ "blimp_principle_A_domain_2": {
239
+ "acc,none": 0.872,
240
+ "acc_stderr,none": 0.010570133761108656,
241
+ "alias": " - blimp_principle_A_domain_2"
242
+ },
243
+ "blimp_principle_A_domain_3": {
244
+ "acc,none": 0.702,
245
+ "acc_stderr,none": 0.014470846741134713,
246
+ "alias": " - blimp_principle_A_domain_3"
247
+ },
248
+ "blimp_principle_A_reconstruction": {
249
+ "acc,none": 0.5,
250
+ "acc_stderr,none": 0.015819299929208316,
251
+ "alias": " - blimp_principle_A_reconstruction"
252
+ },
253
+ "blimp_regular_plural_subject_verb_agreement_1": {
254
+ "acc,none": 0.968,
255
+ "acc_stderr,none": 0.005568393575081357,
256
+ "alias": " - blimp_regular_plural_subject_verb_agreement_1"
257
+ },
258
+ "blimp_regular_plural_subject_verb_agreement_2": {
259
+ "acc,none": 0.928,
260
+ "acc_stderr,none": 0.008178195576218681,
261
+ "alias": " - blimp_regular_plural_subject_verb_agreement_2"
262
+ },
263
+ "blimp_sentential_negation_npi_licensor_present": {
264
+ "acc,none": 0.986,
265
+ "acc_stderr,none": 0.0037172325482565894,
266
+ "alias": " - blimp_sentential_negation_npi_licensor_present"
267
+ },
268
+ "blimp_sentential_negation_npi_scope": {
269
+ "acc,none": 0.758,
270
+ "acc_stderr,none": 0.01355063170555594,
271
+ "alias": " - blimp_sentential_negation_npi_scope"
272
+ },
273
+ "blimp_sentential_subject_island": {
274
+ "acc,none": 0.523,
275
+ "acc_stderr,none": 0.0158025542467261,
276
+ "alias": " - blimp_sentential_subject_island"
277
+ },
278
+ "blimp_superlative_quantifiers_1": {
279
+ "acc,none": 0.875,
280
+ "acc_stderr,none": 0.010463483381956722,
281
+ "alias": " - blimp_superlative_quantifiers_1"
282
+ },
283
+ "blimp_superlative_quantifiers_2": {
284
+ "acc,none": 0.884,
285
+ "acc_stderr,none": 0.010131468138756991,
286
+ "alias": " - blimp_superlative_quantifiers_2"
287
+ },
288
+ "blimp_tough_vs_raising_1": {
289
+ "acc,none": 0.713,
290
+ "acc_stderr,none": 0.014312087053809961,
291
+ "alias": " - blimp_tough_vs_raising_1"
292
+ },
293
+ "blimp_tough_vs_raising_2": {
294
+ "acc,none": 0.905,
295
+ "acc_stderr,none": 0.009276910103103306,
296
+ "alias": " - blimp_tough_vs_raising_2"
297
+ },
298
+ "blimp_transitive": {
299
+ "acc,none": 0.917,
300
+ "acc_stderr,none": 0.008728527206074789,
301
+ "alias": " - blimp_transitive"
302
+ },
303
+ "blimp_wh_island": {
304
+ "acc,none": 0.778,
305
+ "acc_stderr,none": 0.013148721948877364,
306
+ "alias": " - blimp_wh_island"
307
+ },
308
+ "blimp_wh_questions_object_gap": {
309
+ "acc,none": 0.869,
310
+ "acc_stderr,none": 0.010674874844837956,
311
+ "alias": " - blimp_wh_questions_object_gap"
312
+ },
313
+ "blimp_wh_questions_subject_gap": {
314
+ "acc,none": 0.937,
315
+ "acc_stderr,none": 0.00768700787628644,
316
+ "alias": " - blimp_wh_questions_subject_gap"
317
+ },
318
+ "blimp_wh_questions_subject_gap_long_distance": {
319
+ "acc,none": 0.889,
320
+ "acc_stderr,none": 0.009938701010583726,
321
+ "alias": " - blimp_wh_questions_subject_gap_long_distance"
322
+ },
323
+ "blimp_wh_vs_that_no_gap": {
324
+ "acc,none": 0.974,
325
+ "acc_stderr,none": 0.005034813735318246,
326
+ "alias": " - blimp_wh_vs_that_no_gap"
327
+ },
328
+ "blimp_wh_vs_that_no_gap_long_distance": {
329
+ "acc,none": 0.974,
330
+ "acc_stderr,none": 0.005034813735318229,
331
+ "alias": " - blimp_wh_vs_that_no_gap_long_distance"
332
+ },
333
+ "blimp_wh_vs_that_with_gap": {
334
+ "acc,none": 0.508,
335
+ "acc_stderr,none": 0.015817274929209008,
336
+ "alias": " - blimp_wh_vs_that_with_gap"
337
+ },
338
+ "blimp_wh_vs_that_with_gap_long_distance": {
339
+ "acc,none": 0.401,
340
+ "acc_stderr,none": 0.015506109745498322,
341
+ "alias": " - blimp_wh_vs_that_with_gap_long_distance"
342
+ }
343
+ },
344
+ "groups": {
345
+ "blimp": {
346
+ "acc,none": 0.8387313432835821,
347
+ "acc_stderr,none": 0.14773171752507513,
348
+ "alias": "blimp"
349
+ }
350
+ },
351
+ "configs": {
352
+ "blimp_adjunct_island": {
353
+ "task": "blimp_adjunct_island",
354
+ "group": "blimp",
355
+ "dataset_path": "blimp",
356
+ "dataset_name": "adjunct_island",
357
+ "validation_split": "train",
358
+ "doc_to_text": "",
359
+ "doc_to_target": 0,
360
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
361
+ "description": "",
362
+ "target_delimiter": " ",
363
+ "fewshot_delimiter": "\n\n",
364
+ "num_fewshot": 0,
365
+ "metric_list": [
366
+ {
367
+ "metric": "acc"
368
+ }
369
+ ],
370
+ "output_type": "multiple_choice",
371
+ "repeats": 1,
372
+ "should_decontaminate": true,
373
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
374
+ "metadata": {
375
+ "version": 1.0
376
+ }
377
+ },
378
+ "blimp_anaphor_gender_agreement": {
379
+ "task": "blimp_anaphor_gender_agreement",
380
+ "group": "blimp",
381
+ "dataset_path": "blimp",
382
+ "dataset_name": "anaphor_gender_agreement",
383
+ "validation_split": "train",
384
+ "doc_to_text": "",
385
+ "doc_to_target": 0,
386
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
387
+ "description": "",
388
+ "target_delimiter": " ",
389
+ "fewshot_delimiter": "\n\n",
390
+ "num_fewshot": 0,
391
+ "metric_list": [
392
+ {
393
+ "metric": "acc"
394
+ }
395
+ ],
396
+ "output_type": "multiple_choice",
397
+ "repeats": 1,
398
+ "should_decontaminate": true,
399
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
400
+ "metadata": {
401
+ "version": 1.0
402
+ }
403
+ },
404
+ "blimp_anaphor_number_agreement": {
405
+ "task": "blimp_anaphor_number_agreement",
406
+ "group": "blimp",
407
+ "dataset_path": "blimp",
408
+ "dataset_name": "anaphor_number_agreement",
409
+ "validation_split": "train",
410
+ "doc_to_text": "",
411
+ "doc_to_target": 0,
412
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
413
+ "description": "",
414
+ "target_delimiter": " ",
415
+ "fewshot_delimiter": "\n\n",
416
+ "num_fewshot": 0,
417
+ "metric_list": [
418
+ {
419
+ "metric": "acc"
420
+ }
421
+ ],
422
+ "output_type": "multiple_choice",
423
+ "repeats": 1,
424
+ "should_decontaminate": true,
425
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
426
+ "metadata": {
427
+ "version": 1.0
428
+ }
429
+ },
430
+ "blimp_animate_subject_passive": {
431
+ "task": "blimp_animate_subject_passive",
432
+ "group": "blimp",
433
+ "dataset_path": "blimp",
434
+ "dataset_name": "animate_subject_passive",
435
+ "validation_split": "train",
436
+ "doc_to_text": "",
437
+ "doc_to_target": 0,
438
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
439
+ "description": "",
440
+ "target_delimiter": " ",
441
+ "fewshot_delimiter": "\n\n",
442
+ "num_fewshot": 0,
443
+ "metric_list": [
444
+ {
445
+ "metric": "acc"
446
+ }
447
+ ],
448
+ "output_type": "multiple_choice",
449
+ "repeats": 1,
450
+ "should_decontaminate": true,
451
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
452
+ "metadata": {
453
+ "version": 1.0
454
+ }
455
+ },
456
+ "blimp_animate_subject_trans": {
457
+ "task": "blimp_animate_subject_trans",
458
+ "group": "blimp",
459
+ "dataset_path": "blimp",
460
+ "dataset_name": "animate_subject_trans",
461
+ "validation_split": "train",
462
+ "doc_to_text": "",
463
+ "doc_to_target": 0,
464
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
465
+ "description": "",
466
+ "target_delimiter": " ",
467
+ "fewshot_delimiter": "\n\n",
468
+ "num_fewshot": 0,
469
+ "metric_list": [
470
+ {
471
+ "metric": "acc"
472
+ }
473
+ ],
474
+ "output_type": "multiple_choice",
475
+ "repeats": 1,
476
+ "should_decontaminate": true,
477
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
478
+ "metadata": {
479
+ "version": 1.0
480
+ }
481
+ },
482
+ "blimp_causative": {
483
+ "task": "blimp_causative",
484
+ "group": "blimp",
485
+ "dataset_path": "blimp",
486
+ "dataset_name": "causative",
487
+ "validation_split": "train",
488
+ "doc_to_text": "",
489
+ "doc_to_target": 0,
490
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
491
+ "description": "",
492
+ "target_delimiter": " ",
493
+ "fewshot_delimiter": "\n\n",
494
+ "num_fewshot": 0,
495
+ "metric_list": [
496
+ {
497
+ "metric": "acc"
498
+ }
499
+ ],
500
+ "output_type": "multiple_choice",
501
+ "repeats": 1,
502
+ "should_decontaminate": true,
503
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
504
+ "metadata": {
505
+ "version": 1.0
506
+ }
507
+ },
508
+ "blimp_complex_NP_island": {
509
+ "task": "blimp_complex_NP_island",
510
+ "group": "blimp",
511
+ "dataset_path": "blimp",
512
+ "dataset_name": "complex_NP_island",
513
+ "validation_split": "train",
514
+ "doc_to_text": "",
515
+ "doc_to_target": 0,
516
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
517
+ "description": "",
518
+ "target_delimiter": " ",
519
+ "fewshot_delimiter": "\n\n",
520
+ "num_fewshot": 0,
521
+ "metric_list": [
522
+ {
523
+ "metric": "acc"
524
+ }
525
+ ],
526
+ "output_type": "multiple_choice",
527
+ "repeats": 1,
528
+ "should_decontaminate": true,
529
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
530
+ "metadata": {
531
+ "version": 1.0
532
+ }
533
+ },
534
+ "blimp_coordinate_structure_constraint_complex_left_branch": {
535
+ "task": "blimp_coordinate_structure_constraint_complex_left_branch",
536
+ "group": "blimp",
537
+ "dataset_path": "blimp",
538
+ "dataset_name": "coordinate_structure_constraint_complex_left_branch",
539
+ "validation_split": "train",
540
+ "doc_to_text": "",
541
+ "doc_to_target": 0,
542
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
543
+ "description": "",
544
+ "target_delimiter": " ",
545
+ "fewshot_delimiter": "\n\n",
546
+ "num_fewshot": 0,
547
+ "metric_list": [
548
+ {
549
+ "metric": "acc"
550
+ }
551
+ ],
552
+ "output_type": "multiple_choice",
553
+ "repeats": 1,
554
+ "should_decontaminate": true,
555
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
556
+ "metadata": {
557
+ "version": 1.0
558
+ }
559
+ },
560
+ "blimp_coordinate_structure_constraint_object_extraction": {
561
+ "task": "blimp_coordinate_structure_constraint_object_extraction",
562
+ "group": "blimp",
563
+ "dataset_path": "blimp",
564
+ "dataset_name": "coordinate_structure_constraint_object_extraction",
565
+ "validation_split": "train",
566
+ "doc_to_text": "",
567
+ "doc_to_target": 0,
568
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
569
+ "description": "",
570
+ "target_delimiter": " ",
571
+ "fewshot_delimiter": "\n\n",
572
+ "num_fewshot": 0,
573
+ "metric_list": [
574
+ {
575
+ "metric": "acc"
576
+ }
577
+ ],
578
+ "output_type": "multiple_choice",
579
+ "repeats": 1,
580
+ "should_decontaminate": true,
581
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
582
+ "metadata": {
583
+ "version": 1.0
584
+ }
585
+ },
586
+ "blimp_determiner_noun_agreement_1": {
587
+ "task": "blimp_determiner_noun_agreement_1",
588
+ "group": "blimp",
589
+ "dataset_path": "blimp",
590
+ "dataset_name": "determiner_noun_agreement_1",
591
+ "validation_split": "train",
592
+ "doc_to_text": "",
593
+ "doc_to_target": 0,
594
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
595
+ "description": "",
596
+ "target_delimiter": " ",
597
+ "fewshot_delimiter": "\n\n",
598
+ "num_fewshot": 0,
599
+ "metric_list": [
600
+ {
601
+ "metric": "acc"
602
+ }
603
+ ],
604
+ "output_type": "multiple_choice",
605
+ "repeats": 1,
606
+ "should_decontaminate": true,
607
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
608
+ "metadata": {
609
+ "version": 1.0
610
+ }
611
+ },
612
+ "blimp_determiner_noun_agreement_2": {
613
+ "task": "blimp_determiner_noun_agreement_2",
614
+ "group": "blimp",
615
+ "dataset_path": "blimp",
616
+ "dataset_name": "determiner_noun_agreement_2",
617
+ "validation_split": "train",
618
+ "doc_to_text": "",
619
+ "doc_to_target": 0,
620
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
621
+ "description": "",
622
+ "target_delimiter": " ",
623
+ "fewshot_delimiter": "\n\n",
624
+ "num_fewshot": 0,
625
+ "metric_list": [
626
+ {
627
+ "metric": "acc"
628
+ }
629
+ ],
630
+ "output_type": "multiple_choice",
631
+ "repeats": 1,
632
+ "should_decontaminate": true,
633
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
634
+ "metadata": {
635
+ "version": 1.0
636
+ }
637
+ },
638
+ "blimp_determiner_noun_agreement_irregular_1": {
639
+ "task": "blimp_determiner_noun_agreement_irregular_1",
640
+ "group": "blimp",
641
+ "dataset_path": "blimp",
642
+ "dataset_name": "determiner_noun_agreement_irregular_1",
643
+ "validation_split": "train",
644
+ "doc_to_text": "",
645
+ "doc_to_target": 0,
646
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
647
+ "description": "",
648
+ "target_delimiter": " ",
649
+ "fewshot_delimiter": "\n\n",
650
+ "num_fewshot": 0,
651
+ "metric_list": [
652
+ {
653
+ "metric": "acc"
654
+ }
655
+ ],
656
+ "output_type": "multiple_choice",
657
+ "repeats": 1,
658
+ "should_decontaminate": true,
659
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
660
+ "metadata": {
661
+ "version": 1.0
662
+ }
663
+ },
664
+ "blimp_determiner_noun_agreement_irregular_2": {
665
+ "task": "blimp_determiner_noun_agreement_irregular_2",
666
+ "group": "blimp",
667
+ "dataset_path": "blimp",
668
+ "dataset_name": "determiner_noun_agreement_irregular_2",
669
+ "validation_split": "train",
670
+ "doc_to_text": "",
671
+ "doc_to_target": 0,
672
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
673
+ "description": "",
674
+ "target_delimiter": " ",
675
+ "fewshot_delimiter": "\n\n",
676
+ "num_fewshot": 0,
677
+ "metric_list": [
678
+ {
679
+ "metric": "acc"
680
+ }
681
+ ],
682
+ "output_type": "multiple_choice",
683
+ "repeats": 1,
684
+ "should_decontaminate": true,
685
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
686
+ "metadata": {
687
+ "version": 1.0
688
+ }
689
+ },
690
+ "blimp_determiner_noun_agreement_with_adj_2": {
691
+ "task": "blimp_determiner_noun_agreement_with_adj_2",
692
+ "group": "blimp",
693
+ "dataset_path": "blimp",
694
+ "dataset_name": "determiner_noun_agreement_with_adj_2",
695
+ "validation_split": "train",
696
+ "doc_to_text": "",
697
+ "doc_to_target": 0,
698
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
699
+ "description": "",
700
+ "target_delimiter": " ",
701
+ "fewshot_delimiter": "\n\n",
702
+ "num_fewshot": 0,
703
+ "metric_list": [
704
+ {
705
+ "metric": "acc"
706
+ }
707
+ ],
708
+ "output_type": "multiple_choice",
709
+ "repeats": 1,
710
+ "should_decontaminate": true,
711
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
712
+ "metadata": {
713
+ "version": 1.0
714
+ }
715
+ },
716
+ "blimp_determiner_noun_agreement_with_adj_irregular_1": {
717
+ "task": "blimp_determiner_noun_agreement_with_adj_irregular_1",
718
+ "group": "blimp",
719
+ "dataset_path": "blimp",
720
+ "dataset_name": "determiner_noun_agreement_with_adj_irregular_1",
721
+ "validation_split": "train",
722
+ "doc_to_text": "",
723
+ "doc_to_target": 0,
724
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
725
+ "description": "",
726
+ "target_delimiter": " ",
727
+ "fewshot_delimiter": "\n\n",
728
+ "num_fewshot": 0,
729
+ "metric_list": [
730
+ {
731
+ "metric": "acc"
732
+ }
733
+ ],
734
+ "output_type": "multiple_choice",
735
+ "repeats": 1,
736
+ "should_decontaminate": true,
737
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
738
+ "metadata": {
739
+ "version": 1.0
740
+ }
741
+ },
742
+ "blimp_determiner_noun_agreement_with_adj_irregular_2": {
743
+ "task": "blimp_determiner_noun_agreement_with_adj_irregular_2",
744
+ "group": "blimp",
745
+ "dataset_path": "blimp",
746
+ "dataset_name": "determiner_noun_agreement_with_adj_irregular_2",
747
+ "validation_split": "train",
748
+ "doc_to_text": "",
749
+ "doc_to_target": 0,
750
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
751
+ "description": "",
752
+ "target_delimiter": " ",
753
+ "fewshot_delimiter": "\n\n",
754
+ "num_fewshot": 0,
755
+ "metric_list": [
756
+ {
757
+ "metric": "acc"
758
+ }
759
+ ],
760
+ "output_type": "multiple_choice",
761
+ "repeats": 1,
762
+ "should_decontaminate": true,
763
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
764
+ "metadata": {
765
+ "version": 1.0
766
+ }
767
+ },
768
+ "blimp_determiner_noun_agreement_with_adjective_1": {
769
+ "task": "blimp_determiner_noun_agreement_with_adjective_1",
770
+ "group": "blimp",
771
+ "dataset_path": "blimp",
772
+ "dataset_name": "determiner_noun_agreement_with_adjective_1",
773
+ "validation_split": "train",
774
+ "doc_to_text": "",
775
+ "doc_to_target": 0,
776
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
777
+ "description": "",
778
+ "target_delimiter": " ",
779
+ "fewshot_delimiter": "\n\n",
780
+ "num_fewshot": 0,
781
+ "metric_list": [
782
+ {
783
+ "metric": "acc"
784
+ }
785
+ ],
786
+ "output_type": "multiple_choice",
787
+ "repeats": 1,
788
+ "should_decontaminate": true,
789
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
790
+ "metadata": {
791
+ "version": 1.0
792
+ }
793
+ },
794
+ "blimp_distractor_agreement_relational_noun": {
795
+ "task": "blimp_distractor_agreement_relational_noun",
796
+ "group": "blimp",
797
+ "dataset_path": "blimp",
798
+ "dataset_name": "distractor_agreement_relational_noun",
799
+ "validation_split": "train",
800
+ "doc_to_text": "",
801
+ "doc_to_target": 0,
802
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
803
+ "description": "",
804
+ "target_delimiter": " ",
805
+ "fewshot_delimiter": "\n\n",
806
+ "num_fewshot": 0,
807
+ "metric_list": [
808
+ {
809
+ "metric": "acc"
810
+ }
811
+ ],
812
+ "output_type": "multiple_choice",
813
+ "repeats": 1,
814
+ "should_decontaminate": true,
815
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
816
+ "metadata": {
817
+ "version": 1.0
818
+ }
819
+ },
820
+ "blimp_distractor_agreement_relative_clause": {
821
+ "task": "blimp_distractor_agreement_relative_clause",
822
+ "group": "blimp",
823
+ "dataset_path": "blimp",
824
+ "dataset_name": "distractor_agreement_relative_clause",
825
+ "validation_split": "train",
826
+ "doc_to_text": "",
827
+ "doc_to_target": 0,
828
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
829
+ "description": "",
830
+ "target_delimiter": " ",
831
+ "fewshot_delimiter": "\n\n",
832
+ "num_fewshot": 0,
833
+ "metric_list": [
834
+ {
835
+ "metric": "acc"
836
+ }
837
+ ],
838
+ "output_type": "multiple_choice",
839
+ "repeats": 1,
840
+ "should_decontaminate": true,
841
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
842
+ "metadata": {
843
+ "version": 1.0
844
+ }
845
+ },
846
+ "blimp_drop_argument": {
847
+ "task": "blimp_drop_argument",
848
+ "group": "blimp",
849
+ "dataset_path": "blimp",
850
+ "dataset_name": "drop_argument",
851
+ "validation_split": "train",
852
+ "doc_to_text": "",
853
+ "doc_to_target": 0,
854
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
855
+ "description": "",
856
+ "target_delimiter": " ",
857
+ "fewshot_delimiter": "\n\n",
858
+ "num_fewshot": 0,
859
+ "metric_list": [
860
+ {
861
+ "metric": "acc"
862
+ }
863
+ ],
864
+ "output_type": "multiple_choice",
865
+ "repeats": 1,
866
+ "should_decontaminate": true,
867
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
868
+ "metadata": {
869
+ "version": 1.0
870
+ }
871
+ },
872
+ "blimp_ellipsis_n_bar_1": {
873
+ "task": "blimp_ellipsis_n_bar_1",
874
+ "group": "blimp",
875
+ "dataset_path": "blimp",
876
+ "dataset_name": "ellipsis_n_bar_1",
877
+ "validation_split": "train",
878
+ "doc_to_text": "",
879
+ "doc_to_target": 0,
880
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
881
+ "description": "",
882
+ "target_delimiter": " ",
883
+ "fewshot_delimiter": "\n\n",
884
+ "num_fewshot": 0,
885
+ "metric_list": [
886
+ {
887
+ "metric": "acc"
888
+ }
889
+ ],
890
+ "output_type": "multiple_choice",
891
+ "repeats": 1,
892
+ "should_decontaminate": true,
893
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
894
+ "metadata": {
895
+ "version": 1.0
896
+ }
897
+ },
898
+ "blimp_ellipsis_n_bar_2": {
899
+ "task": "blimp_ellipsis_n_bar_2",
900
+ "group": "blimp",
901
+ "dataset_path": "blimp",
902
+ "dataset_name": "ellipsis_n_bar_2",
903
+ "validation_split": "train",
904
+ "doc_to_text": "",
905
+ "doc_to_target": 0,
906
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
907
+ "description": "",
908
+ "target_delimiter": " ",
909
+ "fewshot_delimiter": "\n\n",
910
+ "num_fewshot": 0,
911
+ "metric_list": [
912
+ {
913
+ "metric": "acc"
914
+ }
915
+ ],
916
+ "output_type": "multiple_choice",
917
+ "repeats": 1,
918
+ "should_decontaminate": true,
919
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
920
+ "metadata": {
921
+ "version": 1.0
922
+ }
923
+ },
924
+ "blimp_existential_there_object_raising": {
925
+ "task": "blimp_existential_there_object_raising",
926
+ "group": "blimp",
927
+ "dataset_path": "blimp",
928
+ "dataset_name": "existential_there_object_raising",
929
+ "validation_split": "train",
930
+ "doc_to_text": "",
931
+ "doc_to_target": 0,
932
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
933
+ "description": "",
934
+ "target_delimiter": " ",
935
+ "fewshot_delimiter": "\n\n",
936
+ "num_fewshot": 0,
937
+ "metric_list": [
938
+ {
939
+ "metric": "acc"
940
+ }
941
+ ],
942
+ "output_type": "multiple_choice",
943
+ "repeats": 1,
944
+ "should_decontaminate": true,
945
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
946
+ "metadata": {
947
+ "version": 1.0
948
+ }
949
+ },
950
+ "blimp_existential_there_quantifiers_1": {
951
+ "task": "blimp_existential_there_quantifiers_1",
952
+ "group": "blimp",
953
+ "dataset_path": "blimp",
954
+ "dataset_name": "existential_there_quantifiers_1",
955
+ "validation_split": "train",
956
+ "doc_to_text": "",
957
+ "doc_to_target": 0,
958
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
959
+ "description": "",
960
+ "target_delimiter": " ",
961
+ "fewshot_delimiter": "\n\n",
962
+ "num_fewshot": 0,
963
+ "metric_list": [
964
+ {
965
+ "metric": "acc"
966
+ }
967
+ ],
968
+ "output_type": "multiple_choice",
969
+ "repeats": 1,
970
+ "should_decontaminate": true,
971
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
972
+ "metadata": {
973
+ "version": 1.0
974
+ }
975
+ },
976
+ "blimp_existential_there_quantifiers_2": {
977
+ "task": "blimp_existential_there_quantifiers_2",
978
+ "group": "blimp",
979
+ "dataset_path": "blimp",
980
+ "dataset_name": "existential_there_quantifiers_2",
981
+ "validation_split": "train",
982
+ "doc_to_text": "",
983
+ "doc_to_target": 0,
984
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
985
+ "description": "",
986
+ "target_delimiter": " ",
987
+ "fewshot_delimiter": "\n\n",
988
+ "num_fewshot": 0,
989
+ "metric_list": [
990
+ {
991
+ "metric": "acc"
992
+ }
993
+ ],
994
+ "output_type": "multiple_choice",
995
+ "repeats": 1,
996
+ "should_decontaminate": true,
997
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
998
+ "metadata": {
999
+ "version": 1.0
1000
+ }
1001
+ },
1002
+ "blimp_existential_there_subject_raising": {
1003
+ "task": "blimp_existential_there_subject_raising",
1004
+ "group": "blimp",
1005
+ "dataset_path": "blimp",
1006
+ "dataset_name": "existential_there_subject_raising",
1007
+ "validation_split": "train",
1008
+ "doc_to_text": "",
1009
+ "doc_to_target": 0,
1010
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
1011
+ "description": "",
1012
+ "target_delimiter": " ",
1013
+ "fewshot_delimiter": "\n\n",
1014
+ "num_fewshot": 0,
1015
+ "metric_list": [
1016
+ {
1017
+ "metric": "acc"
1018
+ }
1019
+ ],
1020
+ "output_type": "multiple_choice",
1021
+ "repeats": 1,
1022
+ "should_decontaminate": true,
1023
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
1024
+ "metadata": {
1025
+ "version": 1.0
1026
+ }
1027
+ },
1028
+ "blimp_expletive_it_object_raising": {
1029
+ "task": "blimp_expletive_it_object_raising",
1030
+ "group": "blimp",
1031
+ "dataset_path": "blimp",
1032
+ "dataset_name": "expletive_it_object_raising",
1033
+ "validation_split": "train",
1034
+ "doc_to_text": "",
1035
+ "doc_to_target": 0,
1036
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
1037
+ "description": "",
1038
+ "target_delimiter": " ",
1039
+ "fewshot_delimiter": "\n\n",
1040
+ "num_fewshot": 0,
1041
+ "metric_list": [
1042
+ {
1043
+ "metric": "acc"
1044
+ }
1045
+ ],
1046
+ "output_type": "multiple_choice",
1047
+ "repeats": 1,
1048
+ "should_decontaminate": true,
1049
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
1050
+ "metadata": {
1051
+ "version": 1.0
1052
+ }
1053
+ },
1054
+ "blimp_inchoative": {
1055
+ "task": "blimp_inchoative",
1056
+ "group": "blimp",
1057
+ "dataset_path": "blimp",
1058
+ "dataset_name": "inchoative",
1059
+ "validation_split": "train",
1060
+ "doc_to_text": "",
1061
+ "doc_to_target": 0,
1062
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
1063
+ "description": "",
1064
+ "target_delimiter": " ",
1065
+ "fewshot_delimiter": "\n\n",
1066
+ "num_fewshot": 0,
1067
+ "metric_list": [
1068
+ {
1069
+ "metric": "acc"
1070
+ }
1071
+ ],
1072
+ "output_type": "multiple_choice",
1073
+ "repeats": 1,
1074
+ "should_decontaminate": true,
1075
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
1076
+ "metadata": {
1077
+ "version": 1.0
1078
+ }
1079
+ },
1080
+ "blimp_intransitive": {
1081
+ "task": "blimp_intransitive",
1082
+ "group": "blimp",
1083
+ "dataset_path": "blimp",
1084
+ "dataset_name": "intransitive",
1085
+ "validation_split": "train",
1086
+ "doc_to_text": "",
1087
+ "doc_to_target": 0,
1088
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
1089
+ "description": "",
1090
+ "target_delimiter": " ",
1091
+ "fewshot_delimiter": "\n\n",
1092
+ "num_fewshot": 0,
1093
+ "metric_list": [
1094
+ {
1095
+ "metric": "acc"
1096
+ }
1097
+ ],
1098
+ "output_type": "multiple_choice",
1099
+ "repeats": 1,
1100
+ "should_decontaminate": true,
1101
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
1102
+ "metadata": {
1103
+ "version": 1.0
1104
+ }
1105
+ },
1106
+ "blimp_irregular_past_participle_adjectives": {
1107
+ "task": "blimp_irregular_past_participle_adjectives",
1108
+ "group": "blimp",
1109
+ "dataset_path": "blimp",
1110
+ "dataset_name": "irregular_past_participle_adjectives",
1111
+ "validation_split": "train",
1112
+ "doc_to_text": "",
1113
+ "doc_to_target": 0,
1114
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
1115
+ "description": "",
1116
+ "target_delimiter": " ",
1117
+ "fewshot_delimiter": "\n\n",
1118
+ "num_fewshot": 0,
1119
+ "metric_list": [
1120
+ {
1121
+ "metric": "acc"
1122
+ }
1123
+ ],
1124
+ "output_type": "multiple_choice",
1125
+ "repeats": 1,
1126
+ "should_decontaminate": true,
1127
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
1128
+ "metadata": {
1129
+ "version": 1.0
1130
+ }
1131
+ },
1132
+ "blimp_irregular_past_participle_verbs": {
1133
+ "task": "blimp_irregular_past_participle_verbs",
1134
+ "group": "blimp",
1135
+ "dataset_path": "blimp",
1136
+ "dataset_name": "irregular_past_participle_verbs",
1137
+ "validation_split": "train",
1138
+ "doc_to_text": "",
1139
+ "doc_to_target": 0,
1140
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
1141
+ "description": "",
1142
+ "target_delimiter": " ",
1143
+ "fewshot_delimiter": "\n\n",
1144
+ "num_fewshot": 0,
1145
+ "metric_list": [
1146
+ {
1147
+ "metric": "acc"
1148
+ }
1149
+ ],
1150
+ "output_type": "multiple_choice",
1151
+ "repeats": 1,
1152
+ "should_decontaminate": true,
1153
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
1154
+ "metadata": {
1155
+ "version": 1.0
1156
+ }
1157
+ },
1158
+ "blimp_irregular_plural_subject_verb_agreement_1": {
1159
+ "task": "blimp_irregular_plural_subject_verb_agreement_1",
1160
+ "group": "blimp",
1161
+ "dataset_path": "blimp",
1162
+ "dataset_name": "irregular_plural_subject_verb_agreement_1",
1163
+ "validation_split": "train",
1164
+ "doc_to_text": "",
1165
+ "doc_to_target": 0,
1166
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
1167
+ "description": "",
1168
+ "target_delimiter": " ",
1169
+ "fewshot_delimiter": "\n\n",
1170
+ "num_fewshot": 0,
1171
+ "metric_list": [
1172
+ {
1173
+ "metric": "acc"
1174
+ }
1175
+ ],
1176
+ "output_type": "multiple_choice",
1177
+ "repeats": 1,
1178
+ "should_decontaminate": true,
1179
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
1180
+ "metadata": {
1181
+ "version": 1.0
1182
+ }
1183
+ },
1184
+ "blimp_irregular_plural_subject_verb_agreement_2": {
1185
+ "task": "blimp_irregular_plural_subject_verb_agreement_2",
1186
+ "group": "blimp",
1187
+ "dataset_path": "blimp",
1188
+ "dataset_name": "irregular_plural_subject_verb_agreement_2",
1189
+ "validation_split": "train",
1190
+ "doc_to_text": "",
1191
+ "doc_to_target": 0,
1192
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
1193
+ "description": "",
1194
+ "target_delimiter": " ",
1195
+ "fewshot_delimiter": "\n\n",
1196
+ "num_fewshot": 0,
1197
+ "metric_list": [
1198
+ {
1199
+ "metric": "acc"
1200
+ }
1201
+ ],
1202
+ "output_type": "multiple_choice",
1203
+ "repeats": 1,
1204
+ "should_decontaminate": true,
1205
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
1206
+ "metadata": {
1207
+ "version": 1.0
1208
+ }
1209
+ },
1210
+ "blimp_left_branch_island_echo_question": {
1211
+ "task": "blimp_left_branch_island_echo_question",
1212
+ "group": "blimp",
1213
+ "dataset_path": "blimp",
1214
+ "dataset_name": "left_branch_island_echo_question",
1215
+ "validation_split": "train",
1216
+ "doc_to_text": "",
1217
+ "doc_to_target": 0,
1218
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
1219
+ "description": "",
1220
+ "target_delimiter": " ",
1221
+ "fewshot_delimiter": "\n\n",
1222
+ "num_fewshot": 0,
1223
+ "metric_list": [
1224
+ {
1225
+ "metric": "acc"
1226
+ }
1227
+ ],
1228
+ "output_type": "multiple_choice",
1229
+ "repeats": 1,
1230
+ "should_decontaminate": true,
1231
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
1232
+ "metadata": {
1233
+ "version": 1.0
1234
+ }
1235
+ },
1236
+ "blimp_left_branch_island_simple_question": {
1237
+ "task": "blimp_left_branch_island_simple_question",
1238
+ "group": "blimp",
1239
+ "dataset_path": "blimp",
1240
+ "dataset_name": "left_branch_island_simple_question",
1241
+ "validation_split": "train",
1242
+ "doc_to_text": "",
1243
+ "doc_to_target": 0,
1244
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
1245
+ "description": "",
1246
+ "target_delimiter": " ",
1247
+ "fewshot_delimiter": "\n\n",
1248
+ "num_fewshot": 0,
1249
+ "metric_list": [
1250
+ {
1251
+ "metric": "acc"
1252
+ }
1253
+ ],
1254
+ "output_type": "multiple_choice",
1255
+ "repeats": 1,
1256
+ "should_decontaminate": true,
1257
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
1258
+ "metadata": {
1259
+ "version": 1.0
1260
+ }
1261
+ },
1262
+ "blimp_matrix_question_npi_licensor_present": {
1263
+ "task": "blimp_matrix_question_npi_licensor_present",
1264
+ "group": "blimp",
1265
+ "dataset_path": "blimp",
1266
+ "dataset_name": "matrix_question_npi_licensor_present",
1267
+ "validation_split": "train",
1268
+ "doc_to_text": "",
1269
+ "doc_to_target": 0,
1270
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
1271
+ "description": "",
1272
+ "target_delimiter": " ",
1273
+ "fewshot_delimiter": "\n\n",
1274
+ "num_fewshot": 0,
1275
+ "metric_list": [
1276
+ {
1277
+ "metric": "acc"
1278
+ }
1279
+ ],
1280
+ "output_type": "multiple_choice",
1281
+ "repeats": 1,
1282
+ "should_decontaminate": true,
1283
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
1284
+ "metadata": {
1285
+ "version": 1.0
1286
+ }
1287
+ },
1288
+ "blimp_npi_present_1": {
1289
+ "task": "blimp_npi_present_1",
1290
+ "group": "blimp",
1291
+ "dataset_path": "blimp",
1292
+ "dataset_name": "npi_present_1",
1293
+ "validation_split": "train",
1294
+ "doc_to_text": "",
1295
+ "doc_to_target": 0,
1296
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
1297
+ "description": "",
1298
+ "target_delimiter": " ",
1299
+ "fewshot_delimiter": "\n\n",
1300
+ "num_fewshot": 0,
1301
+ "metric_list": [
1302
+ {
1303
+ "metric": "acc"
1304
+ }
1305
+ ],
1306
+ "output_type": "multiple_choice",
1307
+ "repeats": 1,
1308
+ "should_decontaminate": true,
1309
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
1310
+ "metadata": {
1311
+ "version": 1.0
1312
+ }
1313
+ },
1314
+ "blimp_npi_present_2": {
1315
+ "task": "blimp_npi_present_2",
1316
+ "group": "blimp",
1317
+ "dataset_path": "blimp",
1318
+ "dataset_name": "npi_present_2",
1319
+ "validation_split": "train",
1320
+ "doc_to_text": "",
1321
+ "doc_to_target": 0,
1322
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
1323
+ "description": "",
1324
+ "target_delimiter": " ",
1325
+ "fewshot_delimiter": "\n\n",
1326
+ "num_fewshot": 0,
1327
+ "metric_list": [
1328
+ {
1329
+ "metric": "acc"
1330
+ }
1331
+ ],
1332
+ "output_type": "multiple_choice",
1333
+ "repeats": 1,
1334
+ "should_decontaminate": true,
1335
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
1336
+ "metadata": {
1337
+ "version": 1.0
1338
+ }
1339
+ },
1340
+ "blimp_only_npi_licensor_present": {
1341
+ "task": "blimp_only_npi_licensor_present",
1342
+ "group": "blimp",
1343
+ "dataset_path": "blimp",
1344
+ "dataset_name": "only_npi_licensor_present",
1345
+ "validation_split": "train",
1346
+ "doc_to_text": "",
1347
+ "doc_to_target": 0,
1348
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
1349
+ "description": "",
1350
+ "target_delimiter": " ",
1351
+ "fewshot_delimiter": "\n\n",
1352
+ "num_fewshot": 0,
1353
+ "metric_list": [
1354
+ {
1355
+ "metric": "acc"
1356
+ }
1357
+ ],
1358
+ "output_type": "multiple_choice",
1359
+ "repeats": 1,
1360
+ "should_decontaminate": true,
1361
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
1362
+ "metadata": {
1363
+ "version": 1.0
1364
+ }
1365
+ },
1366
+ "blimp_only_npi_scope": {
1367
+ "task": "blimp_only_npi_scope",
1368
+ "group": "blimp",
1369
+ "dataset_path": "blimp",
1370
+ "dataset_name": "only_npi_scope",
1371
+ "validation_split": "train",
1372
+ "doc_to_text": "",
1373
+ "doc_to_target": 0,
1374
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
1375
+ "description": "",
1376
+ "target_delimiter": " ",
1377
+ "fewshot_delimiter": "\n\n",
1378
+ "num_fewshot": 0,
1379
+ "metric_list": [
1380
+ {
1381
+ "metric": "acc"
1382
+ }
1383
+ ],
1384
+ "output_type": "multiple_choice",
1385
+ "repeats": 1,
1386
+ "should_decontaminate": true,
1387
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
1388
+ "metadata": {
1389
+ "version": 1.0
1390
+ }
1391
+ },
1392
+ "blimp_passive_1": {
1393
+ "task": "blimp_passive_1",
1394
+ "group": "blimp",
1395
+ "dataset_path": "blimp",
1396
+ "dataset_name": "passive_1",
1397
+ "validation_split": "train",
1398
+ "doc_to_text": "",
1399
+ "doc_to_target": 0,
1400
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
1401
+ "description": "",
1402
+ "target_delimiter": " ",
1403
+ "fewshot_delimiter": "\n\n",
1404
+ "num_fewshot": 0,
1405
+ "metric_list": [
1406
+ {
1407
+ "metric": "acc"
1408
+ }
1409
+ ],
1410
+ "output_type": "multiple_choice",
1411
+ "repeats": 1,
1412
+ "should_decontaminate": true,
1413
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
1414
+ "metadata": {
1415
+ "version": 1.0
1416
+ }
1417
+ },
1418
+ "blimp_passive_2": {
1419
+ "task": "blimp_passive_2",
1420
+ "group": "blimp",
1421
+ "dataset_path": "blimp",
1422
+ "dataset_name": "passive_2",
1423
+ "validation_split": "train",
1424
+ "doc_to_text": "",
1425
+ "doc_to_target": 0,
1426
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
1427
+ "description": "",
1428
+ "target_delimiter": " ",
1429
+ "fewshot_delimiter": "\n\n",
1430
+ "num_fewshot": 0,
1431
+ "metric_list": [
1432
+ {
1433
+ "metric": "acc"
1434
+ }
1435
+ ],
1436
+ "output_type": "multiple_choice",
1437
+ "repeats": 1,
1438
+ "should_decontaminate": true,
1439
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
1440
+ "metadata": {
1441
+ "version": 1.0
1442
+ }
1443
+ },
1444
+ "blimp_principle_A_c_command": {
1445
+ "task": "blimp_principle_A_c_command",
1446
+ "group": "blimp",
1447
+ "dataset_path": "blimp",
1448
+ "dataset_name": "principle_A_c_command",
1449
+ "validation_split": "train",
1450
+ "doc_to_text": "",
1451
+ "doc_to_target": 0,
1452
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
1453
+ "description": "",
1454
+ "target_delimiter": " ",
1455
+ "fewshot_delimiter": "\n\n",
1456
+ "num_fewshot": 0,
1457
+ "metric_list": [
1458
+ {
1459
+ "metric": "acc"
1460
+ }
1461
+ ],
1462
+ "output_type": "multiple_choice",
1463
+ "repeats": 1,
1464
+ "should_decontaminate": true,
1465
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
1466
+ "metadata": {
1467
+ "version": 1.0
1468
+ }
1469
+ },
1470
+ "blimp_principle_A_case_1": {
1471
+ "task": "blimp_principle_A_case_1",
1472
+ "group": "blimp",
1473
+ "dataset_path": "blimp",
1474
+ "dataset_name": "principle_A_case_1",
1475
+ "validation_split": "train",
1476
+ "doc_to_text": "",
1477
+ "doc_to_target": 0,
1478
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
1479
+ "description": "",
1480
+ "target_delimiter": " ",
1481
+ "fewshot_delimiter": "\n\n",
1482
+ "num_fewshot": 0,
1483
+ "metric_list": [
1484
+ {
1485
+ "metric": "acc"
1486
+ }
1487
+ ],
1488
+ "output_type": "multiple_choice",
1489
+ "repeats": 1,
1490
+ "should_decontaminate": true,
1491
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
1492
+ "metadata": {
1493
+ "version": 1.0
1494
+ }
1495
+ },
1496
+ "blimp_principle_A_case_2": {
1497
+ "task": "blimp_principle_A_case_2",
1498
+ "group": "blimp",
1499
+ "dataset_path": "blimp",
1500
+ "dataset_name": "principle_A_case_2",
1501
+ "validation_split": "train",
1502
+ "doc_to_text": "",
1503
+ "doc_to_target": 0,
1504
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
1505
+ "description": "",
1506
+ "target_delimiter": " ",
1507
+ "fewshot_delimiter": "\n\n",
1508
+ "num_fewshot": 0,
1509
+ "metric_list": [
1510
+ {
1511
+ "metric": "acc"
1512
+ }
1513
+ ],
1514
+ "output_type": "multiple_choice",
1515
+ "repeats": 1,
1516
+ "should_decontaminate": true,
1517
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
1518
+ "metadata": {
1519
+ "version": 1.0
1520
+ }
1521
+ },
1522
+ "blimp_principle_A_domain_1": {
1523
+ "task": "blimp_principle_A_domain_1",
1524
+ "group": "blimp",
1525
+ "dataset_path": "blimp",
1526
+ "dataset_name": "principle_A_domain_1",
1527
+ "validation_split": "train",
1528
+ "doc_to_text": "",
1529
+ "doc_to_target": 0,
1530
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
1531
+ "description": "",
1532
+ "target_delimiter": " ",
1533
+ "fewshot_delimiter": "\n\n",
1534
+ "num_fewshot": 0,
1535
+ "metric_list": [
1536
+ {
1537
+ "metric": "acc"
1538
+ }
1539
+ ],
1540
+ "output_type": "multiple_choice",
1541
+ "repeats": 1,
1542
+ "should_decontaminate": true,
1543
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
1544
+ "metadata": {
1545
+ "version": 1.0
1546
+ }
1547
+ },
1548
+ "blimp_principle_A_domain_2": {
1549
+ "task": "blimp_principle_A_domain_2",
1550
+ "group": "blimp",
1551
+ "dataset_path": "blimp",
1552
+ "dataset_name": "principle_A_domain_2",
1553
+ "validation_split": "train",
1554
+ "doc_to_text": "",
1555
+ "doc_to_target": 0,
1556
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
1557
+ "description": "",
1558
+ "target_delimiter": " ",
1559
+ "fewshot_delimiter": "\n\n",
1560
+ "num_fewshot": 0,
1561
+ "metric_list": [
1562
+ {
1563
+ "metric": "acc"
1564
+ }
1565
+ ],
1566
+ "output_type": "multiple_choice",
1567
+ "repeats": 1,
1568
+ "should_decontaminate": true,
1569
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
1570
+ "metadata": {
1571
+ "version": 1.0
1572
+ }
1573
+ },
1574
+ "blimp_principle_A_domain_3": {
1575
+ "task": "blimp_principle_A_domain_3",
1576
+ "group": "blimp",
1577
+ "dataset_path": "blimp",
1578
+ "dataset_name": "principle_A_domain_3",
1579
+ "validation_split": "train",
1580
+ "doc_to_text": "",
1581
+ "doc_to_target": 0,
1582
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
1583
+ "description": "",
1584
+ "target_delimiter": " ",
1585
+ "fewshot_delimiter": "\n\n",
1586
+ "num_fewshot": 0,
1587
+ "metric_list": [
1588
+ {
1589
+ "metric": "acc"
1590
+ }
1591
+ ],
1592
+ "output_type": "multiple_choice",
1593
+ "repeats": 1,
1594
+ "should_decontaminate": true,
1595
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
1596
+ "metadata": {
1597
+ "version": 1.0
1598
+ }
1599
+ },
1600
+ "blimp_principle_A_reconstruction": {
1601
+ "task": "blimp_principle_A_reconstruction",
1602
+ "group": "blimp",
1603
+ "dataset_path": "blimp",
1604
+ "dataset_name": "principle_A_reconstruction",
1605
+ "validation_split": "train",
1606
+ "doc_to_text": "",
1607
+ "doc_to_target": 0,
1608
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
1609
+ "description": "",
1610
+ "target_delimiter": " ",
1611
+ "fewshot_delimiter": "\n\n",
1612
+ "num_fewshot": 0,
1613
+ "metric_list": [
1614
+ {
1615
+ "metric": "acc"
1616
+ }
1617
+ ],
1618
+ "output_type": "multiple_choice",
1619
+ "repeats": 1,
1620
+ "should_decontaminate": true,
1621
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
1622
+ "metadata": {
1623
+ "version": 1.0
1624
+ }
1625
+ },
1626
+ "blimp_regular_plural_subject_verb_agreement_1": {
1627
+ "task": "blimp_regular_plural_subject_verb_agreement_1",
1628
+ "group": "blimp",
1629
+ "dataset_path": "blimp",
1630
+ "dataset_name": "regular_plural_subject_verb_agreement_1",
1631
+ "validation_split": "train",
1632
+ "doc_to_text": "",
1633
+ "doc_to_target": 0,
1634
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
1635
+ "description": "",
1636
+ "target_delimiter": " ",
1637
+ "fewshot_delimiter": "\n\n",
1638
+ "num_fewshot": 0,
1639
+ "metric_list": [
1640
+ {
1641
+ "metric": "acc"
1642
+ }
1643
+ ],
1644
+ "output_type": "multiple_choice",
1645
+ "repeats": 1,
1646
+ "should_decontaminate": true,
1647
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
1648
+ "metadata": {
1649
+ "version": 1.0
1650
+ }
1651
+ },
1652
+ "blimp_regular_plural_subject_verb_agreement_2": {
1653
+ "task": "blimp_regular_plural_subject_verb_agreement_2",
1654
+ "group": "blimp",
1655
+ "dataset_path": "blimp",
1656
+ "dataset_name": "regular_plural_subject_verb_agreement_2",
1657
+ "validation_split": "train",
1658
+ "doc_to_text": "",
1659
+ "doc_to_target": 0,
1660
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
1661
+ "description": "",
1662
+ "target_delimiter": " ",
1663
+ "fewshot_delimiter": "\n\n",
1664
+ "num_fewshot": 0,
1665
+ "metric_list": [
1666
+ {
1667
+ "metric": "acc"
1668
+ }
1669
+ ],
1670
+ "output_type": "multiple_choice",
1671
+ "repeats": 1,
1672
+ "should_decontaminate": true,
1673
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
1674
+ "metadata": {
1675
+ "version": 1.0
1676
+ }
1677
+ },
1678
+ "blimp_sentential_negation_npi_licensor_present": {
1679
+ "task": "blimp_sentential_negation_npi_licensor_present",
1680
+ "group": "blimp",
1681
+ "dataset_path": "blimp",
1682
+ "dataset_name": "sentential_negation_npi_licensor_present",
1683
+ "validation_split": "train",
1684
+ "doc_to_text": "",
1685
+ "doc_to_target": 0,
1686
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
1687
+ "description": "",
1688
+ "target_delimiter": " ",
1689
+ "fewshot_delimiter": "\n\n",
1690
+ "num_fewshot": 0,
1691
+ "metric_list": [
1692
+ {
1693
+ "metric": "acc"
1694
+ }
1695
+ ],
1696
+ "output_type": "multiple_choice",
1697
+ "repeats": 1,
1698
+ "should_decontaminate": true,
1699
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
1700
+ "metadata": {
1701
+ "version": 1.0
1702
+ }
1703
+ },
1704
+ "blimp_sentential_negation_npi_scope": {
1705
+ "task": "blimp_sentential_negation_npi_scope",
1706
+ "group": "blimp",
1707
+ "dataset_path": "blimp",
1708
+ "dataset_name": "sentential_negation_npi_scope",
1709
+ "validation_split": "train",
1710
+ "doc_to_text": "",
1711
+ "doc_to_target": 0,
1712
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
1713
+ "description": "",
1714
+ "target_delimiter": " ",
1715
+ "fewshot_delimiter": "\n\n",
1716
+ "num_fewshot": 0,
1717
+ "metric_list": [
1718
+ {
1719
+ "metric": "acc"
1720
+ }
1721
+ ],
1722
+ "output_type": "multiple_choice",
1723
+ "repeats": 1,
1724
+ "should_decontaminate": true,
1725
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
1726
+ "metadata": {
1727
+ "version": 1.0
1728
+ }
1729
+ },
1730
+ "blimp_sentential_subject_island": {
1731
+ "task": "blimp_sentential_subject_island",
1732
+ "group": "blimp",
1733
+ "dataset_path": "blimp",
1734
+ "dataset_name": "sentential_subject_island",
1735
+ "validation_split": "train",
1736
+ "doc_to_text": "",
1737
+ "doc_to_target": 0,
1738
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
1739
+ "description": "",
1740
+ "target_delimiter": " ",
1741
+ "fewshot_delimiter": "\n\n",
1742
+ "num_fewshot": 0,
1743
+ "metric_list": [
1744
+ {
1745
+ "metric": "acc"
1746
+ }
1747
+ ],
1748
+ "output_type": "multiple_choice",
1749
+ "repeats": 1,
1750
+ "should_decontaminate": true,
1751
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
1752
+ "metadata": {
1753
+ "version": 1.0
1754
+ }
1755
+ },
1756
+ "blimp_superlative_quantifiers_1": {
1757
+ "task": "blimp_superlative_quantifiers_1",
1758
+ "group": "blimp",
1759
+ "dataset_path": "blimp",
1760
+ "dataset_name": "superlative_quantifiers_1",
1761
+ "validation_split": "train",
1762
+ "doc_to_text": "",
1763
+ "doc_to_target": 0,
1764
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
1765
+ "description": "",
1766
+ "target_delimiter": " ",
1767
+ "fewshot_delimiter": "\n\n",
1768
+ "num_fewshot": 0,
1769
+ "metric_list": [
1770
+ {
1771
+ "metric": "acc"
1772
+ }
1773
+ ],
1774
+ "output_type": "multiple_choice",
1775
+ "repeats": 1,
1776
+ "should_decontaminate": true,
1777
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
1778
+ "metadata": {
1779
+ "version": 1.0
1780
+ }
1781
+ },
1782
+ "blimp_superlative_quantifiers_2": {
1783
+ "task": "blimp_superlative_quantifiers_2",
1784
+ "group": "blimp",
1785
+ "dataset_path": "blimp",
1786
+ "dataset_name": "superlative_quantifiers_2",
1787
+ "validation_split": "train",
1788
+ "doc_to_text": "",
1789
+ "doc_to_target": 0,
1790
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
1791
+ "description": "",
1792
+ "target_delimiter": " ",
1793
+ "fewshot_delimiter": "\n\n",
1794
+ "num_fewshot": 0,
1795
+ "metric_list": [
1796
+ {
1797
+ "metric": "acc"
1798
+ }
1799
+ ],
1800
+ "output_type": "multiple_choice",
1801
+ "repeats": 1,
1802
+ "should_decontaminate": true,
1803
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
1804
+ "metadata": {
1805
+ "version": 1.0
1806
+ }
1807
+ },
1808
+ "blimp_tough_vs_raising_1": {
1809
+ "task": "blimp_tough_vs_raising_1",
1810
+ "group": "blimp",
1811
+ "dataset_path": "blimp",
1812
+ "dataset_name": "tough_vs_raising_1",
1813
+ "validation_split": "train",
1814
+ "doc_to_text": "",
1815
+ "doc_to_target": 0,
1816
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
1817
+ "description": "",
1818
+ "target_delimiter": " ",
1819
+ "fewshot_delimiter": "\n\n",
1820
+ "num_fewshot": 0,
1821
+ "metric_list": [
1822
+ {
1823
+ "metric": "acc"
1824
+ }
1825
+ ],
1826
+ "output_type": "multiple_choice",
1827
+ "repeats": 1,
1828
+ "should_decontaminate": true,
1829
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
1830
+ "metadata": {
1831
+ "version": 1.0
1832
+ }
1833
+ },
1834
+ "blimp_tough_vs_raising_2": {
1835
+ "task": "blimp_tough_vs_raising_2",
1836
+ "group": "blimp",
1837
+ "dataset_path": "blimp",
1838
+ "dataset_name": "tough_vs_raising_2",
1839
+ "validation_split": "train",
1840
+ "doc_to_text": "",
1841
+ "doc_to_target": 0,
1842
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
1843
+ "description": "",
1844
+ "target_delimiter": " ",
1845
+ "fewshot_delimiter": "\n\n",
1846
+ "num_fewshot": 0,
1847
+ "metric_list": [
1848
+ {
1849
+ "metric": "acc"
1850
+ }
1851
+ ],
1852
+ "output_type": "multiple_choice",
1853
+ "repeats": 1,
1854
+ "should_decontaminate": true,
1855
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
1856
+ "metadata": {
1857
+ "version": 1.0
1858
+ }
1859
+ },
1860
+ "blimp_transitive": {
1861
+ "task": "blimp_transitive",
1862
+ "group": "blimp",
1863
+ "dataset_path": "blimp",
1864
+ "dataset_name": "transitive",
1865
+ "validation_split": "train",
1866
+ "doc_to_text": "",
1867
+ "doc_to_target": 0,
1868
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
1869
+ "description": "",
1870
+ "target_delimiter": " ",
1871
+ "fewshot_delimiter": "\n\n",
1872
+ "num_fewshot": 0,
1873
+ "metric_list": [
1874
+ {
1875
+ "metric": "acc"
1876
+ }
1877
+ ],
1878
+ "output_type": "multiple_choice",
1879
+ "repeats": 1,
1880
+ "should_decontaminate": true,
1881
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
1882
+ "metadata": {
1883
+ "version": 1.0
1884
+ }
1885
+ },
1886
+ "blimp_wh_island": {
1887
+ "task": "blimp_wh_island",
1888
+ "group": "blimp",
1889
+ "dataset_path": "blimp",
1890
+ "dataset_name": "wh_island",
1891
+ "validation_split": "train",
1892
+ "doc_to_text": "",
1893
+ "doc_to_target": 0,
1894
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
1895
+ "description": "",
1896
+ "target_delimiter": " ",
1897
+ "fewshot_delimiter": "\n\n",
1898
+ "num_fewshot": 0,
1899
+ "metric_list": [
1900
+ {
1901
+ "metric": "acc"
1902
+ }
1903
+ ],
1904
+ "output_type": "multiple_choice",
1905
+ "repeats": 1,
1906
+ "should_decontaminate": true,
1907
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
1908
+ "metadata": {
1909
+ "version": 1.0
1910
+ }
1911
+ },
1912
+ "blimp_wh_questions_object_gap": {
1913
+ "task": "blimp_wh_questions_object_gap",
1914
+ "group": "blimp",
1915
+ "dataset_path": "blimp",
1916
+ "dataset_name": "wh_questions_object_gap",
1917
+ "validation_split": "train",
1918
+ "doc_to_text": "",
1919
+ "doc_to_target": 0,
1920
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
1921
+ "description": "",
1922
+ "target_delimiter": " ",
1923
+ "fewshot_delimiter": "\n\n",
1924
+ "num_fewshot": 0,
1925
+ "metric_list": [
1926
+ {
1927
+ "metric": "acc"
1928
+ }
1929
+ ],
1930
+ "output_type": "multiple_choice",
1931
+ "repeats": 1,
1932
+ "should_decontaminate": true,
1933
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
1934
+ "metadata": {
1935
+ "version": 1.0
1936
+ }
1937
+ },
1938
+ "blimp_wh_questions_subject_gap": {
1939
+ "task": "blimp_wh_questions_subject_gap",
1940
+ "group": "blimp",
1941
+ "dataset_path": "blimp",
1942
+ "dataset_name": "wh_questions_subject_gap",
1943
+ "validation_split": "train",
1944
+ "doc_to_text": "",
1945
+ "doc_to_target": 0,
1946
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
1947
+ "description": "",
1948
+ "target_delimiter": " ",
1949
+ "fewshot_delimiter": "\n\n",
1950
+ "num_fewshot": 0,
1951
+ "metric_list": [
1952
+ {
1953
+ "metric": "acc"
1954
+ }
1955
+ ],
1956
+ "output_type": "multiple_choice",
1957
+ "repeats": 1,
1958
+ "should_decontaminate": true,
1959
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
1960
+ "metadata": {
1961
+ "version": 1.0
1962
+ }
1963
+ },
1964
+ "blimp_wh_questions_subject_gap_long_distance": {
1965
+ "task": "blimp_wh_questions_subject_gap_long_distance",
1966
+ "group": "blimp",
1967
+ "dataset_path": "blimp",
1968
+ "dataset_name": "wh_questions_subject_gap_long_distance",
1969
+ "validation_split": "train",
1970
+ "doc_to_text": "",
1971
+ "doc_to_target": 0,
1972
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
1973
+ "description": "",
1974
+ "target_delimiter": " ",
1975
+ "fewshot_delimiter": "\n\n",
1976
+ "num_fewshot": 0,
1977
+ "metric_list": [
1978
+ {
1979
+ "metric": "acc"
1980
+ }
1981
+ ],
1982
+ "output_type": "multiple_choice",
1983
+ "repeats": 1,
1984
+ "should_decontaminate": true,
1985
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
1986
+ "metadata": {
1987
+ "version": 1.0
1988
+ }
1989
+ },
1990
+ "blimp_wh_vs_that_no_gap": {
1991
+ "task": "blimp_wh_vs_that_no_gap",
1992
+ "group": "blimp",
1993
+ "dataset_path": "blimp",
1994
+ "dataset_name": "wh_vs_that_no_gap",
1995
+ "validation_split": "train",
1996
+ "doc_to_text": "",
1997
+ "doc_to_target": 0,
1998
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
1999
+ "description": "",
2000
+ "target_delimiter": " ",
2001
+ "fewshot_delimiter": "\n\n",
2002
+ "num_fewshot": 0,
2003
+ "metric_list": [
2004
+ {
2005
+ "metric": "acc"
2006
+ }
2007
+ ],
2008
+ "output_type": "multiple_choice",
2009
+ "repeats": 1,
2010
+ "should_decontaminate": true,
2011
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
2012
+ "metadata": {
2013
+ "version": 1.0
2014
+ }
2015
+ },
2016
+ "blimp_wh_vs_that_no_gap_long_distance": {
2017
+ "task": "blimp_wh_vs_that_no_gap_long_distance",
2018
+ "group": "blimp",
2019
+ "dataset_path": "blimp",
2020
+ "dataset_name": "wh_vs_that_no_gap_long_distance",
2021
+ "validation_split": "train",
2022
+ "doc_to_text": "",
2023
+ "doc_to_target": 0,
2024
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
2025
+ "description": "",
2026
+ "target_delimiter": " ",
2027
+ "fewshot_delimiter": "\n\n",
2028
+ "num_fewshot": 0,
2029
+ "metric_list": [
2030
+ {
2031
+ "metric": "acc"
2032
+ }
2033
+ ],
2034
+ "output_type": "multiple_choice",
2035
+ "repeats": 1,
2036
+ "should_decontaminate": true,
2037
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
2038
+ "metadata": {
2039
+ "version": 1.0
2040
+ }
2041
+ },
2042
+ "blimp_wh_vs_that_with_gap": {
2043
+ "task": "blimp_wh_vs_that_with_gap",
2044
+ "group": "blimp",
2045
+ "dataset_path": "blimp",
2046
+ "dataset_name": "wh_vs_that_with_gap",
2047
+ "validation_split": "train",
2048
+ "doc_to_text": "",
2049
+ "doc_to_target": 0,
2050
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
2051
+ "description": "",
2052
+ "target_delimiter": " ",
2053
+ "fewshot_delimiter": "\n\n",
2054
+ "num_fewshot": 0,
2055
+ "metric_list": [
2056
+ {
2057
+ "metric": "acc"
2058
+ }
2059
+ ],
2060
+ "output_type": "multiple_choice",
2061
+ "repeats": 1,
2062
+ "should_decontaminate": true,
2063
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
2064
+ "metadata": {
2065
+ "version": 1.0
2066
+ }
2067
+ },
2068
+ "blimp_wh_vs_that_with_gap_long_distance": {
2069
+ "task": "blimp_wh_vs_that_with_gap_long_distance",
2070
+ "group": "blimp",
2071
+ "dataset_path": "blimp",
2072
+ "dataset_name": "wh_vs_that_with_gap_long_distance",
2073
+ "validation_split": "train",
2074
+ "doc_to_text": "",
2075
+ "doc_to_target": 0,
2076
+ "doc_to_choice": "{{[sentence_good, sentence_bad]}}",
2077
+ "description": "",
2078
+ "target_delimiter": " ",
2079
+ "fewshot_delimiter": "\n\n",
2080
+ "num_fewshot": 0,
2081
+ "metric_list": [
2082
+ {
2083
+ "metric": "acc"
2084
+ }
2085
+ ],
2086
+ "output_type": "multiple_choice",
2087
+ "repeats": 1,
2088
+ "should_decontaminate": true,
2089
+ "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}",
2090
+ "metadata": {
2091
+ "version": 1.0
2092
+ }
2093
+ }
2094
+ },
2095
+ "versions": {
2096
+ "blimp": "N/A",
2097
+ "blimp_adjunct_island": 1.0,
2098
+ "blimp_anaphor_gender_agreement": 1.0,
2099
+ "blimp_anaphor_number_agreement": 1.0,
2100
+ "blimp_animate_subject_passive": 1.0,
2101
+ "blimp_animate_subject_trans": 1.0,
2102
+ "blimp_causative": 1.0,
2103
+ "blimp_complex_NP_island": 1.0,
2104
+ "blimp_coordinate_structure_constraint_complex_left_branch": 1.0,
2105
+ "blimp_coordinate_structure_constraint_object_extraction": 1.0,
2106
+ "blimp_determiner_noun_agreement_1": 1.0,
2107
+ "blimp_determiner_noun_agreement_2": 1.0,
2108
+ "blimp_determiner_noun_agreement_irregular_1": 1.0,
2109
+ "blimp_determiner_noun_agreement_irregular_2": 1.0,
2110
+ "blimp_determiner_noun_agreement_with_adj_2": 1.0,
2111
+ "blimp_determiner_noun_agreement_with_adj_irregular_1": 1.0,
2112
+ "blimp_determiner_noun_agreement_with_adj_irregular_2": 1.0,
2113
+ "blimp_determiner_noun_agreement_with_adjective_1": 1.0,
2114
+ "blimp_distractor_agreement_relational_noun": 1.0,
2115
+ "blimp_distractor_agreement_relative_clause": 1.0,
2116
+ "blimp_drop_argument": 1.0,
2117
+ "blimp_ellipsis_n_bar_1": 1.0,
2118
+ "blimp_ellipsis_n_bar_2": 1.0,
2119
+ "blimp_existential_there_object_raising": 1.0,
2120
+ "blimp_existential_there_quantifiers_1": 1.0,
2121
+ "blimp_existential_there_quantifiers_2": 1.0,
2122
+ "blimp_existential_there_subject_raising": 1.0,
2123
+ "blimp_expletive_it_object_raising": 1.0,
2124
+ "blimp_inchoative": 1.0,
2125
+ "blimp_intransitive": 1.0,
2126
+ "blimp_irregular_past_participle_adjectives": 1.0,
2127
+ "blimp_irregular_past_participle_verbs": 1.0,
2128
+ "blimp_irregular_plural_subject_verb_agreement_1": 1.0,
2129
+ "blimp_irregular_plural_subject_verb_agreement_2": 1.0,
2130
+ "blimp_left_branch_island_echo_question": 1.0,
2131
+ "blimp_left_branch_island_simple_question": 1.0,
2132
+ "blimp_matrix_question_npi_licensor_present": 1.0,
2133
+ "blimp_npi_present_1": 1.0,
2134
+ "blimp_npi_present_2": 1.0,
2135
+ "blimp_only_npi_licensor_present": 1.0,
2136
+ "blimp_only_npi_scope": 1.0,
2137
+ "blimp_passive_1": 1.0,
2138
+ "blimp_passive_2": 1.0,
2139
+ "blimp_principle_A_c_command": 1.0,
2140
+ "blimp_principle_A_case_1": 1.0,
2141
+ "blimp_principle_A_case_2": 1.0,
2142
+ "blimp_principle_A_domain_1": 1.0,
2143
+ "blimp_principle_A_domain_2": 1.0,
2144
+ "blimp_principle_A_domain_3": 1.0,
2145
+ "blimp_principle_A_reconstruction": 1.0,
2146
+ "blimp_regular_plural_subject_verb_agreement_1": 1.0,
2147
+ "blimp_regular_plural_subject_verb_agreement_2": 1.0,
2148
+ "blimp_sentential_negation_npi_licensor_present": 1.0,
2149
+ "blimp_sentential_negation_npi_scope": 1.0,
2150
+ "blimp_sentential_subject_island": 1.0,
2151
+ "blimp_superlative_quantifiers_1": 1.0,
2152
+ "blimp_superlative_quantifiers_2": 1.0,
2153
+ "blimp_tough_vs_raising_1": 1.0,
2154
+ "blimp_tough_vs_raising_2": 1.0,
2155
+ "blimp_transitive": 1.0,
2156
+ "blimp_wh_island": 1.0,
2157
+ "blimp_wh_questions_object_gap": 1.0,
2158
+ "blimp_wh_questions_subject_gap": 1.0,
2159
+ "blimp_wh_questions_subject_gap_long_distance": 1.0,
2160
+ "blimp_wh_vs_that_no_gap": 1.0,
2161
+ "blimp_wh_vs_that_no_gap_long_distance": 1.0,
2162
+ "blimp_wh_vs_that_with_gap": 1.0,
2163
+ "blimp_wh_vs_that_with_gap_long_distance": 1.0
2164
+ },
2165
+ "n-shot": {
2166
+ "blimp": 0,
2167
+ "blimp_adjunct_island": 0,
2168
+ "blimp_anaphor_gender_agreement": 0,
2169
+ "blimp_anaphor_number_agreement": 0,
2170
+ "blimp_animate_subject_passive": 0,
2171
+ "blimp_animate_subject_trans": 0,
2172
+ "blimp_causative": 0,
2173
+ "blimp_complex_NP_island": 0,
2174
+ "blimp_coordinate_structure_constraint_complex_left_branch": 0,
2175
+ "blimp_coordinate_structure_constraint_object_extraction": 0,
2176
+ "blimp_determiner_noun_agreement_1": 0,
2177
+ "blimp_determiner_noun_agreement_2": 0,
2178
+ "blimp_determiner_noun_agreement_irregular_1": 0,
2179
+ "blimp_determiner_noun_agreement_irregular_2": 0,
2180
+ "blimp_determiner_noun_agreement_with_adj_2": 0,
2181
+ "blimp_determiner_noun_agreement_with_adj_irregular_1": 0,
2182
+ "blimp_determiner_noun_agreement_with_adj_irregular_2": 0,
2183
+ "blimp_determiner_noun_agreement_with_adjective_1": 0,
2184
+ "blimp_distractor_agreement_relational_noun": 0,
2185
+ "blimp_distractor_agreement_relative_clause": 0,
2186
+ "blimp_drop_argument": 0,
2187
+ "blimp_ellipsis_n_bar_1": 0,
2188
+ "blimp_ellipsis_n_bar_2": 0,
2189
+ "blimp_existential_there_object_raising": 0,
2190
+ "blimp_existential_there_quantifiers_1": 0,
2191
+ "blimp_existential_there_quantifiers_2": 0,
2192
+ "blimp_existential_there_subject_raising": 0,
2193
+ "blimp_expletive_it_object_raising": 0,
2194
+ "blimp_inchoative": 0,
2195
+ "blimp_intransitive": 0,
2196
+ "blimp_irregular_past_participle_adjectives": 0,
2197
+ "blimp_irregular_past_participle_verbs": 0,
2198
+ "blimp_irregular_plural_subject_verb_agreement_1": 0,
2199
+ "blimp_irregular_plural_subject_verb_agreement_2": 0,
2200
+ "blimp_left_branch_island_echo_question": 0,
2201
+ "blimp_left_branch_island_simple_question": 0,
2202
+ "blimp_matrix_question_npi_licensor_present": 0,
2203
+ "blimp_npi_present_1": 0,
2204
+ "blimp_npi_present_2": 0,
2205
+ "blimp_only_npi_licensor_present": 0,
2206
+ "blimp_only_npi_scope": 0,
2207
+ "blimp_passive_1": 0,
2208
+ "blimp_passive_2": 0,
2209
+ "blimp_principle_A_c_command": 0,
2210
+ "blimp_principle_A_case_1": 0,
2211
+ "blimp_principle_A_case_2": 0,
2212
+ "blimp_principle_A_domain_1": 0,
2213
+ "blimp_principle_A_domain_2": 0,
2214
+ "blimp_principle_A_domain_3": 0,
2215
+ "blimp_principle_A_reconstruction": 0,
2216
+ "blimp_regular_plural_subject_verb_agreement_1": 0,
2217
+ "blimp_regular_plural_subject_verb_agreement_2": 0,
2218
+ "blimp_sentential_negation_npi_licensor_present": 0,
2219
+ "blimp_sentential_negation_npi_scope": 0,
2220
+ "blimp_sentential_subject_island": 0,
2221
+ "blimp_superlative_quantifiers_1": 0,
2222
+ "blimp_superlative_quantifiers_2": 0,
2223
+ "blimp_tough_vs_raising_1": 0,
2224
+ "blimp_tough_vs_raising_2": 0,
2225
+ "blimp_transitive": 0,
2226
+ "blimp_wh_island": 0,
2227
+ "blimp_wh_questions_object_gap": 0,
2228
+ "blimp_wh_questions_subject_gap": 0,
2229
+ "blimp_wh_questions_subject_gap_long_distance": 0,
2230
+ "blimp_wh_vs_that_no_gap": 0,
2231
+ "blimp_wh_vs_that_no_gap_long_distance": 0,
2232
+ "blimp_wh_vs_that_with_gap": 0,
2233
+ "blimp_wh_vs_that_with_gap_long_distance": 0
2234
+ },
2235
+ "config": {
2236
+ "model": "hf",
2237
+ "model_args": "pretrained=RWKV/rwkv-4-world-1b5,dtype=bfloat16,trust_remote_code=True",
2238
+ "batch_size": "auto",
2239
+ "batch_sizes": [
2240
+ 64
2241
+ ],
2242
+ "device": null,
2243
+ "use_cache": null,
2244
+ "limit": null,
2245
+ "bootstrap_iters": 100000,
2246
+ "gen_kwargs": null
2247
+ },
2248
+ "git_hash": "99f5004"
2249
+ }
lm-eval-output/RWKV/rwkv-4-world-1b5/blimp/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f46e2a97227219bcf9932b4050748b96cb78b74e78c43fb6f9cb552d58822b69
3
+ size 259425
lm-eval-output/RWKV/rwkv-4-world-1b5/boolq/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:359bf45c42d237db982203f895eb689624c6ec906863e2cd6bb83af00c75db22
3
+ size 1143482
lm-eval-output/RWKV/rwkv-4-world-1b5/boolq/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json ADDED
@@ -0,0 +1,62 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "boolq": {
4
+ "acc,none": 0.5712538226299694,
5
+ "acc_stderr,none": 0.008655800332760226,
6
+ "alias": "boolq"
7
+ }
8
+ },
9
+ "configs": {
10
+ "boolq": {
11
+ "task": "boolq",
12
+ "group": [
13
+ "super-glue-lm-eval-v1"
14
+ ],
15
+ "dataset_path": "super_glue",
16
+ "dataset_name": "boolq",
17
+ "training_split": "train",
18
+ "validation_split": "validation",
19
+ "doc_to_text": "{{passage}}\nQuestion: {{question}}?\nAnswer:",
20
+ "doc_to_target": "label",
21
+ "doc_to_choice": [
22
+ "no",
23
+ "yes"
24
+ ],
25
+ "description": "",
26
+ "target_delimiter": " ",
27
+ "fewshot_delimiter": "\n\n",
28
+ "metric_list": [
29
+ {
30
+ "metric": "acc"
31
+ }
32
+ ],
33
+ "output_type": "multiple_choice",
34
+ "repeats": 1,
35
+ "should_decontaminate": true,
36
+ "doc_to_decontamination_query": "passage",
37
+ "metadata": {
38
+ "version": 2.0
39
+ }
40
+ }
41
+ },
42
+ "versions": {
43
+ "boolq": 2.0
44
+ },
45
+ "n-shot": {
46
+ "boolq": 0
47
+ },
48
+ "config": {
49
+ "model": "hf",
50
+ "model_args": "pretrained=RWKV/rwkv-4-world-1b5,dtype=bfloat16,trust_remote_code=True",
51
+ "batch_size": "auto",
52
+ "batch_sizes": [
53
+ 32
54
+ ],
55
+ "device": null,
56
+ "use_cache": null,
57
+ "limit": null,
58
+ "bootstrap_iters": 100000,
59
+ "gen_kwargs": null
60
+ },
61
+ "git_hash": "99f5004"
62
+ }
lm-eval-output/RWKV/rwkv-4-world-1b5/boolq/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:04dd49321d659819f4b01c56c03cc2a447b8fbb402e4bcd708934f9814ac5ff9
3
+ size 14327
lm-eval-output/RWKV/rwkv-4-world-1b5/cb/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c180968c265e8e3a06bba65a5ab53f8f7e45c6bf53a10e8042bfdc3012e8a641
3
+ size 14178
lm-eval-output/RWKV/rwkv-4-world-1b5/cb/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json ADDED
@@ -0,0 +1,68 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "cb": {
4
+ "acc,none": 0.35714285714285715,
5
+ "acc_stderr,none": 0.0646095738380922,
6
+ "f1,none": 0.31203703703703706,
7
+ "f1_stderr,none": "N/A",
8
+ "alias": "cb"
9
+ }
10
+ },
11
+ "configs": {
12
+ "cb": {
13
+ "task": "cb",
14
+ "group": [
15
+ "super-glue-lm-eval-v1"
16
+ ],
17
+ "dataset_path": "super_glue",
18
+ "dataset_name": "cb",
19
+ "training_split": "train",
20
+ "validation_split": "validation",
21
+ "doc_to_text": "{{premise}}\nQuestion: {{hypothesis}}. True, False, or Neither?\nAnswer:",
22
+ "doc_to_target": "label",
23
+ "doc_to_choice": [
24
+ "True",
25
+ "False",
26
+ "Neither"
27
+ ],
28
+ "description": "",
29
+ "target_delimiter": " ",
30
+ "fewshot_delimiter": "\n\n",
31
+ "metric_list": [
32
+ {
33
+ "metric": "acc"
34
+ },
35
+ {
36
+ "metric": "f1",
37
+ "aggregation": "def cb_multi_fi(items):\n preds, golds = zip(*items)\n preds = np.array(preds)\n golds = np.array(golds)\n f11 = sklearn.metrics.f1_score(y_true=golds == 0, y_pred=preds == 0)\n f12 = sklearn.metrics.f1_score(y_true=golds == 1, y_pred=preds == 1)\n f13 = sklearn.metrics.f1_score(y_true=golds == 2, y_pred=preds == 2)\n avg_f1 = np.mean([f11, f12, f13])\n return avg_f1\n"
38
+ }
39
+ ],
40
+ "output_type": "multiple_choice",
41
+ "repeats": 1,
42
+ "should_decontaminate": false,
43
+ "metadata": {
44
+ "version": 1.0
45
+ }
46
+ }
47
+ },
48
+ "versions": {
49
+ "cb": 1.0
50
+ },
51
+ "n-shot": {
52
+ "cb": 0
53
+ },
54
+ "config": {
55
+ "model": "hf",
56
+ "model_args": "pretrained=RWKV/rwkv-4-world-1b5,dtype=bfloat16,trust_remote_code=True",
57
+ "batch_size": "auto",
58
+ "batch_sizes": [
59
+ 64
60
+ ],
61
+ "device": null,
62
+ "use_cache": null,
63
+ "limit": null,
64
+ "bootstrap_iters": 100000,
65
+ "gen_kwargs": null
66
+ },
67
+ "git_hash": "99f5004"
68
+ }
lm-eval-output/RWKV/rwkv-4-world-1b5/cb/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6cedb9d13addc428c5a27fa69a76b871810dc574f87cc1b85435ef4656ce27c7
3
+ size 16568
lm-eval-output/RWKV/rwkv-4-world-1b5/ceval-valid/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a09ed4326e30bb9b0b4afe547f66ca473ebc1d1709ba2530ea5334bcbf17a6f3
3
+ size 323020
lm-eval-output/RWKV/rwkv-4-world-1b5/ceval-valid/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json ADDED
@@ -0,0 +1,2590 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "ceval-valid": {
4
+ "acc,none": 0.2310549777117385,
5
+ "acc_stderr,none": 0.10765310032289531,
6
+ "acc_norm,none": 0.2310549777117385,
7
+ "acc_norm_stderr,none": 0.10765310032289531,
8
+ "alias": "ceval-valid"
9
+ },
10
+ "ceval-valid_accountant": {
11
+ "acc,none": 0.22448979591836735,
12
+ "acc_stderr,none": 0.06022425581505364,
13
+ "acc_norm,none": 0.22448979591836735,
14
+ "acc_norm_stderr,none": 0.06022425581505364,
15
+ "alias": " - ceval-valid_accountant"
16
+ },
17
+ "ceval-valid_advanced_mathematics": {
18
+ "acc,none": 0.3157894736842105,
19
+ "acc_stderr,none": 0.10956136839295433,
20
+ "acc_norm,none": 0.3157894736842105,
21
+ "acc_norm_stderr,none": 0.10956136839295433,
22
+ "alias": " - ceval-valid_advanced_mathematics"
23
+ },
24
+ "ceval-valid_art_studies": {
25
+ "acc,none": 0.45454545454545453,
26
+ "acc_stderr,none": 0.08802234877744129,
27
+ "acc_norm,none": 0.45454545454545453,
28
+ "acc_norm_stderr,none": 0.08802234877744129,
29
+ "alias": " - ceval-valid_art_studies"
30
+ },
31
+ "ceval-valid_basic_medicine": {
32
+ "acc,none": 0.05263157894736842,
33
+ "acc_stderr,none": 0.05263157894736841,
34
+ "acc_norm,none": 0.05263157894736842,
35
+ "acc_norm_stderr,none": 0.05263157894736841,
36
+ "alias": " - ceval-valid_basic_medicine"
37
+ },
38
+ "ceval-valid_business_administration": {
39
+ "acc,none": 0.24242424242424243,
40
+ "acc_stderr,none": 0.07575757575757577,
41
+ "acc_norm,none": 0.24242424242424243,
42
+ "acc_norm_stderr,none": 0.07575757575757577,
43
+ "alias": " - ceval-valid_business_administration"
44
+ },
45
+ "ceval-valid_chinese_language_and_literature": {
46
+ "acc,none": 0.21739130434782608,
47
+ "acc_stderr,none": 0.08793911249520547,
48
+ "acc_norm,none": 0.21739130434782608,
49
+ "acc_norm_stderr,none": 0.08793911249520547,
50
+ "alias": " - ceval-valid_chinese_language_and_literature"
51
+ },
52
+ "ceval-valid_civil_servant": {
53
+ "acc,none": 0.2553191489361702,
54
+ "acc_stderr,none": 0.06429065810876616,
55
+ "acc_norm,none": 0.2553191489361702,
56
+ "acc_norm_stderr,none": 0.06429065810876616,
57
+ "alias": " - ceval-valid_civil_servant"
58
+ },
59
+ "ceval-valid_clinical_medicine": {
60
+ "acc,none": 0.22727272727272727,
61
+ "acc_stderr,none": 0.09144861547306321,
62
+ "acc_norm,none": 0.22727272727272727,
63
+ "acc_norm_stderr,none": 0.09144861547306321,
64
+ "alias": " - ceval-valid_clinical_medicine"
65
+ },
66
+ "ceval-valid_college_chemistry": {
67
+ "acc,none": 0.125,
68
+ "acc_stderr,none": 0.06895966054592131,
69
+ "acc_norm,none": 0.125,
70
+ "acc_norm_stderr,none": 0.06895966054592131,
71
+ "alias": " - ceval-valid_college_chemistry"
72
+ },
73
+ "ceval-valid_college_economics": {
74
+ "acc,none": 0.2909090909090909,
75
+ "acc_stderr,none": 0.06180629713445796,
76
+ "acc_norm,none": 0.2909090909090909,
77
+ "acc_norm_stderr,none": 0.06180629713445796,
78
+ "alias": " - ceval-valid_college_economics"
79
+ },
80
+ "ceval-valid_college_physics": {
81
+ "acc,none": 0.21052631578947367,
82
+ "acc_stderr,none": 0.0960916767552923,
83
+ "acc_norm,none": 0.21052631578947367,
84
+ "acc_norm_stderr,none": 0.0960916767552923,
85
+ "alias": " - ceval-valid_college_physics"
86
+ },
87
+ "ceval-valid_college_programming": {
88
+ "acc,none": 0.24324324324324326,
89
+ "acc_stderr,none": 0.07150679219093488,
90
+ "acc_norm,none": 0.24324324324324326,
91
+ "acc_norm_stderr,none": 0.07150679219093488,
92
+ "alias": " - ceval-valid_college_programming"
93
+ },
94
+ "ceval-valid_computer_architecture": {
95
+ "acc,none": 0.2857142857142857,
96
+ "acc_stderr,none": 0.10101525445522108,
97
+ "acc_norm,none": 0.2857142857142857,
98
+ "acc_norm_stderr,none": 0.10101525445522108,
99
+ "alias": " - ceval-valid_computer_architecture"
100
+ },
101
+ "ceval-valid_computer_network": {
102
+ "acc,none": 0.21052631578947367,
103
+ "acc_stderr,none": 0.0960916767552923,
104
+ "acc_norm,none": 0.21052631578947367,
105
+ "acc_norm_stderr,none": 0.0960916767552923,
106
+ "alias": " - ceval-valid_computer_network"
107
+ },
108
+ "ceval-valid_discrete_mathematics": {
109
+ "acc,none": 0.375,
110
+ "acc_stderr,none": 0.125,
111
+ "acc_norm,none": 0.375,
112
+ "acc_norm_stderr,none": 0.125,
113
+ "alias": " - ceval-valid_discrete_mathematics"
114
+ },
115
+ "ceval-valid_education_science": {
116
+ "acc,none": 0.2413793103448276,
117
+ "acc_stderr,none": 0.080869237238335,
118
+ "acc_norm,none": 0.2413793103448276,
119
+ "acc_norm_stderr,none": 0.080869237238335,
120
+ "alias": " - ceval-valid_education_science"
121
+ },
122
+ "ceval-valid_electrical_engineer": {
123
+ "acc,none": 0.21621621621621623,
124
+ "acc_stderr,none": 0.06861056852129647,
125
+ "acc_norm,none": 0.21621621621621623,
126
+ "acc_norm_stderr,none": 0.06861056852129647,
127
+ "alias": " - ceval-valid_electrical_engineer"
128
+ },
129
+ "ceval-valid_environmental_impact_assessment_engineer": {
130
+ "acc,none": 0.16129032258064516,
131
+ "acc_stderr,none": 0.06715051611181073,
132
+ "acc_norm,none": 0.16129032258064516,
133
+ "acc_norm_stderr,none": 0.06715051611181073,
134
+ "alias": " - ceval-valid_environmental_impact_assessment_engineer"
135
+ },
136
+ "ceval-valid_fire_engineer": {
137
+ "acc,none": 0.25806451612903225,
138
+ "acc_stderr,none": 0.0798889274021794,
139
+ "acc_norm,none": 0.25806451612903225,
140
+ "acc_norm_stderr,none": 0.0798889274021794,
141
+ "alias": " - ceval-valid_fire_engineer"
142
+ },
143
+ "ceval-valid_high_school_biology": {
144
+ "acc,none": 0.3684210526315789,
145
+ "acc_stderr,none": 0.11369720523522558,
146
+ "acc_norm,none": 0.3684210526315789,
147
+ "acc_norm_stderr,none": 0.11369720523522558,
148
+ "alias": " - ceval-valid_high_school_biology"
149
+ },
150
+ "ceval-valid_high_school_chemistry": {
151
+ "acc,none": 0.21052631578947367,
152
+ "acc_stderr,none": 0.0960916767552923,
153
+ "acc_norm,none": 0.21052631578947367,
154
+ "acc_norm_stderr,none": 0.0960916767552923,
155
+ "alias": " - ceval-valid_high_school_chemistry"
156
+ },
157
+ "ceval-valid_high_school_chinese": {
158
+ "acc,none": 0.21052631578947367,
159
+ "acc_stderr,none": 0.0960916767552923,
160
+ "acc_norm,none": 0.21052631578947367,
161
+ "acc_norm_stderr,none": 0.0960916767552923,
162
+ "alias": " - ceval-valid_high_school_chinese"
163
+ },
164
+ "ceval-valid_high_school_geography": {
165
+ "acc,none": 0.21052631578947367,
166
+ "acc_stderr,none": 0.0960916767552923,
167
+ "acc_norm,none": 0.21052631578947367,
168
+ "acc_norm_stderr,none": 0.0960916767552923,
169
+ "alias": " - ceval-valid_high_school_geography"
170
+ },
171
+ "ceval-valid_high_school_history": {
172
+ "acc,none": 0.3,
173
+ "acc_stderr,none": 0.10513149660756933,
174
+ "acc_norm,none": 0.3,
175
+ "acc_norm_stderr,none": 0.10513149660756933,
176
+ "alias": " - ceval-valid_high_school_history"
177
+ },
178
+ "ceval-valid_high_school_mathematics": {
179
+ "acc,none": 0.2222222222222222,
180
+ "acc_stderr,none": 0.10083169033033672,
181
+ "acc_norm,none": 0.2222222222222222,
182
+ "acc_norm_stderr,none": 0.10083169033033672,
183
+ "alias": " - ceval-valid_high_school_mathematics"
184
+ },
185
+ "ceval-valid_high_school_physics": {
186
+ "acc,none": 0.21052631578947367,
187
+ "acc_stderr,none": 0.0960916767552923,
188
+ "acc_norm,none": 0.21052631578947367,
189
+ "acc_norm_stderr,none": 0.0960916767552923,
190
+ "alias": " - ceval-valid_high_school_physics"
191
+ },
192
+ "ceval-valid_high_school_politics": {
193
+ "acc,none": 0.21052631578947367,
194
+ "acc_stderr,none": 0.0960916767552923,
195
+ "acc_norm,none": 0.21052631578947367,
196
+ "acc_norm_stderr,none": 0.0960916767552923,
197
+ "alias": " - ceval-valid_high_school_politics"
198
+ },
199
+ "ceval-valid_ideological_and_moral_cultivation": {
200
+ "acc,none": 0.2631578947368421,
201
+ "acc_stderr,none": 0.10379087338771256,
202
+ "acc_norm,none": 0.2631578947368421,
203
+ "acc_norm_stderr,none": 0.10379087338771256,
204
+ "alias": " - ceval-valid_ideological_and_moral_cultivation"
205
+ },
206
+ "ceval-valid_law": {
207
+ "acc,none": 0.25,
208
+ "acc_stderr,none": 0.09028938981432691,
209
+ "acc_norm,none": 0.25,
210
+ "acc_norm_stderr,none": 0.09028938981432691,
211
+ "alias": " - ceval-valid_law"
212
+ },
213
+ "ceval-valid_legal_professional": {
214
+ "acc,none": 0.043478260869565216,
215
+ "acc_stderr,none": 0.04347826086956523,
216
+ "acc_norm,none": 0.043478260869565216,
217
+ "acc_norm_stderr,none": 0.04347826086956523,
218
+ "alias": " - ceval-valid_legal_professional"
219
+ },
220
+ "ceval-valid_logic": {
221
+ "acc,none": 0.18181818181818182,
222
+ "acc_stderr,none": 0.08416546361568647,
223
+ "acc_norm,none": 0.18181818181818182,
224
+ "acc_norm_stderr,none": 0.08416546361568647,
225
+ "alias": " - ceval-valid_logic"
226
+ },
227
+ "ceval-valid_mao_zedong_thought": {
228
+ "acc,none": 0.3333333333333333,
229
+ "acc_stderr,none": 0.0982946374365981,
230
+ "acc_norm,none": 0.3333333333333333,
231
+ "acc_norm_stderr,none": 0.0982946374365981,
232
+ "alias": " - ceval-valid_mao_zedong_thought"
233
+ },
234
+ "ceval-valid_marxism": {
235
+ "acc,none": 0.2631578947368421,
236
+ "acc_stderr,none": 0.10379087338771256,
237
+ "acc_norm,none": 0.2631578947368421,
238
+ "acc_norm_stderr,none": 0.10379087338771256,
239
+ "alias": " - ceval-valid_marxism"
240
+ },
241
+ "ceval-valid_metrology_engineer": {
242
+ "acc,none": 0.125,
243
+ "acc_stderr,none": 0.06895966054592131,
244
+ "acc_norm,none": 0.125,
245
+ "acc_norm_stderr,none": 0.06895966054592131,
246
+ "alias": " - ceval-valid_metrology_engineer"
247
+ },
248
+ "ceval-valid_middle_school_biology": {
249
+ "acc,none": 0.14285714285714285,
250
+ "acc_stderr,none": 0.07824607964359517,
251
+ "acc_norm,none": 0.14285714285714285,
252
+ "acc_norm_stderr,none": 0.07824607964359517,
253
+ "alias": " - ceval-valid_middle_school_biology"
254
+ },
255
+ "ceval-valid_middle_school_chemistry": {
256
+ "acc,none": 0.15,
257
+ "acc_stderr,none": 0.0819178021909125,
258
+ "acc_norm,none": 0.15,
259
+ "acc_norm_stderr,none": 0.0819178021909125,
260
+ "alias": " - ceval-valid_middle_school_chemistry"
261
+ },
262
+ "ceval-valid_middle_school_geography": {
263
+ "acc,none": 0.08333333333333333,
264
+ "acc_stderr,none": 0.08333333333333331,
265
+ "acc_norm,none": 0.08333333333333333,
266
+ "acc_norm_stderr,none": 0.08333333333333331,
267
+ "alias": " - ceval-valid_middle_school_geography"
268
+ },
269
+ "ceval-valid_middle_school_history": {
270
+ "acc,none": 0.18181818181818182,
271
+ "acc_stderr,none": 0.08416546361568647,
272
+ "acc_norm,none": 0.18181818181818182,
273
+ "acc_norm_stderr,none": 0.08416546361568647,
274
+ "alias": " - ceval-valid_middle_school_history"
275
+ },
276
+ "ceval-valid_middle_school_mathematics": {
277
+ "acc,none": 0.15789473684210525,
278
+ "acc_stderr,none": 0.08594700851870798,
279
+ "acc_norm,none": 0.15789473684210525,
280
+ "acc_norm_stderr,none": 0.08594700851870798,
281
+ "alias": " - ceval-valid_middle_school_mathematics"
282
+ },
283
+ "ceval-valid_middle_school_physics": {
284
+ "acc,none": 0.21052631578947367,
285
+ "acc_stderr,none": 0.0960916767552923,
286
+ "acc_norm,none": 0.21052631578947367,
287
+ "acc_norm_stderr,none": 0.0960916767552923,
288
+ "alias": " - ceval-valid_middle_school_physics"
289
+ },
290
+ "ceval-valid_middle_school_politics": {
291
+ "acc,none": 0.2857142857142857,
292
+ "acc_stderr,none": 0.10101525445522108,
293
+ "acc_norm,none": 0.2857142857142857,
294
+ "acc_norm_stderr,none": 0.10101525445522108,
295
+ "alias": " - ceval-valid_middle_school_politics"
296
+ },
297
+ "ceval-valid_modern_chinese_history": {
298
+ "acc,none": 0.17391304347826086,
299
+ "acc_stderr,none": 0.08081046758996392,
300
+ "acc_norm,none": 0.17391304347826086,
301
+ "acc_norm_stderr,none": 0.08081046758996392,
302
+ "alias": " - ceval-valid_modern_chinese_history"
303
+ },
304
+ "ceval-valid_operating_system": {
305
+ "acc,none": 0.15789473684210525,
306
+ "acc_stderr,none": 0.08594700851870798,
307
+ "acc_norm,none": 0.15789473684210525,
308
+ "acc_norm_stderr,none": 0.08594700851870798,
309
+ "alias": " - ceval-valid_operating_system"
310
+ },
311
+ "ceval-valid_physician": {
312
+ "acc,none": 0.2653061224489796,
313
+ "acc_stderr,none": 0.06372446937141223,
314
+ "acc_norm,none": 0.2653061224489796,
315
+ "acc_norm_stderr,none": 0.06372446937141223,
316
+ "alias": " - ceval-valid_physician"
317
+ },
318
+ "ceval-valid_plant_protection": {
319
+ "acc,none": 0.3181818181818182,
320
+ "acc_stderr,none": 0.10163945352271772,
321
+ "acc_norm,none": 0.3181818181818182,
322
+ "acc_norm_stderr,none": 0.10163945352271772,
323
+ "alias": " - ceval-valid_plant_protection"
324
+ },
325
+ "ceval-valid_probability_and_statistics": {
326
+ "acc,none": 0.1111111111111111,
327
+ "acc_stderr,none": 0.07622159339667062,
328
+ "acc_norm,none": 0.1111111111111111,
329
+ "acc_norm_stderr,none": 0.07622159339667062,
330
+ "alias": " - ceval-valid_probability_and_statistics"
331
+ },
332
+ "ceval-valid_professional_tour_guide": {
333
+ "acc,none": 0.3448275862068966,
334
+ "acc_stderr,none": 0.08982552969857373,
335
+ "acc_norm,none": 0.3448275862068966,
336
+ "acc_norm_stderr,none": 0.08982552969857373,
337
+ "alias": " - ceval-valid_professional_tour_guide"
338
+ },
339
+ "ceval-valid_sports_science": {
340
+ "acc,none": 0.10526315789473684,
341
+ "acc_stderr,none": 0.07233518641434489,
342
+ "acc_norm,none": 0.10526315789473684,
343
+ "acc_norm_stderr,none": 0.07233518641434489,
344
+ "alias": " - ceval-valid_sports_science"
345
+ },
346
+ "ceval-valid_tax_accountant": {
347
+ "acc,none": 0.20408163265306123,
348
+ "acc_stderr,none": 0.05817221556628254,
349
+ "acc_norm,none": 0.20408163265306123,
350
+ "acc_norm_stderr,none": 0.05817221556628254,
351
+ "alias": " - ceval-valid_tax_accountant"
352
+ },
353
+ "ceval-valid_teacher_qualification": {
354
+ "acc,none": 0.29545454545454547,
355
+ "acc_stderr,none": 0.06957698714453991,
356
+ "acc_norm,none": 0.29545454545454547,
357
+ "acc_norm_stderr,none": 0.06957698714453991,
358
+ "alias": " - ceval-valid_teacher_qualification"
359
+ },
360
+ "ceval-valid_urban_and_rural_planner": {
361
+ "acc,none": 0.21739130434782608,
362
+ "acc_stderr,none": 0.06148754619013454,
363
+ "acc_norm,none": 0.21739130434782608,
364
+ "acc_norm_stderr,none": 0.06148754619013454,
365
+ "alias": " - ceval-valid_urban_and_rural_planner"
366
+ },
367
+ "ceval-valid_veterinary_medicine": {
368
+ "acc,none": 0.21739130434782608,
369
+ "acc_stderr,none": 0.08793911249520547,
370
+ "acc_norm,none": 0.21739130434782608,
371
+ "acc_norm_stderr,none": 0.08793911249520547,
372
+ "alias": " - ceval-valid_veterinary_medicine"
373
+ }
374
+ },
375
+ "groups": {
376
+ "ceval-valid": {
377
+ "acc,none": 0.2310549777117385,
378
+ "acc_stderr,none": 0.10765310032289531,
379
+ "acc_norm,none": 0.2310549777117385,
380
+ "acc_norm_stderr,none": 0.10765310032289531,
381
+ "alias": "ceval-valid"
382
+ }
383
+ },
384
+ "configs": {
385
+ "ceval-valid_accountant": {
386
+ "task": "ceval-valid_accountant",
387
+ "group": "ceval-valid",
388
+ "dataset_path": "ceval/ceval-exam",
389
+ "dataset_name": "accountant",
390
+ "validation_split": "val",
391
+ "fewshot_split": "dev",
392
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
393
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
394
+ "doc_to_choice": [
395
+ "A",
396
+ "B",
397
+ "C",
398
+ "D"
399
+ ],
400
+ "description": "以下是中国关于注册会计师的单项选择题,请选出其中的正确答案。\n\n",
401
+ "target_delimiter": " ",
402
+ "fewshot_delimiter": "\n\n",
403
+ "fewshot_config": {
404
+ "sampler": "first_n"
405
+ },
406
+ "metric_list": [
407
+ {
408
+ "metric": "acc",
409
+ "aggregation": "mean",
410
+ "higher_is_better": true
411
+ },
412
+ {
413
+ "metric": "acc_norm",
414
+ "aggregation": "mean",
415
+ "higher_is_better": true
416
+ }
417
+ ],
418
+ "output_type": "multiple_choice",
419
+ "repeats": 1,
420
+ "should_decontaminate": false,
421
+ "metadata": {
422
+ "version": 1.0
423
+ }
424
+ },
425
+ "ceval-valid_advanced_mathematics": {
426
+ "task": "ceval-valid_advanced_mathematics",
427
+ "group": "ceval-valid",
428
+ "dataset_path": "ceval/ceval-exam",
429
+ "dataset_name": "advanced_mathematics",
430
+ "validation_split": "val",
431
+ "fewshot_split": "dev",
432
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
433
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
434
+ "doc_to_choice": [
435
+ "A",
436
+ "B",
437
+ "C",
438
+ "D"
439
+ ],
440
+ "description": "以下是中国关于高等数学的单项选择题,请选出其中的正确答案。\n\n",
441
+ "target_delimiter": " ",
442
+ "fewshot_delimiter": "\n\n",
443
+ "fewshot_config": {
444
+ "sampler": "first_n"
445
+ },
446
+ "metric_list": [
447
+ {
448
+ "metric": "acc",
449
+ "aggregation": "mean",
450
+ "higher_is_better": true
451
+ },
452
+ {
453
+ "metric": "acc_norm",
454
+ "aggregation": "mean",
455
+ "higher_is_better": true
456
+ }
457
+ ],
458
+ "output_type": "multiple_choice",
459
+ "repeats": 1,
460
+ "should_decontaminate": false,
461
+ "metadata": {
462
+ "version": 1.0
463
+ }
464
+ },
465
+ "ceval-valid_art_studies": {
466
+ "task": "ceval-valid_art_studies",
467
+ "group": "ceval-valid",
468
+ "dataset_path": "ceval/ceval-exam",
469
+ "dataset_name": "art_studies",
470
+ "validation_split": "val",
471
+ "fewshot_split": "dev",
472
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
473
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
474
+ "doc_to_choice": [
475
+ "A",
476
+ "B",
477
+ "C",
478
+ "D"
479
+ ],
480
+ "description": "以下是中国关于艺术学的单项选择题,请选出其中的正确答案。\n\n",
481
+ "target_delimiter": " ",
482
+ "fewshot_delimiter": "\n\n",
483
+ "fewshot_config": {
484
+ "sampler": "first_n"
485
+ },
486
+ "metric_list": [
487
+ {
488
+ "metric": "acc",
489
+ "aggregation": "mean",
490
+ "higher_is_better": true
491
+ },
492
+ {
493
+ "metric": "acc_norm",
494
+ "aggregation": "mean",
495
+ "higher_is_better": true
496
+ }
497
+ ],
498
+ "output_type": "multiple_choice",
499
+ "repeats": 1,
500
+ "should_decontaminate": false,
501
+ "metadata": {
502
+ "version": 1.0
503
+ }
504
+ },
505
+ "ceval-valid_basic_medicine": {
506
+ "task": "ceval-valid_basic_medicine",
507
+ "group": "ceval-valid",
508
+ "dataset_path": "ceval/ceval-exam",
509
+ "dataset_name": "basic_medicine",
510
+ "validation_split": "val",
511
+ "fewshot_split": "dev",
512
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
513
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
514
+ "doc_to_choice": [
515
+ "A",
516
+ "B",
517
+ "C",
518
+ "D"
519
+ ],
520
+ "description": "以下是中国关于基础医学的单项选择题,请选出其中的正确答案。\n\n",
521
+ "target_delimiter": " ",
522
+ "fewshot_delimiter": "\n\n",
523
+ "fewshot_config": {
524
+ "sampler": "first_n"
525
+ },
526
+ "metric_list": [
527
+ {
528
+ "metric": "acc",
529
+ "aggregation": "mean",
530
+ "higher_is_better": true
531
+ },
532
+ {
533
+ "metric": "acc_norm",
534
+ "aggregation": "mean",
535
+ "higher_is_better": true
536
+ }
537
+ ],
538
+ "output_type": "multiple_choice",
539
+ "repeats": 1,
540
+ "should_decontaminate": false,
541
+ "metadata": {
542
+ "version": 1.0
543
+ }
544
+ },
545
+ "ceval-valid_business_administration": {
546
+ "task": "ceval-valid_business_administration",
547
+ "group": "ceval-valid",
548
+ "dataset_path": "ceval/ceval-exam",
549
+ "dataset_name": "business_administration",
550
+ "validation_split": "val",
551
+ "fewshot_split": "dev",
552
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
553
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
554
+ "doc_to_choice": [
555
+ "A",
556
+ "B",
557
+ "C",
558
+ "D"
559
+ ],
560
+ "description": "以下是中国关于工商管理的单项选择题,请选出其中的正确答案。\n\n",
561
+ "target_delimiter": " ",
562
+ "fewshot_delimiter": "\n\n",
563
+ "fewshot_config": {
564
+ "sampler": "first_n"
565
+ },
566
+ "metric_list": [
567
+ {
568
+ "metric": "acc",
569
+ "aggregation": "mean",
570
+ "higher_is_better": true
571
+ },
572
+ {
573
+ "metric": "acc_norm",
574
+ "aggregation": "mean",
575
+ "higher_is_better": true
576
+ }
577
+ ],
578
+ "output_type": "multiple_choice",
579
+ "repeats": 1,
580
+ "should_decontaminate": false,
581
+ "metadata": {
582
+ "version": 1.0
583
+ }
584
+ },
585
+ "ceval-valid_chinese_language_and_literature": {
586
+ "task": "ceval-valid_chinese_language_and_literature",
587
+ "group": "ceval-valid",
588
+ "dataset_path": "ceval/ceval-exam",
589
+ "dataset_name": "chinese_language_and_literature",
590
+ "validation_split": "val",
591
+ "fewshot_split": "dev",
592
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
593
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
594
+ "doc_to_choice": [
595
+ "A",
596
+ "B",
597
+ "C",
598
+ "D"
599
+ ],
600
+ "description": "以下是中国关于中国语言文学的单项选择题,请选出其中的正确答案。\n\n",
601
+ "target_delimiter": " ",
602
+ "fewshot_delimiter": "\n\n",
603
+ "fewshot_config": {
604
+ "sampler": "first_n"
605
+ },
606
+ "metric_list": [
607
+ {
608
+ "metric": "acc",
609
+ "aggregation": "mean",
610
+ "higher_is_better": true
611
+ },
612
+ {
613
+ "metric": "acc_norm",
614
+ "aggregation": "mean",
615
+ "higher_is_better": true
616
+ }
617
+ ],
618
+ "output_type": "multiple_choice",
619
+ "repeats": 1,
620
+ "should_decontaminate": false,
621
+ "metadata": {
622
+ "version": 1.0
623
+ }
624
+ },
625
+ "ceval-valid_civil_servant": {
626
+ "task": "ceval-valid_civil_servant",
627
+ "group": "ceval-valid",
628
+ "dataset_path": "ceval/ceval-exam",
629
+ "dataset_name": "civil_servant",
630
+ "validation_split": "val",
631
+ "fewshot_split": "dev",
632
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
633
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
634
+ "doc_to_choice": [
635
+ "A",
636
+ "B",
637
+ "C",
638
+ "D"
639
+ ],
640
+ "description": "以下是中国关于公务员的单项选择题,请选出其中的正确答案。\n\n",
641
+ "target_delimiter": " ",
642
+ "fewshot_delimiter": "\n\n",
643
+ "fewshot_config": {
644
+ "sampler": "first_n"
645
+ },
646
+ "metric_list": [
647
+ {
648
+ "metric": "acc",
649
+ "aggregation": "mean",
650
+ "higher_is_better": true
651
+ },
652
+ {
653
+ "metric": "acc_norm",
654
+ "aggregation": "mean",
655
+ "higher_is_better": true
656
+ }
657
+ ],
658
+ "output_type": "multiple_choice",
659
+ "repeats": 1,
660
+ "should_decontaminate": false,
661
+ "metadata": {
662
+ "version": 1.0
663
+ }
664
+ },
665
+ "ceval-valid_clinical_medicine": {
666
+ "task": "ceval-valid_clinical_medicine",
667
+ "group": "ceval-valid",
668
+ "dataset_path": "ceval/ceval-exam",
669
+ "dataset_name": "clinical_medicine",
670
+ "validation_split": "val",
671
+ "fewshot_split": "dev",
672
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
673
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
674
+ "doc_to_choice": [
675
+ "A",
676
+ "B",
677
+ "C",
678
+ "D"
679
+ ],
680
+ "description": "以下是中国关于临床医学的单项选择题,请选出其中的正确答案。\n\n",
681
+ "target_delimiter": " ",
682
+ "fewshot_delimiter": "\n\n",
683
+ "fewshot_config": {
684
+ "sampler": "first_n"
685
+ },
686
+ "metric_list": [
687
+ {
688
+ "metric": "acc",
689
+ "aggregation": "mean",
690
+ "higher_is_better": true
691
+ },
692
+ {
693
+ "metric": "acc_norm",
694
+ "aggregation": "mean",
695
+ "higher_is_better": true
696
+ }
697
+ ],
698
+ "output_type": "multiple_choice",
699
+ "repeats": 1,
700
+ "should_decontaminate": false,
701
+ "metadata": {
702
+ "version": 1.0
703
+ }
704
+ },
705
+ "ceval-valid_college_chemistry": {
706
+ "task": "ceval-valid_college_chemistry",
707
+ "group": "ceval-valid",
708
+ "dataset_path": "ceval/ceval-exam",
709
+ "dataset_name": "college_chemistry",
710
+ "validation_split": "val",
711
+ "fewshot_split": "dev",
712
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
713
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
714
+ "doc_to_choice": [
715
+ "A",
716
+ "B",
717
+ "C",
718
+ "D"
719
+ ],
720
+ "description": "以下是中国关于大学化学的单项选择题,请选出其中的正确答案���\n\n",
721
+ "target_delimiter": " ",
722
+ "fewshot_delimiter": "\n\n",
723
+ "fewshot_config": {
724
+ "sampler": "first_n"
725
+ },
726
+ "metric_list": [
727
+ {
728
+ "metric": "acc",
729
+ "aggregation": "mean",
730
+ "higher_is_better": true
731
+ },
732
+ {
733
+ "metric": "acc_norm",
734
+ "aggregation": "mean",
735
+ "higher_is_better": true
736
+ }
737
+ ],
738
+ "output_type": "multiple_choice",
739
+ "repeats": 1,
740
+ "should_decontaminate": false,
741
+ "metadata": {
742
+ "version": 1.0
743
+ }
744
+ },
745
+ "ceval-valid_college_economics": {
746
+ "task": "ceval-valid_college_economics",
747
+ "group": "ceval-valid",
748
+ "dataset_path": "ceval/ceval-exam",
749
+ "dataset_name": "college_economics",
750
+ "validation_split": "val",
751
+ "fewshot_split": "dev",
752
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
753
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
754
+ "doc_to_choice": [
755
+ "A",
756
+ "B",
757
+ "C",
758
+ "D"
759
+ ],
760
+ "description": "以下是中国关于大学经济学的单项选择题,请选出其中的正确答案。\n\n",
761
+ "target_delimiter": " ",
762
+ "fewshot_delimiter": "\n\n",
763
+ "fewshot_config": {
764
+ "sampler": "first_n"
765
+ },
766
+ "metric_list": [
767
+ {
768
+ "metric": "acc",
769
+ "aggregation": "mean",
770
+ "higher_is_better": true
771
+ },
772
+ {
773
+ "metric": "acc_norm",
774
+ "aggregation": "mean",
775
+ "higher_is_better": true
776
+ }
777
+ ],
778
+ "output_type": "multiple_choice",
779
+ "repeats": 1,
780
+ "should_decontaminate": false,
781
+ "metadata": {
782
+ "version": 1.0
783
+ }
784
+ },
785
+ "ceval-valid_college_physics": {
786
+ "task": "ceval-valid_college_physics",
787
+ "group": "ceval-valid",
788
+ "dataset_path": "ceval/ceval-exam",
789
+ "dataset_name": "college_physics",
790
+ "validation_split": "val",
791
+ "fewshot_split": "dev",
792
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
793
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
794
+ "doc_to_choice": [
795
+ "A",
796
+ "B",
797
+ "C",
798
+ "D"
799
+ ],
800
+ "description": "以下是中国关于大学物理的单项选择题,请选出其中的正确答案。\n\n",
801
+ "target_delimiter": " ",
802
+ "fewshot_delimiter": "\n\n",
803
+ "fewshot_config": {
804
+ "sampler": "first_n"
805
+ },
806
+ "metric_list": [
807
+ {
808
+ "metric": "acc",
809
+ "aggregation": "mean",
810
+ "higher_is_better": true
811
+ },
812
+ {
813
+ "metric": "acc_norm",
814
+ "aggregation": "mean",
815
+ "higher_is_better": true
816
+ }
817
+ ],
818
+ "output_type": "multiple_choice",
819
+ "repeats": 1,
820
+ "should_decontaminate": false,
821
+ "metadata": {
822
+ "version": 1.0
823
+ }
824
+ },
825
+ "ceval-valid_college_programming": {
826
+ "task": "ceval-valid_college_programming",
827
+ "group": "ceval-valid",
828
+ "dataset_path": "ceval/ceval-exam",
829
+ "dataset_name": "college_programming",
830
+ "validation_split": "val",
831
+ "fewshot_split": "dev",
832
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
833
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
834
+ "doc_to_choice": [
835
+ "A",
836
+ "B",
837
+ "C",
838
+ "D"
839
+ ],
840
+ "description": "以下是中国关于大学编程的单项选择题,请选出其中的正确答案。\n\n",
841
+ "target_delimiter": " ",
842
+ "fewshot_delimiter": "\n\n",
843
+ "fewshot_config": {
844
+ "sampler": "first_n"
845
+ },
846
+ "metric_list": [
847
+ {
848
+ "metric": "acc",
849
+ "aggregation": "mean",
850
+ "higher_is_better": true
851
+ },
852
+ {
853
+ "metric": "acc_norm",
854
+ "aggregation": "mean",
855
+ "higher_is_better": true
856
+ }
857
+ ],
858
+ "output_type": "multiple_choice",
859
+ "repeats": 1,
860
+ "should_decontaminate": false,
861
+ "metadata": {
862
+ "version": 1.0
863
+ }
864
+ },
865
+ "ceval-valid_computer_architecture": {
866
+ "task": "ceval-valid_computer_architecture",
867
+ "group": "ceval-valid",
868
+ "dataset_path": "ceval/ceval-exam",
869
+ "dataset_name": "computer_architecture",
870
+ "validation_split": "val",
871
+ "fewshot_split": "dev",
872
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
873
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
874
+ "doc_to_choice": [
875
+ "A",
876
+ "B",
877
+ "C",
878
+ "D"
879
+ ],
880
+ "description": "以下是中国关于计算机组成的单项选择题,请选出其中的正确答案。\n\n",
881
+ "target_delimiter": " ",
882
+ "fewshot_delimiter": "\n\n",
883
+ "fewshot_config": {
884
+ "sampler": "first_n"
885
+ },
886
+ "metric_list": [
887
+ {
888
+ "metric": "acc",
889
+ "aggregation": "mean",
890
+ "higher_is_better": true
891
+ },
892
+ {
893
+ "metric": "acc_norm",
894
+ "aggregation": "mean",
895
+ "higher_is_better": true
896
+ }
897
+ ],
898
+ "output_type": "multiple_choice",
899
+ "repeats": 1,
900
+ "should_decontaminate": false,
901
+ "metadata": {
902
+ "version": 1.0
903
+ }
904
+ },
905
+ "ceval-valid_computer_network": {
906
+ "task": "ceval-valid_computer_network",
907
+ "group": "ceval-valid",
908
+ "dataset_path": "ceval/ceval-exam",
909
+ "dataset_name": "computer_network",
910
+ "validation_split": "val",
911
+ "fewshot_split": "dev",
912
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
913
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
914
+ "doc_to_choice": [
915
+ "A",
916
+ "B",
917
+ "C",
918
+ "D"
919
+ ],
920
+ "description": "以下是中国关于计算机网络的单项选择题,请选出其中的正确答案。\n\n",
921
+ "target_delimiter": " ",
922
+ "fewshot_delimiter": "\n\n",
923
+ "fewshot_config": {
924
+ "sampler": "first_n"
925
+ },
926
+ "metric_list": [
927
+ {
928
+ "metric": "acc",
929
+ "aggregation": "mean",
930
+ "higher_is_better": true
931
+ },
932
+ {
933
+ "metric": "acc_norm",
934
+ "aggregation": "mean",
935
+ "higher_is_better": true
936
+ }
937
+ ],
938
+ "output_type": "multiple_choice",
939
+ "repeats": 1,
940
+ "should_decontaminate": false,
941
+ "metadata": {
942
+ "version": 1.0
943
+ }
944
+ },
945
+ "ceval-valid_discrete_mathematics": {
946
+ "task": "ceval-valid_discrete_mathematics",
947
+ "group": "ceval-valid",
948
+ "dataset_path": "ceval/ceval-exam",
949
+ "dataset_name": "discrete_mathematics",
950
+ "validation_split": "val",
951
+ "fewshot_split": "dev",
952
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
953
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
954
+ "doc_to_choice": [
955
+ "A",
956
+ "B",
957
+ "C",
958
+ "D"
959
+ ],
960
+ "description": "以下是中国关于离散数学的单项选择题,请选出其中的正确答案。\n\n",
961
+ "target_delimiter": " ",
962
+ "fewshot_delimiter": "\n\n",
963
+ "fewshot_config": {
964
+ "sampler": "first_n"
965
+ },
966
+ "metric_list": [
967
+ {
968
+ "metric": "acc",
969
+ "aggregation": "mean",
970
+ "higher_is_better": true
971
+ },
972
+ {
973
+ "metric": "acc_norm",
974
+ "aggregation": "mean",
975
+ "higher_is_better": true
976
+ }
977
+ ],
978
+ "output_type": "multiple_choice",
979
+ "repeats": 1,
980
+ "should_decontaminate": false,
981
+ "metadata": {
982
+ "version": 1.0
983
+ }
984
+ },
985
+ "ceval-valid_education_science": {
986
+ "task": "ceval-valid_education_science",
987
+ "group": "ceval-valid",
988
+ "dataset_path": "ceval/ceval-exam",
989
+ "dataset_name": "education_science",
990
+ "validation_split": "val",
991
+ "fewshot_split": "dev",
992
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
993
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
994
+ "doc_to_choice": [
995
+ "A",
996
+ "B",
997
+ "C",
998
+ "D"
999
+ ],
1000
+ "description": "以下是中国关于教育学的单项选择题,请选出其中的正确答案。\n\n",
1001
+ "target_delimiter": " ",
1002
+ "fewshot_delimiter": "\n\n",
1003
+ "fewshot_config": {
1004
+ "sampler": "first_n"
1005
+ },
1006
+ "metric_list": [
1007
+ {
1008
+ "metric": "acc",
1009
+ "aggregation": "mean",
1010
+ "higher_is_better": true
1011
+ },
1012
+ {
1013
+ "metric": "acc_norm",
1014
+ "aggregation": "mean",
1015
+ "higher_is_better": true
1016
+ }
1017
+ ],
1018
+ "output_type": "multiple_choice",
1019
+ "repeats": 1,
1020
+ "should_decontaminate": false,
1021
+ "metadata": {
1022
+ "version": 1.0
1023
+ }
1024
+ },
1025
+ "ceval-valid_electrical_engineer": {
1026
+ "task": "ceval-valid_electrical_engineer",
1027
+ "group": "ceval-valid",
1028
+ "dataset_path": "ceval/ceval-exam",
1029
+ "dataset_name": "electrical_engineer",
1030
+ "validation_split": "val",
1031
+ "fewshot_split": "dev",
1032
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
1033
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
1034
+ "doc_to_choice": [
1035
+ "A",
1036
+ "B",
1037
+ "C",
1038
+ "D"
1039
+ ],
1040
+ "description": "以下是中国关于注册电气工程师的单项选择题,请选出其中的正确答案。\n\n",
1041
+ "target_delimiter": " ",
1042
+ "fewshot_delimiter": "\n\n",
1043
+ "fewshot_config": {
1044
+ "sampler": "first_n"
1045
+ },
1046
+ "metric_list": [
1047
+ {
1048
+ "metric": "acc",
1049
+ "aggregation": "mean",
1050
+ "higher_is_better": true
1051
+ },
1052
+ {
1053
+ "metric": "acc_norm",
1054
+ "aggregation": "mean",
1055
+ "higher_is_better": true
1056
+ }
1057
+ ],
1058
+ "output_type": "multiple_choice",
1059
+ "repeats": 1,
1060
+ "should_decontaminate": false,
1061
+ "metadata": {
1062
+ "version": 1.0
1063
+ }
1064
+ },
1065
+ "ceval-valid_environmental_impact_assessment_engineer": {
1066
+ "task": "ceval-valid_environmental_impact_assessment_engineer",
1067
+ "group": "ceval-valid",
1068
+ "dataset_path": "ceval/ceval-exam",
1069
+ "dataset_name": "environmental_impact_assessment_engineer",
1070
+ "validation_split": "val",
1071
+ "fewshot_split": "dev",
1072
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
1073
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
1074
+ "doc_to_choice": [
1075
+ "A",
1076
+ "B",
1077
+ "C",
1078
+ "D"
1079
+ ],
1080
+ "description": "以下是中国关于环境影响评价工程师的单项选择题,请选出其中的正确答案。\n\n",
1081
+ "target_delimiter": " ",
1082
+ "fewshot_delimiter": "\n\n",
1083
+ "fewshot_config": {
1084
+ "sampler": "first_n"
1085
+ },
1086
+ "metric_list": [
1087
+ {
1088
+ "metric": "acc",
1089
+ "aggregation": "mean",
1090
+ "higher_is_better": true
1091
+ },
1092
+ {
1093
+ "metric": "acc_norm",
1094
+ "aggregation": "mean",
1095
+ "higher_is_better": true
1096
+ }
1097
+ ],
1098
+ "output_type": "multiple_choice",
1099
+ "repeats": 1,
1100
+ "should_decontaminate": false,
1101
+ "metadata": {
1102
+ "version": 1.0
1103
+ }
1104
+ },
1105
+ "ceval-valid_fire_engineer": {
1106
+ "task": "ceval-valid_fire_engineer",
1107
+ "group": "ceval-valid",
1108
+ "dataset_path": "ceval/ceval-exam",
1109
+ "dataset_name": "fire_engineer",
1110
+ "validation_split": "val",
1111
+ "fewshot_split": "dev",
1112
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
1113
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
1114
+ "doc_to_choice": [
1115
+ "A",
1116
+ "B",
1117
+ "C",
1118
+ "D"
1119
+ ],
1120
+ "description": "以下是中国关于注册消防工程师的单项选择题,请选出其中的正确答案。\n\n",
1121
+ "target_delimiter": " ",
1122
+ "fewshot_delimiter": "\n\n",
1123
+ "fewshot_config": {
1124
+ "sampler": "first_n"
1125
+ },
1126
+ "metric_list": [
1127
+ {
1128
+ "metric": "acc",
1129
+ "aggregation": "mean",
1130
+ "higher_is_better": true
1131
+ },
1132
+ {
1133
+ "metric": "acc_norm",
1134
+ "aggregation": "mean",
1135
+ "higher_is_better": true
1136
+ }
1137
+ ],
1138
+ "output_type": "multiple_choice",
1139
+ "repeats": 1,
1140
+ "should_decontaminate": false,
1141
+ "metadata": {
1142
+ "version": 1.0
1143
+ }
1144
+ },
1145
+ "ceval-valid_high_school_biology": {
1146
+ "task": "ceval-valid_high_school_biology",
1147
+ "group": "ceval-valid",
1148
+ "dataset_path": "ceval/ceval-exam",
1149
+ "dataset_name": "high_school_biology",
1150
+ "validation_split": "val",
1151
+ "fewshot_split": "dev",
1152
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
1153
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
1154
+ "doc_to_choice": [
1155
+ "A",
1156
+ "B",
1157
+ "C",
1158
+ "D"
1159
+ ],
1160
+ "description": "以下是中国关于高中生物的单项选择题,请选出其中的正确答案。\n\n",
1161
+ "target_delimiter": " ",
1162
+ "fewshot_delimiter": "\n\n",
1163
+ "fewshot_config": {
1164
+ "sampler": "first_n"
1165
+ },
1166
+ "metric_list": [
1167
+ {
1168
+ "metric": "acc",
1169
+ "aggregation": "mean",
1170
+ "higher_is_better": true
1171
+ },
1172
+ {
1173
+ "metric": "acc_norm",
1174
+ "aggregation": "mean",
1175
+ "higher_is_better": true
1176
+ }
1177
+ ],
1178
+ "output_type": "multiple_choice",
1179
+ "repeats": 1,
1180
+ "should_decontaminate": false,
1181
+ "metadata": {
1182
+ "version": 1.0
1183
+ }
1184
+ },
1185
+ "ceval-valid_high_school_chemistry": {
1186
+ "task": "ceval-valid_high_school_chemistry",
1187
+ "group": "ceval-valid",
1188
+ "dataset_path": "ceval/ceval-exam",
1189
+ "dataset_name": "high_school_chemistry",
1190
+ "validation_split": "val",
1191
+ "fewshot_split": "dev",
1192
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
1193
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
1194
+ "doc_to_choice": [
1195
+ "A",
1196
+ "B",
1197
+ "C",
1198
+ "D"
1199
+ ],
1200
+ "description": "以下是中国关于高中化学的单项选择题,请选出其中的正确答案。\n\n",
1201
+ "target_delimiter": " ",
1202
+ "fewshot_delimiter": "\n\n",
1203
+ "fewshot_config": {
1204
+ "sampler": "first_n"
1205
+ },
1206
+ "metric_list": [
1207
+ {
1208
+ "metric": "acc",
1209
+ "aggregation": "mean",
1210
+ "higher_is_better": true
1211
+ },
1212
+ {
1213
+ "metric": "acc_norm",
1214
+ "aggregation": "mean",
1215
+ "higher_is_better": true
1216
+ }
1217
+ ],
1218
+ "output_type": "multiple_choice",
1219
+ "repeats": 1,
1220
+ "should_decontaminate": false,
1221
+ "metadata": {
1222
+ "version": 1.0
1223
+ }
1224
+ },
1225
+ "ceval-valid_high_school_chinese": {
1226
+ "task": "ceval-valid_high_school_chinese",
1227
+ "group": "ceval-valid",
1228
+ "dataset_path": "ceval/ceval-exam",
1229
+ "dataset_name": "high_school_chinese",
1230
+ "validation_split": "val",
1231
+ "fewshot_split": "dev",
1232
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
1233
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
1234
+ "doc_to_choice": [
1235
+ "A",
1236
+ "B",
1237
+ "C",
1238
+ "D"
1239
+ ],
1240
+ "description": "以下是中国关于高中语文的单项选择题,请选出其中的正确答案。\n\n",
1241
+ "target_delimiter": " ",
1242
+ "fewshot_delimiter": "\n\n",
1243
+ "fewshot_config": {
1244
+ "sampler": "first_n"
1245
+ },
1246
+ "metric_list": [
1247
+ {
1248
+ "metric": "acc",
1249
+ "aggregation": "mean",
1250
+ "higher_is_better": true
1251
+ },
1252
+ {
1253
+ "metric": "acc_norm",
1254
+ "aggregation": "mean",
1255
+ "higher_is_better": true
1256
+ }
1257
+ ],
1258
+ "output_type": "multiple_choice",
1259
+ "repeats": 1,
1260
+ "should_decontaminate": false,
1261
+ "metadata": {
1262
+ "version": 1.0
1263
+ }
1264
+ },
1265
+ "ceval-valid_high_school_geography": {
1266
+ "task": "ceval-valid_high_school_geography",
1267
+ "group": "ceval-valid",
1268
+ "dataset_path": "ceval/ceval-exam",
1269
+ "dataset_name": "high_school_geography",
1270
+ "validation_split": "val",
1271
+ "fewshot_split": "dev",
1272
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
1273
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
1274
+ "doc_to_choice": [
1275
+ "A",
1276
+ "B",
1277
+ "C",
1278
+ "D"
1279
+ ],
1280
+ "description": "以下是中国关于高中地理的单项选择题,请选出其中的正确答案。\n\n",
1281
+ "target_delimiter": " ",
1282
+ "fewshot_delimiter": "\n\n",
1283
+ "fewshot_config": {
1284
+ "sampler": "first_n"
1285
+ },
1286
+ "metric_list": [
1287
+ {
1288
+ "metric": "acc",
1289
+ "aggregation": "mean",
1290
+ "higher_is_better": true
1291
+ },
1292
+ {
1293
+ "metric": "acc_norm",
1294
+ "aggregation": "mean",
1295
+ "higher_is_better": true
1296
+ }
1297
+ ],
1298
+ "output_type": "multiple_choice",
1299
+ "repeats": 1,
1300
+ "should_decontaminate": false,
1301
+ "metadata": {
1302
+ "version": 1.0
1303
+ }
1304
+ },
1305
+ "ceval-valid_high_school_history": {
1306
+ "task": "ceval-valid_high_school_history",
1307
+ "group": "ceval-valid",
1308
+ "dataset_path": "ceval/ceval-exam",
1309
+ "dataset_name": "high_school_history",
1310
+ "validation_split": "val",
1311
+ "fewshot_split": "dev",
1312
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
1313
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
1314
+ "doc_to_choice": [
1315
+ "A",
1316
+ "B",
1317
+ "C",
1318
+ "D"
1319
+ ],
1320
+ "description": "以下是中国关于高中历史的单项选择题,请选出其中的正确答案。\n\n",
1321
+ "target_delimiter": " ",
1322
+ "fewshot_delimiter": "\n\n",
1323
+ "fewshot_config": {
1324
+ "sampler": "first_n"
1325
+ },
1326
+ "metric_list": [
1327
+ {
1328
+ "metric": "acc",
1329
+ "aggregation": "mean",
1330
+ "higher_is_better": true
1331
+ },
1332
+ {
1333
+ "metric": "acc_norm",
1334
+ "aggregation": "mean",
1335
+ "higher_is_better": true
1336
+ }
1337
+ ],
1338
+ "output_type": "multiple_choice",
1339
+ "repeats": 1,
1340
+ "should_decontaminate": false,
1341
+ "metadata": {
1342
+ "version": 1.0
1343
+ }
1344
+ },
1345
+ "ceval-valid_high_school_mathematics": {
1346
+ "task": "ceval-valid_high_school_mathematics",
1347
+ "group": "ceval-valid",
1348
+ "dataset_path": "ceval/ceval-exam",
1349
+ "dataset_name": "high_school_mathematics",
1350
+ "validation_split": "val",
1351
+ "fewshot_split": "dev",
1352
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
1353
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
1354
+ "doc_to_choice": [
1355
+ "A",
1356
+ "B",
1357
+ "C",
1358
+ "D"
1359
+ ],
1360
+ "description": "以下是中国关于高中数学的单项选择题,请选出其中的正确答案。\n\n",
1361
+ "target_delimiter": " ",
1362
+ "fewshot_delimiter": "\n\n",
1363
+ "fewshot_config": {
1364
+ "sampler": "first_n"
1365
+ },
1366
+ "metric_list": [
1367
+ {
1368
+ "metric": "acc",
1369
+ "aggregation": "mean",
1370
+ "higher_is_better": true
1371
+ },
1372
+ {
1373
+ "metric": "acc_norm",
1374
+ "aggregation": "mean",
1375
+ "higher_is_better": true
1376
+ }
1377
+ ],
1378
+ "output_type": "multiple_choice",
1379
+ "repeats": 1,
1380
+ "should_decontaminate": false,
1381
+ "metadata": {
1382
+ "version": 1.0
1383
+ }
1384
+ },
1385
+ "ceval-valid_high_school_physics": {
1386
+ "task": "ceval-valid_high_school_physics",
1387
+ "group": "ceval-valid",
1388
+ "dataset_path": "ceval/ceval-exam",
1389
+ "dataset_name": "high_school_physics",
1390
+ "validation_split": "val",
1391
+ "fewshot_split": "dev",
1392
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
1393
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
1394
+ "doc_to_choice": [
1395
+ "A",
1396
+ "B",
1397
+ "C",
1398
+ "D"
1399
+ ],
1400
+ "description": "以下是中国关于高中物理的单项选择题,请选出其中的正确答案。\n\n",
1401
+ "target_delimiter": " ",
1402
+ "fewshot_delimiter": "\n\n",
1403
+ "fewshot_config": {
1404
+ "sampler": "first_n"
1405
+ },
1406
+ "metric_list": [
1407
+ {
1408
+ "metric": "acc",
1409
+ "aggregation": "mean",
1410
+ "higher_is_better": true
1411
+ },
1412
+ {
1413
+ "metric": "acc_norm",
1414
+ "aggregation": "mean",
1415
+ "higher_is_better": true
1416
+ }
1417
+ ],
1418
+ "output_type": "multiple_choice",
1419
+ "repeats": 1,
1420
+ "should_decontaminate": false,
1421
+ "metadata": {
1422
+ "version": 1.0
1423
+ }
1424
+ },
1425
+ "ceval-valid_high_school_politics": {
1426
+ "task": "ceval-valid_high_school_politics",
1427
+ "group": "ceval-valid",
1428
+ "dataset_path": "ceval/ceval-exam",
1429
+ "dataset_name": "high_school_politics",
1430
+ "validation_split": "val",
1431
+ "fewshot_split": "dev",
1432
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
1433
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
1434
+ "doc_to_choice": [
1435
+ "A",
1436
+ "B",
1437
+ "C",
1438
+ "D"
1439
+ ],
1440
+ "description": "以下是中国关于高中政治的单项选择题,请选出其中的正确答案。\n\n",
1441
+ "target_delimiter": " ",
1442
+ "fewshot_delimiter": "\n\n",
1443
+ "fewshot_config": {
1444
+ "sampler": "first_n"
1445
+ },
1446
+ "metric_list": [
1447
+ {
1448
+ "metric": "acc",
1449
+ "aggregation": "mean",
1450
+ "higher_is_better": true
1451
+ },
1452
+ {
1453
+ "metric": "acc_norm",
1454
+ "aggregation": "mean",
1455
+ "higher_is_better": true
1456
+ }
1457
+ ],
1458
+ "output_type": "multiple_choice",
1459
+ "repeats": 1,
1460
+ "should_decontaminate": false,
1461
+ "metadata": {
1462
+ "version": 1.0
1463
+ }
1464
+ },
1465
+ "ceval-valid_ideological_and_moral_cultivation": {
1466
+ "task": "ceval-valid_ideological_and_moral_cultivation",
1467
+ "group": "ceval-valid",
1468
+ "dataset_path": "ceval/ceval-exam",
1469
+ "dataset_name": "ideological_and_moral_cultivation",
1470
+ "validation_split": "val",
1471
+ "fewshot_split": "dev",
1472
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
1473
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
1474
+ "doc_to_choice": [
1475
+ "A",
1476
+ "B",
1477
+ "C",
1478
+ "D"
1479
+ ],
1480
+ "description": "以下是中国关于思想道德修养与法律基础的单项选择题,请选出其中的正确答案。\n\n",
1481
+ "target_delimiter": " ",
1482
+ "fewshot_delimiter": "\n\n",
1483
+ "fewshot_config": {
1484
+ "sampler": "first_n"
1485
+ },
1486
+ "metric_list": [
1487
+ {
1488
+ "metric": "acc",
1489
+ "aggregation": "mean",
1490
+ "higher_is_better": true
1491
+ },
1492
+ {
1493
+ "metric": "acc_norm",
1494
+ "aggregation": "mean",
1495
+ "higher_is_better": true
1496
+ }
1497
+ ],
1498
+ "output_type": "multiple_choice",
1499
+ "repeats": 1,
1500
+ "should_decontaminate": false,
1501
+ "metadata": {
1502
+ "version": 1.0
1503
+ }
1504
+ },
1505
+ "ceval-valid_law": {
1506
+ "task": "ceval-valid_law",
1507
+ "group": "ceval-valid",
1508
+ "dataset_path": "ceval/ceval-exam",
1509
+ "dataset_name": "law",
1510
+ "validation_split": "val",
1511
+ "fewshot_split": "dev",
1512
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
1513
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
1514
+ "doc_to_choice": [
1515
+ "A",
1516
+ "B",
1517
+ "C",
1518
+ "D"
1519
+ ],
1520
+ "description": "以下是中国关于法学的单项选择题,请选出其中的正确答案。\n\n",
1521
+ "target_delimiter": " ",
1522
+ "fewshot_delimiter": "\n\n",
1523
+ "fewshot_config": {
1524
+ "sampler": "first_n"
1525
+ },
1526
+ "metric_list": [
1527
+ {
1528
+ "metric": "acc",
1529
+ "aggregation": "mean",
1530
+ "higher_is_better": true
1531
+ },
1532
+ {
1533
+ "metric": "acc_norm",
1534
+ "aggregation": "mean",
1535
+ "higher_is_better": true
1536
+ }
1537
+ ],
1538
+ "output_type": "multiple_choice",
1539
+ "repeats": 1,
1540
+ "should_decontaminate": false,
1541
+ "metadata": {
1542
+ "version": 1.0
1543
+ }
1544
+ },
1545
+ "ceval-valid_legal_professional": {
1546
+ "task": "ceval-valid_legal_professional",
1547
+ "group": "ceval-valid",
1548
+ "dataset_path": "ceval/ceval-exam",
1549
+ "dataset_name": "legal_professional",
1550
+ "validation_split": "val",
1551
+ "fewshot_split": "dev",
1552
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
1553
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
1554
+ "doc_to_choice": [
1555
+ "A",
1556
+ "B",
1557
+ "C",
1558
+ "D"
1559
+ ],
1560
+ "description": "以下是中国关于法律职业资格的单项选择题,请选出其中的正确答案。\n\n",
1561
+ "target_delimiter": " ",
1562
+ "fewshot_delimiter": "\n\n",
1563
+ "fewshot_config": {
1564
+ "sampler": "first_n"
1565
+ },
1566
+ "metric_list": [
1567
+ {
1568
+ "metric": "acc",
1569
+ "aggregation": "mean",
1570
+ "higher_is_better": true
1571
+ },
1572
+ {
1573
+ "metric": "acc_norm",
1574
+ "aggregation": "mean",
1575
+ "higher_is_better": true
1576
+ }
1577
+ ],
1578
+ "output_type": "multiple_choice",
1579
+ "repeats": 1,
1580
+ "should_decontaminate": false,
1581
+ "metadata": {
1582
+ "version": 1.0
1583
+ }
1584
+ },
1585
+ "ceval-valid_logic": {
1586
+ "task": "ceval-valid_logic",
1587
+ "group": "ceval-valid",
1588
+ "dataset_path": "ceval/ceval-exam",
1589
+ "dataset_name": "logic",
1590
+ "validation_split": "val",
1591
+ "fewshot_split": "dev",
1592
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
1593
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
1594
+ "doc_to_choice": [
1595
+ "A",
1596
+ "B",
1597
+ "C",
1598
+ "D"
1599
+ ],
1600
+ "description": "以下是中国关于逻辑学的单项选择题,请选出其中的正确答案。\n\n",
1601
+ "target_delimiter": " ",
1602
+ "fewshot_delimiter": "\n\n",
1603
+ "fewshot_config": {
1604
+ "sampler": "first_n"
1605
+ },
1606
+ "metric_list": [
1607
+ {
1608
+ "metric": "acc",
1609
+ "aggregation": "mean",
1610
+ "higher_is_better": true
1611
+ },
1612
+ {
1613
+ "metric": "acc_norm",
1614
+ "aggregation": "mean",
1615
+ "higher_is_better": true
1616
+ }
1617
+ ],
1618
+ "output_type": "multiple_choice",
1619
+ "repeats": 1,
1620
+ "should_decontaminate": false,
1621
+ "metadata": {
1622
+ "version": 1.0
1623
+ }
1624
+ },
1625
+ "ceval-valid_mao_zedong_thought": {
1626
+ "task": "ceval-valid_mao_zedong_thought",
1627
+ "group": "ceval-valid",
1628
+ "dataset_path": "ceval/ceval-exam",
1629
+ "dataset_name": "mao_zedong_thought",
1630
+ "validation_split": "val",
1631
+ "fewshot_split": "dev",
1632
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
1633
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
1634
+ "doc_to_choice": [
1635
+ "A",
1636
+ "B",
1637
+ "C",
1638
+ "D"
1639
+ ],
1640
+ "description": "以下是中国关于毛泽东思想和中国特色社会主义理论体系概论的单项选择题,请选出其中的正确答案。\n\n",
1641
+ "target_delimiter": " ",
1642
+ "fewshot_delimiter": "\n\n",
1643
+ "fewshot_config": {
1644
+ "sampler": "first_n"
1645
+ },
1646
+ "metric_list": [
1647
+ {
1648
+ "metric": "acc",
1649
+ "aggregation": "mean",
1650
+ "higher_is_better": true
1651
+ },
1652
+ {
1653
+ "metric": "acc_norm",
1654
+ "aggregation": "mean",
1655
+ "higher_is_better": true
1656
+ }
1657
+ ],
1658
+ "output_type": "multiple_choice",
1659
+ "repeats": 1,
1660
+ "should_decontaminate": false,
1661
+ "metadata": {
1662
+ "version": 1.0
1663
+ }
1664
+ },
1665
+ "ceval-valid_marxism": {
1666
+ "task": "ceval-valid_marxism",
1667
+ "group": "ceval-valid",
1668
+ "dataset_path": "ceval/ceval-exam",
1669
+ "dataset_name": "marxism",
1670
+ "validation_split": "val",
1671
+ "fewshot_split": "dev",
1672
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
1673
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
1674
+ "doc_to_choice": [
1675
+ "A",
1676
+ "B",
1677
+ "C",
1678
+ "D"
1679
+ ],
1680
+ "description": "以下是中国关于马克思主义基本原理的单项选择题,请选出其中的正确答案。\n\n",
1681
+ "target_delimiter": " ",
1682
+ "fewshot_delimiter": "\n\n",
1683
+ "fewshot_config": {
1684
+ "sampler": "first_n"
1685
+ },
1686
+ "metric_list": [
1687
+ {
1688
+ "metric": "acc",
1689
+ "aggregation": "mean",
1690
+ "higher_is_better": true
1691
+ },
1692
+ {
1693
+ "metric": "acc_norm",
1694
+ "aggregation": "mean",
1695
+ "higher_is_better": true
1696
+ }
1697
+ ],
1698
+ "output_type": "multiple_choice",
1699
+ "repeats": 1,
1700
+ "should_decontaminate": false,
1701
+ "metadata": {
1702
+ "version": 1.0
1703
+ }
1704
+ },
1705
+ "ceval-valid_metrology_engineer": {
1706
+ "task": "ceval-valid_metrology_engineer",
1707
+ "group": "ceval-valid",
1708
+ "dataset_path": "ceval/ceval-exam",
1709
+ "dataset_name": "metrology_engineer",
1710
+ "validation_split": "val",
1711
+ "fewshot_split": "dev",
1712
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
1713
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
1714
+ "doc_to_choice": [
1715
+ "A",
1716
+ "B",
1717
+ "C",
1718
+ "D"
1719
+ ],
1720
+ "description": "以下是中国关于注册计量师的单项选择题,请选出其中的正确答案。\n\n",
1721
+ "target_delimiter": " ",
1722
+ "fewshot_delimiter": "\n\n",
1723
+ "fewshot_config": {
1724
+ "sampler": "first_n"
1725
+ },
1726
+ "metric_list": [
1727
+ {
1728
+ "metric": "acc",
1729
+ "aggregation": "mean",
1730
+ "higher_is_better": true
1731
+ },
1732
+ {
1733
+ "metric": "acc_norm",
1734
+ "aggregation": "mean",
1735
+ "higher_is_better": true
1736
+ }
1737
+ ],
1738
+ "output_type": "multiple_choice",
1739
+ "repeats": 1,
1740
+ "should_decontaminate": false,
1741
+ "metadata": {
1742
+ "version": 1.0
1743
+ }
1744
+ },
1745
+ "ceval-valid_middle_school_biology": {
1746
+ "task": "ceval-valid_middle_school_biology",
1747
+ "group": "ceval-valid",
1748
+ "dataset_path": "ceval/ceval-exam",
1749
+ "dataset_name": "middle_school_biology",
1750
+ "validation_split": "val",
1751
+ "fewshot_split": "dev",
1752
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
1753
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
1754
+ "doc_to_choice": [
1755
+ "A",
1756
+ "B",
1757
+ "C",
1758
+ "D"
1759
+ ],
1760
+ "description": "以下是中国关于初中生物的单项选择题,请选出其中的正确答案。\n\n",
1761
+ "target_delimiter": " ",
1762
+ "fewshot_delimiter": "\n\n",
1763
+ "fewshot_config": {
1764
+ "sampler": "first_n"
1765
+ },
1766
+ "metric_list": [
1767
+ {
1768
+ "metric": "acc",
1769
+ "aggregation": "mean",
1770
+ "higher_is_better": true
1771
+ },
1772
+ {
1773
+ "metric": "acc_norm",
1774
+ "aggregation": "mean",
1775
+ "higher_is_better": true
1776
+ }
1777
+ ],
1778
+ "output_type": "multiple_choice",
1779
+ "repeats": 1,
1780
+ "should_decontaminate": false,
1781
+ "metadata": {
1782
+ "version": 1.0
1783
+ }
1784
+ },
1785
+ "ceval-valid_middle_school_chemistry": {
1786
+ "task": "ceval-valid_middle_school_chemistry",
1787
+ "group": "ceval-valid",
1788
+ "dataset_path": "ceval/ceval-exam",
1789
+ "dataset_name": "middle_school_chemistry",
1790
+ "validation_split": "val",
1791
+ "fewshot_split": "dev",
1792
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
1793
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
1794
+ "doc_to_choice": [
1795
+ "A",
1796
+ "B",
1797
+ "C",
1798
+ "D"
1799
+ ],
1800
+ "description": "以下是中国关于初中化学的单项选择题,请选出其中的正确答案。\n\n",
1801
+ "target_delimiter": " ",
1802
+ "fewshot_delimiter": "\n\n",
1803
+ "fewshot_config": {
1804
+ "sampler": "first_n"
1805
+ },
1806
+ "metric_list": [
1807
+ {
1808
+ "metric": "acc",
1809
+ "aggregation": "mean",
1810
+ "higher_is_better": true
1811
+ },
1812
+ {
1813
+ "metric": "acc_norm",
1814
+ "aggregation": "mean",
1815
+ "higher_is_better": true
1816
+ }
1817
+ ],
1818
+ "output_type": "multiple_choice",
1819
+ "repeats": 1,
1820
+ "should_decontaminate": false,
1821
+ "metadata": {
1822
+ "version": 1.0
1823
+ }
1824
+ },
1825
+ "ceval-valid_middle_school_geography": {
1826
+ "task": "ceval-valid_middle_school_geography",
1827
+ "group": "ceval-valid",
1828
+ "dataset_path": "ceval/ceval-exam",
1829
+ "dataset_name": "middle_school_geography",
1830
+ "validation_split": "val",
1831
+ "fewshot_split": "dev",
1832
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
1833
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
1834
+ "doc_to_choice": [
1835
+ "A",
1836
+ "B",
1837
+ "C",
1838
+ "D"
1839
+ ],
1840
+ "description": "以下是中国关于初中地理的单项选择题,请选出其中的正确答案。\n\n",
1841
+ "target_delimiter": " ",
1842
+ "fewshot_delimiter": "\n\n",
1843
+ "fewshot_config": {
1844
+ "sampler": "first_n"
1845
+ },
1846
+ "metric_list": [
1847
+ {
1848
+ "metric": "acc",
1849
+ "aggregation": "mean",
1850
+ "higher_is_better": true
1851
+ },
1852
+ {
1853
+ "metric": "acc_norm",
1854
+ "aggregation": "mean",
1855
+ "higher_is_better": true
1856
+ }
1857
+ ],
1858
+ "output_type": "multiple_choice",
1859
+ "repeats": 1,
1860
+ "should_decontaminate": false,
1861
+ "metadata": {
1862
+ "version": 1.0
1863
+ }
1864
+ },
1865
+ "ceval-valid_middle_school_history": {
1866
+ "task": "ceval-valid_middle_school_history",
1867
+ "group": "ceval-valid",
1868
+ "dataset_path": "ceval/ceval-exam",
1869
+ "dataset_name": "middle_school_history",
1870
+ "validation_split": "val",
1871
+ "fewshot_split": "dev",
1872
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
1873
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
1874
+ "doc_to_choice": [
1875
+ "A",
1876
+ "B",
1877
+ "C",
1878
+ "D"
1879
+ ],
1880
+ "description": "以下是中国关于初中历史的单项选择题,请选出其中的正确答案。\n\n",
1881
+ "target_delimiter": " ",
1882
+ "fewshot_delimiter": "\n\n",
1883
+ "fewshot_config": {
1884
+ "sampler": "first_n"
1885
+ },
1886
+ "metric_list": [
1887
+ {
1888
+ "metric": "acc",
1889
+ "aggregation": "mean",
1890
+ "higher_is_better": true
1891
+ },
1892
+ {
1893
+ "metric": "acc_norm",
1894
+ "aggregation": "mean",
1895
+ "higher_is_better": true
1896
+ }
1897
+ ],
1898
+ "output_type": "multiple_choice",
1899
+ "repeats": 1,
1900
+ "should_decontaminate": false,
1901
+ "metadata": {
1902
+ "version": 1.0
1903
+ }
1904
+ },
1905
+ "ceval-valid_middle_school_mathematics": {
1906
+ "task": "ceval-valid_middle_school_mathematics",
1907
+ "group": "ceval-valid",
1908
+ "dataset_path": "ceval/ceval-exam",
1909
+ "dataset_name": "middle_school_mathematics",
1910
+ "validation_split": "val",
1911
+ "fewshot_split": "dev",
1912
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
1913
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
1914
+ "doc_to_choice": [
1915
+ "A",
1916
+ "B",
1917
+ "C",
1918
+ "D"
1919
+ ],
1920
+ "description": "以下是中国关于初中数学的单项选择题,请选出其中的正确答案。\n\n",
1921
+ "target_delimiter": " ",
1922
+ "fewshot_delimiter": "\n\n",
1923
+ "fewshot_config": {
1924
+ "sampler": "first_n"
1925
+ },
1926
+ "metric_list": [
1927
+ {
1928
+ "metric": "acc",
1929
+ "aggregation": "mean",
1930
+ "higher_is_better": true
1931
+ },
1932
+ {
1933
+ "metric": "acc_norm",
1934
+ "aggregation": "mean",
1935
+ "higher_is_better": true
1936
+ }
1937
+ ],
1938
+ "output_type": "multiple_choice",
1939
+ "repeats": 1,
1940
+ "should_decontaminate": false,
1941
+ "metadata": {
1942
+ "version": 1.0
1943
+ }
1944
+ },
1945
+ "ceval-valid_middle_school_physics": {
1946
+ "task": "ceval-valid_middle_school_physics",
1947
+ "group": "ceval-valid",
1948
+ "dataset_path": "ceval/ceval-exam",
1949
+ "dataset_name": "middle_school_physics",
1950
+ "validation_split": "val",
1951
+ "fewshot_split": "dev",
1952
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
1953
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
1954
+ "doc_to_choice": [
1955
+ "A",
1956
+ "B",
1957
+ "C",
1958
+ "D"
1959
+ ],
1960
+ "description": "以下是中国关于初中物理的单项选择题,请选出其中的正确答案。\n\n",
1961
+ "target_delimiter": " ",
1962
+ "fewshot_delimiter": "\n\n",
1963
+ "fewshot_config": {
1964
+ "sampler": "first_n"
1965
+ },
1966
+ "metric_list": [
1967
+ {
1968
+ "metric": "acc",
1969
+ "aggregation": "mean",
1970
+ "higher_is_better": true
1971
+ },
1972
+ {
1973
+ "metric": "acc_norm",
1974
+ "aggregation": "mean",
1975
+ "higher_is_better": true
1976
+ }
1977
+ ],
1978
+ "output_type": "multiple_choice",
1979
+ "repeats": 1,
1980
+ "should_decontaminate": false,
1981
+ "metadata": {
1982
+ "version": 1.0
1983
+ }
1984
+ },
1985
+ "ceval-valid_middle_school_politics": {
1986
+ "task": "ceval-valid_middle_school_politics",
1987
+ "group": "ceval-valid",
1988
+ "dataset_path": "ceval/ceval-exam",
1989
+ "dataset_name": "middle_school_politics",
1990
+ "validation_split": "val",
1991
+ "fewshot_split": "dev",
1992
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
1993
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
1994
+ "doc_to_choice": [
1995
+ "A",
1996
+ "B",
1997
+ "C",
1998
+ "D"
1999
+ ],
2000
+ "description": "以下是中国关于初中政治的单项选择题,请选出其中的正确答案。\n\n",
2001
+ "target_delimiter": " ",
2002
+ "fewshot_delimiter": "\n\n",
2003
+ "fewshot_config": {
2004
+ "sampler": "first_n"
2005
+ },
2006
+ "metric_list": [
2007
+ {
2008
+ "metric": "acc",
2009
+ "aggregation": "mean",
2010
+ "higher_is_better": true
2011
+ },
2012
+ {
2013
+ "metric": "acc_norm",
2014
+ "aggregation": "mean",
2015
+ "higher_is_better": true
2016
+ }
2017
+ ],
2018
+ "output_type": "multiple_choice",
2019
+ "repeats": 1,
2020
+ "should_decontaminate": false,
2021
+ "metadata": {
2022
+ "version": 1.0
2023
+ }
2024
+ },
2025
+ "ceval-valid_modern_chinese_history": {
2026
+ "task": "ceval-valid_modern_chinese_history",
2027
+ "group": "ceval-valid",
2028
+ "dataset_path": "ceval/ceval-exam",
2029
+ "dataset_name": "modern_chinese_history",
2030
+ "validation_split": "val",
2031
+ "fewshot_split": "dev",
2032
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
2033
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
2034
+ "doc_to_choice": [
2035
+ "A",
2036
+ "B",
2037
+ "C",
2038
+ "D"
2039
+ ],
2040
+ "description": "以下是中国关于近代史纲要的单项选择题,请选出其中的正确答案。\n\n",
2041
+ "target_delimiter": " ",
2042
+ "fewshot_delimiter": "\n\n",
2043
+ "fewshot_config": {
2044
+ "sampler": "first_n"
2045
+ },
2046
+ "metric_list": [
2047
+ {
2048
+ "metric": "acc",
2049
+ "aggregation": "mean",
2050
+ "higher_is_better": true
2051
+ },
2052
+ {
2053
+ "metric": "acc_norm",
2054
+ "aggregation": "mean",
2055
+ "higher_is_better": true
2056
+ }
2057
+ ],
2058
+ "output_type": "multiple_choice",
2059
+ "repeats": 1,
2060
+ "should_decontaminate": false,
2061
+ "metadata": {
2062
+ "version": 1.0
2063
+ }
2064
+ },
2065
+ "ceval-valid_operating_system": {
2066
+ "task": "ceval-valid_operating_system",
2067
+ "group": "ceval-valid",
2068
+ "dataset_path": "ceval/ceval-exam",
2069
+ "dataset_name": "operating_system",
2070
+ "validation_split": "val",
2071
+ "fewshot_split": "dev",
2072
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
2073
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
2074
+ "doc_to_choice": [
2075
+ "A",
2076
+ "B",
2077
+ "C",
2078
+ "D"
2079
+ ],
2080
+ "description": "以下是中国关于操作系统的单项选择题,请选出其中的正确答案。\n\n",
2081
+ "target_delimiter": " ",
2082
+ "fewshot_delimiter": "\n\n",
2083
+ "fewshot_config": {
2084
+ "sampler": "first_n"
2085
+ },
2086
+ "metric_list": [
2087
+ {
2088
+ "metric": "acc",
2089
+ "aggregation": "mean",
2090
+ "higher_is_better": true
2091
+ },
2092
+ {
2093
+ "metric": "acc_norm",
2094
+ "aggregation": "mean",
2095
+ "higher_is_better": true
2096
+ }
2097
+ ],
2098
+ "output_type": "multiple_choice",
2099
+ "repeats": 1,
2100
+ "should_decontaminate": false,
2101
+ "metadata": {
2102
+ "version": 1.0
2103
+ }
2104
+ },
2105
+ "ceval-valid_physician": {
2106
+ "task": "ceval-valid_physician",
2107
+ "group": "ceval-valid",
2108
+ "dataset_path": "ceval/ceval-exam",
2109
+ "dataset_name": "physician",
2110
+ "validation_split": "val",
2111
+ "fewshot_split": "dev",
2112
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
2113
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
2114
+ "doc_to_choice": [
2115
+ "A",
2116
+ "B",
2117
+ "C",
2118
+ "D"
2119
+ ],
2120
+ "description": "以下是中国关于医师资格的单项选择题,请选出其中的正确答案。\n\n",
2121
+ "target_delimiter": " ",
2122
+ "fewshot_delimiter": "\n\n",
2123
+ "fewshot_config": {
2124
+ "sampler": "first_n"
2125
+ },
2126
+ "metric_list": [
2127
+ {
2128
+ "metric": "acc",
2129
+ "aggregation": "mean",
2130
+ "higher_is_better": true
2131
+ },
2132
+ {
2133
+ "metric": "acc_norm",
2134
+ "aggregation": "mean",
2135
+ "higher_is_better": true
2136
+ }
2137
+ ],
2138
+ "output_type": "multiple_choice",
2139
+ "repeats": 1,
2140
+ "should_decontaminate": false,
2141
+ "metadata": {
2142
+ "version": 1.0
2143
+ }
2144
+ },
2145
+ "ceval-valid_plant_protection": {
2146
+ "task": "ceval-valid_plant_protection",
2147
+ "group": "ceval-valid",
2148
+ "dataset_path": "ceval/ceval-exam",
2149
+ "dataset_name": "plant_protection",
2150
+ "validation_split": "val",
2151
+ "fewshot_split": "dev",
2152
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
2153
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
2154
+ "doc_to_choice": [
2155
+ "A",
2156
+ "B",
2157
+ "C",
2158
+ "D"
2159
+ ],
2160
+ "description": "以下是中国关于植物保护的单项选择题,请选出其中的正确答案。\n\n",
2161
+ "target_delimiter": " ",
2162
+ "fewshot_delimiter": "\n\n",
2163
+ "fewshot_config": {
2164
+ "sampler": "first_n"
2165
+ },
2166
+ "metric_list": [
2167
+ {
2168
+ "metric": "acc",
2169
+ "aggregation": "mean",
2170
+ "higher_is_better": true
2171
+ },
2172
+ {
2173
+ "metric": "acc_norm",
2174
+ "aggregation": "mean",
2175
+ "higher_is_better": true
2176
+ }
2177
+ ],
2178
+ "output_type": "multiple_choice",
2179
+ "repeats": 1,
2180
+ "should_decontaminate": false,
2181
+ "metadata": {
2182
+ "version": 1.0
2183
+ }
2184
+ },
2185
+ "ceval-valid_probability_and_statistics": {
2186
+ "task": "ceval-valid_probability_and_statistics",
2187
+ "group": "ceval-valid",
2188
+ "dataset_path": "ceval/ceval-exam",
2189
+ "dataset_name": "probability_and_statistics",
2190
+ "validation_split": "val",
2191
+ "fewshot_split": "dev",
2192
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
2193
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
2194
+ "doc_to_choice": [
2195
+ "A",
2196
+ "B",
2197
+ "C",
2198
+ "D"
2199
+ ],
2200
+ "description": "以下是中国关于概率统计的单项选择题,请选出其中的正确答案。\n\n",
2201
+ "target_delimiter": " ",
2202
+ "fewshot_delimiter": "\n\n",
2203
+ "fewshot_config": {
2204
+ "sampler": "first_n"
2205
+ },
2206
+ "metric_list": [
2207
+ {
2208
+ "metric": "acc",
2209
+ "aggregation": "mean",
2210
+ "higher_is_better": true
2211
+ },
2212
+ {
2213
+ "metric": "acc_norm",
2214
+ "aggregation": "mean",
2215
+ "higher_is_better": true
2216
+ }
2217
+ ],
2218
+ "output_type": "multiple_choice",
2219
+ "repeats": 1,
2220
+ "should_decontaminate": false,
2221
+ "metadata": {
2222
+ "version": 1.0
2223
+ }
2224
+ },
2225
+ "ceval-valid_professional_tour_guide": {
2226
+ "task": "ceval-valid_professional_tour_guide",
2227
+ "group": "ceval-valid",
2228
+ "dataset_path": "ceval/ceval-exam",
2229
+ "dataset_name": "professional_tour_guide",
2230
+ "validation_split": "val",
2231
+ "fewshot_split": "dev",
2232
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
2233
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
2234
+ "doc_to_choice": [
2235
+ "A",
2236
+ "B",
2237
+ "C",
2238
+ "D"
2239
+ ],
2240
+ "description": "以下是中国关于导游资格的单项选择题,请选出其中的正确答案。\n\n",
2241
+ "target_delimiter": " ",
2242
+ "fewshot_delimiter": "\n\n",
2243
+ "fewshot_config": {
2244
+ "sampler": "first_n"
2245
+ },
2246
+ "metric_list": [
2247
+ {
2248
+ "metric": "acc",
2249
+ "aggregation": "mean",
2250
+ "higher_is_better": true
2251
+ },
2252
+ {
2253
+ "metric": "acc_norm",
2254
+ "aggregation": "mean",
2255
+ "higher_is_better": true
2256
+ }
2257
+ ],
2258
+ "output_type": "multiple_choice",
2259
+ "repeats": 1,
2260
+ "should_decontaminate": false,
2261
+ "metadata": {
2262
+ "version": 1.0
2263
+ }
2264
+ },
2265
+ "ceval-valid_sports_science": {
2266
+ "task": "ceval-valid_sports_science",
2267
+ "group": "ceval-valid",
2268
+ "dataset_path": "ceval/ceval-exam",
2269
+ "dataset_name": "sports_science",
2270
+ "validation_split": "val",
2271
+ "fewshot_split": "dev",
2272
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
2273
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
2274
+ "doc_to_choice": [
2275
+ "A",
2276
+ "B",
2277
+ "C",
2278
+ "D"
2279
+ ],
2280
+ "description": "以下是中国关于体育学的单项选择题,请选出其中的正确答案。\n\n",
2281
+ "target_delimiter": " ",
2282
+ "fewshot_delimiter": "\n\n",
2283
+ "fewshot_config": {
2284
+ "sampler": "first_n"
2285
+ },
2286
+ "metric_list": [
2287
+ {
2288
+ "metric": "acc",
2289
+ "aggregation": "mean",
2290
+ "higher_is_better": true
2291
+ },
2292
+ {
2293
+ "metric": "acc_norm",
2294
+ "aggregation": "mean",
2295
+ "higher_is_better": true
2296
+ }
2297
+ ],
2298
+ "output_type": "multiple_choice",
2299
+ "repeats": 1,
2300
+ "should_decontaminate": false,
2301
+ "metadata": {
2302
+ "version": 1.0
2303
+ }
2304
+ },
2305
+ "ceval-valid_tax_accountant": {
2306
+ "task": "ceval-valid_tax_accountant",
2307
+ "group": "ceval-valid",
2308
+ "dataset_path": "ceval/ceval-exam",
2309
+ "dataset_name": "tax_accountant",
2310
+ "validation_split": "val",
2311
+ "fewshot_split": "dev",
2312
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
2313
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
2314
+ "doc_to_choice": [
2315
+ "A",
2316
+ "B",
2317
+ "C",
2318
+ "D"
2319
+ ],
2320
+ "description": "以下是中国关于税务师的单项选择题,请选出其中的正确答案。\n\n",
2321
+ "target_delimiter": " ",
2322
+ "fewshot_delimiter": "\n\n",
2323
+ "fewshot_config": {
2324
+ "sampler": "first_n"
2325
+ },
2326
+ "metric_list": [
2327
+ {
2328
+ "metric": "acc",
2329
+ "aggregation": "mean",
2330
+ "higher_is_better": true
2331
+ },
2332
+ {
2333
+ "metric": "acc_norm",
2334
+ "aggregation": "mean",
2335
+ "higher_is_better": true
2336
+ }
2337
+ ],
2338
+ "output_type": "multiple_choice",
2339
+ "repeats": 1,
2340
+ "should_decontaminate": false,
2341
+ "metadata": {
2342
+ "version": 1.0
2343
+ }
2344
+ },
2345
+ "ceval-valid_teacher_qualification": {
2346
+ "task": "ceval-valid_teacher_qualification",
2347
+ "group": "ceval-valid",
2348
+ "dataset_path": "ceval/ceval-exam",
2349
+ "dataset_name": "teacher_qualification",
2350
+ "validation_split": "val",
2351
+ "fewshot_split": "dev",
2352
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
2353
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
2354
+ "doc_to_choice": [
2355
+ "A",
2356
+ "B",
2357
+ "C",
2358
+ "D"
2359
+ ],
2360
+ "description": "以下是中国关于教师资格的单项选择题,请选出其中的正确答案。\n\n",
2361
+ "target_delimiter": " ",
2362
+ "fewshot_delimiter": "\n\n",
2363
+ "fewshot_config": {
2364
+ "sampler": "first_n"
2365
+ },
2366
+ "metric_list": [
2367
+ {
2368
+ "metric": "acc",
2369
+ "aggregation": "mean",
2370
+ "higher_is_better": true
2371
+ },
2372
+ {
2373
+ "metric": "acc_norm",
2374
+ "aggregation": "mean",
2375
+ "higher_is_better": true
2376
+ }
2377
+ ],
2378
+ "output_type": "multiple_choice",
2379
+ "repeats": 1,
2380
+ "should_decontaminate": false,
2381
+ "metadata": {
2382
+ "version": 1.0
2383
+ }
2384
+ },
2385
+ "ceval-valid_urban_and_rural_planner": {
2386
+ "task": "ceval-valid_urban_and_rural_planner",
2387
+ "group": "ceval-valid",
2388
+ "dataset_path": "ceval/ceval-exam",
2389
+ "dataset_name": "urban_and_rural_planner",
2390
+ "validation_split": "val",
2391
+ "fewshot_split": "dev",
2392
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
2393
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
2394
+ "doc_to_choice": [
2395
+ "A",
2396
+ "B",
2397
+ "C",
2398
+ "D"
2399
+ ],
2400
+ "description": "以下是中国关于注册城乡规划师的单项选择题,请选出其中的正确答案。\n\n",
2401
+ "target_delimiter": " ",
2402
+ "fewshot_delimiter": "\n\n",
2403
+ "fewshot_config": {
2404
+ "sampler": "first_n"
2405
+ },
2406
+ "metric_list": [
2407
+ {
2408
+ "metric": "acc",
2409
+ "aggregation": "mean",
2410
+ "higher_is_better": true
2411
+ },
2412
+ {
2413
+ "metric": "acc_norm",
2414
+ "aggregation": "mean",
2415
+ "higher_is_better": true
2416
+ }
2417
+ ],
2418
+ "output_type": "multiple_choice",
2419
+ "repeats": 1,
2420
+ "should_decontaminate": false,
2421
+ "metadata": {
2422
+ "version": 1.0
2423
+ }
2424
+ },
2425
+ "ceval-valid_veterinary_medicine": {
2426
+ "task": "ceval-valid_veterinary_medicine",
2427
+ "group": "ceval-valid",
2428
+ "dataset_path": "ceval/ceval-exam",
2429
+ "dataset_name": "veterinary_medicine",
2430
+ "validation_split": "val",
2431
+ "fewshot_split": "dev",
2432
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
2433
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
2434
+ "doc_to_choice": [
2435
+ "A",
2436
+ "B",
2437
+ "C",
2438
+ "D"
2439
+ ],
2440
+ "description": "以下是中国关于兽医学的单项选择题,请选出其中的正确答案。\n\n",
2441
+ "target_delimiter": " ",
2442
+ "fewshot_delimiter": "\n\n",
2443
+ "fewshot_config": {
2444
+ "sampler": "first_n"
2445
+ },
2446
+ "metric_list": [
2447
+ {
2448
+ "metric": "acc",
2449
+ "aggregation": "mean",
2450
+ "higher_is_better": true
2451
+ },
2452
+ {
2453
+ "metric": "acc_norm",
2454
+ "aggregation": "mean",
2455
+ "higher_is_better": true
2456
+ }
2457
+ ],
2458
+ "output_type": "multiple_choice",
2459
+ "repeats": 1,
2460
+ "should_decontaminate": false,
2461
+ "metadata": {
2462
+ "version": 1.0
2463
+ }
2464
+ }
2465
+ },
2466
+ "versions": {
2467
+ "ceval-valid": "N/A",
2468
+ "ceval-valid_accountant": 1.0,
2469
+ "ceval-valid_advanced_mathematics": 1.0,
2470
+ "ceval-valid_art_studies": 1.0,
2471
+ "ceval-valid_basic_medicine": 1.0,
2472
+ "ceval-valid_business_administration": 1.0,
2473
+ "ceval-valid_chinese_language_and_literature": 1.0,
2474
+ "ceval-valid_civil_servant": 1.0,
2475
+ "ceval-valid_clinical_medicine": 1.0,
2476
+ "ceval-valid_college_chemistry": 1.0,
2477
+ "ceval-valid_college_economics": 1.0,
2478
+ "ceval-valid_college_physics": 1.0,
2479
+ "ceval-valid_college_programming": 1.0,
2480
+ "ceval-valid_computer_architecture": 1.0,
2481
+ "ceval-valid_computer_network": 1.0,
2482
+ "ceval-valid_discrete_mathematics": 1.0,
2483
+ "ceval-valid_education_science": 1.0,
2484
+ "ceval-valid_electrical_engineer": 1.0,
2485
+ "ceval-valid_environmental_impact_assessment_engineer": 1.0,
2486
+ "ceval-valid_fire_engineer": 1.0,
2487
+ "ceval-valid_high_school_biology": 1.0,
2488
+ "ceval-valid_high_school_chemistry": 1.0,
2489
+ "ceval-valid_high_school_chinese": 1.0,
2490
+ "ceval-valid_high_school_geography": 1.0,
2491
+ "ceval-valid_high_school_history": 1.0,
2492
+ "ceval-valid_high_school_mathematics": 1.0,
2493
+ "ceval-valid_high_school_physics": 1.0,
2494
+ "ceval-valid_high_school_politics": 1.0,
2495
+ "ceval-valid_ideological_and_moral_cultivation": 1.0,
2496
+ "ceval-valid_law": 1.0,
2497
+ "ceval-valid_legal_professional": 1.0,
2498
+ "ceval-valid_logic": 1.0,
2499
+ "ceval-valid_mao_zedong_thought": 1.0,
2500
+ "ceval-valid_marxism": 1.0,
2501
+ "ceval-valid_metrology_engineer": 1.0,
2502
+ "ceval-valid_middle_school_biology": 1.0,
2503
+ "ceval-valid_middle_school_chemistry": 1.0,
2504
+ "ceval-valid_middle_school_geography": 1.0,
2505
+ "ceval-valid_middle_school_history": 1.0,
2506
+ "ceval-valid_middle_school_mathematics": 1.0,
2507
+ "ceval-valid_middle_school_physics": 1.0,
2508
+ "ceval-valid_middle_school_politics": 1.0,
2509
+ "ceval-valid_modern_chinese_history": 1.0,
2510
+ "ceval-valid_operating_system": 1.0,
2511
+ "ceval-valid_physician": 1.0,
2512
+ "ceval-valid_plant_protection": 1.0,
2513
+ "ceval-valid_probability_and_statistics": 1.0,
2514
+ "ceval-valid_professional_tour_guide": 1.0,
2515
+ "ceval-valid_sports_science": 1.0,
2516
+ "ceval-valid_tax_accountant": 1.0,
2517
+ "ceval-valid_teacher_qualification": 1.0,
2518
+ "ceval-valid_urban_and_rural_planner": 1.0,
2519
+ "ceval-valid_veterinary_medicine": 1.0
2520
+ },
2521
+ "n-shot": {
2522
+ "ceval-valid": 0,
2523
+ "ceval-valid_accountant": 0,
2524
+ "ceval-valid_advanced_mathematics": 0,
2525
+ "ceval-valid_art_studies": 0,
2526
+ "ceval-valid_basic_medicine": 0,
2527
+ "ceval-valid_business_administration": 0,
2528
+ "ceval-valid_chinese_language_and_literature": 0,
2529
+ "ceval-valid_civil_servant": 0,
2530
+ "ceval-valid_clinical_medicine": 0,
2531
+ "ceval-valid_college_chemistry": 0,
2532
+ "ceval-valid_college_economics": 0,
2533
+ "ceval-valid_college_physics": 0,
2534
+ "ceval-valid_college_programming": 0,
2535
+ "ceval-valid_computer_architecture": 0,
2536
+ "ceval-valid_computer_network": 0,
2537
+ "ceval-valid_discrete_mathematics": 0,
2538
+ "ceval-valid_education_science": 0,
2539
+ "ceval-valid_electrical_engineer": 0,
2540
+ "ceval-valid_environmental_impact_assessment_engineer": 0,
2541
+ "ceval-valid_fire_engineer": 0,
2542
+ "ceval-valid_high_school_biology": 0,
2543
+ "ceval-valid_high_school_chemistry": 0,
2544
+ "ceval-valid_high_school_chinese": 0,
2545
+ "ceval-valid_high_school_geography": 0,
2546
+ "ceval-valid_high_school_history": 0,
2547
+ "ceval-valid_high_school_mathematics": 0,
2548
+ "ceval-valid_high_school_physics": 0,
2549
+ "ceval-valid_high_school_politics": 0,
2550
+ "ceval-valid_ideological_and_moral_cultivation": 0,
2551
+ "ceval-valid_law": 0,
2552
+ "ceval-valid_legal_professional": 0,
2553
+ "ceval-valid_logic": 0,
2554
+ "ceval-valid_mao_zedong_thought": 0,
2555
+ "ceval-valid_marxism": 0,
2556
+ "ceval-valid_metrology_engineer": 0,
2557
+ "ceval-valid_middle_school_biology": 0,
2558
+ "ceval-valid_middle_school_chemistry": 0,
2559
+ "ceval-valid_middle_school_geography": 0,
2560
+ "ceval-valid_middle_school_history": 0,
2561
+ "ceval-valid_middle_school_mathematics": 0,
2562
+ "ceval-valid_middle_school_physics": 0,
2563
+ "ceval-valid_middle_school_politics": 0,
2564
+ "ceval-valid_modern_chinese_history": 0,
2565
+ "ceval-valid_operating_system": 0,
2566
+ "ceval-valid_physician": 0,
2567
+ "ceval-valid_plant_protection": 0,
2568
+ "ceval-valid_probability_and_statistics": 0,
2569
+ "ceval-valid_professional_tour_guide": 0,
2570
+ "ceval-valid_sports_science": 0,
2571
+ "ceval-valid_tax_accountant": 0,
2572
+ "ceval-valid_teacher_qualification": 0,
2573
+ "ceval-valid_urban_and_rural_planner": 0,
2574
+ "ceval-valid_veterinary_medicine": 0
2575
+ },
2576
+ "config": {
2577
+ "model": "hf",
2578
+ "model_args": "pretrained=RWKV/rwkv-4-world-1b5,dtype=bfloat16,trust_remote_code=True",
2579
+ "batch_size": "auto",
2580
+ "batch_sizes": [
2581
+ 64
2582
+ ],
2583
+ "device": null,
2584
+ "use_cache": null,
2585
+ "limit": null,
2586
+ "bootstrap_iters": 100000,
2587
+ "gen_kwargs": null
2588
+ },
2589
+ "git_hash": "99f5004"
2590
+ }
lm-eval-output/RWKV/rwkv-4-world-1b5/ceval-valid/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3f40034048c8e13209a93a47135ce9d54d611671ce82d5ccb8472d516ef8e50f
3
+ size 59475
lm-eval-output/RWKV/rwkv-4-world-1b5/cmmlu/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:efe2d9e165ddbcc33c7dec06e372a3c5d054622ef9a57db183e366ccdef65567
3
+ size 2315246
lm-eval-output/RWKV/rwkv-4-world-1b5/cmmlu/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json ADDED
The diff for this file is too large to render. See raw diff
 
lm-eval-output/RWKV/rwkv-4-world-1b5/cmmlu/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4846f0328eca8774ef3d95f7ac6f4b6dd7350a2ba6826703f1836db5363781f2
3
+ size 86779
lm-eval-output/RWKV/rwkv-4-world-1b5/cola/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0403ef298376c18efa336676e4fbaf417633fb40af04044170858c789e08cbc8
3
+ size 60794
lm-eval-output/RWKV/rwkv-4-world-1b5/cola/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "cola": {
4
+ "mcc,none": 0.005746920543176395,
5
+ "mcc_stderr,none": 0.03167442197384713,
6
+ "alias": "cola"
7
+ }
8
+ },
9
+ "configs": {
10
+ "cola": {
11
+ "task": "cola",
12
+ "group": "glue",
13
+ "dataset_path": "glue",
14
+ "dataset_name": "cola",
15
+ "training_split": "train",
16
+ "validation_split": "validation",
17
+ "doc_to_text": "{{sentence}}\nQuestion: Does this sentence make sense?\nAnswer:",
18
+ "doc_to_target": "label",
19
+ "doc_to_choice": [
20
+ "no",
21
+ "yes"
22
+ ],
23
+ "description": "",
24
+ "target_delimiter": " ",
25
+ "fewshot_delimiter": "\n\n",
26
+ "metric_list": [
27
+ {
28
+ "metric": "mcc"
29
+ }
30
+ ],
31
+ "output_type": "multiple_choice",
32
+ "repeats": 1,
33
+ "should_decontaminate": true,
34
+ "doc_to_decontamination_query": "sentence",
35
+ "metadata": {
36
+ "version": 1.0
37
+ }
38
+ }
39
+ },
40
+ "versions": {
41
+ "cola": 1.0
42
+ },
43
+ "n-shot": {
44
+ "cola": 0
45
+ },
46
+ "config": {
47
+ "model": "hf",
48
+ "model_args": "pretrained=RWKV/rwkv-4-world-1b5,dtype=bfloat16,trust_remote_code=True",
49
+ "batch_size": "auto",
50
+ "batch_sizes": [
51
+ 64
52
+ ],
53
+ "device": null,
54
+ "use_cache": null,
55
+ "limit": null,
56
+ "bootstrap_iters": 100000,
57
+ "gen_kwargs": null
58
+ },
59
+ "git_hash": "99f5004"
60
+ }
lm-eval-output/RWKV/rwkv-4-world-1b5/cola/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:72cdb771c5916159fe3a996e1c5af351400990a52fdcc052c07594b2952385b3
3
+ size 14934
lm-eval-output/RWKV/rwkv-4-world-1b5/copa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b86be926ca902ba5cd26fa55c1f2a4c3061b3be927d85c377ea1949dad68e2a4
3
+ size 10206
lm-eval-output/RWKV/rwkv-4-world-1b5/copa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "copa": {
4
+ "acc,none": 0.78,
5
+ "acc_stderr,none": 0.04163331998932261,
6
+ "alias": "copa"
7
+ }
8
+ },
9
+ "configs": {
10
+ "copa": {
11
+ "task": "copa",
12
+ "group": [
13
+ "super-glue-lm-eval-v1"
14
+ ],
15
+ "dataset_path": "super_glue",
16
+ "dataset_name": "copa",
17
+ "training_split": "train",
18
+ "validation_split": "validation",
19
+ "doc_to_text": "def doc_to_text(doc):\n # Drop the period\n connector = {\n \"cause\": \"because\",\n \"effect\": \"therefore\",\n }[doc[\"question\"]]\n return doc[\"premise\"].strip()[:-1] + f\" {connector}\"\n",
20
+ "doc_to_target": "def doc_to_target(doc):\n correct_choice = doc[\"choice1\"] if doc[\"label\"] == 0 else doc[\"choice2\"]\n # Connect the sentences\n return \" \" + convert_choice(correct_choice)\n",
21
+ "doc_to_choice": "def doc_to_choice(doc):\n return [\" \" + convert_choice(doc[\"choice1\"]), \" \" + convert_choice(doc[\"choice2\"])]\n",
22
+ "description": "",
23
+ "target_delimiter": " ",
24
+ "fewshot_delimiter": "\n\n",
25
+ "metric_list": [
26
+ {
27
+ "metric": "acc"
28
+ }
29
+ ],
30
+ "output_type": "multiple_choice",
31
+ "repeats": 1,
32
+ "should_decontaminate": false,
33
+ "metadata": {
34
+ "version": 1.0
35
+ }
36
+ }
37
+ },
38
+ "versions": {
39
+ "copa": 1.0
40
+ },
41
+ "n-shot": {
42
+ "copa": 0
43
+ },
44
+ "config": {
45
+ "model": "hf",
46
+ "model_args": "pretrained=RWKV/rwkv-4-world-1b5,dtype=bfloat16,trust_remote_code=True",
47
+ "batch_size": "auto",
48
+ "batch_sizes": [
49
+ 64
50
+ ],
51
+ "device": null,
52
+ "use_cache": null,
53
+ "limit": null,
54
+ "bootstrap_iters": 100000,
55
+ "gen_kwargs": null
56
+ },
57
+ "git_hash": "99f5004"
58
+ }