PEFT
heegyu commited on
Commit
6b7b667
·
2 Parent(s): 87001f8 06212cc

steps-12500

Browse files
README.md DELETED
@@ -1,32 +0,0 @@
1
- ---
2
- datasets:
3
- - kyujinpy/KOpen-platypus
4
- - ehartford/dolphin
5
- ---
6
-
7
-
8
- # steps-12500
9
- | Task |Version| Metric |Value | |Stderr|
10
- |----------------|------:|--------|-----:|---|-----:|
11
- |kobest_boolq | 0|acc |0.5064|± |0.0133|
12
- | | |macro_f1|0.3669|± |0.0094|
13
- |kobest_copa | 0|acc |0.7260|± |0.0141|
14
- | | |macro_f1|0.7254|± |0.0141|
15
- |kobest_hellaswag| 0|acc |0.4380|± |0.0222|
16
- | | |acc_norm|0.5460|± |0.0223|
17
- | | |macro_f1|0.4345|± |0.0222|
18
- |kobest_sentineg | 0|acc |0.8161|± |0.0195|
19
- | | |macro_f1|0.8134|± |0.0199|
20
-
21
- # steps-25000
22
- | Task |Version| Metric |Value | |Stderr|
23
- |----------------|------:|--------|-----:|---|-----:|
24
- |kobest_boolq | 0|acc |0.5028|± |0.0133|
25
- | | |macro_f1|0.3409|± |0.0069|
26
- |kobest_copa | 0|acc |0.7170|± |0.0143|
27
- | | |macro_f1|0.7164|± |0.0142|
28
- |kobest_hellaswag| 0|acc |0.4380|± |0.0222|
29
- | | |acc_norm|0.5380|± |0.0223|
30
- | | |macro_f1|0.4329|± |0.0221|
31
- |kobest_sentineg | 0|acc |0.7809|± |0.0208|
32
- | | |macro_f1|0.7746|± |0.0214|
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
adapter_config.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "auto_mapping": null,
3
+ "base_model_name_or_path": "42dot/42dot_LLM-PLM-1.3B",
4
+ "bias": "none",
5
+ "fan_in_fan_out": false,
6
+ "inference_mode": true,
7
+ "init_lora_weights": true,
8
+ "layers_pattern": null,
9
+ "layers_to_transform": null,
10
+ "lora_alpha": 32,
11
+ "lora_dropout": 0.1,
12
+ "modules_to_save": null,
13
+ "peft_type": "LORA",
14
+ "r": 32,
15
+ "revision": null,
16
+ "target_modules": [
17
+ "q_proj",
18
+ "k_proj",
19
+ "v_proj",
20
+ "o_proj",
21
+ "gate_proj",
22
+ "up_proj",
23
+ "down_proj"
24
+ ],
25
+ "task_type": "CAUSAL_LM"
26
+ }
adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bf94e1c7a225df4ba405a10281ae88eeb8554c5d1d3e183f2c33d7ded03b34f2
3
+ size 121227274
added_tokens.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "<|endoftext|>": 50256,
3
+ "<||bos||>": 50257,
4
+ "<||pad||>": 50258,
5
+ "<||unk||>": 50259
6
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
special_tokens_map.json ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "<||bos||>",
3
+ "eos_token": {
4
+ "content": "<|endoftext|>",
5
+ "lstrip": false,
6
+ "normalized": true,
7
+ "rstrip": false,
8
+ "single_word": false
9
+ },
10
+ "pad_token": "<||pad||>",
11
+ "unk_token": "<||unk||>"
12
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "bos_token": {
5
+ "__type": "AddedToken",
6
+ "content": "<|endoftext|>",
7
+ "lstrip": false,
8
+ "normalized": true,
9
+ "rstrip": false,
10
+ "single_word": false
11
+ },
12
+ "clean_up_tokenization_spaces": true,
13
+ "eos_token": {
14
+ "__type": "AddedToken",
15
+ "content": "<|endoftext|>",
16
+ "lstrip": false,
17
+ "normalized": true,
18
+ "rstrip": false,
19
+ "single_word": false
20
+ },
21
+ "errors": "replace",
22
+ "model_max_length": 8192,
23
+ "pad_token": null,
24
+ "tokenizer_class": "GPT2Tokenizer",
25
+ "unk_token": {
26
+ "__type": "AddedToken",
27
+ "content": "<|endoftext|>",
28
+ "lstrip": false,
29
+ "normalized": true,
30
+ "rstrip": false,
31
+ "single_word": false
32
+ }
33
+ }
vocab.json ADDED
The diff for this file is too large to render. See raw diff