kyujinpy commited on
Commit
d92f86e
·
1 Parent(s): 138d111

Upload README.md

Browse files
Files changed (1) hide show
  1. README.md +70 -0
README.md CHANGED
@@ -1,3 +1,73 @@
1
  ---
 
 
 
 
2
  license: cc-by-nc-sa-4.0
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ language:
3
+ - en
4
+ - ko
5
+ pipeline_tag: text-generation
6
  license: cc-by-nc-sa-4.0
7
  ---
8
+
9
+ # **SOLAR-tail-10.7B-Merge-v1.0**
10
+
11
+ ## Model Details
12
+
13
+ **Model Developers** Kyujin Han (kyujinpy)
14
+
15
+ **Method**
16
+ Using [Mergekit](https://github.com/cg123/mergekit).
17
+ - [upstage/SOLAR-10.7B-v1.0](https://huggingface.co/upstage/SOLAR-10.7B-v1.0)
18
+ - [Yhyu13/LMCocktail-10.7B-v1](Yhyu13/LMCocktail-10.7B-v1)
19
+
20
+ **Merge config**
21
+ ```
22
+ slices:
23
+ - sources:
24
+ - model: upstage/SOLAR-10.7B-v1.0
25
+ layer_range: [0, 48]
26
+ - model: Yhyu13/LMCocktail-10.7B-v1
27
+ layer_range: [0, 48]
28
+
29
+ merge_method: slerp
30
+ base_model: upstage/SOLAR-10.7B-v1.0
31
+
32
+ parameters:
33
+ t:
34
+ - filter: self_attn
35
+ value: [0, 0.5, 0.3, 0.7, 1]
36
+ - filter: mlp
37
+ value: [1, 0.5, 0.7, 0.3, 0]
38
+ - value: 0.5 # fallback for rest of tensors
39
+ tokenizer_source: union
40
+
41
+ dtype: float16
42
+ ```
43
+
44
+ # **Model Benchmark**
45
+
46
+ ## Open leaderboard
47
+ - Follow up as [link](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).
48
+
49
+ | Model | Average | ARC | HellaSwag | MMLU | TruthfulQA | Winogrande | GSM8K |
50
+ | --- | --- | --- | --- | --- | --- | --- | --- |
51
+ | PracticeLLM/SOLAR-tail-10.7B-Merge-v1.0 | NaN | NaN | NaN | NaN | NaN | NaN | NaN |
52
+ | jjourney1125/M-SOLAR-10.7B-v1.0 | 55.15 | 49.57 | 60.12 | 54.60 | 49.23 | 62.22 |
53
+ | beomi/Yi-Ko-6B | 48.79 | 41.04 | 53.39 | 46.28 | 41.64 | 61.63 |
54
+ | mistralai/Mistral-7B-v0.1 | 46.89 | 38.14 | 48.19 | 45.20 | 46.13 | 56.79 |
55
+
56
+
57
+ # Implementation Code
58
+ ```python
59
+ ### KO-Platypus
60
+ from transformers import AutoModelForCausalLM, AutoTokenizer
61
+ import torch
62
+
63
+ repo = "PracticeLLM/SOLAR-tail-10.7B-Merge-v1.0"
64
+ OpenOrca = AutoModelForCausalLM.from_pretrained(
65
+ repo,
66
+ return_dict=True,
67
+ torch_dtype=torch.float16,
68
+ device_map='auto'
69
+ )
70
+ OpenOrca_tokenizer = AutoTokenizer.from_pretrained(repo)
71
+ ```
72
+
73
+ ---