DarqueDante commited on
Commit
1dad186
·
verified ·
1 Parent(s): c40ca4f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +124 -3
README.md CHANGED
@@ -1,12 +1,99 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  base_model:
3
- - DarqueDante/LLama-3-Dolphin-2.9.1-120b
4
  library_name: transformers
5
  tags:
6
  - mergekit
7
  - merge
8
-
 
 
 
 
 
 
 
 
 
 
 
 
9
  ---
 
 
10
  # merge
11
 
12
  This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
@@ -20,7 +107,6 @@ This model was merged using the passthrough merge method.
20
 
21
  The following models were included in the merge:
22
  * [cognitivecomputations/dolphin-2.9.1-llama-3-70b](https://huggingface.co/cognitivecomputations/dolphin-2.9.1-llama-3-70b)
23
- * [cognitivecomputations/dolphin-2.9.1-llama-3-70b](https://huggingface.co/cognitivecomputations/dolphin-2.9.1-llama-3-70b)
24
 
25
  ### Configuration
26
 
@@ -52,3 +138,38 @@ slices:
52
  merge_method: passthrough
53
  dtype: float16
54
  ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Hugging Face's logo
2
+ Hugging Face
3
+ Search models, datasets, users...
4
+ Models
5
+ Datasets
6
+ Spaces
7
+ Posts
8
+ Docs
9
+ Solutions
10
+ Pricing
11
+
12
+
13
+
14
+
15
+ DarqueDante
16
+ /
17
+ LLama-3-Mega-Dolphin-2.9.1-120b
18
+
19
+ like
20
+ 0
21
+ Text Generation
22
+ Transformers
23
+ Safetensors
24
+
25
+ cognitivecomputations/Dolphin-2.9
26
+
27
+ teknium/OpenHermes-2.5
28
+
29
+ m-a-p/CodeFeedback-Filtered-Instruction
30
+
31
+ cognitivecomputations/dolphin-coder
32
+
33
+ cognitivecomputations/samantha-data
34
+
35
+ HuggingFaceH4/ultrachat_200k
36
+
37
+ microsoft/orca-math-word-problems-200k
38
+
39
+ abacusai/SystemChat-1.1
40
+
41
+ Locutusque/function-calling-chatml
42
+
43
+ internlm/Agent-FLAN
44
+ English
45
+ llama
46
+ mergekit
47
+ Merge
48
+ conversational
49
+ Inference Endpoints
50
+ text-generation-inference
51
+ Model card
52
+ Files and versions
53
+ Community
54
+ Settings
55
+ LLama-3-Mega-Dolphin-2.9.1-120b
56
+ /
57
+ README.md
58
+
59
+ DarqueDante's picture
60
+ DarqueDante
61
+ Update README.md
62
+ 9f0a74b
63
+ VERIFIED
64
+ 3 minutes ago
65
+ Preview
66
+ Code
67
+ |
68
+ raw
69
+ history
70
+ blame
71
+ edit
72
+ delete
73
+ 2.68 kB
74
  ---
75
  base_model:
76
+ - cognitivecomputations/dolphin-2.9.1-llama-3-70b
77
  library_name: transformers
78
  tags:
79
  - mergekit
80
  - merge
81
+ language:
82
+ - en
83
+ datasets:
84
+ - cognitivecomputations/Dolphin-2.9
85
+ - teknium/OpenHermes-2.5
86
+ - m-a-p/CodeFeedback-Filtered-Instruction
87
+ - cognitivecomputations/dolphin-coder
88
+ - cognitivecomputations/samantha-data
89
+ - HuggingFaceH4/ultrachat_200k
90
+ - microsoft/orca-math-word-problems-200k
91
+ - abacusai/SystemChat-1.1
92
+ - Locutusque/function-calling-chatml
93
+ - internlm/Agent-FLAN
94
  ---
95
+ ![image/jpeg](https://huggingface.co/DarqueDante/LLama-3-Dolphin-2.9.1-120b/blob/main/mega_dolphin.png)
96
+
97
  # merge
98
 
99
  This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
 
107
 
108
  The following models were included in the merge:
109
  * [cognitivecomputations/dolphin-2.9.1-llama-3-70b](https://huggingface.co/cognitivecomputations/dolphin-2.9.1-llama-3-70b)
 
110
 
111
  ### Configuration
112
 
 
138
  merge_method: passthrough
139
  dtype: float16
140
  ```
141
+
142
+ This model uses ChatML prompt template format.
143
+
144
+ example:
145
+
146
+ ```
147
+ <|im_start|>system
148
+ You are Dolphin, a helpful AI assistant.<|im_end|>
149
+ <|im_start|>user
150
+ {prompt}<|im_end|>
151
+ <|im_start|>assistant
152
+
153
+
154
+ ## 💻 Usage
155
+ ```python
156
+ !pip install -qU transformers accelerate
157
+
158
+ from transformers import AutoTokenizer
159
+ import transformers
160
+ import torch
161
+
162
+ model = "DarqueDante/LLama-3-Dolphin-2.9.1-120b"
163
+ messages = [{"role": "user", "content": "Who is Andrej Karpathy?"}]
164
+
165
+ tokenizer = AutoTokenizer.from_pretrained(model)
166
+ prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
167
+ pipeline = transformers.pipeline(
168
+ "text-generation",
169
+ model=model,
170
+ torch_dtype=torch.float16,
171
+ device_map="auto",
172
+ )
173
+ outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
174
+ print(outputs[0]["generated_text"])
175
+ ```