flyingfishinwater commited on
Commit
c973f13
·
verified ·
1 Parent(s): 446ae9a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +59 -20
README.md CHANGED
@@ -22,10 +22,11 @@ Llama 3 is the latest and most advanced LLM trained over 15T tokens, which impro
22
 
23
  **Prompt Format:**
24
  ```
25
- <|begin_of_text|><|start_header_id|>user<|end_header_id|>
26
 
27
  {{prompt}}<|eot_id|><|start_header_id|>assistant<|end_header_id|>
28
 
 
29
 
30
  ```
31
 
@@ -97,7 +98,7 @@ The TinyLlama project aims to pretrain a 1.1B Llama model on 3 trillion tokens.
97
 
98
  **Prompt Format:**
99
  ```
100
- <|system|>You are a friendly chatbot who always responds in the style of a pirate.</s><|user|>{{prompt}}</s><|assistant|>
101
  ```
102
 
103
  **Template Name:** TinyLlama
@@ -146,12 +147,12 @@ The Mistral-7B-v0.2 Large Language Model (LLM) is a pretrained generative text m
146
 
147
  ---
148
 
149
- # OpenChat 3.5
150
  OpenChat is an innovative library of open-source language models, fine-tuned with C-RLFT - a strategy inspired by offline reinforcement learning. Our models learn from mixed-quality data without preference labels, delivering exceptional performance on par with ChatGPT, even with a 7B model. Despite our simple approach, we are committed to developing a high-performance, commercially viable, open-source large language model, and we continue to make significant strides toward this vision.
151
 
152
  **Model Intention:** It's a 7B large model and performs really good for Q&A. But it requires a high-end device to run.
153
 
154
- **Model URL:** [https://huggingface.co/flyingfishinwater/goodmodels/resolve/main/openchat-3.5-1210.Q8_0.gguf?download=true](https://huggingface.co/flyingfishinwater/goodmodels/resolve/main/openchat-3.5-1210.Q8_0.gguf?download=true)
155
 
156
  **Model Info URL:** [https://huggingface.co/openchat/openchat_3.5](https://huggingface.co/openchat/openchat_3.5)
157
 
@@ -161,13 +162,13 @@ OpenChat is an innovative library of open-source language models, fine-tuned wit
161
 
162
  **Developer:** [https://openchat.team/](https://openchat.team/)
163
 
164
- **File Size:** 7695 MB
165
 
166
- **Context Length:** 4096 tokens
167
 
168
  **Prompt Format:**
169
  ```
170
- <s>[INST]{{prompt}}[/INST]</s>
171
  ```
172
 
173
  **Template Name:** Mistral
@@ -186,7 +187,7 @@ Phi-2 is a Transformer with 2.7 billion parameters. It was trained using the sam
186
 
187
  **Model Intention:** It's a 2.7B model and is intended for QA, chat, and code purposes
188
 
189
- **Model URL:** [https://huggingface.co/ggml-org/models/resolve/main/phi-2/ggml-model-q8_0.gguf?download=true](https://huggingface.co/ggml-org/models/resolve/main/phi-2/ggml-model-q8_0.gguf?download=true)
190
 
191
  **Model Info URL:** [https://huggingface.co/microsoft/phi-2](https://huggingface.co/microsoft/phi-2)
192
 
@@ -196,9 +197,9 @@ Phi-2 is a Transformer with 2.7 billion parameters. It was trained using the sam
196
 
197
  **Developer:** [https://huggingface.co/microsoft/phi-2](https://huggingface.co/microsoft/phi-2)
198
 
199
- **File Size:** 2960 MB
200
 
201
- **Context Length:** 4096 tokens
202
 
203
  **Prompt Format:**
204
  ```
@@ -222,7 +223,7 @@ The Yi series models are the next generation of open-source large language model
222
 
223
  **Model Intention:** It's a 6B model and can understand English and Chinese. It's good for QA and Chat
224
 
225
- **Model URL:** [https://huggingface.co/flyingfishinwater/goodmodels/resolve/main/yi-6b-chat-Q8_0.gguf?download=true](https://huggingface.co/flyingfishinwater/goodmodels/resolve/main/yi-6b-chat-Q8_0.gguf?download=true)
226
 
227
  **Model Info URL:** [https://huggingface.co/01-ai/Yi-6B-Chat](https://huggingface.co/01-ai/Yi-6B-Chat)
228
 
@@ -232,9 +233,9 @@ The Yi series models are the next generation of open-source large language model
232
 
233
  **Developer:** [https://01.ai/](https://01.ai/)
234
 
235
- **File Size:** 6440 MB
236
 
237
- **Context Length:** 200000 tokens
238
 
239
  **Prompt Format:**
240
  ```
@@ -297,7 +298,7 @@ Gemma is a family of lightweight, state-of-the-art open models built from the sa
297
  # StarCoder2 3B
298
  StarCoder2-3B model is a 3B parameter model trained on 17 programming languages from The Stack v2, with opt-out requests excluded. The model uses Grouped Query Attention, a context window of 16,384 tokens with a sliding window attention of 4,096 tokens, and was trained using the Fill-in-the-Middle objective on 3+ trillion tokens
299
 
300
- **Model Intention:** The model is good at 17 programming languages. It can help you resolve programming requirements
301
 
302
  **Model URL:** [https://huggingface.co/flyingfishinwater/goodmodels/resolve/main/starcoder2-3b-instruct-gguf_Q8_0.gguf?download=true](https://huggingface.co/flyingfishinwater/goodmodels/resolve/main/starcoder2-3b-instruct-gguf_Q8_0.gguf?download=true)
303
 
@@ -311,12 +312,11 @@ StarCoder2-3B model is a 3B parameter model trained on 17 programming languages
311
 
312
  **File Size:** 3220 MB
313
 
314
- **Context Length:** 8192 tokens
315
 
316
  **Prompt Format:**
317
  ```
318
- ### Instruction
319
- {{prompt}}### Response
320
 
321
  ```
322
 
@@ -368,6 +368,45 @@ Chinese Tiny LLM 2B 是首个以中文为中心的大型语言模型,主要在
368
  **Parse Special Tokens:** Yes
369
 
370
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
371
  ---
372
 
373
  # Dophin 2.8 Mistralv02 7B
@@ -387,11 +426,11 @@ This model is based on Mistral-7b-v0.2 with 16k context lengths. It's a uncensor
387
 
388
  **File Size:** 2728 MB
389
 
390
- **Context Length:** 16384 tokens
391
 
392
  **Prompt Format:**
393
  ```
394
- <|im_start|>user
395
  {{prompt}}
396
  <|im_end|>
397
  <|im_start|>assistant
@@ -444,4 +483,4 @@ ASSISTANT:
444
  **Parse Special Tokens:** Yes
445
 
446
 
447
- ---
 
22
 
23
  **Prompt Format:**
24
  ```
25
+ <|start_header_id|>user<|end_header_id|>
26
 
27
  {{prompt}}<|eot_id|><|start_header_id|>assistant<|end_header_id|>
28
 
29
+ assistant
30
 
31
  ```
32
 
 
98
 
99
  **Prompt Format:**
100
  ```
101
+ <|user|>{{prompt}}</s><|assistant|>
102
  ```
103
 
104
  **Template Name:** TinyLlama
 
147
 
148
  ---
149
 
150
+ # OpenChat 3.5(0106)
151
  OpenChat is an innovative library of open-source language models, fine-tuned with C-RLFT - a strategy inspired by offline reinforcement learning. Our models learn from mixed-quality data without preference labels, delivering exceptional performance on par with ChatGPT, even with a 7B model. Despite our simple approach, we are committed to developing a high-performance, commercially viable, open-source large language model, and we continue to make significant strides toward this vision.
152
 
153
  **Model Intention:** It's a 7B large model and performs really good for Q&A. But it requires a high-end device to run.
154
 
155
+ **Model URL:** [https://huggingface.co/flyingfishinwater/goodmodels/resolve/main/openchat-3.5-0106.Q3_K_M.gguf?download=true](https://huggingface.co/flyingfishinwater/goodmodels/resolve/main/openchat-3.5-0106.Q3_K_M.gguf?download=true)
156
 
157
  **Model Info URL:** [https://huggingface.co/openchat/openchat_3.5](https://huggingface.co/openchat/openchat_3.5)
158
 
 
162
 
163
  **Developer:** [https://openchat.team/](https://openchat.team/)
164
 
165
+ **File Size:** 3520 MB
166
 
167
+ **Context Length:** 8192 tokens
168
 
169
  **Prompt Format:**
170
  ```
171
+ GPT4 Correct User: {{prompt}}<|end_of_turn|>GPT4 Correct Assistant:
172
  ```
173
 
174
  **Template Name:** Mistral
 
187
 
188
  **Model Intention:** It's a 2.7B model and is intended for QA, chat, and code purposes
189
 
190
+ **Model URL:** [https://huggingface.co/ggml-org/models/resolve/main/phi-2.Q5_K_M.gguf?download=true](https://huggingface.co/ggml-org/models/resolve/main/phi-2.Q5_K_M.gguf?download=true)
191
 
192
  **Model Info URL:** [https://huggingface.co/microsoft/phi-2](https://huggingface.co/microsoft/phi-2)
193
 
 
197
 
198
  **Developer:** [https://huggingface.co/microsoft/phi-2](https://huggingface.co/microsoft/phi-2)
199
 
200
+ **File Size:** 2070 MB
201
 
202
+ **Context Length:** 2048 tokens
203
 
204
  **Prompt Format:**
205
  ```
 
223
 
224
  **Model Intention:** It's a 6B model and can understand English and Chinese. It's good for QA and Chat
225
 
226
+ **Model URL:** [https://huggingface.co/flyingfishinwater/goodmodels/resolve/main/yi-chat-6b.Q4_K_M.gguf?download=true](https://huggingface.co/flyingfishinwater/goodmodels/resolve/main/yi-chat-6b.Q4_K_M.gguf?download=true)
227
 
228
  **Model Info URL:** [https://huggingface.co/01-ai/Yi-6B-Chat](https://huggingface.co/01-ai/Yi-6B-Chat)
229
 
 
233
 
234
  **Developer:** [https://01.ai/](https://01.ai/)
235
 
236
+ **File Size:** 3670 MB
237
 
238
+ **Context Length:** 4096 tokens
239
 
240
  **Prompt Format:**
241
  ```
 
298
  # StarCoder2 3B
299
  StarCoder2-3B model is a 3B parameter model trained on 17 programming languages from The Stack v2, with opt-out requests excluded. The model uses Grouped Query Attention, a context window of 16,384 tokens with a sliding window attention of 4,096 tokens, and was trained using the Fill-in-the-Middle objective on 3+ trillion tokens
300
 
301
+ **Model Intention:** The model is good at 17 programming languages. By just start with your codes, the model will finish it.
302
 
303
  **Model URL:** [https://huggingface.co/flyingfishinwater/goodmodels/resolve/main/starcoder2-3b-instruct-gguf_Q8_0.gguf?download=true](https://huggingface.co/flyingfishinwater/goodmodels/resolve/main/starcoder2-3b-instruct-gguf_Q8_0.gguf?download=true)
304
 
 
312
 
313
  **File Size:** 3220 MB
314
 
315
+ **Context Length:** 16384 tokens
316
 
317
  **Prompt Format:**
318
  ```
319
+ {{prompt}}
 
320
 
321
  ```
322
 
 
368
  **Parse Special Tokens:** Yes
369
 
370
 
371
+ ---
372
+
373
+ # Qwen1.5 4B Chat
374
+ Qwen is the large language model and large multimodal model series of the Qwen Team, Alibaba Group. It supports both Chinese and English. 通义千问是阿里巴巴公司开发的大大预言模型,支持中英文双语。
375
+
376
+ **Model Intention:** It's one of the best LLM that supports Chinese and English. 这是支持中英双语的最佳的大语言模型之一。
377
+
378
+ **Model URL:** [https://huggingface.co/flyingfishinwater/goodmodels/resolve/main/qwen1_5-4b-chat-q4_k_m.gguf?download=true](https://huggingface.co/flyingfishinwater/goodmodels/resolve/main/qwen1_5-4b-chat-q4_k_m.gguf?download=true)
379
+
380
+ **Model Info URL:** [https://huggingface.co/Qwen/Qwen1.5-4B-Chat-GGUF](https://huggingface.co/Qwen/Qwen1.5-4B-Chat-GGUF)
381
+
382
+ **Model License:** [License Info](https://huggingface.co/Qwen/Qwen1.5-4B-Chat/raw/main/LICENSE)
383
+
384
+ **Model Description:** Qwen is the large language model and large multimodal model series of the Qwen Team, Alibaba Group. It supports both Chinese and English. 通义千问是阿里巴巴公司开发的大大预言模型,支持中英文双语。
385
+
386
+ **Developer:** [https://qwenlm.github.io/](https://qwenlm.github.io/)
387
+
388
+ **File Size:** 2460 MB
389
+
390
+ **Context Length:** 32768 tokens
391
+
392
+ **Prompt Format:**
393
+ ```
394
+ <|im_start|>user
395
+ {{prompt}}
396
+ <|im_end|>
397
+ <|im_start|>assistant
398
+
399
+ ```
400
+
401
+ **Template Name:** chatml
402
+
403
+ **Add BOS Token:** Yes
404
+
405
+ **Add EOS Token:** No
406
+
407
+ **Parse Special Tokens:** Yes
408
+
409
+
410
  ---
411
 
412
  # Dophin 2.8 Mistralv02 7B
 
426
 
427
  **File Size:** 2728 MB
428
 
429
+ **Context Length:** 32768 tokens
430
 
431
  **Prompt Format:**
432
  ```
433
+ <s><|im_start|>user
434
  {{prompt}}
435
  <|im_end|>
436
  <|im_start|>assistant
 
483
  **Parse Special Tokens:** Yes
484
 
485
 
486
+ ---