GGUF
TensorBlock
GGUF
Inference Endpoints
morriszms commited on
Commit
9aab72b
·
verified ·
1 Parent(s): b75f168

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,15 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ Orca-2-13b-SFT-v4-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
37
+ Orca-2-13b-SFT-v4-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
38
+ Orca-2-13b-SFT-v4-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
39
+ Orca-2-13b-SFT-v4-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
40
+ Orca-2-13b-SFT-v4-Q4_0.gguf filter=lfs diff=lfs merge=lfs -text
41
+ Orca-2-13b-SFT-v4-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
42
+ Orca-2-13b-SFT-v4-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
43
+ Orca-2-13b-SFT-v4-Q5_0.gguf filter=lfs diff=lfs merge=lfs -text
44
+ Orca-2-13b-SFT-v4-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
45
+ Orca-2-13b-SFT-v4-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
46
+ Orca-2-13b-SFT-v4-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
47
+ Orca-2-13b-SFT-v4-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
Orca-2-13b-SFT-v4-Q2_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e3d9d2829275e7882cb4db774673f4a382a8ee4a5617da6894e69cd8870b927
3
+ size 4854288576
Orca-2-13b-SFT-v4-Q3_K_L.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9fb9b16f1a035ca1cb3e97d42cacd9111dc8d786170c9e7fe037d888b9b23bc8
3
+ size 6929579744
Orca-2-13b-SFT-v4-Q3_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:699b97da26cf8a8fc9477bc29138e5bd4f5fdabd63eb3b3ae0f197891a20fdfa
3
+ size 6337789664
Orca-2-13b-SFT-v4-Q3_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:25d4202205c6baa852e1919074da1bd9a35ac0c684693b40eea6dcb52b1aaaa7
3
+ size 5659000544
Orca-2-13b-SFT-v4-Q4_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ff6b5f95a28f5a7a5ab0d0ce129ca151ce1af76aea18d976a12ffd865afcd643
3
+ size 7365856960
Orca-2-13b-SFT-v4-Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69dbd5f4e9aa8d4deb32d056f64722f210f218d2160c277d825bb157ee3d0e9e
3
+ size 7865978560
Orca-2-13b-SFT-v4-Q4_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2b97968171a8c6da5c0e9593d4e7d3a1fbed2a09b5f5c39b0d36ff0ada7bf27c
3
+ size 7423200960
Orca-2-13b-SFT-v4-Q5_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b142e3516f14a006ec1f8c48428f1eb3e398ae71d2eda1f6ae4b3b3542e151e8
3
+ size 8972310080
Orca-2-13b-SFT-v4-Q5_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7b3456c94369541d6295963ad1a0ea875048bb6cd5cac57fabe14f3f39d56bf4
3
+ size 9229948480
Orca-2-13b-SFT-v4-Q5_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b3f8d3469739ca050de0576485e8fd1cfa910e8e4669ea79302c2b2a0218061d
3
+ size 8972310080
Orca-2-13b-SFT-v4-Q6_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:152ad6fca26bf941f8fae25275c42b2a4ba6aa6278e4cbd89aad6edd9a527f9a
3
+ size 10679166528
Orca-2-13b-SFT-v4-Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ac30d773472429e84f381837ec2ee4e217bc8300ac60ef3cb091fad0868aef10
3
+ size 13831353152
README.md ADDED
@@ -0,0 +1,142 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ language:
4
+ - en
5
+ - fr
6
+ - es
7
+ - hi
8
+ - zh
9
+ - code
10
+ base_model: Locutusque/Orca-2-13b-SFT-v4
11
+ datasets:
12
+ - HuggingFaceH4/no_robots
13
+ - mlabonne/guanaco-llama2-1k
14
+ - OpenAssistant/oasst_top1_2023-08-25
15
+ - totally-not-an-llm/EverythingLM-data-V3
16
+ widget:
17
+ - example_title: folklore
18
+ text: '<|im_start|>user
19
+
20
+ Are the stories of Nahanni Valley, like the decapitated men, true?
21
+
22
+ <|im_end|>
23
+
24
+ <|im_start|>assistant
25
+
26
+ '
27
+ output:
28
+ text: 'The Nahanni Valley in Canada is a remote and rugged region that has long
29
+ been shrouded in mystery. It is home to the South Nahanni River, which was designated
30
+ as a UNESCO World Heritage Site due to its exceptional natural beauty and biodiversity.
31
+ The valley''s history includes tales of gold prospectors, fur traders, and indigenous
32
+ populations, but it also features some unexplained events and legends. One such
33
+ story involves two prospectors who allegedly encountered an unknown creature
34
+ that led them into the forest before returning without their heads. This tale
35
+ has been passed down through generations, but there are no definitive answers
36
+ regarding its veracity.
37
+
38
+
39
+ It is important to note that these stories may have originated from local folklore
40
+ or were exaggerated over time. There is no concrete evidence to support the
41
+ existence of supernatural creatures or paranormal occurrences within the Nahanni
42
+ Valley. Nonetheless, the area remains a captivating destination for those seeking
43
+ adventure and exploration. If you have any questions about this topic or need
44
+ assistance related to traveling to the Nahanni Valley, please feel free to ask!'
45
+ - example_title: Alan Turing
46
+ text: '<|im_start|>user
47
+
48
+ Why was Alan Turing castrated?
49
+
50
+ <|im_end|>
51
+
52
+ <|im_start|>assistant
53
+
54
+ '
55
+ output:
56
+ text: 'Alan Turing, a renowned mathematician and computer scientist who played
57
+ a crucial role in breaking the German Enigma code during World War II, faced
58
+ significant challenges due to his homosexuality. In 1952, he was convicted of
59
+ "gross indecency" for engaging in a consensual relationship with another man,
60
+ which led to him being sentenced to hormone therapy designed to reduce his sex
61
+ drive. This treatment included the administration of estrogen, which had the
62
+ side effect of causing breast growth.
63
+
64
+
65
+ In an effort to avoid this unwanted physical change, Turing requested that his
66
+ doctors cease the hormone treatments. However, they ignored his request and
67
+ continued the therapy without his consent. As a result, Turing''s breasts began
68
+ to grow, leading him to undergo surgery to remove them. The operation, known
69
+ as a mastectomy, caused considerable pain and discomfort, further contributing
70
+ to his emotional turmoil.
71
+
72
+
73
+ It is important to note that Turing''s castration was not directly carried out
74
+ by the authorities but rather resulted from the unintended consequences of the
75
+ forced hormone therapy.'
76
+ tags:
77
+ - TensorBlock
78
+ - GGUF
79
+ ---
80
+
81
+ <div style="width: auto; margin-left: auto; margin-right: auto">
82
+ <img src="https://i.imgur.com/jC7kdl8.jpeg" alt="TensorBlock" style="width: 100%; min-width: 400px; display: block; margin: auto;">
83
+ </div>
84
+ <div style="display: flex; justify-content: space-between; width: 100%;">
85
+ <div style="display: flex; flex-direction: column; align-items: flex-start;">
86
+ <p style="margin-top: 0.5em; margin-bottom: 0em;">
87
+ Feedback and support: TensorBlock's <a href="https://x.com/tensorblock_aoi">Twitter/X</a>, <a href="https://t.me/TensorBlock">Telegram Group</a> and <a href="https://x.com/tensorblock_aoi">Discord server</a>
88
+ </p>
89
+ </div>
90
+ </div>
91
+
92
+ ## Locutusque/Orca-2-13b-SFT-v4 - GGUF
93
+
94
+ This repo contains GGUF format model files for [Locutusque/Orca-2-13b-SFT-v4](https://huggingface.co/Locutusque/Orca-2-13b-SFT-v4).
95
+
96
+ The files were quantized using machines provided by [TensorBlock](https://tensorblock.co/), and they are compatible with llama.cpp as of [commit b4011](https://github.com/ggerganov/llama.cpp/commit/a6744e43e80f4be6398fc7733a01642c846dce1d).
97
+
98
+ ## Prompt template
99
+
100
+ ```
101
+
102
+ ```
103
+
104
+ ## Model file specification
105
+
106
+ | Filename | Quant type | File Size | Description |
107
+ | -------- | ---------- | --------- | ----------- |
108
+ | [Orca-2-13b-SFT-v4-Q2_K.gguf](https://huggingface.co/tensorblock/Orca-2-13b-SFT-v4-GGUF/tree/main/Orca-2-13b-SFT-v4-Q2_K.gguf) | Q2_K | 4.521 GB | smallest, significant quality loss - not recommended for most purposes |
109
+ | [Orca-2-13b-SFT-v4-Q3_K_S.gguf](https://huggingface.co/tensorblock/Orca-2-13b-SFT-v4-GGUF/tree/main/Orca-2-13b-SFT-v4-Q3_K_S.gguf) | Q3_K_S | 5.270 GB | very small, high quality loss |
110
+ | [Orca-2-13b-SFT-v4-Q3_K_M.gguf](https://huggingface.co/tensorblock/Orca-2-13b-SFT-v4-GGUF/tree/main/Orca-2-13b-SFT-v4-Q3_K_M.gguf) | Q3_K_M | 5.903 GB | very small, high quality loss |
111
+ | [Orca-2-13b-SFT-v4-Q3_K_L.gguf](https://huggingface.co/tensorblock/Orca-2-13b-SFT-v4-GGUF/tree/main/Orca-2-13b-SFT-v4-Q3_K_L.gguf) | Q3_K_L | 6.454 GB | small, substantial quality loss |
112
+ | [Orca-2-13b-SFT-v4-Q4_0.gguf](https://huggingface.co/tensorblock/Orca-2-13b-SFT-v4-GGUF/tree/main/Orca-2-13b-SFT-v4-Q4_0.gguf) | Q4_0 | 6.860 GB | legacy; small, very high quality loss - prefer using Q3_K_M |
113
+ | [Orca-2-13b-SFT-v4-Q4_K_S.gguf](https://huggingface.co/tensorblock/Orca-2-13b-SFT-v4-GGUF/tree/main/Orca-2-13b-SFT-v4-Q4_K_S.gguf) | Q4_K_S | 6.913 GB | small, greater quality loss |
114
+ | [Orca-2-13b-SFT-v4-Q4_K_M.gguf](https://huggingface.co/tensorblock/Orca-2-13b-SFT-v4-GGUF/tree/main/Orca-2-13b-SFT-v4-Q4_K_M.gguf) | Q4_K_M | 7.326 GB | medium, balanced quality - recommended |
115
+ | [Orca-2-13b-SFT-v4-Q5_0.gguf](https://huggingface.co/tensorblock/Orca-2-13b-SFT-v4-GGUF/tree/main/Orca-2-13b-SFT-v4-Q5_0.gguf) | Q5_0 | 8.356 GB | legacy; medium, balanced quality - prefer using Q4_K_M |
116
+ | [Orca-2-13b-SFT-v4-Q5_K_S.gguf](https://huggingface.co/tensorblock/Orca-2-13b-SFT-v4-GGUF/tree/main/Orca-2-13b-SFT-v4-Q5_K_S.gguf) | Q5_K_S | 8.356 GB | large, low quality loss - recommended |
117
+ | [Orca-2-13b-SFT-v4-Q5_K_M.gguf](https://huggingface.co/tensorblock/Orca-2-13b-SFT-v4-GGUF/tree/main/Orca-2-13b-SFT-v4-Q5_K_M.gguf) | Q5_K_M | 8.596 GB | large, very low quality loss - recommended |
118
+ | [Orca-2-13b-SFT-v4-Q6_K.gguf](https://huggingface.co/tensorblock/Orca-2-13b-SFT-v4-GGUF/tree/main/Orca-2-13b-SFT-v4-Q6_K.gguf) | Q6_K | 9.946 GB | very large, extremely low quality loss |
119
+ | [Orca-2-13b-SFT-v4-Q8_0.gguf](https://huggingface.co/tensorblock/Orca-2-13b-SFT-v4-GGUF/tree/main/Orca-2-13b-SFT-v4-Q8_0.gguf) | Q8_0 | 12.881 GB | very large, extremely low quality loss - not recommended |
120
+
121
+
122
+ ## Downloading instruction
123
+
124
+ ### Command line
125
+
126
+ Firstly, install Huggingface Client
127
+
128
+ ```shell
129
+ pip install -U "huggingface_hub[cli]"
130
+ ```
131
+
132
+ Then, downoad the individual model file the a local directory
133
+
134
+ ```shell
135
+ huggingface-cli download tensorblock/Orca-2-13b-SFT-v4-GGUF --include "Orca-2-13b-SFT-v4-Q2_K.gguf" --local-dir MY_LOCAL_DIR
136
+ ```
137
+
138
+ If you wanna download multiple model files with a pattern (e.g., `*Q4_K*gguf`), you can try:
139
+
140
+ ```shell
141
+ huggingface-cli download tensorblock/Orca-2-13b-SFT-v4-GGUF --local-dir MY_LOCAL_DIR --local-dir-use-symlinks False --include='*Q4_K*gguf'
142
+ ```