mantafloppy
commited on
Upload folder using huggingface_hub
Browse files- README.md +3 -15
- whiterabbitneo-33b-v1-q4_k.gguf +3 -0
- whiterabbitneo-33b-v1-q8_0.gguf +3 -0
README.md
CHANGED
@@ -4,18 +4,7 @@ license_name: deepseek
|
|
4 |
license_link: https://huggingface.co/deepseek-ai/deepseek-coder-33b-base/blob/main/LICENSE
|
5 |
---
|
6 |
|
7 |
-
***Currently uploading
|
8 |
-
|
9 |
-
# WhiteRabbitNeo 33B v1 - GGUF
|
10 |
-
- Model creator: [WhiteRabbitNeo](https://huggingface.co/whiterabbitneo)
|
11 |
-
- Original model: [WhiteRabbitNeo 33B v1](https://huggingface.co/whiterabbitneo/WhiteRabbitNeo-33B-v1)
|
12 |
-
|
13 |
-
<!-- description start -->
|
14 |
-
## Description
|
15 |
-
|
16 |
-
This repo contains GGUF format model files for [WhiteRabbitNeo's WhiteRabbitNeo 33B v1](https://huggingface.co/whiterabbitneo/WhiteRabbitNeo-33B-v1).
|
17 |
-
|
18 |
-
<!-- description end -->
|
19 |
|
20 |
See https://huggingface.co/TheBloke/WhiteRabbitNeo-33B-v1-GGUF to see how to run.
|
21 |
|
@@ -30,15 +19,14 @@ snapshot_download(repo_id=model_id, local_dir="whiterabbitneo-hf",
|
|
30 |
```
|
31 |
|
32 |
brew install gh
|
33 |
-
|
34 |
gh auth login
|
35 |
-
|
36 |
gh pr checkout 3633
|
37 |
|
38 |
python3 llama.cpp/convert.py whiterabbitneo-hf --outfile whiterabbitneo-33b-v1-q8_0.gguf --outtype q8_0 --padvocab
|
39 |
|
40 |
-
python3 llama.cpp/convert.py whiterabbitneo-hf --outfile whiterabbitneo-f16.gguf --outtype f16 --padvocab
|
41 |
|
|
|
|
|
42 |
llama.cpp/quantize whiterabbitneo-f16.gguf whiterabbitneo-q4_k.gguf q4_k
|
43 |
|
44 |
```
|
|
|
4 |
license_link: https://huggingface.co/deepseek-ai/deepseek-coder-33b-base/blob/main/LICENSE
|
5 |
---
|
6 |
|
7 |
+
***Currently uploading q4_k_m : Upload started at 15:13 January 13 2024. 15mbps connection. Should take 4 hour.***
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
8 |
|
9 |
See https://huggingface.co/TheBloke/WhiteRabbitNeo-33B-v1-GGUF to see how to run.
|
10 |
|
|
|
19 |
```
|
20 |
|
21 |
brew install gh
|
|
|
22 |
gh auth login
|
|
|
23 |
gh pr checkout 3633
|
24 |
|
25 |
python3 llama.cpp/convert.py whiterabbitneo-hf --outfile whiterabbitneo-33b-v1-q8_0.gguf --outtype q8_0 --padvocab
|
26 |
|
|
|
27 |
|
28 |
+
|
29 |
+
python3 llama.cpp/convert.py whiterabbitneo-hf --outfile whiterabbitneo-f16.gguf --outtype f16 --padvocab
|
30 |
llama.cpp/quantize whiterabbitneo-f16.gguf whiterabbitneo-q4_k.gguf q4_k
|
31 |
|
32 |
```
|
whiterabbitneo-33b-v1-q4_k.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b78315b5ab07dc687bc5650089ee224aa640468b39f2e5728000495a6061d874
|
3 |
+
size 19940659296
|
whiterabbitneo-33b-v1-q8_0.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:423633292e85c037e8c98806a9fecd22b6028a3911f57a92cc9c5796998893bb
|
3 |
+
size 35430879296
|