mantafloppy commited on
Commit
27d0546
·
verified ·
1 Parent(s): a12cb45

Upload folder using huggingface_hub

Browse files
README.md CHANGED
@@ -4,18 +4,7 @@ license_name: deepseek
4
  license_link: https://huggingface.co/deepseek-ai/deepseek-coder-33b-base/blob/main/LICENSE
5
  ---
6
 
7
- ***Currently uploading q8 : Upload started at 15:40 January 13 2024. 15mbps connection.***
8
-
9
- # WhiteRabbitNeo 33B v1 - GGUF
10
- - Model creator: [WhiteRabbitNeo](https://huggingface.co/whiterabbitneo)
11
- - Original model: [WhiteRabbitNeo 33B v1](https://huggingface.co/whiterabbitneo/WhiteRabbitNeo-33B-v1)
12
-
13
- <!-- description start -->
14
- ## Description
15
-
16
- This repo contains GGUF format model files for [WhiteRabbitNeo's WhiteRabbitNeo 33B v1](https://huggingface.co/whiterabbitneo/WhiteRabbitNeo-33B-v1).
17
-
18
- <!-- description end -->
19
 
20
  See https://huggingface.co/TheBloke/WhiteRabbitNeo-33B-v1-GGUF to see how to run.
21
 
@@ -30,15 +19,14 @@ snapshot_download(repo_id=model_id, local_dir="whiterabbitneo-hf",
30
  ```
31
 
32
  brew install gh
33
-
34
  gh auth login
35
-
36
  gh pr checkout 3633
37
 
38
  python3 llama.cpp/convert.py whiterabbitneo-hf --outfile whiterabbitneo-33b-v1-q8_0.gguf --outtype q8_0 --padvocab
39
 
40
- python3 llama.cpp/convert.py whiterabbitneo-hf --outfile whiterabbitneo-f16.gguf --outtype f16 --padvocab
41
 
 
 
42
  llama.cpp/quantize whiterabbitneo-f16.gguf whiterabbitneo-q4_k.gguf q4_k
43
 
44
  ```
 
4
  license_link: https://huggingface.co/deepseek-ai/deepseek-coder-33b-base/blob/main/LICENSE
5
  ---
6
 
7
+ ***Currently uploading q4_k_m : Upload started at 15:13 January 13 2024. 15mbps connection. Should take 4 hour.***
 
 
 
 
 
 
 
 
 
 
 
8
 
9
  See https://huggingface.co/TheBloke/WhiteRabbitNeo-33B-v1-GGUF to see how to run.
10
 
 
19
  ```
20
 
21
  brew install gh
 
22
  gh auth login
 
23
  gh pr checkout 3633
24
 
25
  python3 llama.cpp/convert.py whiterabbitneo-hf --outfile whiterabbitneo-33b-v1-q8_0.gguf --outtype q8_0 --padvocab
26
 
 
27
 
28
+
29
+ python3 llama.cpp/convert.py whiterabbitneo-hf --outfile whiterabbitneo-f16.gguf --outtype f16 --padvocab
30
  llama.cpp/quantize whiterabbitneo-f16.gguf whiterabbitneo-q4_k.gguf q4_k
31
 
32
  ```
whiterabbitneo-33b-v1-q4_k.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b78315b5ab07dc687bc5650089ee224aa640468b39f2e5728000495a6061d874
3
+ size 19940659296
whiterabbitneo-33b-v1-q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:423633292e85c037e8c98806a9fecd22b6028a3911f57a92cc9c5796998893bb
3
+ size 35430879296