roleplaiapp commited on
Commit
3096dd8
·
verified ·
1 Parent(s): 624c5b7

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +45 -0
README.md ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ pipeline_tag: text-generation
5
+ tags:
6
+ - llama-cpp
7
+ - AceInstruct-72B
8
+ - gguf
9
+ - Q4_K_M
10
+ - 72b
11
+ - 4-bit
12
+ - AceInstruct
13
+ - llama-cpp
14
+ - nvidia
15
+ - code
16
+ - math
17
+ - chat
18
+ - roleplay
19
+ - text-generation
20
+ - safetensors
21
+ - nlp
22
+ - code
23
+ library_name: transformers
24
+ base_model: nvidia/AceInstruct-72B
25
+ ---
26
+
27
+ # roleplaiapp/AceInstruct-72B-Q4_K_M-GGUF
28
+
29
+ **Repo:** `roleplaiapp/AceInstruct-72B-Q4_K_M-GGUF`
30
+ **Original Model:** `AceInstruct-72B`
31
+ **Organization:** `nvidia`
32
+ **Quantized File:** `aceinstruct-72b-q4_k_m.gguf`
33
+ **Quantization:** `GGUF`
34
+ **Quantization Method:** `Q4_K_M`
35
+ **Use Imatrix:** `False`
36
+ **Split Model:** `False`
37
+
38
+ ## Overview
39
+ This is an GGUF Q4_K_M quantized version of [AceInstruct-72B](https://huggingface.co/nvidia/AceInstruct-72B).
40
+
41
+ ## Quantization By
42
+ I often have idle A100 GPUs while building/testing and training the RP app, so I put them to use quantizing models.
43
+ I hope the community finds these quantizations useful.
44
+
45
+ Andrew Webby @ [RolePlai](https://roleplai.app/)