roleplaiapp commited on
Commit
3517c6d
·
verified ·
1 Parent(s): f18e1b6

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +53 -0
README.md ADDED
@@ -0,0 +1,53 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ language:
4
+ - en
5
+ - fr
6
+ - it
7
+ - pt
8
+ - hi
9
+ - es
10
+ - th
11
+ - de
12
+ base_model:
13
+ - meta-llama/Llama-3.1-70B
14
+ tags:
15
+ - llama-cpp
16
+ - Llama-3.3-70B-Instruct
17
+ - gguf
18
+ - llama
19
+ - 70b
20
+ - Q6_K
21
+ - llama-cpp
22
+ - gguf
23
+ - meta-llama
24
+ - code
25
+ - math
26
+ - chat
27
+ - roleplay
28
+ - text-generation
29
+ - safetensors
30
+ - nlp
31
+ - code
32
+ pipeline_tag: text-generation
33
+ ---
34
+
35
+ # Llama-3.3-70B-Instruct-Q6_K-GGUF
36
+
37
+ **Repo:** `roleplaiapp/Llama-3.3-70B-Instruct-Q6_K-GGUF`
38
+ **Original Model:** `Llama-3.3-70B-Instruct`
39
+ **Organization:** `meta-llama`
40
+ **Quantized File:** `llama-3.3-70b-instruct-q6_k.gguf`
41
+ **Quantization:** `GGUF`
42
+ **Quantization Method:** `Q6_K`
43
+ **Use Imatrix:** `False`
44
+ **Split Model:** `True`
45
+
46
+ ## Overview
47
+ This is an GGUF Q6_K quantized version of [Llama-3.3-70B-Instruct](https://huggingface.co/meta-llama/Llama-3.3-70B-Instruct).
48
+
49
+ ## Quantization By
50
+ I often have idle A100 GPUs while building/testing and training the RP app, so I put them to use quantizing models.
51
+ I hope the community finds these quantizations useful.
52
+
53
+ Andrew Webby @ [RolePlai](https://roleplai.app/)