OpenSourceRonin
commited on
Upload model Mistral-Large-Instruct-2407-v16-k65536-256-woft
Browse files
README.md
ADDED
@@ -0,0 +1,46 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
|
2 |
+
---
|
3 |
+
license: other
|
4 |
+
license_name: mrl
|
5 |
+
license_link: https://mistral.ai/licenses/MRL-0.1.md
|
6 |
+
base_model:
|
7 |
+
- mistralai/Mistral-Large-Instruct-2407
|
8 |
+
base_model_relation: quantized
|
9 |
+
tags:
|
10 |
+
- VPTQ
|
11 |
+
- Quantized
|
12 |
+
- Quantization
|
13 |
+
---
|
14 |
+
**Disclaimer**:
|
15 |
+
|
16 |
+
The model is reproduced based on the paper *VPTQ: Extreme Low-bit Vector Post-Training Quantization for Large Language Models* [github](https://github.com/microsoft/vptq) and [arXiv](https://arxiv.org/abs/2409.17066)
|
17 |
+
|
18 |
+
The model itself is sourced from a community release.
|
19 |
+
|
20 |
+
It is intended only for experimental purposes.
|
21 |
+
|
22 |
+
Users are responsible for any consequences arising from the use of this model.
|
23 |
+
|
24 |
+
**Note**:
|
25 |
+
|
26 |
+
The PPL test results are for reference only and were collected using GPTQ testing script.
|
27 |
+
|
28 |
+
```json
|
29 |
+
{
|
30 |
+
"ctx_2048": {
|
31 |
+
"wikitext2": 16.465784072875977,
|
32 |
+
"c4": 23.362998962402344,
|
33 |
+
"c4-new": 28.253490447998047
|
34 |
+
},
|
35 |
+
"ctx_4096": {
|
36 |
+
"wikitext2": 14.906161308288574,
|
37 |
+
"c4": 22.091983795166016,
|
38 |
+
"c4-new": 26.17787742614746
|
39 |
+
},
|
40 |
+
"ctx_8192": {
|
41 |
+
"wikitext2": 14.027444839477539,
|
42 |
+
"c4": 15.35131549835205,
|
43 |
+
"c4-new": 25.94689178466797
|
44 |
+
}
|
45 |
+
}
|
46 |
+
```
|