InferenceIllusionist
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -8,11 +8,14 @@ license: cc-by-nc-4.0
|
|
8 |
- Model creator: [Sao10K](https://huggingface.co/Sao10K/)
|
9 |
- Original model: [Fimbulvetr-11B-v2](https://huggingface.co/Sao10K/Fimbulvetr-11B-v2)
|
10 |
|
11 |
-
All credits to Sao10K for the original model. This is just a
|
12 |
|
13 |
-
|
|
|
|
|
14 |
|
15 |
See original model card details below.
|
|
|
16 |
|
17 |
![Fox1](https://huggingface.co/Sao10K/Fimbulvetr-11B-v2/resolve/main/cute1.jpg)
|
18 |
|
|
|
8 |
- Model creator: [Sao10K](https://huggingface.co/Sao10K/)
|
9 |
- Original model: [Fimbulvetr-11B-v2](https://huggingface.co/Sao10K/Fimbulvetr-11B-v2)
|
10 |
|
11 |
+
All credits to Sao10K for the original model. This is just a quick test of the new quantization types such as IQ_3S in an attempt to further reduce VRAM requirements.
|
12 |
|
13 |
+
Looking for Q3/Q4/Q5 quants? See the link in the model card below.
|
14 |
+
|
15 |
+
Quantized from fp16 with love. Importance matrix file [Fimbulvetr-11B-v2-imatrix.dat](https://huggingface.co/InferenceIllusionist/Fimbulvetr-11B-v2-iMat-GGUF/blob/main/Fimbulvetr-11B-v2-imatrix.dat) was calculated using Q8_0.
|
16 |
|
17 |
See original model card details below.
|
18 |
+
---
|
19 |
|
20 |
![Fox1](https://huggingface.co/Sao10K/Fimbulvetr-11B-v2/resolve/main/cute1.jpg)
|
21 |
|