Triangle104 commited on
Commit
ce44941
·
verified ·
1 Parent(s): 38f39a9

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +45 -0
README.md CHANGED
@@ -14,6 +14,51 @@ language:
14
  This model was converted to GGUF format from [`allura-org/MN-12b-RP-Ink`](https://huggingface.co/allura-org/MN-12b-RP-Ink) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
15
  Refer to the [original model card](https://huggingface.co/allura-org/MN-12b-RP-Ink) for more details on the model.
16
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
17
  ## Use with llama.cpp
18
  Install llama.cpp through brew (works on Mac and Linux)
19
 
 
14
  This model was converted to GGUF format from [`allura-org/MN-12b-RP-Ink`](https://huggingface.co/allura-org/MN-12b-RP-Ink) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
15
  Refer to the [original model card](https://huggingface.co/allura-org/MN-12b-RP-Ink) for more details on the model.
16
 
17
+ ---
18
+ Model details:
19
+ -
20
+ A roleplay-focused LoRA finetune of Mistral Nemo Instruct. Methodology and hyperparams inspired by SorcererLM and Slush.
21
+ Renamed to Ink to distinguish from [insert every other rp tune ever], but it's the same data as was used in the Teleut RP model.
22
+ Dataset
23
+
24
+ The worst mix of data you've ever seen. Like, seriously, you do not want to see the things that went into this model. It's bad.
25
+
26
+ "this is like washing down an adderall with a bottle of methylated rotgut" - inflatebot
27
+ Quants
28
+
29
+ Static GGUFs
30
+
31
+ Recommended Settings
32
+
33
+ Chat template: Mistral v3-Tekken
34
+ Recommended samplers (not the be-all-end-all, try some on your own!):
35
+
36
+ Temp 1.25 / MinP 0.1
37
+ Temp 1.03 / TopK 200 / MinP 0.05 / TopA 0.2
38
+
39
+ Hyperparams
40
+ General
41
+
42
+ Epochs = 2
43
+ LR = 6e-5
44
+ LR Scheduler = Cosine
45
+ Optimizer = Paged AdamW 8bit
46
+ Effective batch size = 12
47
+
48
+ LoRA
49
+
50
+ Rank = 16
51
+ Alpha = 32
52
+ Dropout = 0.25 (Inspiration: Slush)
53
+
54
+ Credits
55
+
56
+ Humongous thanks to the people who created the data. I would credit you all, but that would be cheating ;)
57
+ Big thanks to all Allura members, especially Toasty, for testing and emotional support ilya /platonic
58
+ Also special thanks to Bot for making the model card image here :3
59
+ NO thanks to Infermatic. They suck at hosting models
60
+
61
+ ---
62
  ## Use with llama.cpp
63
  Install llama.cpp through brew (works on Mac and Linux)
64