Transformers
GGUF
llama-cpp
gguf-my-repo
Inference Endpoints
conversational
Triangle104 commited on
Commit
b2d71c0
·
verified ·
1 Parent(s): dd5f7e2

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +55 -0
README.md CHANGED
@@ -15,6 +15,61 @@ tags:
15
  This model was converted to GGUF format from [`nbeerbower/mistral-nemo-kartoffel-12B`](https://huggingface.co/nbeerbower/mistral-nemo-kartoffel-12B) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
16
  Refer to the [original model card](https://huggingface.co/nbeerbower/mistral-nemo-kartoffel-12B) for more details on the model.
17
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
18
  ## Use with llama.cpp
19
  Install llama.cpp through brew (works on Mac and Linux)
20
 
 
15
  This model was converted to GGUF format from [`nbeerbower/mistral-nemo-kartoffel-12B`](https://huggingface.co/nbeerbower/mistral-nemo-kartoffel-12B) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
16
  Refer to the [original model card](https://huggingface.co/nbeerbower/mistral-nemo-kartoffel-12B) for more details on the model.
17
 
18
+ ---
19
+ Model details:
20
+ -
21
+ Mahou-1.5-mistral-nemo-12B-lorablated finetuned on various datasets.
22
+ Method
23
+
24
+ ORPO tuned with 8x A100 for 2 epochs.
25
+
26
+ QLoRA config:
27
+
28
+ # QLoRA config
29
+ bnb_config = BitsAndBytesConfig(
30
+ load_in_4bit=True,
31
+ bnb_4bit_quant_type="nf4",
32
+ bnb_4bit_compute_dtype=torch_dtype,
33
+ bnb_4bit_use_double_quant=True,
34
+ )
35
+
36
+ # LoRA config
37
+ peft_config = LoraConfig(
38
+ r=16,
39
+ lora_alpha=32,
40
+ lora_dropout=0.05,
41
+ bias="none",
42
+ task_type="CAUSAL_LM",
43
+ target_modules=['up_proj', 'down_proj', 'gate_proj', 'k_proj', 'q_proj', 'v_proj', 'o_proj']
44
+ )
45
+
46
+ Training config:
47
+
48
+ orpo_args = ORPOConfig(
49
+ run_name=new_model,
50
+ learning_rate=8e-6,
51
+ lr_scheduler_type="linear",
52
+ max_length=2048,
53
+ max_prompt_length=1024,
54
+ max_completion_length=1024,
55
+ beta=0.1,
56
+ per_device_train_batch_size=4,
57
+ per_device_eval_batch_size=4,
58
+ gradient_accumulation_steps=1,
59
+ optim="paged_adamw_8bit",
60
+ num_train_epochs=2,
61
+ evaluation_strategy="steps",
62
+ eval_steps=0.2,
63
+ logging_steps=1,
64
+ warmup_steps=10,
65
+ max_grad_norm=10,
66
+ report_to="wandb",
67
+ output_dir="./results/",
68
+ bf16=True,
69
+ gradient_checkpointing=True,
70
+ )
71
+
72
+ ---
73
  ## Use with llama.cpp
74
  Install llama.cpp through brew (works on Mac and Linux)
75