Safetensors
English
llama
trollek commited on
Commit
b5eb743
·
verified ·
1 Parent(s): 14bd26c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +51 -4
README.md CHANGED
@@ -3,6 +3,7 @@ license: apache-2.0
3
  datasets:
4
  - SkunkworksAI/reasoning-0.01
5
  - trollek/ThoughtfulAssistant-v01
 
6
  base_model: trollek/LittleInstructionMaker-4B-v0.2
7
  ---
8
  # ThoughtStream-4B-v0.2
@@ -30,10 +31,56 @@ With this second version I have tried to have 2 more ways of thinking than just
30
 
31
  ## Config
32
 
33
- ```jinja
34
-
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
35
 
36
  ```
37
 
38
- ## Training results
39
-
 
3
  datasets:
4
  - SkunkworksAI/reasoning-0.01
5
  - trollek/ThoughtfulAssistant-v01
6
+ - trollek/ThoughtfulAssistant-v02
7
  base_model: trollek/LittleInstructionMaker-4B-v0.2
8
  ---
9
  # ThoughtStream-4B-v0.2
 
31
 
32
  ## Config
33
 
34
+ ```yaml
35
+ ### model
36
+ model_name_or_path: lim-v02-thought
37
+
38
+ ### method
39
+ stage: sft
40
+ do_train: true
41
+ finetuning_type: lora
42
+ lora_target: all
43
+ loraplus_lr_ratio: 12.0
44
+ lora_rank: 16
45
+ lora_alpha: 16
46
+ use_unsloth: true
47
+ quantization_bit: 4
48
+ upcast_layernorm: true
49
+ seed: 127
50
+ optim: lion_8bit
51
+ additional_target: embed_tokens
52
+
53
+ ### dataset
54
+ dataset: reasoning_assistant,thoughtful_v01,thoughtful_v02
55
+ template: ninja_chatml
56
+ cutoff_len: 8192
57
+ overwrite_cache: false
58
+ preprocessing_num_workers: 12
59
+
60
+ ### output
61
+ output_dir: /home/trolle/Documents/Projects/trollek/danube3/merges/lim-v02-thought/loras/reasoning
62
+ logging_steps: 5
63
+ save_steps: 1
64
+ save_strategy: epoch
65
+ plot_loss: true
66
+ overwrite_output_dir: false
67
+
68
+ ### train
69
+ per_device_train_batch_size: 2
70
+ gradient_accumulation_steps: 4
71
+ learning_rate: 0.000002
72
+ num_train_epochs: 2
73
+ lr_scheduler_type: constant_with_warmup
74
+ warmup_ratio: 0.01
75
+ bf16: true
76
+ flash_attn: fa2
77
+
78
+ ### eval
79
+ val_size: 0.01
80
+ per_device_eval_batch_size: 1
81
+ eval_strategy: steps
82
+ eval_steps: 1000
83
 
84
  ```
85
 
86
+ ## Training results