Update README.md
Browse files
README.md
CHANGED
@@ -3,6 +3,7 @@ license: apache-2.0
|
|
3 |
datasets:
|
4 |
- SkunkworksAI/reasoning-0.01
|
5 |
- trollek/ThoughtfulAssistant-v01
|
|
|
6 |
base_model: trollek/LittleInstructionMaker-4B-v0.2
|
7 |
---
|
8 |
# ThoughtStream-4B-v0.2
|
@@ -30,10 +31,56 @@ With this second version I have tried to have 2 more ways of thinking than just
|
|
30 |
|
31 |
## Config
|
32 |
|
33 |
-
```
|
34 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
35 |
|
36 |
```
|
37 |
|
38 |
-
## Training results
|
39 |
-
|
|
|
3 |
datasets:
|
4 |
- SkunkworksAI/reasoning-0.01
|
5 |
- trollek/ThoughtfulAssistant-v01
|
6 |
+
- trollek/ThoughtfulAssistant-v02
|
7 |
base_model: trollek/LittleInstructionMaker-4B-v0.2
|
8 |
---
|
9 |
# ThoughtStream-4B-v0.2
|
|
|
31 |
|
32 |
## Config
|
33 |
|
34 |
+
```yaml
|
35 |
+
### model
|
36 |
+
model_name_or_path: lim-v02-thought
|
37 |
+
|
38 |
+
### method
|
39 |
+
stage: sft
|
40 |
+
do_train: true
|
41 |
+
finetuning_type: lora
|
42 |
+
lora_target: all
|
43 |
+
loraplus_lr_ratio: 12.0
|
44 |
+
lora_rank: 16
|
45 |
+
lora_alpha: 16
|
46 |
+
use_unsloth: true
|
47 |
+
quantization_bit: 4
|
48 |
+
upcast_layernorm: true
|
49 |
+
seed: 127
|
50 |
+
optim: lion_8bit
|
51 |
+
additional_target: embed_tokens
|
52 |
+
|
53 |
+
### dataset
|
54 |
+
dataset: reasoning_assistant,thoughtful_v01,thoughtful_v02
|
55 |
+
template: ninja_chatml
|
56 |
+
cutoff_len: 8192
|
57 |
+
overwrite_cache: false
|
58 |
+
preprocessing_num_workers: 12
|
59 |
+
|
60 |
+
### output
|
61 |
+
output_dir: /home/trolle/Documents/Projects/trollek/danube3/merges/lim-v02-thought/loras/reasoning
|
62 |
+
logging_steps: 5
|
63 |
+
save_steps: 1
|
64 |
+
save_strategy: epoch
|
65 |
+
plot_loss: true
|
66 |
+
overwrite_output_dir: false
|
67 |
+
|
68 |
+
### train
|
69 |
+
per_device_train_batch_size: 2
|
70 |
+
gradient_accumulation_steps: 4
|
71 |
+
learning_rate: 0.000002
|
72 |
+
num_train_epochs: 2
|
73 |
+
lr_scheduler_type: constant_with_warmup
|
74 |
+
warmup_ratio: 0.01
|
75 |
+
bf16: true
|
76 |
+
flash_attn: fa2
|
77 |
+
|
78 |
+
### eval
|
79 |
+
val_size: 0.01
|
80 |
+
per_device_eval_batch_size: 1
|
81 |
+
eval_strategy: steps
|
82 |
+
eval_steps: 1000
|
83 |
|
84 |
```
|
85 |
|
86 |
+
## Training results
|
|