Hawoly18 commited on
Commit
1dfedbe
·
verified ·
1 Parent(s): 23cca59

End of training

Browse files
README.md CHANGED
@@ -1,7 +1,7 @@
1
  ---
2
- base_model: unsloth/mistral-7b-instruct-v0.3-bnb-4bit
3
  library_name: peft
4
- license: apache-2.0
5
  tags:
6
  - trl
7
  - sft
@@ -17,7 +17,7 @@ should probably proofread and complete it, then remove this comment. -->
17
 
18
  # outputs
19
 
20
- This model is a fine-tuned version of [unsloth/mistral-7b-instruct-v0.3-bnb-4bit](https://huggingface.co/unsloth/mistral-7b-instruct-v0.3-bnb-4bit) on the None dataset.
21
 
22
  ## Model description
23
 
@@ -54,8 +54,8 @@ The following hyperparameters were used during training:
54
 
55
  ### Framework versions
56
 
57
- - PEFT 0.12.0
58
- - Transformers 4.44.0
59
- - Pytorch 2.3.1+cu121
60
- - Datasets 2.20.0
61
  - Tokenizers 0.19.1
 
1
  ---
2
+ base_model: unsloth/meta-llama-3.1-8b-bnb-4bit
3
  library_name: peft
4
+ license: llama3.1
5
  tags:
6
  - trl
7
  - sft
 
17
 
18
  # outputs
19
 
20
+ This model is a fine-tuned version of [unsloth/meta-llama-3.1-8b-bnb-4bit](https://huggingface.co/unsloth/meta-llama-3.1-8b-bnb-4bit) on the None dataset.
21
 
22
  ## Model description
23
 
 
54
 
55
  ### Framework versions
56
 
57
+ - PEFT 0.13.3.dev0
58
+ - Transformers 4.44.2
59
+ - Pytorch 2.4.1+cu121
60
+ - Datasets 3.0.1
61
  - Tokenizers 0.19.1
adapter_config.json CHANGED
@@ -1,8 +1,9 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "unsloth/mistral-7b-instruct-v0.3-bnb-4bit",
5
  "bias": "none",
 
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
8
  "init_lora_weights": true,
@@ -20,13 +21,13 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "up_proj",
24
- "q_proj",
25
  "k_proj",
 
26
  "v_proj",
27
- "down_proj",
28
  "gate_proj",
29
- "o_proj"
 
 
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "unsloth/meta-llama-3.1-8b-bnb-4bit",
5
  "bias": "none",
6
+ "exclude_modules": null,
7
  "fan_in_fan_out": false,
8
  "inference_mode": true,
9
  "init_lora_weights": true,
 
21
  "rank_pattern": {},
22
  "revision": null,
23
  "target_modules": [
 
 
24
  "k_proj",
25
+ "o_proj",
26
  "v_proj",
 
27
  "gate_proj",
28
+ "down_proj",
29
+ "q_proj",
30
+ "up_proj"
31
  ],
32
  "task_type": "CAUSAL_LM",
33
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d44550386a74be584d68fb7a912cb3a10ec39f6562f84aeb643f52f43815e8b5
3
  size 167832240
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a99194740e22320bcc9dc66b656c6e73f06196d51802f5def4098520ca9809ac
3
  size 167832240
runs/Oct15_11-15-10_f6a5a5cff894/events.out.tfevents.1728990944.f6a5a5cff894.279.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:05aba6f86c7550098306b91807015a00b9fb4670968b65b2ab00d433bec76e4c
3
+ size 18635
special_tokens_map.json CHANGED
@@ -1,27 +1,20 @@
1
  {
2
  "bos_token": {
3
- "content": "<s>",
4
  "lstrip": false,
5
  "normalized": false,
6
  "rstrip": false,
7
  "single_word": false
8
  },
9
  "eos_token": {
10
- "content": "</s>",
11
  "lstrip": false,
12
  "normalized": false,
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
  "pad_token": {
17
- "content": "[control_768]",
18
- "lstrip": false,
19
- "normalized": false,
20
- "rstrip": false,
21
- "single_word": false
22
- },
23
- "unk_token": {
24
- "content": "<unk>",
25
  "lstrip": false,
26
  "normalized": false,
27
  "rstrip": false,
 
1
  {
2
  "bos_token": {
3
+ "content": "<|begin_of_text|>",
4
  "lstrip": false,
5
  "normalized": false,
6
  "rstrip": false,
7
  "single_word": false
8
  },
9
  "eos_token": {
10
+ "content": "<|end_of_text|>",
11
  "lstrip": false,
12
  "normalized": false,
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
  "pad_token": {
17
+ "content": "<|finetune_right_pad_id|>",
 
 
 
 
 
 
 
18
  "lstrip": false,
19
  "normalized": false,
20
  "rstrip": false,
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
The diff for this file is too large to render. See raw diff
 
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cfb8b1aa7694fe8996b9dc9fe84059913f11683a94e4fbba5aba2c7f9200071b
3
- size 5176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ff5ca8de5d04ad67adeb9cab3f0fc38d6d92854b7c8d3951ca847d16714f5eaf
3
+ size 5432