wilmerhenao
commited on
Commit
·
fb7211b
1
Parent(s):
2bb239d
Introducing Olinguito: A Language Model Fine-tuned with LORA Algorithm on Alpaca-cleaned Data
Browse filesThis commit adds Olinguito, a new language model derived from Dolly, which has been fine-tuned using the LORA (Low-Rank Adaptation of Large Language Models ) algorithm. Olinguito's training data has undergone meticulous cleaning, specifically extracted from Alpaca sources. By applying LORA, Olinguito aims to provide enhanced performance, accuracy, and robustness in natural language processing tasks. This commit lays the foundation for incorporating Olinguito into our Hugging Face repository, enabling users to access and utilize this refined language model for various applications.
- README.md +42 -1
- adapter_config.json +20 -0
- adapter_model.bin +3 -0
README.md
CHANGED
@@ -1,3 +1,44 @@
|
|
1 |
---
|
2 |
-
|
3 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
+
library_name: peft
|
3 |
---
|
4 |
+
## Training procedure
|
5 |
+
|
6 |
+
|
7 |
+
The following `bitsandbytes` quantization config was used during training:
|
8 |
+
- load_in_8bit: True
|
9 |
+
- load_in_4bit: False
|
10 |
+
- llm_int8_threshold: 6.0
|
11 |
+
- llm_int8_skip_modules: None
|
12 |
+
- llm_int8_enable_fp32_cpu_offload: False
|
13 |
+
- llm_int8_has_fp16_weight: False
|
14 |
+
- bnb_4bit_quant_type: fp4
|
15 |
+
- bnb_4bit_use_double_quant: False
|
16 |
+
- bnb_4bit_compute_dtype: float32
|
17 |
+
|
18 |
+
The following `bitsandbytes` quantization config was used during training:
|
19 |
+
- load_in_8bit: True
|
20 |
+
- load_in_4bit: False
|
21 |
+
- llm_int8_threshold: 6.0
|
22 |
+
- llm_int8_skip_modules: None
|
23 |
+
- llm_int8_enable_fp32_cpu_offload: False
|
24 |
+
- llm_int8_has_fp16_weight: False
|
25 |
+
- bnb_4bit_quant_type: fp4
|
26 |
+
- bnb_4bit_use_double_quant: False
|
27 |
+
- bnb_4bit_compute_dtype: float32
|
28 |
+
|
29 |
+
The following `bitsandbytes` quantization config was used during training:
|
30 |
+
- load_in_8bit: True
|
31 |
+
- load_in_4bit: False
|
32 |
+
- llm_int8_threshold: 6.0
|
33 |
+
- llm_int8_skip_modules: None
|
34 |
+
- llm_int8_enable_fp32_cpu_offload: False
|
35 |
+
- llm_int8_has_fp16_weight: False
|
36 |
+
- bnb_4bit_quant_type: fp4
|
37 |
+
- bnb_4bit_use_double_quant: False
|
38 |
+
- bnb_4bit_compute_dtype: float32
|
39 |
+
### Framework versions
|
40 |
+
|
41 |
+
- PEFT 0.4.0.dev0
|
42 |
+
- PEFT 0.4.0.dev0
|
43 |
+
|
44 |
+
- PEFT 0.4.0.dev0
|
adapter_config.json
ADDED
@@ -0,0 +1,20 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"base_model_name_or_path": "EleutherAI/gpt-j-6B",
|
3 |
+
"bias": "none",
|
4 |
+
"fan_in_fan_out": false,
|
5 |
+
"inference_mode": true,
|
6 |
+
"init_lora_weights": true,
|
7 |
+
"layers_pattern": null,
|
8 |
+
"layers_to_transform": null,
|
9 |
+
"lora_alpha": 16,
|
10 |
+
"lora_dropout": 0.05,
|
11 |
+
"modules_to_save": null,
|
12 |
+
"peft_type": "LORA",
|
13 |
+
"r": 4,
|
14 |
+
"revision": null,
|
15 |
+
"target_modules": [
|
16 |
+
"q_proj",
|
17 |
+
"v_proj"
|
18 |
+
],
|
19 |
+
"task_type": "CAUSAL_LM"
|
20 |
+
}
|
adapter_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2f486b4780c854cb07b402669a80fff69bf52d382be990154f4019ae53a13a25
|
3 |
+
size 7379597
|