brishtiteveja
commited on
End of training
Browse files- README.md +14 -1
- adapter_model.bin +3 -0
README.md
CHANGED
@@ -110,7 +110,9 @@ special_tokens:
|
|
110 |
[<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/banglallm/banglallm-training/runs/balglallm-training-llama3.2-1b-finetuning-5Oct2024_10_00_AM-id-1)
|
111 |
# BanglaLLama-3.2-1b-bangla-alpaca-orca-instruct-v0.0.1
|
112 |
|
113 |
-
This model is a fine-tuned version of [BanglaLLM/BanglaLLama-3.2-1b-unolp-culturax-base-v0.0.1](https://huggingface.co/BanglaLLM/BanglaLLama-3.2-1b-unolp-culturax-base-v0.0.1) on
|
|
|
|
|
114 |
|
115 |
## Model description
|
116 |
|
@@ -140,6 +142,17 @@ The following hyperparameters were used during training:
|
|
140 |
- lr_scheduler_warmup_steps: 10
|
141 |
- num_epochs: 1
|
142 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
143 |
### Framework versions
|
144 |
|
145 |
- PEFT 0.12.0
|
|
|
110 |
[<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/banglallm/banglallm-training/runs/balglallm-training-llama3.2-1b-finetuning-5Oct2024_10_00_AM-id-1)
|
111 |
# BanglaLLama-3.2-1b-bangla-alpaca-orca-instruct-v0.0.1
|
112 |
|
113 |
+
This model is a fine-tuned version of [BanglaLLM/BanglaLLama-3.2-1b-unolp-culturax-base-v0.0.1](https://huggingface.co/BanglaLLM/BanglaLLama-3.2-1b-unolp-culturax-base-v0.0.1) on the None dataset.
|
114 |
+
It achieves the following results on the evaluation set:
|
115 |
+
- Loss: 0.4511
|
116 |
|
117 |
## Model description
|
118 |
|
|
|
142 |
- lr_scheduler_warmup_steps: 10
|
143 |
- num_epochs: 1
|
144 |
|
145 |
+
### Training results
|
146 |
+
|
147 |
+
| Training Loss | Epoch | Step | Validation Loss |
|
148 |
+
|:-------------:|:------:|:----:|:---------------:|
|
149 |
+
| 0.8315 | 0.0002 | 1 | 0.9496 |
|
150 |
+
| 0.4253 | 0.2501 | 1424 | 0.5142 |
|
151 |
+
| 0.487 | 0.5003 | 2848 | 0.4754 |
|
152 |
+
| 0.4383 | 0.7504 | 4272 | 0.4550 |
|
153 |
+
| 0.478 | 1.0006 | 5696 | 0.4511 |
|
154 |
+
|
155 |
+
|
156 |
### Framework versions
|
157 |
|
158 |
- PEFT 0.12.0
|
adapter_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1f95a44ce90cdd17173f748d0a8d1c261d758aceb7fd5e52075407bfdd15771d
|
3 |
+
size 1140932242
|