Samsul Rahmadani commited on
Commit
b50a7e8
·
1 Parent(s): 121b9d2

add readme

Browse files
Files changed (2) hide show
  1. .gitattributes +0 -1
  2. README.md +37 -3
.gitattributes CHANGED
@@ -40,4 +40,3 @@ generation_config.json filter=lfs diff=lfs merge=lfs -text
40
  special_tokens_map.json filter=lfs diff=lfs merge=lfs -text
41
  tokenizer.json filter=lfs diff=lfs merge=lfs -text
42
  tokenizer_config.json filter=lfs diff=lfs merge=lfs -text
43
- README.md filter=lfs diff=lfs merge=lfs -text
 
40
  special_tokens_map.json filter=lfs diff=lfs merge=lfs -text
41
  tokenizer.json filter=lfs diff=lfs merge=lfs -text
42
  tokenizer_config.json filter=lfs diff=lfs merge=lfs -text
 
README.md CHANGED
@@ -1,3 +1,37 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:bec9f2cd81f7396a64ee474145117d81139906443a21480cbe4993c97cb499aa
3
- size 5238
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: bigscience-bloom-rail-1.0
3
+ ---
4
+ # BahasaGPT-1 Fine-Tuning Documentation Summary
5
+
6
+ ## Introduction
7
+
8
+ This document provides an overview of the BahasaGPT-1 model, which is a fine-tuned model for a specific task in the Indonesian language. The model is based on the Bloomz-7B-mt architecture and is fine-tuned using a dataset of over 70,000 Indonesian instructions.
9
+
10
+ ## Model Details
11
+
12
+ **Model Name:** BahasaGPT-1
13
+
14
+ **Model Source:** Bloomz-7B-mt
15
+
16
+ **Dataset for Fine-Tuning:** Over 70k Indonesia Instruct Dataset generated using the Alpaca method from the following sources:
17
+
18
+ - [Stanford Alpaca](https://github.com/tatsu-lab/stanford_alpaca)
19
+ - Translated instructions from OA ([Anh/data at main · LAION-AI/Anh](https://github.com/LAION-AI/Anh))
20
+
21
+ ## Fine-Tuning Process
22
+
23
+ The BahasaGPT-1 model was fine-tuned using a dataset of over 70,000 Indonesian instructions, which were generated using the Alpaca method from Stanford and translated instructions from OA. This combination of datasets allowed the model to be better adapted to the specific needs of Indonesian language tasks.
24
+
25
+ The fine-tuning process involved adjusting the model's weights and biases based on the input dataset. This was done iteratively to optimize the model's performance for the specific task in the Indonesian language.
26
+
27
+ ## Known Limitations
28
+
29
+ Despite the successful fine-tuning, the BahasaGPT-1 model still has some limitations:
30
+
31
+ 1. **Hallucination:** The model sometimes generates outputs that may seem plausible but are not based on the input data. This may lead to incorrect or nonsensical responses in some cases.
32
+
33
+ 2. **Repeated Tokens:** The model occasionally produces repeated tokens in the output, which may affect the overall coherence and readability of the generated text.
34
+
35
+ ## Conclusion
36
+
37
+ The BahasaGPT-1 model is a fine-tuned language model for Indonesian language tasks, based on the Bloomz-7B-mt architecture. The model was trained on a dataset of over 70,000 Indonesian instructions generated using the Alpaca method and translated instructions from OA. Despite some limitations, such as occasional hallucination and repeated tokens, the model provides a valuable tool for working with Indonesian language tasks.