andythetechnerd03
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -11,7 +11,7 @@ Here's a little example:
|
|
11 |
``` python
|
12 |
from transformers import AutoTokenizer, AutoModelForCausalLM
|
13 |
|
14 |
-
model_name = "
|
15 |
tokenizer = AutoTokenizer.from_pretrained(model_name, device_map="auto")
|
16 |
model = AutoModelForCausalLM.from_pretrained(model_name, device_map="auto")
|
17 |
|
@@ -66,7 +66,7 @@ This is not easy. Such data that takes the input as a long text (newspaper artic
|
|
66 |
|
67 |
After all, we have about 72,101 samples with a ratio of 0.05 (68495 on the train set and 3606 on the test set)
|
68 |
|
69 |
-
We published the dataset at [here](https://huggingface.co/datasets/
|
70 |
|
71 |
### Custom Evaluation Data
|
72 |
As part of the final evaluation for benchmark, we gathered around 27 Vietnamese children's stories and divided into many samples, accumulating to 118 samples. The dataset can be found [here](/data/eval_set.json)
|
@@ -94,7 +94,7 @@ As mentioned earlier, we use [Vistral-7B-Chat](https://huggingface.co/Viet-Mistr
|
|
94 |
- Optimizer: Adamw bnb 8bit
|
95 |
- Sequence Len: 1096
|
96 |
|
97 |
-
The weights can be found [here](https://huggingface.co/
|
98 |
|
99 |
The notebook for training can be found at `notebook/Fine_tune_LLMs_with_Axolotl.ipynb`
|
100 |
|
|
|
11 |
``` python
|
12 |
from transformers import AutoTokenizer, AutoModelForCausalLM
|
13 |
|
14 |
+
model_name = "andythetechnerd03/VistralPoem5"
|
15 |
tokenizer = AutoTokenizer.from_pretrained(model_name, device_map="auto")
|
16 |
model = AutoModelForCausalLM.from_pretrained(model_name, device_map="auto")
|
17 |
|
|
|
66 |
|
67 |
After all, we have about 72,101 samples with a ratio of 0.05 (68495 on the train set and 3606 on the test set)
|
68 |
|
69 |
+
We published the dataset at [here](https://huggingface.co/datasets/andythetechnerd03/Vietnamese-Poem-5words)
|
70 |
|
71 |
### Custom Evaluation Data
|
72 |
As part of the final evaluation for benchmark, we gathered around 27 Vietnamese children's stories and divided into many samples, accumulating to 118 samples. The dataset can be found [here](/data/eval_set.json)
|
|
|
94 |
- Optimizer: Adamw bnb 8bit
|
95 |
- Sequence Len: 1096
|
96 |
|
97 |
+
The weights can be found [here](https://huggingface.co/andythetechnerd03/VistralPoem5)
|
98 |
|
99 |
The notebook for training can be found at `notebook/Fine_tune_LLMs_with_Axolotl.ipynb`
|
100 |
|