andythetechnerd03 commited on
Commit
eb9e559
·
verified ·
1 Parent(s): 463733c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -11,7 +11,7 @@ Here's a little example:
11
  ``` python
12
  from transformers import AutoTokenizer, AutoModelForCausalLM
13
 
14
- model_name = "pphuc25/VistralPoem5"
15
  tokenizer = AutoTokenizer.from_pretrained(model_name, device_map="auto")
16
  model = AutoModelForCausalLM.from_pretrained(model_name, device_map="auto")
17
 
@@ -66,7 +66,7 @@ This is not easy. Such data that takes the input as a long text (newspaper artic
66
 
67
  After all, we have about 72,101 samples with a ratio of 0.05 (68495 on the train set and 3606 on the test set)
68
 
69
- We published the dataset at [here](https://huggingface.co/datasets/pphuc25/poem-5-words-vietnamese)
70
 
71
  ### Custom Evaluation Data
72
  As part of the final evaluation for benchmark, we gathered around 27 Vietnamese children's stories and divided into many samples, accumulating to 118 samples. The dataset can be found [here](/data/eval_set.json)
@@ -94,7 +94,7 @@ As mentioned earlier, we use [Vistral-7B-Chat](https://huggingface.co/Viet-Mistr
94
  - Optimizer: Adamw bnb 8bit
95
  - Sequence Len: 1096
96
 
97
- The weights can be found [here](https://huggingface.co/pphuc25/poem-vistral)
98
 
99
  The notebook for training can be found at `notebook/Fine_tune_LLMs_with_Axolotl.ipynb`
100
 
 
11
  ``` python
12
  from transformers import AutoTokenizer, AutoModelForCausalLM
13
 
14
+ model_name = "andythetechnerd03/VistralPoem5"
15
  tokenizer = AutoTokenizer.from_pretrained(model_name, device_map="auto")
16
  model = AutoModelForCausalLM.from_pretrained(model_name, device_map="auto")
17
 
 
66
 
67
  After all, we have about 72,101 samples with a ratio of 0.05 (68495 on the train set and 3606 on the test set)
68
 
69
+ We published the dataset at [here](https://huggingface.co/datasets/andythetechnerd03/Vietnamese-Poem-5words)
70
 
71
  ### Custom Evaluation Data
72
  As part of the final evaluation for benchmark, we gathered around 27 Vietnamese children's stories and divided into many samples, accumulating to 118 samples. The dataset can be found [here](/data/eval_set.json)
 
94
  - Optimizer: Adamw bnb 8bit
95
  - Sequence Len: 1096
96
 
97
+ The weights can be found [here](https://huggingface.co/andythetechnerd03/VistralPoem5)
98
 
99
  The notebook for training can be found at `notebook/Fine_tune_LLMs_with_Axolotl.ipynb`
100