Ivan444410 commited on
Commit
0b2713f
·
verified ·
1 Parent(s): 4f84a13

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -13,7 +13,7 @@ mathPiX and then trained an LLM to match the metadata of a document with the con
13
 
14
 
15
  1) Download LLamaFactory (I recommend on this point - https://github.com/hiyouga/LLaMA-Factory/tree/36039b0fe01c17ae30dba60e247d7ba8a1beb20a).
16
- 2) Save in data folder your dataset, update dataset_info.
17
  3) Upload this model.
18
  4) Run
19
- python3 -u LLaMA-Factory/src/train.py --stage sft --model_name_or_path DeepKarkhanis/Mistral-Passthrough-8L-10B --adapter_name_or_path way_to_this_midel --finetuning_type lora --template default --dataset_dir LLaMA-Factory/data --eval_dataset dataset_name --cutoff_len 10000 --max_samples 100000 --per_device_eval_batch_size 1 --predict_with_generate True --max_new_tokens 8000 --top_p 0.7 --temperature 0.95 --output_dir output_dir --do_predict True
 
13
 
14
 
15
  1) Download LLamaFactory (I recommend on this point - https://github.com/hiyouga/LLaMA-Factory/tree/36039b0fe01c17ae30dba60e247d7ba8a1beb20a).
16
+ 2) Save in data folder your dataset, update dataset_info (check README.md and data/README.md).
17
  3) Upload this model.
18
  4) Run
19
+ python3 -u LLaMA-Factory/src/train.py --stage sft --model_name_or_path DeepKarkhanis/Mistral-Passthrough-8L-10B --adapter_name_or_path {way_to_this_midel} --finetuning_type lora --template default --dataset_dir LLaMA-Factory/data --eval_dataset {your_dataset_name} --cutoff_len 16000 --max_samples 100000 --per_device_eval_batch_size 1 --predict_with_generate True --max_new_tokens 16000 --top_p 0.7 --temperature 0.95 --output_dir {your_output_dir} --do_predict True