Simon Park commited on
Commit
e276010
·
1 Parent(s): f2ce313

Updated Model Card

Browse files
Files changed (1) hide show
  1. README.md +70 -0
README.md CHANGED
@@ -1,3 +1,73 @@
1
  ---
 
 
 
2
  license: llama3
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ base_model:
3
+ - meta-llama/Meta-Llama-3-8B
4
+ library_name: transformers
5
  license: llama3
6
  ---
7
+
8
+ # Model Card for Llama-3-8B-Instruct-SkillMix
9
+ This model was SFT-ed from [meta-llama/Meta-Llama-3-8B](https://huggingface.co/meta-llama/Meta-Llama-3-8B) with data generated by the Seed-Dataset Agnostic version of the Instruct-SkillMix pipeline.
10
+
11
+ ## Training Details
12
+ We used 4000 examples from Instruct-SkillMix-SDA(k=2) (data available at [PrincetonPLI/Instruct-SkillMix-SDA](https://huggingface.co/datasets/PrincetonPLI/Instruct-SkillMix-SDA/blob/main/data/ism_sda_k2_4K.json)).
13
+
14
+ - LR: 2e-5
15
+ - Linear Warmup Ratio: 0.03
16
+ - Decay: Cosine Decay to 0
17
+ - Batch Size: 128
18
+ - epoch: 7 / 15
19
+ - Optimizer: AdamW
20
+ - Sequence Length: 1024
21
+
22
+ ## Evaluation Details
23
+ We provide the set of generation configuration used for evaluation.
24
+
25
+ ### AlpacaEval
26
+ - model_kwargs:
27
+ - torch_dtype: 'bfloat16'
28
+ - max_new_tokens: 2048
29
+ - temperature: 0.9
30
+ - top_p: 1.0
31
+ - do_sample: True
32
+ - stop_token_ids:
33
+ - 128001
34
+ - 128009
35
+
36
+ ### MTBench
37
+ - model_kwargs:
38
+ - torch_dtype: 'bfloat16'
39
+ - max_new_tokens: 1024
40
+ - temperature: 0.7
41
+ - stop_token_ids:
42
+ - 128001
43
+ - 128009
44
+
45
+ ### WildBench
46
+ - model_kwargs:
47
+ - torch_dtype: 'bfloat16'
48
+ - max_new_tokens: 4096
49
+ - temperature: 0.9
50
+ - top_p: 1.0
51
+ - do_sample: True
52
+ - stop_token_ids:
53
+ - 128001
54
+ - 128009
55
+
56
+ ## Citation
57
+ Paper: [Instruct-SkillMix](https://www.arxiv.org/abs/2408.14774)
58
+ ```
59
+ @misc{kaur2024instructskillmixpowerfulpipelinellm,
60
+ title={Instruct-SkillMix: A Powerful Pipeline for LLM Instruction Tuning},
61
+ author={Simran Kaur and Simon Park and Anirudh Goyal and Sanjeev Arora},
62
+ year={2024},
63
+ eprint={2408.14774},
64
+ archivePrefix={arXiv},
65
+ primaryClass={cs.LG},
66
+ url={https://arxiv.org/abs/2408.14774},
67
+ }
68
+ ```
69
+
70
+ ## Contact
71
+ Simran Kaur, Princeton University
72
+ Simon Park, Princeton University
73
+ {skaur, juhyunp} 'at' princeton 'dot' edu