Simon Park
commited on
Commit
·
e276010
1
Parent(s):
f2ce313
Updated Model Card
Browse files
README.md
CHANGED
@@ -1,3 +1,73 @@
|
|
1 |
---
|
|
|
|
|
|
|
2 |
license: llama3
|
3 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
+
base_model:
|
3 |
+
- meta-llama/Meta-Llama-3-8B
|
4 |
+
library_name: transformers
|
5 |
license: llama3
|
6 |
---
|
7 |
+
|
8 |
+
# Model Card for Llama-3-8B-Instruct-SkillMix
|
9 |
+
This model was SFT-ed from [meta-llama/Meta-Llama-3-8B](https://huggingface.co/meta-llama/Meta-Llama-3-8B) with data generated by the Seed-Dataset Agnostic version of the Instruct-SkillMix pipeline.
|
10 |
+
|
11 |
+
## Training Details
|
12 |
+
We used 4000 examples from Instruct-SkillMix-SDA(k=2) (data available at [PrincetonPLI/Instruct-SkillMix-SDA](https://huggingface.co/datasets/PrincetonPLI/Instruct-SkillMix-SDA/blob/main/data/ism_sda_k2_4K.json)).
|
13 |
+
|
14 |
+
- LR: 2e-5
|
15 |
+
- Linear Warmup Ratio: 0.03
|
16 |
+
- Decay: Cosine Decay to 0
|
17 |
+
- Batch Size: 128
|
18 |
+
- epoch: 7 / 15
|
19 |
+
- Optimizer: AdamW
|
20 |
+
- Sequence Length: 1024
|
21 |
+
|
22 |
+
## Evaluation Details
|
23 |
+
We provide the set of generation configuration used for evaluation.
|
24 |
+
|
25 |
+
### AlpacaEval
|
26 |
+
- model_kwargs:
|
27 |
+
- torch_dtype: 'bfloat16'
|
28 |
+
- max_new_tokens: 2048
|
29 |
+
- temperature: 0.9
|
30 |
+
- top_p: 1.0
|
31 |
+
- do_sample: True
|
32 |
+
- stop_token_ids:
|
33 |
+
- 128001
|
34 |
+
- 128009
|
35 |
+
|
36 |
+
### MTBench
|
37 |
+
- model_kwargs:
|
38 |
+
- torch_dtype: 'bfloat16'
|
39 |
+
- max_new_tokens: 1024
|
40 |
+
- temperature: 0.7
|
41 |
+
- stop_token_ids:
|
42 |
+
- 128001
|
43 |
+
- 128009
|
44 |
+
|
45 |
+
### WildBench
|
46 |
+
- model_kwargs:
|
47 |
+
- torch_dtype: 'bfloat16'
|
48 |
+
- max_new_tokens: 4096
|
49 |
+
- temperature: 0.9
|
50 |
+
- top_p: 1.0
|
51 |
+
- do_sample: True
|
52 |
+
- stop_token_ids:
|
53 |
+
- 128001
|
54 |
+
- 128009
|
55 |
+
|
56 |
+
## Citation
|
57 |
+
Paper: [Instruct-SkillMix](https://www.arxiv.org/abs/2408.14774)
|
58 |
+
```
|
59 |
+
@misc{kaur2024instructskillmixpowerfulpipelinellm,
|
60 |
+
title={Instruct-SkillMix: A Powerful Pipeline for LLM Instruction Tuning},
|
61 |
+
author={Simran Kaur and Simon Park and Anirudh Goyal and Sanjeev Arora},
|
62 |
+
year={2024},
|
63 |
+
eprint={2408.14774},
|
64 |
+
archivePrefix={arXiv},
|
65 |
+
primaryClass={cs.LG},
|
66 |
+
url={https://arxiv.org/abs/2408.14774},
|
67 |
+
}
|
68 |
+
```
|
69 |
+
|
70 |
+
## Contact
|
71 |
+
Simran Kaur, Princeton University
|
72 |
+
Simon Park, Princeton University
|
73 |
+
{skaur, juhyunp} 'at' princeton 'dot' edu
|