Wanfq commited on
Commit
b91b3c5
·
1 Parent(s): f6b444f

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +49 -0
README.md ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: llama2
3
+ datasets:
4
+ - meta-math/MetaMathQA
5
+ ---
6
+
7
+ ## Model Details
8
+
9
+ metamath-llemma-34b is Q-Lora fine-tuned on the MetaMathQA datasets and based on the powerful llemma-34b model.
10
+
11
+ ## Installation
12
+
13
+ ```
14
+ pip install transformers==4.35.0
15
+ pip install torch==2.0.1
16
+ pip install sentencepiece==0.1.99
17
+ pip install tokenizers==0.13.3
18
+ pip install accelerate==0.21.0
19
+ pip install bitsandbytes==0.40.0
20
+ pip install vllm
21
+ pip install fraction
22
+ pip install protobuf
23
+ ```
24
+
25
+ ## Model Usage
26
+
27
+ prompting template:
28
+
29
+ '''
30
+
31
+ "Below is an instruction that describes a task. "
32
+ "Write a response that appropriately completes the request.\n\n"
33
+ "### Instruction:\n{instruction}\n\n### Response: Let's think step by step."
34
+
35
+ '''
36
+
37
+ where you need to use your query question to replace the {instruction}
38
+
39
+ ## Experiments
40
+
41
+ | Model | GSM8k Pass@1 | MATH Pass@1 |
42
+ |---------------------|--------------|-------------|
43
+ | MetaMath-7B | 66.5 | 19.8 |
44
+ | MetaMath-13B | 72.3 | 22.4 |
45
+ | MetaMath-Llemma-7B | 69.2 | 30.0 |
46
+ | MetaMath-Mistral-7B | **77.7** | 28.2 |
47
+ | 🔥 metamath-llemma-34B | 76.1 | **31.4** |
48
+
49
+ *Model card adapted from [MetaMath-70B-V1.0](https://huggingface.co/meta-math/MetaMath-70B-V1.0/blob/main/README.md)*