Update README.md
Browse files
README.md
CHANGED
@@ -43,7 +43,7 @@ Previous experiments have demonstrated that orca-like datasets yield substantial
|
|
43 |
| 2048 | 5.22 | 5.38 | 5.87 | 5.23 | **5.07** |
|
44 |
| 4096 | 4.90 | 5.08 | 5.50 | 4.91 | **4.77** |
|
45 |
| 8192 | **4.71** | 4.90 | 5.32 | Not Tested | 57.1 |
|
46 |
-
| 12000 |
|
47 |
|
48 |
- This model is very competitive with the Llama-1 33b extended context variants. In fact, it outperforms bhenrym14/airoboros-33b-gpt4-1.4.1-lxctx-PI-16384-fp16 everywhere <=8192 tokens.
|
49 |
- Not presented here, but this model outperforms the base llama-2-13b on MMLU-fs with a score of 58.3. If this score ends up being be replicated on the HF LLM leaderboard, **this would place this model at 2nd or 3rd overall for MMLU among 13b models (and the #1 for extended context)**
|
|
|
43 |
| 2048 | 5.22 | 5.38 | 5.87 | 5.23 | **5.07** |
|
44 |
| 4096 | 4.90 | 5.08 | 5.50 | 4.91 | **4.77** |
|
45 |
| 8192 | **4.71** | 4.90 | 5.32 | Not Tested | 57.1 |
|
46 |
+
| 12000 | 55 | **4.82** | 56.1 | Not Tested | Not Tested |
|
47 |
|
48 |
- This model is very competitive with the Llama-1 33b extended context variants. In fact, it outperforms bhenrym14/airoboros-33b-gpt4-1.4.1-lxctx-PI-16384-fp16 everywhere <=8192 tokens.
|
49 |
- Not presented here, but this model outperforms the base llama-2-13b on MMLU-fs with a score of 58.3. If this score ends up being be replicated on the HF LLM leaderboard, **this would place this model at 2nd or 3rd overall for MMLU among 13b models (and the #1 for extended context)**
|