File size: 1,495 Bytes
392b0ab
 
 
 
 
 
 
 
 
8d769f3
392b0ab
 
8d769f3
 
 
 
1b4185a
8d769f3
8d011ab
392b0ab
8d769f3
acb3b56
392b0ab
98a74ef
392b0ab
98a74ef
 
 
9edf4e3
 
392b0ab
b04e50d
 
 
 
8d769f3
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
---
base_model:
- appvoid/palmer-002-32k
- raidhon/coven_tiny_1.1b_32k_orpo_alpha
- appvoid/palmer-003
library_name: transformers
tags:
- mergekit
- merge
license: apache-2.0
---

<style>
@import url('https://fonts.googleapis.com/css2?family=Vollkorn:ital,wght@0,400..900;1,400..900&display=swap');
</style>

<div style="background-color: transparent; border-radius: .5rem; padding: 2rem; font-family: monospace; font-size: .85rem; text-align: justify;">
  
![palmer-004](https://huggingface.co/appvoid/palmer-004-original/resolve/main/palmer-004.jpeg)

  
#### june update

This model has improved overall performance at the expense of small degradation on winogrande. As all palmer models, the model is biased to respond to answers without using any specific prompt, feel free to further fine-tune it for your specific use case.

| Model                          | MMLU  | ARC-C | HellaSwag | PIQA   | Winogrande | Average |
|--------------------------------|-------|-------|-----------|--------|------------|---------|
| tinyllama-3t                   | 0.2577| 0.3029| 0.5935    | 0.7329 | 0.5959     | 0.4966  |
| palmer-004-old                 | 0.2601| 0.3456| 0.6138    | 0.7443 | **0.6511** | 0.5229  |
| palmer-004                 | **0.2661**|**0.3490**|**0.6173**|**0.7481**| 0.6417   | **0.5244**|

#### note

Even though palmer-003 is only 2k context size, it's **0.5257** on average so if you don't need that much context size (32k) you are better off with the former one.

</div>