File size: 1,495 Bytes
392b0ab 8d769f3 392b0ab 8d769f3 1b4185a 8d769f3 8d011ab 392b0ab 8d769f3 acb3b56 392b0ab 98a74ef 392b0ab 98a74ef 9edf4e3 392b0ab b04e50d 8d769f3 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 |
---
base_model:
- appvoid/palmer-002-32k
- raidhon/coven_tiny_1.1b_32k_orpo_alpha
- appvoid/palmer-003
library_name: transformers
tags:
- mergekit
- merge
license: apache-2.0
---
<style>
@import url('https://fonts.googleapis.com/css2?family=Vollkorn:ital,wght@0,400..900;1,400..900&display=swap');
</style>
<div style="background-color: transparent; border-radius: .5rem; padding: 2rem; font-family: monospace; font-size: .85rem; text-align: justify;">
![palmer-004](https://huggingface.co/appvoid/palmer-004-original/resolve/main/palmer-004.jpeg)
#### june update
This model has improved overall performance at the expense of small degradation on winogrande. As all palmer models, the model is biased to respond to answers without using any specific prompt, feel free to further fine-tune it for your specific use case.
| Model | MMLU | ARC-C | HellaSwag | PIQA | Winogrande | Average |
|--------------------------------|-------|-------|-----------|--------|------------|---------|
| tinyllama-3t | 0.2577| 0.3029| 0.5935 | 0.7329 | 0.5959 | 0.4966 |
| palmer-004-old | 0.2601| 0.3456| 0.6138 | 0.7443 | **0.6511** | 0.5229 |
| palmer-004 | **0.2661**|**0.3490**|**0.6173**|**0.7481**| 0.6417 | **0.5244**|
#### note
Even though palmer-003 is only 2k context size, it's **0.5257** on average so if you don't need that much context size (32k) you are better off with the former one.
</div> |