base_model: jondurbin/bagel-34b-v0.2 | |
license: apache-2.0 | |
![image/png](/static-proxy?url=https%3A%2F%2Fcdn-uploads.huggingface.co%2Fproduction%2Fuploads%2F64c14f6b02e1f8f67c73bd05%2Fpf4d6FA7DriRtVq5HCkxd.png%3C%2Fspan%3E)%3C!-- HTML_TAG_END --> | |
![image/png](/static-proxy?url=https%3A%2F%2Fcdn-uploads.huggingface.co%2Fproduction%2Fuploads%2F64c14f6b02e1f8f67c73bd05%2Fe4u8VYfDBh11u60rFYJHF.png%3C%2Fspan%3E)%3C!-- HTML_TAG_END --> | |
This model is a finetune of jondurbin's excellent [bagel](https://huggingface.co/jondurbin/bagel-34b-v0.2) model. | |
It has been trained with new datasets and a new technique, which we will share to the community soon. | |
This model has not utilised any form of merging. | |
### Evaluation Results | |
| Average | ARC | HellaSwag | MMLU | TruthfulQA | Winogrande | GSM8K | | |
| --- | --- | --- | --- | --- | --- | --- | | |
| 77.29 | 74.23 | 86.76 | 76.66 | 70.22 | 83.66 | 72.18 | | |
### Contamination Results | |
With reference model jondurbin/bagel-34b-v0.2: | |
| ARC | TruthfulQA | GSM8K | | |
| --- | --- | --- | | |
| 0.08| 0.38| 0.88| |