6bpw exl2 quant of: https://huggingface.co/flammenai/Flammades-Mistral-Nemo-12B


image/png

Flammades-Mistral-Nemo-12B

nbeerbower/Mistral-Nemo-Gutenberg-Doppel-12B-v2 finetuned on flammenai/Date-DPO-NoAsterisks and jondurbin/truthy-dpo-v0.1.

Method

ORPO tuned with 2x RTX 3090 for 3 epochs.

Open LLM Leaderboard Evaluation Results

Detailed results can be found here

Metric Value
Avg. 22.34
IFEval (0-Shot) 38.42
BBH (3-Shot) 32.39
MATH Lvl 5 (4-Shot) 6.19
GPQA (0-shot) 7.16
MuSR (0-shot) 20.31
MMLU-PRO (5-shot) 29.57
Downloads last month
2
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for Jellon/Flammades-Mistral-Nemo-12B-exl2-6bpw

Datasets used to train Jellon/Flammades-Mistral-Nemo-12B-exl2-6bpw

Evaluation results