miscii-14b-1225
Image source: Rrhar'il | Phigros
Prompting & Usage
See miscii-14b-1028 for more details.
Training Details
Coming soon
Merge Details
This is a merge of pre-trained language models created using mergekit.
Merge Method
This model was merged using the TIES merge method using miscii-14b-1028 as a base.
Models Merged
The following models were included in the merge:
- sthenno/exp-002
- sthenno/miscii-1218
Configuration
The following YAML configuration was used to produce this model:
tokenizer_source: "base"
chat_template: "chatml"
merge_method: ties
dtype: bfloat16
parameters:
normalize: true
base_model: sthenno-com/miscii-14b-1028
models:
- model: sthenno-com/miscii-14b-1028
parameters:
weight: 1
density: 0.5
- model: sthenno/miscii-1218
parameters:
weight: 1
density: 0.5
- model: sthenno/exp-002
parameters:
weight: 0.9
density: 0.5
- model: sthenno/miscii-1218
parameters:
weight: 0.6
density: 0.5
Open LLM Leaderboard Evaluation Results
Congratulations to the miscii series models for surpassing 40 points for the first time! As of December 25, 2024, this should be the best-performing 14B model in the tests, right?
Metric | Value |
---|---|
Avg. | 40.08 |
IFEval (0-Shot) | 78.78 |
BBH (3-Shot) | 50.91 |
MATH Lvl 5 (4-Shot) | 31.57 |
GPQA (0-shot) | 17.00 |
MuSR (0-shot) | 14.77 |
MMLU-PRO (5-shot) | 47.46 |
- Downloads last month
- 264
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.
Model tree for sthenno-com/miscii-14b-1225
Base model
Qwen/Qwen2.5-14B
Finetuned
Qwen/Qwen2.5-14B-Instruct
Finetuned
sthenno-com/miscii-14b-1028
Space using sthenno-com/miscii-14b-1225 1
Evaluation results
- strict accuracy on IFEval (0-Shot)Open LLM Leaderboard78.780
- normalized accuracy on BBH (3-Shot)Open LLM Leaderboard50.910
- exact match on MATH Lvl 5 (4-Shot)Open LLM Leaderboard31.570
- acc_norm on GPQA (0-shot)Open LLM Leaderboard17.000
- acc_norm on MuSR (0-shot)Open LLM Leaderboard14.770
- accuracy on MMLU-PRO (5-shot)test set Open LLM Leaderboard47.460