Trained on 3 epochs of the totally-not-an-llm/EverythingLM-data-V2-sharegpt
dataset.
### HUMAN:
{prompt}
### RESPONSE:
<leave a newline for the model to answer>
note: Changed a few of the finetuning parameters this time around. I have no idea if its any good but Feel free to give it a try!
Open LLM Leaderboard Evaluation Results
Detailed results can be found here
Metric | Value |
---|---|
Avg. | 36.29 |
ARC (25-shot) | 42.83 |
HellaSwag (10-shot) | 73.28 |
MMLU (5-shot) | 26.87 |
TruthfulQA (0-shot) | 37.26 |
Winogrande (5-shot) | 66.61 |
GSM8K (5-shot) | 1.59 |
DROP (3-shot) | 5.61 |
- Downloads last month
- 1,157
Inference Providers
NEW
This model is not currently available via any of the supported third-party Inference Providers, and
the model is not deployed on the HF Inference API.