Trained on 3 epochs of the totally-not-an-llm/EverythingLM-data-V2-sharegpt dataset.

### HUMAN:
{prompt}

### RESPONSE:
<leave a newline for the model to answer>

note: Changed a few of the finetuning parameters this time around. I have no idea if its any good but Feel free to give it a try!

Built with Axolotl

Open LLM Leaderboard Evaluation Results

Detailed results can be found here

Metric Value
Avg. 36.29
ARC (25-shot) 42.83
HellaSwag (10-shot) 73.28
MMLU (5-shot) 26.87
TruthfulQA (0-shot) 37.26
Winogrande (5-shot) 66.61
GSM8K (5-shot) 1.59
DROP (3-shot) 5.61
Downloads last month
1,157
Safetensors
Model size
3.43B params
Tensor type
F32
·
FP16
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for harborwater/open-llama-3b-everything-v2

Quantizations
1 model

Dataset used to train harborwater/open-llama-3b-everything-v2