metadata
language:
- en
license: apache-2.0
tags:
- code
- unsloth
- trl
- sft
This is a fine-tune of Llama3-8B base model using an Alpaca-like instruction-tuning dataset generated with GPT4. My aim with doing this was to compare the instruction-tuned performance of this model with the one Llama has available.