Half the data was geared towards better reasoning (EvolKit-20k and reasoning-base-20k), the other half will help to de-censor the model (WizardLM data set).

Uploaded model

  • Developed by: theprint
  • License: apache-2.0
  • Finetuned from model : unsloth/llama-3.2-1b-instruct-bnb-4bit

This llama model was trained 2x faster with Unsloth and Huggingface's TRL library.

Downloads last month
338
Safetensors
Model size
1.24B params
Tensor type
BF16
·
Inference API
Unable to determine this model’s pipeline type. Check the docs .

Collection including theprint/ReWiz-Llama-3.2-1B