QuantFactory Banner

QuantFactory/QwQ-14B-Math-v0.2-GGUF

This is quantized version of qingy2024/QwQ-14B-Math-v0.2 created using llama.cpp

Original Model Card

Uploaded model

  • Developed by: qingy2024
  • License: apache-2.0
  • Finetuned from model : unsloth/qwen2.5-14b-bnb-4bit

This model is a fine-tuned version of Qwen 2.5-14B, trained on QwQ 32B Preview's responses to questions from the NuminaMathCoT dataset.

Note: This model uses the standard ChatML template.

At 500 steps, the loss was plateauing so I decided to stop training to prevent excessive overfitting.


Training Details

  • Base Model: Qwen 2.5-14B
  • Fine-Tuning Dataset: Verified subset of NuminaMathCoT using Qwen 2.5 3B Instruct as a judge. (the sharegpt-verified-cleaned subset from my dataset).
  • QLoRA Configuration:
    • Rank: 32
    • Rank Stabilization: Enabled
  • Optimization Settings:
    • Batch Size: 8
    • Gradient Accumulation Steps: 2 (Effective Batch Size: 16)
    • Warm-Up Steps: 5
    • Weight Decay: 0.01
  • Training Steps: 500 steps
  • Hardware Information: A100-80GB

Downloads last month
331
GGUF
Model size
14.8B params
Architecture
qwen2

2-bit

3-bit

4-bit

5-bit

6-bit

8-bit

Inference API
Unable to determine this model’s pipeline type. Check the docs .

Dataset used to train QuantFactory/QwQ-14B-Math-v0.2-GGUF