Uploaded model

  • Developed by: qingy2024
  • License: apache-2.0
  • Finetuned from model : unsloth/qwen2.5-14b-bnb-4bit

This model is a fine-tuned version of Qwen 2.5-14B, trained on QwQ 32B Preview's responses to questions from the NuminaMathCoT dataset.

Note: This model uses the standard ChatML template.

At 500 steps, the loss was plateauing so I decided to stop training to prevent excessive overfitting.


Training Details

  • Base Model: Qwen 2.5-14B
  • Fine-Tuning Dataset: Verified subset of NuminaMathCoT using Qwen 2.5 3B Instruct as a judge. (the sharegpt-verified-cleaned subset from my dataset).
  • QLoRA Configuration:
    • Rank: 32
    • Rank Stabilization: Enabled
  • Optimization Settings:
    • Batch Size: 8
    • Gradient Accumulation Steps: 2 (Effective Batch Size: 16)
    • Warm-Up Steps: 5
    • Weight Decay: 0.01
  • Training Steps: 500 steps
  • Hardware Information: A100-80GB

Downloads last month
87
Safetensors
Model size
14.8B params
Tensor type
BF16
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for qingy2024/QwQ-14B-Math-v0.2

Quantizations
5 models

Dataset used to train qingy2024/QwQ-14B-Math-v0.2

Collection including qingy2024/QwQ-14B-Math-v0.2