QwQ 14B Math
Collection
QwQ for the GPU middle-class, specialized in solving math problems step by step.
•
3 items
•
Updated
This model is a fine-tuned version of Qwen 2.5-14B, trained on QwQ 32B Preview's responses to questions from the NuminaMathCoT dataset.
Note: This model uses the standard ChatML template.
At 500 steps, the loss was plateauing so I decided to stop training to prevent excessive overfitting.
sharegpt-verified-cleaned
subset from my dataset).