image/png This is the second in a series of models designed to replicate the prose quality of the Claude 3 models, specifically Sonnet and Opus. This model is fine-tuned on top of Qwen1.5 32B.

Prompting

Model has been Instruct tuned with the ChatML formatting. A typical input would look like this:

"""<|im_start|>user
Hi there!<|im_end|>
<|im_start|>assistant
Nice to meet you!<|im_end|>
<|im_start|>user
Can I ask a question?<|im_end|>
<|im_start|>assistant
"""

Credits

Three new general purpose instruction following datasets were added on top of the original Stheno dataset (which had certain low quality entries purged/removed). The first two were designed specifically for the Magnum series, to better address prompt adherence and coherence:

This model has been a team effort, and the credits goes to all members of Anthracite.

Training

The training was done for 2 epochs with a learning rate of 1e-05. We used 8x NVIDIA H100 Tensor Core GPUs for the full-parameter fine-tuning of the model.

Built with Axolotl

Safety

...

Downloads last month
128
Safetensors
Model size
32.5B params
Tensor type
BF16
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for anthracite-org/magnum-v1-32b

Base model

Qwen/Qwen1.5-32B
Finetuned
(5)
this model
Quantizations
6 models

Datasets used to train anthracite-org/magnum-v1-32b

Space using anthracite-org/magnum-v1-32b 1

Collection including anthracite-org/magnum-v1-32b