my_cool_model

This model is a finetuned version of meta-llama/Llama-3.2-3B-Instruct on the tatsu-lab/alpaca dataset.

Model description

More information needed

Training and evaluation results

More information needed

Training procedure

This model was trained using the torchtune library using the following command:

ppo_full_finetune_single_device.py \
 --config ./target/7B_full_ppo_low_memory_single_device.yaml \
 device=cuda \
 metric_logger._component_=torchtune.utils.metric_logging.WandBLogger \
 metric_logger.project=torchtune_ppo \
 forward_batch_size=2 \
 batch_size=64 \
 ppo_batch_size=32 \
 gradient_accumulation_steps=16 \
 compile=True \
 optimizer._component_=bitsandbytes.optim.PagedAdamW \
 optimizer.lr=3e-4

Framework versions

  • torchtune 0.0.0
  • torchao 0.5.0
  • datasets 2.20.0
  • sentencepiece 0.2.0
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference API
Unable to determine this model's library. Check the docs .

Model tree for smohammadi/torchtune-modelcards-test_20241003183117

Finetuned
(166)
this model

Dataset used to train smohammadi/torchtune-modelcards-test_20241003183117