checkpoint_dir: checkpoints/google/gemma-2-2b | |
out_dir: out/finetune/lora | |
precision: bf16-true | |
quantize: bnb.nf4-dq | |
devices: 1 | |
num_nodes: 1 | |
lora_r: 8 | |
lora_alpha: 16 | |
lora_dropout: 0.05 | |
lora_query: true | |
lora_key: true | |
lora_value: true | |
lora_projection: true | |
lora_mlp: false | |
lora_head: false | |
data: | |
class_path: litgpt.data.JSON | |
init_args: | |
json_path: data/train.json | |
mask_prompt: false | |
val_split_fraction: 0.05 | |
prompt_style: alpaca | |
ignore_index: -100 | |
seed: 42 | |
num_workers: 4 | |
train: | |
save_interval: 1000 | |
log_interval: 1 | |
global_batch_size: 4 | |
micro_batch_size: 1 | |
lr_warmup_steps: 1000 | |
epochs: 1 | |
max_seq_length: 1024 | |
min_lr: 6.0e-05 | |
eval: | |
interval: 1000 | |
max_new_tokens: 100 | |
max_iters: 100 | |
initial_validation: false | |
final_validation: true | |
optimizer: AdamW | |
logger_name: csv | |
seed: 1337 | |