finetuned-qlora-bafoGPT-2 / hyperparameters.yaml
ChallengerSpaceShuttle's picture
Upload folder using huggingface_hub
c1c63aa verified
raw
history blame contribute delete
829 Bytes
checkpoint_dir: checkpoints/google/gemma-2-2b
out_dir: out/finetune/lora
precision: bf16-true
quantize: bnb.nf4-dq
devices: 1
num_nodes: 1
lora_r: 8
lora_alpha: 16
lora_dropout: 0.05
lora_query: true
lora_key: true
lora_value: true
lora_projection: true
lora_mlp: false
lora_head: false
data:
class_path: litgpt.data.JSON
init_args:
json_path: data/train.json
mask_prompt: false
val_split_fraction: 0.05
prompt_style: alpaca
ignore_index: -100
seed: 42
num_workers: 4
train:
save_interval: 1000
log_interval: 1
global_batch_size: 4
micro_batch_size: 1
lr_warmup_steps: 1000
epochs: 1
max_seq_length: 1024
min_lr: 6.0e-05
eval:
interval: 1000
max_new_tokens: 100
max_iters: 100
initial_validation: false
final_validation: true
optimizer: AdamW
logger_name: csv
seed: 1337