--- library_name: transformers tags: [] --- # Just for PEFT testing See PR #1658 ## Reproduction ```python import torch from huggingface_hub import cached_download from transformers import AutoModelForCausalLM from peft import LoraConfig, PeftModel, get_peft_model torch.manual_seed(0) model = AutoModelForCausalLM.from_pretrained("hf-internal-testing/tiny-random-BertModel").eval() lora_config = LoraConfig(init_lora_weights=False) peft_model = get_peft_model(model, lora_config) peft_model.push_to_hub("peft-internal-testing/tiny-random-BertModel-lora") del model, peft_model torch.manual_seed(0) model = AutoModelForCausalLM.from_pretrained("hf-internal-testing/tiny-random-BertModel").eval() torch.manual_seed(1) lora_config = LoraConfig(init_lora_weights=False) peft_model = get_peft_model(model, lora_config) peft_model.push_to_hub("peft-internal-testing/tiny-random-BertModel-lora", revision="v1.2.3") ```