Model Card for Model ID
This model builds upon PepMLM, aimed at generating peptides from receptor sequences. It incorporates the ESM model framework from HuggingFace for its core architecture. The key enhancement in this model is the adoption of LoRA for training, distinguishing it from its predecessor.
Usage:
from transformers import AutoTokenizer, AutoModelForMaskedLM
from peft import PeftConfig
import torch
model_name = "littleworth/esm2_t33_650M_UR50D_pepmlm_lora_adapter_merged"
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForMaskedLM.from_pretrained(model_name).to(device)
Packages version:
{'transformers': '4.36.0', 'peft': '0.9.0', 'torch': '2.0.0'}
Training summary:
{
"train/loss": 1.5091,
"train/grad_norm": 3.6427412033081055,
"train/learning_rate": 6.773224309612687e-7,
"train/epoch": 5,
"train/global_step": 6395,
"_timestamp": 1709229361.5373268,
"_runtime": 25556.57973074913,
"_step": 639,
"train/train_runtime": 25557.6176,
"train/train_samples_per_second": 4.003,
"train/train_steps_per_second": 0.25,
"train/total_flos": 220903283526564960,
"train/train_loss": 1.8436848362317955,
"_wandb": {
"runtime": 25556
}
}
- Downloads last month
- 5
Inference Providers
NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API:
The model has no pipeline_tag.
Model tree for littleworth/esm2_t33_650M_UR50D_pepmlm_lora_adapter_merged
Base model
facebook/esm2_t33_650M_UR50D