Zakia's picture
Updated the examples sections of the model card
1760e13
---
license: apache-2.0
datasets:
- Zakia/drugscom_reviews
language:
- en
metrics:
- rewards mean change
- rewards median change
library_name: transformers
pipeline_tag: text-generation
tags:
- health
- medicine
- patient reviews
- drug reviews
- depression
- text generation
widget:
- text: After starting this new treatment, I felt
example_title: Example 1
- text: I was apprehensive about the side effects of
example_title: Example 2
- text: This medication has changed my life for the better
example_title: Example 3
- text: I've had a terrible experience with this medication
example_title: Example 4
- text: Since I began taking L-methylfolate, my experience has been
example_title: Example 5
---
# Model Card for Zakia/gpt2-drugscom_depression_reviews-hq-v1
This model is a GPT-2-based language model further refined using Reinforcement Learning with Human Feedback (RLHF) on patient drug reviews related to depression from Drugs.com.
The fine-tuning utilizes the 🤗 Hugging Face [Transformer Reinforcement Learning (TRL)](https://github.com/huggingface/trl) library to enhance the model's ability to generate high-quality synthetic patient reviews.
The dataset used for fine-tuning is the [Zakia/drugscom_reviews](https://huggingface.co./datasets/Zakia/drugscom_reviews) dataset, which is filtered for the condition 'Depression'.
The base model for fine-tuning was the [Zakia/gpt2-drugscom_depression_reviews](https://huggingface.co./Zakia/gpt2-drugscom_depression_reviews).
## Model Details
### Model Description
- Developed by: [Zakia](https://huggingface.co./Zakia)
- Model type: Text Generation with RLHF
- Language(s) (NLP): English
- License: Apache 2.0
- Base model: [Zakia/gpt2-drugscom_depression_reviews](https://huggingface.co./Zakia/gpt2-drugscom_depression_reviews)
- Reward model: [Zakia/distilbert-drugscom_depression_reviews](https://huggingface.co./Zakia/distilbert-drugscom_depression_reviews)
## Uses
### Direct Use
This model generates synthetic patient reviews of depression medications. It is intended for research, educational purposes, or to support professional healthcare insights.
### Out-of-Scope Use
Not intended for clinical use or to diagnose or treat health conditions.
## Bias, Risks, and Limitations
The model's outputs reflect patterns in the training data and should not be considered clinical advice.
Biases present in the training data could be amplified.
### Recommendations
Use the model as a tool for generating synthetic patient reviews and for NLP research.
## How to Get Started with the Model
Use the code below to generate synthetic high quality drug reviews for depression with the model.
```python
from transformers import GPT2LMHeadModel, GPT2Tokenizer
import torch
model_name = "Zakia/gpt2-drugscom_depression_reviews-hq-v1"
model = GPT2LMHeadModel.from_pretrained(model_name)
tokenizer = GPT2Tokenizer.from_pretrained(model_name)
# Function to generate high-quality text
def generate_high_quality_review(prompt, model, tokenizer):
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs)
return tokenizer.decode(outputs[0], skip_special_tokens=True)
# Example usage for various scenarios
prompts = [
"After starting this new treatment, I felt",
"I was apprehensive about the side effects of",
"This medication has changed my life for the better",
"I've had a terrible experience with this medication",
"Since I began taking L-methylfolate, my experience has been"
]
for prompt in prompts:
print(f"Prompt: {prompt}")
print(generate_high_quality_review(prompt, model, tokenizer))
print()
```
## Training Details
### Training Data
The model was fine-tuned on patient reviews related to depression, filtered from Drugs.com.
This dataset is accessible from [Zakia/drugscom_reviews](https://huggingface.co./datasets/Zakia/drugscom_reviews) on Hugging Face datasets (condition = 'Depression') for 'train'.
Number of records in train dataset: 9069 rows.
### Training Procedure
#### Preprocessing
The reviews were cleaned and preprocessed to remove quotes, HTML tags and decode HTML entities.
#### Training Hyperparameters
- Learning Rate: 1.41e-5
- Batch Size: 128
## Evaluation
- Rewards before and after RLHF
#### Metrics
The model's performance was evaluated based on rewards before and after RLHF.
### Results
### Evaluation Results
The RLHF fine-tuning was conducted using a dataset of patient reviews for depression.
The model showed significant improvement in the synthetic reviews' quality.
| Metric | Before RLHF | After RLHF |
|:----------------------|--------------:|-------------:|
| Rewards Mean Change | -1.622 | 1.416 |
| Rewards Median Change | -1.828 | 2.063 |
The positive shift in rewards suggests that the model is now more adept at generating reviews that align
with high-quality patient feedback.
## Technical Specifications
### Model Architecture and Objective
The GPT-2 architecture was enhanced through RLHF to produce text that closely resembles authentic patient experiences.
### Compute Infrastructure
The model was trained using a T4 GPU on Google Colab.
#### Hardware
T4 GPU via Google Colab.
## Citation
If you use this model, please cite both the original GPT-2 and DistilBERT papers:
**GPT-2 BibTeX:**
```bibtex
@article{radford2019language,
title={Language Models are Unsupervised Multitask Learners},
author={Radford, Alec and others},
year={2019}
}
```
**DistilBERT BibTeX:**
```bibtex
@article{sanh2019distilbert,
title={DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter},
author={Sanh, Victor and Debut, Lysandre and Chaumond, Julien and Wolf, Thomas},
journal={arXiv preprint arXiv:1910.01108},
year={2019}
}
```
**APA:**
- Radford, A., et al. (2019). Language Models are Unsupervised Multitask Learners.
- Sanh, V., Debut, L., Chaumond, J., & Wolf, T. (2019). DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter. arXiv preprint arXiv:1910.01108.
## More Information
For further queries or issues with the model, please use the [discussions section on this model's Hugging Face page](https://huggingface.co./Zakia/gpt2-drugscom_depression_reviews-hq-v1/discussions).
## Model Card Authors
- [Zakia](https://huggingface.co./Zakia)
## Model Card Contact
For more information or inquiries regarding this model, please use the [discussions section on this model's Hugging Face page](https://huggingface.co./Zakia/gpt2-drugscom_depression_reviews-hq-v1/discussions).