A newer version of this model is available: google/flan-t5-small

If you already know T5, FLAN-T5 is just better at everything. For the same number of parameters, these models have been fine-tuned on more than 1000 additional tasks covering also more languages. As mentioned in the first few lines of the abstract :

Flan-PaLM 540B achieves state-of-the-art performance on several benchmarks, such as 75.2% on five-shot MMLU. We also publicly release Flan-T5 checkpoints,1 which achieve strong few-shot performance even compared to much larger models, such as PaLM 62B. Overall, instruction finetuning is a general method for improving the performance and usability of pretrained language models.

Disclaimer: Content from this model card has been written by the Hugging Face team, and parts of it were copy pasted from the T5 model card.

Limitations:

  • The model may sometimes generate irrelevant keywords
  • Performance may vary depending on the length and complexity of the input text
    • For best results, use long clean texts
    • Length limit is 512 tokens due to Flan-T5 architecture
  • The model is trained on English text and may not perform well on other languages

Usage

# Use a pipeline as a high-level helper
from transformers import pipeline

pipe = pipeline("text2text-generation", model="saidsef/flan-t5-small-tuned-tech-docs")

# Load model directly
from transformers import AutoTokenizer, AutoModelForSeq2SeqLM

tokenizer = AutoTokenizer.from_pretrained("saidsef/flan-t5-small-tuned-tech-docs")
model = AutoModelForSeq2SeqLM.from_pretrained("saidsef/flan-t5-small-tuned-tech-docs")

Framework versions

  • Transformers 4.45.1
  • Pytorch 2.4.1+cu121
  • Datasets 3.0.1
  • Tokenizers 0.20.0

Ethical Considerations

When using this model, consider the potential impact of automated keyword extraction on content creation and SEO practices. Ensure that the use of this model complies with relevant guidelines and does not contribute to the creation of misleading or spammy content.

Downloads last month
53
Safetensors
Model size
77M params
Tensor type
F32
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for saidsef/flan-t5-small-tuned-tech-docs

Finetuned
(298)
this model

Dataset used to train saidsef/flan-t5-small-tuned-tech-docs