saidsef's picture
chore: update model information
6055835 verified
metadata
library_name: transformers
tags:
  - code
  - tech-docs
license: apache-2.0
datasets:
  - saidsef/tech-docs
language:
  - en
base_model:
  - google/flan-t5-small
new_version: google/flan-t5-small
pipeline_tag: text2text-generation

If you already know T5, FLAN-T5 is just better at everything. For the same number of parameters, these models have been fine-tuned on more than 1000 additional tasks covering also more languages. As mentioned in the first few lines of the abstract :

Flan-PaLM 540B achieves state-of-the-art performance on several benchmarks, such as 75.2% on five-shot MMLU. We also publicly release Flan-T5 checkpoints,1 which achieve strong few-shot performance even compared to much larger models, such as PaLM 62B. Overall, instruction finetuning is a general method for improving the performance and usability of pretrained language models.

Disclaimer: Content from this model card has been written by the Hugging Face team, and parts of it were copy pasted from the T5 model card.

Limitations:

  • The model may sometimes generate irrelevant keywords
  • Performance may vary depending on the length and complexity of the input text
    • For best results, use long clean texts
    • Length limit is 512 tokens due to Flan-T5 architecture
  • The model is trained on English text and may not perform well on other languages

Usage

# Use a pipeline as a high-level helper
from transformers import pipeline

pipe = pipeline("text2text-generation", model="saidsef/flan-t5-small-tuned-tech-docs")

# Load model directly
from transformers import AutoTokenizer, AutoModelForSeq2SeqLM

tokenizer = AutoTokenizer.from_pretrained("saidsef/flan-t5-small-tuned-tech-docs")
model = AutoModelForSeq2SeqLM.from_pretrained("saidsef/flan-t5-small-tuned-tech-docs")

Framework versions

  • Transformers 4.45.1
  • Pytorch 2.4.1+cu121
  • Datasets 3.0.1
  • Tokenizers 0.20.0

Ethical Considerations

When using this model, consider the potential impact of automated keyword extraction on content creation and SEO practices. Ensure that the use of this model complies with relevant guidelines and does not contribute to the creation of misleading or spammy content.