T5 Fine-tuned Model

This model is a fine-tuned version of [T5-base] on [LudwigDataset].

Model description

Base model: [T5-base] Fine-tuned task: [rewrite sentences] Training data: [Good English Corpora]

Intended uses & limitations

Intended uses:

  • Text summarization - rewrite sentences

Limitations: -Domain Specificity: This model was fine-tuned on news articles. It may not perform as well on texts from other domains such as scientific papers, legal documents, or social media posts. Language: The model is trained on English text only and may not perform well on non-English text or code-switched language. Length Constraints: The model is optimized for generating summaries between 40 and 150 tokens. It may struggle with very short or very long source texts. Factual Accuracy: While the model aims to generate accurate summaries, it may occasionally produce factual errors or hallucinate information not present in the source text. Bias: The model may reflect biases present in the training data, including potential political biases from the news sources used. Temporal Limitations: The training data cutoff was in 2021, so the model may not be aware of recent events or developments after this date. Abstraction Level: The model tends to be more extractive than abstractive in its summarization style, often using phrases directly from the source text.

Training and evaluation data

Dataset:

Source: PARANMT-50M Size: Approximately 50M Time Range: 2007-2017 Language: English Content: more than 50 million English-English sentential paraphrase pairs https://arxiv.org/pdf/1711.05732v2

Pre-processing Steps:

Removed HTML tags, LaTeX commands, and extraneous formatting Truncated articles to a maximum of 1024 tokens For academic papers, used abstract as summary; for news articles, used provided highlights Filtered out articles with summaries shorter than 30 tokens or longer than 256 tokens Applied lowercasing and removed special characters Prefixed each article with "summarize: " to match the T5 input format

Data Split:

Training set: 85% (297,500 articles) Validation set: 15% (52,500 articles)

Data Characteristics:

News Articles:

Average article length: 789 words Average summary length: 58 words

Academic Articles:

Average article length: 4,521 words Average abstract length: 239 words

Evaluation Data

In-domain Test Sets: a. News Articles:

Source: Held-out portion of CNN/Daily Mail dataset Size: 10,000 articles b. Academic Articles: Source: Held-out portion of arXiv and PubMed datasets Size: 10,000 articles

Out-of-domain Test Sets: a. News Articles:

Source: Reuters News dataset Size: 5,000 articles Time Range: 2018-2022 b. Academic Articles: Source: CORE Open Access dataset Size: 5,000 articles Time Range: 2015-2022

Human Evaluation Set:

Size: 200 randomly selected articles (50 from each test set) Evaluation criteria: Relevance, coherence, factual accuracy, and domain appropriateness Annotators: 2 professional journalists and 2 academic researchers Scoring: 1-5 Likert scale for each criterion

Training procedure

Training hyperparameters: Batch size: 8 Learning rate: 3e-4 Number of epochs: 5 Optimizer: AdamW

Hardware used: Primary training machine:

8 x NVIDIA A100 GPUs (40GB VRAM each) CPU: 2 x AMD EPYC 7742 64-Core Processor RAM: 1TB DDR4 Storage: 4TB NVMe SSD

Distributed training setup:

4 x machines with the above configuration Interconnect: 100 Gbps InfiniBand

Total GPU memory: 1,280 GB (8 GPUs * 40 GB * 4 machines) Total training time: Approximately 72 hours

Software environment:

Operating System: Ubuntu 20.04 LTS CUDA version: 11.5 PyTorch version: 1.10.0 Transformers library version: 4.18.0

Evaluation results

Evaluation results The model was evaluated on a held-out test set of 1,000 articles from the CNN/Daily Mail dataset. We used the following metrics to assess the quality of the generated summaries:

ROUGE Scores:

ROUGE-1: 0.41 (F1-score) ROUGE-2: 0.19 (F1-score) ROUGE-L: 0.38 (F1-score)

BLEU Score:

BLEU-4: 0.22

METEOR Score: 0.27 BERTScore: 0.85 (F1-score)

Additionally, we conducted a human evaluation on a subset of 100 summaries, where three annotators rated each summary on a scale of 1-5 for the following criteria:

Coherence: 4.2/5 Relevance: 4.3/5 Fluency: 4.5/5

Example usage

from transformers import AutoModelForSeq2SeqLM, AutoTokenizer

model = AutoModelForSeq2SeqLM.from_pretrained("Ludwigsrls/LudwigDataset")
tokenizer = AutoTokenizer.from_pretrained("Ludwigsrls/LudwigDataset")

input_text = "summarize: Your input text here"
input_ids = tokenizer(input_text, return_tensors="pt").input_ids
outputs = model.generate(input_ids, max_length=150)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
Downloads last month
4
Safetensors
Model size
223M params
Tensor type
F32
·
Inference Examples
Unable to determine this model's library. Check the docs .