ailm's picture
Update README.md
cb3502e verified
---
license: apache-2.0
language:
- en
metrics:
- rouge
base_model: google/pegasus-cnn_dailymail
---
### Pegasus-based Text Summarization Model
Model Name: pegsus-text-summarization
### Model Description
This model is a fine-tuned version of the Pegasus model, specifically adapted for the task of text summarization. It is trained on the SAMSum dataset, which is designed for summarizing conversations.
### Usage
This model can be used to generate concise summaries of input text, particularly for conversational text or dialogue-based inputs.
### How to Use
You can use this model with the Hugging Face transformers library. Below is an example code snippet:
```bash
from transformers import PegasusForConditionalGeneration, PegasusTokenizer
# Load the pre-trained model and tokenizer
model_name = "ailm/pegsus-text-summarization"
model = PegasusForConditionalGeneration.from_pretrained(model_name)
tokenizer = PegasusTokenizer.from_pretrained(model_name)
# Define the input text
text = "Your input text here"
# Tokenize the input text
tokens = tokenizer(text, truncation=True, padding="longest", return_tensors="pt")
# Generate the summary
summary = model.generate(**tokens)
# Decode and print the summary
print(tokenizer.decode(summary[0], skip_special_tokens=True))