Pegasus CNN/DailyMail (TensorFlow)
This is a TensorFlow version of the google/pegasus-cnn_dailymail model, converted from PyTorch weights.
Model Description
PEGASUS is a pre-training approach for abstractive text summarization. This model was fine-tuned on the CNN/DailyMail dataset for news summarization tasks.
Key Features:
- π Converted from PyTorch to TensorFlow for better TF.js and TensorFlow ecosystem compatibility
- π° Specialized for news article summarization
- π― Fine-tuned on CNN/DailyMail dataset
Usage
from transformers import TFAutoModelForSeq2SeqLM, AutoTokenizer
# Load model and tokenizer
model = TFAutoModelForSeq2SeqLM.from_pretrained("your-username/pegasus-cnn-dailymail-tf")
tokenizer = AutoTokenizer.from_pretrained("your-username/pegasus-cnn-dailymail-tf")
# Example usage
article = "Your news article text here..."
inputs = tokenizer(article, max_length=1024, return_tensors="tf", truncation=True)
summary_ids = model.generate(inputs["input_ids"], max_length=150, min_length=30, length_penalty=2.0, num_beams=4, early_stopping=True)
summary = tokenizer.decode(summary_ids[0], skip_special_tokens=True)
print(summary)
Model Details
- Model Type: Sequence-to-sequence (Text Summarization)
- Language: English
- License: Apache 2.0
- Framework: TensorFlow
- Base Model: google/pegasus-cnn_dailymail
Training Data
This model was originally trained on the CNN/DailyMail dataset, which contains news articles paired with human-written summaries.
Performance
This TensorFlow model should perform identically to the original PyTorch version, as it was converted directly from the same weights.
Citation
@misc{zhang2019pegasus,
title={PEGASUS: Pre-training with Extracted Gap-sentences for Abstractive Summarization},
author={Jingqing Zhang and Yao Zhao and Mohammad Saleh and Peter J. Liu},
year={2019},
eprint={1912.08777},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
Conversion Notes
This model was converted from PyTorch to TensorFlow using the from_pt=True
parameter in the Transformers library, ensuring weight preservation and identical performance.
- Downloads last month
- 163
Dataset used to train data-plumber/pegasus-cnn-dailymail-tf
Evaluation results
- ROUGE-1 on CNN/DailyMailself-reported21.860
- ROUGE-2 on CNN/DailyMailself-reported8.900
- ROUGE-L on CNN/DailyMailself-reported16.850