Model Card for bwlw3127/clickbait-bart-dailymail

This model fine-tunes facebook/bart-base to generate clickbait-style headlines from DailyMail articles.
It takes an article as input and produces a short, attention-grabbing headline.


Model Details

Model Description

This is a sequence-to-sequence model trained with the 🤗 Transformers library. It was adapted from facebook/bart-base and fine-tuned on scraped DailyMail article/headline pairs.
The goal is to demonstrate how large language models can be steered toward stylistic objectives such as “clickbait” headline generation.

  • Developed by: @bwlw3127
  • Model type: Seq2Seq (Encoder–Decoder) — BART
  • Language(s): English
  • License: MIT (code); base model follows facebook/bart-base license
  • Finetuned from: facebook/bart-base

Model Sources


Uses

Direct Use

  • Generate sensational or catchy headlines from article text.
  • Educational demo for seq2seq fine-tuning with Hugging Face.

Downstream Use

  • Adaptation to other headline-generation tasks (news summarization, marketing, etc.) by re-finetuning on different headline corpora.

Out-of-Scope Use

  • Producing factually reliable news summaries.
  • Sensitive domains where misleading/sensational output may cause harm.

Bias, Risks, and Limitations

  • The model reflects biases in DailyMail content and clickbait style.
  • Headlines may exaggerate, distort, or misrepresent facts.
  • It should not be used in production systems where factual accuracy is critical.

Recommendations

  • Use only in controlled or educational contexts.
  • Always review outputs manually before publishing.

How to Get Started with the Model

from transformers import AutoTokenizer, AutoModelForSeq2SeqLM

model_id = "bwlw3127/clickbait-bart-dailymail"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForSeq2SeqLM.from_pretrained(model_id)

article = "Scientists have discovered a new exoplanet twice the size of Earth..."
inputs = tokenizer("generate clickbait headline: " + article,
                   return_tensors="pt", truncation=True, max_length=512)

outputs = model.generate(**inputs, max_length=24, num_beams=4, early_stopping=True)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
Downloads last month
9
Safetensors
Model size
139M params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support