EN-DE parent ➜ EN-NL fine-tuned on creative corpus

Authors: Niek Holter
Thesis: “Transferring Creativity”

Summary

This model starts from Helsinki-NLP’s MarianMT opus-mt-en-de and is fine-tuned on a 10k-sentence non-creative English–Dutch corpus (Journalistic texts DPC).
It is one of four systems trained for my bachelor’s thesis to study how transfer-learning settings affect MT creativity.

Parent model Fine-tune data BLEU COMET Transformed Creativity Score
en-de Creative 12.7 0.626 0.38

Intended use

  • Research on creative MT and low-resource transfer learning

Training details

  • Hardware  : NVIDIA GTX 1070 (CUDA 12.1)
  • Epochs : Early-stopped ≤ 200 (patience 5)
  • LR / batch : 2 e-5 / 16
  • Script : finetuning.py
  • Env : environment.yml

Data

Downloads last month
12
Safetensors
Model size
73.9M params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Evaluation results