Model Card for Model ID

Model Details

The 4-bit of the base model TinyLlama/TinyLlama-1.1B-Chat-v1.0 was fine-tuned on a tweet sentiment dataset cardiffnlp/tweet_eval in order to determine tweets tonality by positive, neutral or negative.

Model Description

We used a system prompt to instruct the model:

SYSTEM PROMPT:

You are a tweet sentiment classifier. For each tweet input, analyze its sentiment and output exactly one word: "negative", "neutral", or "positive". Do not include any extra text.

But the model is not trained to return only the sentiment name.

The PEFT method applied was QLoRA, the optimizer parameters be offloaded to CPU, when GPU memory is limited but could slow down the learning.

Training Details

The peft framework was applied to the model, k_proj and v_proj layers whas modified with LoRA linear layer rank=8, alpha=16.

This model was trained with SFTTrainer of trl, batch_size=12, gradient_accumulation_steps=128, learning_rate=5e-4, lr_scheduler_type="linear"

The model achieved 0.19 macro f1-score on the test dataset(initial model 0.03)

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for xinyuema/llm-course-hw3-tinyllama-qlora-model

Finetuned
(395)
this model

Dataset used to train xinyuema/llm-course-hw3-tinyllama-qlora-model

Collection including xinyuema/llm-course-hw3-tinyllama-qlora-model