Shijasmon's picture
Update README.md
cacff40 verified
---
license: apache-2.0
tags:
- causal-lm
- llama
- peft
- lora
- fine-tuning
- text-generation
- productivity
library_name: transformers
---
# Daily Tasks Fine Tuned LLaMA 3.2 1B – Weekly & Daily Task Planner
## Model Description
This is a **fine-tuned LLaMA 3.2 1B model** designed to generate structured weekly and daily plans. It can produce:
- Workout routines
- Study schedules
- Meal plans
- Other daily task setups
Fine-tuning was done using **PEFT LoRA** with float16 precision for efficient training on GPU.
---
## Intended Use
This model is intended for **personal productivity, fitness planning, and educational scheduling**. It is **not meant for medical, legal, or critical decision-making**.
---
## Usage
```python
from transformers import pipeline, AutoTokenizer, AutoModelForCausalLM
model_name = "your-username/daily_tasks_fine_tuned_llama3_2_1b"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name)
generator = pipeline(
"text-generation",
model=model,
tokenizer=tokenizer,
device=0 # Use -1 for CPU
)
prompt = "Plan a 7-day workout routine for cardiovascular health."
output = generator(prompt, max_new_tokens=600)
print(output[0]['generated_text'])