metadata
			license: mit
tags:
  - generated-from-train
  - instruction-tuned
  - phi2
  - lora
  - low-resource
  - fine-tuning
datasets:
  - yahma/alpaca-cleaned
base_model: microsoft/phi-2
widget:
  - text: |-
      ### Instruction:
      Explain the concept of gravity.
      ### Response:
π§ phi2-lora-instruct
This is a LoRA fine-tuned version of Microsoftβs Phi-2 model trained on 500 examples from the yahma/alpaca-cleaned instruction dataset.
β Fine-Tuned by:
howtomakepplragequit β working on scalable, efficient LLM training for real-world instruction-following.
ποΈ Model Architecture
- Base model: microsoft/phi-2 (2.7B parameters)
- Adapter: LoRA (Low-Rank Adaptation), trained with PEFT
- Quantization: 4-bit NF4 via bitsandbytesfor efficient memory use
π¦ Dataset
- yahma/alpaca-cleaned
- Instruction-based Q&A for natural language understanding and generation
- Covers topics like science, grammar, everyday tasks, and reasoning
π οΈ Training Details
- Training platform: Google Colab (Free T4 GPU)
- Epochs: 2
- Batch size: 2 (with gradient accumulation)
- Optimizer: AdamW (via Transformers Trainer)
- Training time: ~20β30 mins
π Intended Use
- Ideal for instruction-following tasks, such as:- Explanation
- Summarization
- List generation
- Creative writing
 
- Can be adapted to custom domains (health, code, manufacturing) by adding your own prompts + responses.
π Example Prompt
Instruction: Give three tips to improve time management.
π§ͺ Try it Out
To use this model in your own project:
from transformers import AutoModelForCausalLM, AutoTokenizer
model = AutoModelForCausalLM.from_pretrained("howtomakepplragequit/phi2-lora-instruct")
tokenizer = AutoTokenizer.from_pretrained("howtomakepplragequit/phi2-lora-instruct")
input_text = "### Instruction:\nExplain how machine learning works.\n\n### Response:"
inputs = tokenizer(input_text, return_tensors="pt").to("cuda")
output = model.generate(**inputs, max_new_tokens=100)
print(tokenizer.decode(output[0], skip_special_tokens=True))