--- library_name: transformers tags: [causal-lm, lora, fine-tuned, qwen, deepseek] --- # Model Card for Qwen-1.5B-LoRA-philosophy This model is a LoRA-fine-tuned causal language model based on `deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B`. It was trained on a custom philosophy dataset with fields `"prompt"` and `"completion"`. ## Model Details ### Model Description A parameter-efficient fine-tuning of a 1.5B-parameter Qwen-based model. At inference time, you can feed it a text prompt and it will generate the continuation. - **Developed by:** - **Funded by [optional]:** - **Shared by [optional]:** - **Model type:** Causal Language Model (LM) with LoRA adapters - **Language(s) (NLP):** English - **License:** - **Finetuned from model [optional]:** deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B ### Model Sources [optional] - **Repository:** https://huggingface.co/your-username/small-fine-tunes - **Paper [optional]:** - **Demo [optional]:** ## Uses ### Direct Use This model can be used out-of-the-box for text generation tasks such as chatbots, text completion, and conversational AI workflows. ### Downstream Use [optional] Developers can further fine-tune or adapt the model for domain-specific conversation, question answering, or summarization tasks. ### Out-of-Scope Use - High-stakes decision making without human oversight - Generation of disallowed or sensitive content - Real-time safety-critical systems ## Bias, Risks, and Limitations Since the base model and the fine-tuning data are proprietary or custom, unknown biases may exist. The model may: - Produce incorrect or hallucinatory statements - Reflect biases present in the source data ### Recommendations - Always review generated text for factual accuracy. - Do not rely on this model for safety-critical applications without additional guardrails. ## How to Get Started with the Model ```python from transformers import AutoTokenizer, AutoModelForCausalLM, pipeline tokenizer = AutoTokenizer.from_pretrained("your-username/small-fine-tunes") model = AutoModelForCausalLM.from_pretrained("your-username/small-fine-tunes") generator = pipeline("text-generation", model=model, tokenizer=tokenizer) output = generator("Once upon a time", max_length=100) print(output[0]["generated_text"])