physics-chat / README.md
jedlee2004's picture
Upload fine-tuned model and tokenizer
c0afb6a verified
---
base_model: deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B
library_name: peft
---
# DeepSeek-R1-Distill-Qwen-1.5B Fine-Tuned on Physics
This repository contains a fine-tuned version of the DeepSeek-R1-Distill-Qwen-1.5B base model, adapted specifically for answering physics-related questions with detailed, step-by-step chain-of-thought reasoning. The model has been fine-tuned using Parameter-Efficient Fine-Tuning (PEFT) with LoRA and 4-bit quantization to reduce memory usage while maintaining performance in the physics domain.
## Model Details
### Model Description
The model is specialized for physics tasks through fine-tuning on three curated datasets:
- **camel_physics:** Educational examples with structured prompts and chain-of-thought reasoning.
- **arxiv_physics:** Research-level questions and scholarly excerpts from physics papers.
- **alpaca_physics:** Instruction-based conversational examples in physics.
Fine-tuning was performed using PEFT techniques (LoRA) combined with 4-bit quantization. This configuration enables the model to generate comprehensive and contextually accurate explanations for complex physics problems.
- **Developed by:** Your Organization or Name
- **Funded by:** [Funding Source, if applicable]
- **Shared by:** Your Organization or Name
- **Model type:** Transformer-based causal language model, fine-tuned with PEFT (LoRA)
- **Language(s):** English
- **License:** [Specify License, e.g., Apache-2.0 or MIT]
- **Finetuned from model:** deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B
### Model Sources
- **Repository:** [Link to the model repository on Hugging Face]
- **Paper:** [Link to any associated paper or blog post]
- **Demo:** [Link to a demo, if available]
## Uses
### Direct Use
This model can be used to:
- Answer physics-related questions.
- Generate detailed explanations and step-by-step chain-of-thought reasoning for physics problems.
- Serve as an educational tool for physics and mathematics learners.
### Downstream Use
The model can be integrated into:
- Educational platforms and tutoring applications.
- Research assistance tools in physics.
- Chatbots and virtual assistants with a scientific focus.
### Out-of-Scope Use
The model is not intended for:
- Domains outside of physics, where domain-specific knowledge is critical.
- High-stakes applications without human verification.
- Use cases requiring generation in languages other than English.
## Bias, Risks, and Limitations
- **Bias:** The model is fine-tuned on curated physics datasets and may reflect biases inherent in that data.
- **Risks:** Inaccurate or oversimplified explanations may be generated, especially for advanced or niche physics topics. Users should verify outputs.
- **Limitations:** The model's knowledge is limited to the physics topics covered in the training data and may not generalize to emerging or unconventional topics.
### Recommendations
Users should:
- Verify the generated content for accuracy, particularly in educational or research contexts.
- Use the model as a supportive tool rather than a standalone source.
- Be aware of its domain-specific training and adjust expectations accordingly.
## How to Get Started with the Model
Install the required libraries:
```bash
pip install transformers peft