--- base_model: unsloth/DeepSeek-R1-Distill-Qwen-7B-unsloth-bnb-4bit library_name: transformers license: apache-2.0 datasets: - leonvanbokhorst/friction-overthinking-v2 - leonvanbokhorst/friction-disagreement-v2 - leonvanbokhorst/friction-uncertainty-v2 language: - en tags: - ai-safety - ai-friction - human-like-messiness - ai-overthink - ai-disagreement - ai-reluctance - ai-uncertainty - unsloth pipeline_tag: text-generation --- # Friction Reasoning Model This model is fine-tuned to engage in productive disagreement, overthinking, and reluctance. It's based on DeepSeek-R1-Distill-Qwen-7B and trained on a curated dataset of disagreement, overthinking, and reluctance examples. ## Model Description - **Model Architecture**: DeepSeek-R1-Distill-Qwen-7B with LoRA adapters - **Language(s)**: English - **License**: Apache 2.0 - **Finetuning Approach**: Instruction tuning with friction-based reasoning examples ### Training Procedure - **Hardware**: NVIDIA RTX 4090 (24GB) - **Framework**: Unsloth + PyTorch - **Training Time**: 35 minutes - **Epochs**: 7 (early convergence around epoch 4) - **Batch Size**: 2 per device (effective batch size 8 with gradient accumulation) - **Optimization**: AdamW 8-bit - **Learning Rate**: 2e-4 with cosine schedule - **Weight Decay**: 0.01 - **Gradient Clipping**: 0.5 - **Mixed Precision**: bfloat16 ## Intended Use This model is designed for: - Engaging in productive disagreement - Challenging assumptions constructively - Providing alternative perspectives - Deep analytical thinking - Careful consideration of complex issues ### Limitations The model: - Is not designed for factual question-answering - May sometimes be overly disagreeable - Should not be used for medical, legal, or financial advice - Works best with reflective or analytical queries - May not perform well on objective or factual tasks ### Bias and Risks The model: - May exhibit biases present in the training data - Could potentially reinforce overthinking in certain situations - Might challenge user assumptions in sensitive contexts - Should be used with appropriate content warnings ## Citation If you use this model in your research, please cite: ```bibtex @misc{friction-reasoning-2025, author = {Leon van Bokhorst}, title = {Mixture of Friction: Fine-tuned Language Model for Productive Disagreement, Overthinking, Uncertainty and Reluctance}, year = {2025}, publisher = {HuggingFace}, journal = {HuggingFace Model Hub}, howpublished = {\url{https://huggingface.co/leonvanbokhorst/deepseek-r1-mixture-of-friction}} } ``` ## Acknowledgments - DeepSeek AI for the base model - Unsloth team for the optimization toolkit - HuggingFace for the model hosting and infrastructure