File size: 2,730 Bytes
8017f30 2c90348 7802afd 2c90348 286a871 1975bcf 7802afd 30a9cea d9923a2 286a871 30a9cea 286a871 d9923a2 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 |
---
base_model: unsloth/DeepSeek-R1-Distill-Qwen-7B-unsloth-bnb-4bit
library_name: transformers
license: apache-2.0
datasets:
- leonvanbokhorst/friction-overthinking-v2
- leonvanbokhorst/friction-disagreement-v2
- leonvanbokhorst/friction-uncertainty-v2
language:
- en
tags:
- ai-safety
- ai-friction
- human-like-messiness
- ai-overthink
- ai-disagreement
- ai-reluctance
- ai-uncertainty
- unsloth
pipeline_tag: text-generation
---
# Friction Reasoning Model
This model is fine-tuned to engage in productive disagreement, overthinking, and reluctance. It's based on DeepSeek-R1-Distill-Qwen-7B and trained on a curated dataset of disagreement, overthinking, and reluctance examples.
## Model Description
- **Model Architecture**: DeepSeek-R1-Distill-Qwen-7B with LoRA adapters
- **Language(s)**: English
- **License**: Apache 2.0
- **Finetuning Approach**: Instruction tuning with friction-based reasoning examples
### Training Procedure
- **Hardware**: NVIDIA RTX 4090 (24GB)
- **Framework**: Unsloth + PyTorch
- **Training Time**: 35 minutes
- **Epochs**: 7 (early convergence around epoch 4)
- **Batch Size**: 2 per device (effective batch size 8 with gradient accumulation)
- **Optimization**: AdamW 8-bit
- **Learning Rate**: 2e-4 with cosine schedule
- **Weight Decay**: 0.01
- **Gradient Clipping**: 0.5
- **Mixed Precision**: bfloat16
## Intended Use
This model is designed for:
- Engaging in productive disagreement
- Challenging assumptions constructively
- Providing alternative perspectives
- Deep analytical thinking
- Careful consideration of complex issues
### Limitations
The model:
- Is not designed for factual question-answering
- May sometimes be overly disagreeable
- Should not be used for medical, legal, or financial advice
- Works best with reflective or analytical queries
- May not perform well on objective or factual tasks
### Bias and Risks
The model:
- May exhibit biases present in the training data
- Could potentially reinforce overthinking in certain situations
- Might challenge user assumptions in sensitive contexts
- Should be used with appropriate content warnings
## Citation
If you use this model in your research, please cite:
```bibtex
@misc{friction-reasoning-2025,
author = {Leon van Bokhorst},
title = {Mixture of Friction: Fine-tuned Language Model for Productive Disagreement, Overthinking, Uncertainty and Reluctance},
year = {2025},
publisher = {HuggingFace},
journal = {HuggingFace Model Hub},
howpublished = {\url{https://huggingface.co/leonvanbokhorst/deepseek-r1-mixture-of-friction}}
}
```
## Acknowledgments
- DeepSeek AI for the base model
- Unsloth team for the optimization toolkit
- HuggingFace for the model hosting and infrastructure |