leonvanbokhorst commited on
Commit
30a9cea
·
verified ·
1 Parent(s): 7802afd

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -52
README.md CHANGED
@@ -5,7 +5,6 @@ license: apache-2.0
5
  datasets:
6
  - leonvanbokhorst/friction-overthinking-v2
7
  - leonvanbokhorst/friction-disagreement-v2
8
- - leonvanbokhorst/reluctance-v6.1
9
  - leonvanbokhorst/friction-uncertainty-v2
10
  language:
11
  - en
@@ -17,6 +16,7 @@ tags:
17
  - ai-disagreement
18
  - ai-reluctance
19
  - ai-uncertainty
 
20
  pipeline_tag: text-generation
21
  ---
22
 
@@ -86,57 +86,7 @@ The model:
86
  - Might challenge user assumptions in sensitive contexts
87
  - Should be used with appropriate content warnings
88
 
89
- ## Usage
90
-
91
- Example usage with the Transformers library:
92
-
93
- ```python
94
- from transformers import AutoModelForCausalLM, AutoTokenizer
95
-
96
- # Load model and tokenizer
97
- model_name = "leonvanbokhorst/deepseek-r1-mixture-of-friction"
98
- tokenizer = AutoTokenizer.from_pretrained(model_name)
99
- model = AutoModelForCausalLM.from_pretrained(model_name)
100
-
101
- # Format input with chat template
102
- prompt = """<|im_start|>system
103
- You are a human-like AI assistant.
104
- <|im_end|>
105
- <|im_start|>user
106
- Why do I keep procrastinating important tasks?
107
- <|im_end|>
108
- <|im_start|>assistant"""
109
-
110
- # Generate response
111
- inputs = tokenizer(prompt, return_tensors="pt")
112
- outputs = model.generate(
113
- inputs["input_ids"],
114
- max_length=512,
115
- temperature=0.7,
116
- top_p=0.9
117
- )
118
- response = tokenizer.decode(outputs[0], skip_special_tokens=True)
119
- ```
120
 
121
- ## Training Details
122
-
123
- ### LoRA Configuration
124
- - **Rank**: 16
125
- - **Alpha**: 32
126
- - **Target Modules**:
127
- - q_proj
128
- - k_proj
129
- - v_proj
130
- - o_proj
131
- - gate_proj
132
- - up_proj
133
- - down_proj
134
-
135
- ### Dataset Processing
136
- - Examples stacked up to 4096 tokens
137
- - 90/10 train/validation split
138
- - Consistent seed (42) for reproducibility
139
- - Token-based sampling for balanced training
140
 
141
  ## Citation
142
 
@@ -145,7 +95,7 @@ If you use this model in your research, please cite:
145
  ```bibtex
146
  @misc{friction-reasoning-2025,
147
  author = {Leon van Bokhorst},
148
- title = {Mixture of Friction: Fine-tuned Language Model for Productive Disagreement, Overthinking, and Hesitation},
149
  year = {2025},
150
  publisher = {HuggingFace},
151
  journal = {HuggingFace Model Hub},
 
5
  datasets:
6
  - leonvanbokhorst/friction-overthinking-v2
7
  - leonvanbokhorst/friction-disagreement-v2
 
8
  - leonvanbokhorst/friction-uncertainty-v2
9
  language:
10
  - en
 
16
  - ai-disagreement
17
  - ai-reluctance
18
  - ai-uncertainty
19
+ - unsloth
20
  pipeline_tag: text-generation
21
  ---
22
 
 
86
  - Might challenge user assumptions in sensitive contexts
87
  - Should be used with appropriate content warnings
88
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
89
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
90
 
91
  ## Citation
92
 
 
95
  ```bibtex
96
  @misc{friction-reasoning-2025,
97
  author = {Leon van Bokhorst},
98
+ title = {Mixture of Friction: Fine-tuned Language Model for Productive Disagreement, Overthinking, Uncertainty and Reluctance},
99
  year = {2025},
100
  publisher = {HuggingFace},
101
  journal = {HuggingFace Model Hub},