RekklesAI commited on
Commit
82b5ab0
Β·
verified Β·
1 Parent(s): 7c99a96

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +40 -16
README.md CHANGED
@@ -21,7 +21,30 @@ pipeline_tag: image-text-to-text
21
 
22
  ## Model Description
23
 
24
- LogicFlow-gemma-3-27b-thinking is a fine-tuned **multimodal** version of [google/gemma-3-27b-it](https://huggingface.co/google/gemma-3-27b-it) that has been specifically optimized for logical reasoning, step-by-step thinking, and mathematical problem-solving with both text and image inputs. This model has been trained using LoRA (Low-Rank Adaptation) technique and then merged with the base model for optimal performance.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
25
 
26
  The model demonstrates enhanced capabilities in:
27
  - **🧠 Logical Reasoning**: Improved ability to work through complex logical problems step by step
@@ -48,10 +71,15 @@ The model demonstrates enhanced capabilities in:
48
  ## Training Details
49
 
50
  ### Training Data
51
- The model was fine-tuned on a combination of high-quality datasets:
52
- - **openo1_sft**: Supervised fine-tuning data for reasoning
53
- - **open_thoughts**: Dataset focused on step-by-step thinking processes
54
- - **open_r1_math**: Mathematical reasoning and problem-solving dataset
 
 
 
 
 
55
 
56
  ### Training Configuration
57
 
@@ -88,12 +116,12 @@ The model was fine-tuned on a combination of high-quality datasets:
88
  - **Freeze Multi-modal Projector**: true
89
 
90
  #### Special Features
91
- - **Enable Thinking**: true (enhanced reasoning capability)
92
- - **Template**: gemma
93
- - **Trust Remote Code**: true
94
- - **Preprocessing Workers**: 16
95
- - **Save Steps**: 100
96
- - **Logging Steps**: 5
97
 
98
  ### Training Results
99
 
@@ -122,8 +150,6 @@ The loss curve demonstrates stable convergence with the final training loss reac
122
 
123
  ### Comprehensive Evaluation Results
124
 
125
- Following established AI benchmarking best practices [(Domino AI, 2020)](https://domino.ai/blog/benchmarking-predictive-models), we conducted systematic evaluations across multiple domains to assess both predictive performance and operational characteristics. As emphasized by [(Cohere, 2025)](https://cohere.com/blog/ai-benchmarks-for-business), effective AI evaluation requires testing beyond simple accuracy metrics to capture real-world complexity and business needs.
126
-
127
  | **Benchmark** | **Metric** | **Base Gemma-3-27B-IT** | **LogicFlow-gemma-3-27b-thinking** | **Improvement** |
128
  |---------------|------------|--------------------------|-------------------------------------|-----------------|
129
  | **πŸ“Š Mathematical Reasoning** |
@@ -294,7 +320,6 @@ This model uses the standard Gemma 3 multimodal chat template with optimized for
294
 
295
  #### Text-only Chat
296
  ```python
297
- # Assuming model_name = "RekklesAI/LogicFlow-gemma-3-27b-thinking" is already defined
298
  messages = [
299
  {"role": "system", "content": "You are a helpful AI assistant specialized in logical reasoning and mathematics."},
300
  {"role": "user", "content": "Explain the reasoning behind the Pythagorean theorem and provide a step-by-step proof."}
@@ -319,7 +344,6 @@ print(response)
319
  ```python
320
  from PIL import Image
321
 
322
- # Assuming model and processor for "RekklesAI/LogicFlow-gemma-3-27b-thinking" are already loaded
323
  # Load an image
324
  image = Image.open("path/to/your/image.jpg")
325
 
@@ -502,7 +526,7 @@ For full reproducibility, here is the complete training configuration used:
502
  ```yaml
503
  bf16: true
504
  cutoff_len: 2048
505
- dataset: openo1_sft,open_thoughts,open_r1_math
506
  dataset_dir: data
507
  ddp_timeout: 180000000
508
  do_train: true
 
21
 
22
  ## Model Description
23
 
24
+ LogicFlow-gemma-3-27b-thinking is an advanced **multimodal reasoning model** built upon [google/gemma-3-27b-it](https://huggingface.co/google/gemma-3-27b-it), specifically designed to excel at complex logical reasoning, mathematical problem-solving, and step-by-step analytical thinking. This model represents a significant advancement in AI reasoning capabilities, achieved through careful fine-tuning on three specialized, high-quality datasets using LoRA (Low-Rank Adaptation) technique.
25
+
26
+ ### Training Dataset Foundation
27
+
28
+ Our model has been meticulously trained on three cutting-edge datasets, each contributing unique reasoning capabilities:
29
+
30
+ #### 🧠 **OpenO1-SFT Dataset**
31
+ - **Purpose**: Supervised fine-tuning for advanced reasoning patterns
32
+ - **Content**: High-quality reasoning demonstrations with explicit thought processes
33
+ - **Impact**: Enables the model to break down complex problems systematically and show transparent reasoning chains
34
+
35
+ #### πŸ’­ **Open-Thoughts Dataset**
36
+ - **Purpose**: Step-by-step thinking process modeling
37
+ - **Content**: Detailed internal monologues and reasoning progressions for various problem types
38
+ - **Impact**: Teaches the model to externalize its thinking process, making reasoning transparent and verifiable
39
+
40
+ #### πŸ”’ **OpenR1-Math Dataset**
41
+ - **Purpose**: Mathematical reasoning and problem-solving specialization
42
+ - **Content**: Comprehensive mathematical problems with detailed solution methodologies
43
+ - **Impact**: Significantly enhances performance on mathematical reasoning tasks, from basic arithmetic to advanced competition-level problems
44
+
45
+ ### Key Innovations
46
+
47
+ This unique combination of datasets creates a model that not only provides correct answers but also demonstrates **how** it arrives at those answers, making it particularly valuable for educational applications, research, and any scenario requiring explainable AI reasoning.
48
 
49
  The model demonstrates enhanced capabilities in:
50
  - **🧠 Logical Reasoning**: Improved ability to work through complex logical problems step by step
 
71
  ## Training Details
72
 
73
  ### Training Data
74
+ The model was fine-tuned on three carefully selected, high-quality datasets that form the foundation of its exceptional reasoning capabilities:
75
+
76
+ - **🧠 OpenO1-SFT**: Advanced supervised fine-tuning dataset containing high-quality reasoning demonstrations with explicit thought processes, enabling systematic problem breakdown and transparent reasoning chains
77
+
78
+ - **πŸ’­ Open-Thoughts**: Specialized dataset focused on step-by-step thinking processes, featuring detailed internal monologues and reasoning progressions that teach the model to externalize and structure its thinking
79
+
80
+ - **πŸ”’ OpenR1-Math**: Comprehensive mathematical reasoning dataset with detailed solution methodologies, significantly enhancing performance from basic arithmetic to advanced competition-level mathematical problems
81
+
82
+ This synergistic combination creates a model that excels not only at providing accurate answers but also at demonstrating clear, verifiable reasoning processes.
83
 
84
  ### Training Configuration
85
 
 
116
  - **Freeze Multi-modal Projector**: true
117
 
118
  #### Special Features
119
+ - **Enable Thinking**: true (**Critical** - Activates advanced Chain-of-Thought reasoning from OpenO1-SFT and Open-Thoughts datasets)
120
+ - **Template**: gemma (Optimized for multimodal reasoning tasks)
121
+ - **Trust Remote Code**: true (Required for advanced vision capabilities)
122
+ - **Preprocessing Workers**: 16 (Optimized for multimodal data processing)
123
+ - **Save Steps**: 100 (Frequent checkpointing for training stability)
124
+ - **Logging Steps**: 5 (Detailed training monitoring)
125
 
126
  ### Training Results
127
 
 
150
 
151
  ### Comprehensive Evaluation Results
152
 
 
 
153
  | **Benchmark** | **Metric** | **Base Gemma-3-27B-IT** | **LogicFlow-gemma-3-27b-thinking** | **Improvement** |
154
  |---------------|------------|--------------------------|-------------------------------------|-----------------|
155
  | **πŸ“Š Mathematical Reasoning** |
 
320
 
321
  #### Text-only Chat
322
  ```python
 
323
  messages = [
324
  {"role": "system", "content": "You are a helpful AI assistant specialized in logical reasoning and mathematics."},
325
  {"role": "user", "content": "Explain the reasoning behind the Pythagorean theorem and provide a step-by-step proof."}
 
344
  ```python
345
  from PIL import Image
346
 
 
347
  # Load an image
348
  image = Image.open("path/to/your/image.jpg")
349
 
 
526
  ```yaml
527
  bf16: true
528
  cutoff_len: 2048
529
+ dataset: openo1_sft,open_thoughts,open_r1_math # Three specialized reasoning datasets
530
  dataset_dir: data
531
  ddp_timeout: 180000000
532
  do_train: true