Upload README.md with huggingface_hub
Browse files
README.md
CHANGED
@@ -21,7 +21,30 @@ pipeline_tag: image-text-to-text
|
|
21 |
|
22 |
## Model Description
|
23 |
|
24 |
-
LogicFlow-gemma-3-27b-thinking is
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
25 |
|
26 |
The model demonstrates enhanced capabilities in:
|
27 |
- **π§ Logical Reasoning**: Improved ability to work through complex logical problems step by step
|
@@ -48,10 +71,15 @@ The model demonstrates enhanced capabilities in:
|
|
48 |
## Training Details
|
49 |
|
50 |
### Training Data
|
51 |
-
The model was fine-tuned on
|
52 |
-
|
53 |
-
-
|
54 |
-
|
|
|
|
|
|
|
|
|
|
|
55 |
|
56 |
### Training Configuration
|
57 |
|
@@ -88,12 +116,12 @@ The model was fine-tuned on a combination of high-quality datasets:
|
|
88 |
- **Freeze Multi-modal Projector**: true
|
89 |
|
90 |
#### Special Features
|
91 |
-
- **Enable Thinking**: true (
|
92 |
-
- **Template**: gemma
|
93 |
-
- **Trust Remote Code**: true
|
94 |
-
- **Preprocessing Workers**: 16
|
95 |
-
- **Save Steps**: 100
|
96 |
-
- **Logging Steps**: 5
|
97 |
|
98 |
### Training Results
|
99 |
|
@@ -122,8 +150,6 @@ The loss curve demonstrates stable convergence with the final training loss reac
|
|
122 |
|
123 |
### Comprehensive Evaluation Results
|
124 |
|
125 |
-
Following established AI benchmarking best practices [(Domino AI, 2020)](https://domino.ai/blog/benchmarking-predictive-models), we conducted systematic evaluations across multiple domains to assess both predictive performance and operational characteristics. As emphasized by [(Cohere, 2025)](https://cohere.com/blog/ai-benchmarks-for-business), effective AI evaluation requires testing beyond simple accuracy metrics to capture real-world complexity and business needs.
|
126 |
-
|
127 |
| **Benchmark** | **Metric** | **Base Gemma-3-27B-IT** | **LogicFlow-gemma-3-27b-thinking** | **Improvement** |
|
128 |
|---------------|------------|--------------------------|-------------------------------------|-----------------|
|
129 |
| **π Mathematical Reasoning** |
|
@@ -294,7 +320,6 @@ This model uses the standard Gemma 3 multimodal chat template with optimized for
|
|
294 |
|
295 |
#### Text-only Chat
|
296 |
```python
|
297 |
-
# Assuming model_name = "RekklesAI/LogicFlow-gemma-3-27b-thinking" is already defined
|
298 |
messages = [
|
299 |
{"role": "system", "content": "You are a helpful AI assistant specialized in logical reasoning and mathematics."},
|
300 |
{"role": "user", "content": "Explain the reasoning behind the Pythagorean theorem and provide a step-by-step proof."}
|
@@ -319,7 +344,6 @@ print(response)
|
|
319 |
```python
|
320 |
from PIL import Image
|
321 |
|
322 |
-
# Assuming model and processor for "RekklesAI/LogicFlow-gemma-3-27b-thinking" are already loaded
|
323 |
# Load an image
|
324 |
image = Image.open("path/to/your/image.jpg")
|
325 |
|
@@ -502,7 +526,7 @@ For full reproducibility, here is the complete training configuration used:
|
|
502 |
```yaml
|
503 |
bf16: true
|
504 |
cutoff_len: 2048
|
505 |
-
dataset: openo1_sft,open_thoughts,open_r1_math
|
506 |
dataset_dir: data
|
507 |
ddp_timeout: 180000000
|
508 |
do_train: true
|
|
|
21 |
|
22 |
## Model Description
|
23 |
|
24 |
+
LogicFlow-gemma-3-27b-thinking is an advanced **multimodal reasoning model** built upon [google/gemma-3-27b-it](https://huggingface.co/google/gemma-3-27b-it), specifically designed to excel at complex logical reasoning, mathematical problem-solving, and step-by-step analytical thinking. This model represents a significant advancement in AI reasoning capabilities, achieved through careful fine-tuning on three specialized, high-quality datasets using LoRA (Low-Rank Adaptation) technique.
|
25 |
+
|
26 |
+
### Training Dataset Foundation
|
27 |
+
|
28 |
+
Our model has been meticulously trained on three cutting-edge datasets, each contributing unique reasoning capabilities:
|
29 |
+
|
30 |
+
#### π§ **OpenO1-SFT Dataset**
|
31 |
+
- **Purpose**: Supervised fine-tuning for advanced reasoning patterns
|
32 |
+
- **Content**: High-quality reasoning demonstrations with explicit thought processes
|
33 |
+
- **Impact**: Enables the model to break down complex problems systematically and show transparent reasoning chains
|
34 |
+
|
35 |
+
#### π **Open-Thoughts Dataset**
|
36 |
+
- **Purpose**: Step-by-step thinking process modeling
|
37 |
+
- **Content**: Detailed internal monologues and reasoning progressions for various problem types
|
38 |
+
- **Impact**: Teaches the model to externalize its thinking process, making reasoning transparent and verifiable
|
39 |
+
|
40 |
+
#### π’ **OpenR1-Math Dataset**
|
41 |
+
- **Purpose**: Mathematical reasoning and problem-solving specialization
|
42 |
+
- **Content**: Comprehensive mathematical problems with detailed solution methodologies
|
43 |
+
- **Impact**: Significantly enhances performance on mathematical reasoning tasks, from basic arithmetic to advanced competition-level problems
|
44 |
+
|
45 |
+
### Key Innovations
|
46 |
+
|
47 |
+
This unique combination of datasets creates a model that not only provides correct answers but also demonstrates **how** it arrives at those answers, making it particularly valuable for educational applications, research, and any scenario requiring explainable AI reasoning.
|
48 |
|
49 |
The model demonstrates enhanced capabilities in:
|
50 |
- **π§ Logical Reasoning**: Improved ability to work through complex logical problems step by step
|
|
|
71 |
## Training Details
|
72 |
|
73 |
### Training Data
|
74 |
+
The model was fine-tuned on three carefully selected, high-quality datasets that form the foundation of its exceptional reasoning capabilities:
|
75 |
+
|
76 |
+
- **π§ OpenO1-SFT**: Advanced supervised fine-tuning dataset containing high-quality reasoning demonstrations with explicit thought processes, enabling systematic problem breakdown and transparent reasoning chains
|
77 |
+
|
78 |
+
- **π Open-Thoughts**: Specialized dataset focused on step-by-step thinking processes, featuring detailed internal monologues and reasoning progressions that teach the model to externalize and structure its thinking
|
79 |
+
|
80 |
+
- **π’ OpenR1-Math**: Comprehensive mathematical reasoning dataset with detailed solution methodologies, significantly enhancing performance from basic arithmetic to advanced competition-level mathematical problems
|
81 |
+
|
82 |
+
This synergistic combination creates a model that excels not only at providing accurate answers but also at demonstrating clear, verifiable reasoning processes.
|
83 |
|
84 |
### Training Configuration
|
85 |
|
|
|
116 |
- **Freeze Multi-modal Projector**: true
|
117 |
|
118 |
#### Special Features
|
119 |
+
- **Enable Thinking**: true (**Critical** - Activates advanced Chain-of-Thought reasoning from OpenO1-SFT and Open-Thoughts datasets)
|
120 |
+
- **Template**: gemma (Optimized for multimodal reasoning tasks)
|
121 |
+
- **Trust Remote Code**: true (Required for advanced vision capabilities)
|
122 |
+
- **Preprocessing Workers**: 16 (Optimized for multimodal data processing)
|
123 |
+
- **Save Steps**: 100 (Frequent checkpointing for training stability)
|
124 |
+
- **Logging Steps**: 5 (Detailed training monitoring)
|
125 |
|
126 |
### Training Results
|
127 |
|
|
|
150 |
|
151 |
### Comprehensive Evaluation Results
|
152 |
|
|
|
|
|
153 |
| **Benchmark** | **Metric** | **Base Gemma-3-27B-IT** | **LogicFlow-gemma-3-27b-thinking** | **Improvement** |
|
154 |
|---------------|------------|--------------------------|-------------------------------------|-----------------|
|
155 |
| **π Mathematical Reasoning** |
|
|
|
320 |
|
321 |
#### Text-only Chat
|
322 |
```python
|
|
|
323 |
messages = [
|
324 |
{"role": "system", "content": "You are a helpful AI assistant specialized in logical reasoning and mathematics."},
|
325 |
{"role": "user", "content": "Explain the reasoning behind the Pythagorean theorem and provide a step-by-step proof."}
|
|
|
344 |
```python
|
345 |
from PIL import Image
|
346 |
|
|
|
347 |
# Load an image
|
348 |
image = Image.open("path/to/your/image.jpg")
|
349 |
|
|
|
526 |
```yaml
|
527 |
bf16: true
|
528 |
cutoff_len: 2048
|
529 |
+
dataset: openo1_sft,open_thoughts,open_r1_math # Three specialized reasoning datasets
|
530 |
dataset_dir: data
|
531 |
ddp_timeout: 180000000
|
532 |
do_train: true
|