Bleta-Logjike 27B Albanian Logical Reasoning Model

Model Description

  • Developed by: klei aliaj & Armir Celiku
  • Model type: Bleta-Logjike 27B optimized for Albanian logical reasoning
  • License: apache-2.0
  • Format: Full-precision model (HuggingFace Transformers format)
  • Language: Albanian
  • Base architecture: Based on Gemma 3 27B

This model is the full-precision version of the Bleta-Logjike 27B model, specifically optimized for logical reasoning tasks in the Albanian language. Bleta is an Albanian adaptation based on Google's Gemma 3 architecture, with this version focused on enhancing logical reasoning and problem-solving capabilities for Albanian speakers.

Capabilities & Features

Logical Reasoning Focus

This Albanian language model excels at:

  1. Logical analysis and deduction in Albanian
  2. Step-by-step problem solving
  3. Structured reasoning for complex problems
  4. Understanding logical relationships and dependencies
  5. Mathematical reasoning for grade-school level problems
  6. Conversational reasoning and explanations

Albanian Language Optimization

  • Native support for Albanian grammar and vocabulary
  • Understanding of Albanian cultural context
  • Handling of Albanian-specific logical expressions and constructs
  • Natural conversational abilities in Albanian

Training Methodology

GRPO Approach

This model was fine-tuned using Generative Rejection Policy Optimization (GRPO), a reinforcement learning technique that trains models to optimize for specific reward functions. GRPO allows the model to learn from feedback on its generated responses, improving reasoning quality over time by:

  1. Generating multiple candidate responses
  2. Evaluating responses against specific reward criteria
  3. Learning to prefer high-quality reasoning patterns
  4. Optimizing for step-by-step problem solving

GSM8K Dataset

The training utilized the GSM8K (Grade School Math 8K) dataset, which contains over 8,000 high-quality grade school math problems, requiring step-by-step reasoning to solve. The dataset provides:

  • Diverse mathematical problem types
  • Multi-step reasoning challenges
  • Clear step-by-step solutions
  • Grade-school level complexity

This dataset was adapted for Albanian language training to ensure the model can handle mathematical reasoning tasks in Albanian.

Technical Specifications

Model Architecture

  • 27B parameters
  • Based on Gemma 3 architecture with Albanian adaptations
  • 128K context window
  • QK normalization
  • 5 sliding + 1 global attention pattern
  • 1024 sliding window attention

Usage Requirements

  • Recommended minimum 48GB GPU VRAM for full-precision inference
  • Compatible with Hugging Face Transformers library
  • Can be loaded with 4-bit or 8-bit quantization for lower resource environments

Usage with Transformers

from transformers import AutoModelForCausalLM, AutoTokenizer

model_name = "klei1/bleta-logjike-27b"
model = AutoModelForCausalLM.from_pretrained(model_name, device_map="auto", load_in_8bit=True)
tokenizer = AutoTokenizer.from_pretrained(model_name)

messages = [
    {"role": "user", "content": "Si llogaritet sipërfaqja e një trekëndëshi?"}
]

text = tokenizer.apply_chat_template(messages, add_generation_prompt=True, tokenize=False)
inputs = tokenizer(text, return_tensors="pt").to(model.device)

outputs = model.generate(**inputs, max_new_tokens=512, temperature=0.7, top_p=0.95)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))

Limitations

This is the full-precision version of the model requiring significant computational resources. For deployment on consumer hardware, consider using the 8-bit quantized GGUF version available at klei1/bleta-logjike-27b-finetune.

Acknowledgments

  • Google for developing the Gemma 3 architecture
  • OpenAI for the GSM8K dataset
  • Hugging Face for their TRL library and GRPO implementation
Downloads last month
15
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Collection including klei1/bleta-logjike-27b