
LLMSIEM/logem
LLMSIEM/logem is a specialized language model fine-tuned for Security Information and Event Management (SIEM) tasks, particularly excelling at structured field extraction from security logs and events.
Model Details
Model Description
LLMSIEM/logem is a fine-tuned version of Qwen3-0.6B, specifically optimized for cybersecurity applications. The model demonstrates that targeted fine-tuning can dramatically improve performance on domain-specific tasks, achieving superior results compared to much larger general-purpose models.
- Developed by: [Hassan Shehata]
- Model type: Causal Language Model (Fine-tuned)
- Language(s): English
- License: Apache 2.0
- Finetuned from model: Qwen/Qwen3-0.6B
- Model size: 1.2 GB (FP16), 396 MB (Q4_K_M quantized)
- Parameters: 0.6B
Model Sources
- Blog Post: [LinkedIn/Blog Series Link]
Performance Highlights
๐ Best-in-class performance for SIEM field extraction tasks:
- 66.7% perfect matches (FP16 version)
- 0.833 F1 score - outperforms 12B parameter models
- 1.00s average response time - 3x faster than larger alternatives
- Zero complete failures on standardized test suite
Uses
Direct Use
The model is designed for cybersecurity professionals and SIEM engineers who need to:
- Extract structured fields from security logs
- Parse and normalize security event data
- Automate log analysis workflows
- Generate structured outputs from unstructured security data
Example Use Cases
# Example: Extract fields from a security log
input_text = "Extract fields from: Failed login attempt from 192.168.1.100 for user admin at 2024-01-15T10:30:45Z"
# Model will output structured JSON with relevant fields:
# {
# "event_type": "failed_login",
# "source_ip": "192.168.1.100",
# "username": "admin",
# "timestamp": "2024-01-15T10:30:45Z"
# }
Downstream Use
- Integration into SIEM platforms (Splunk, ELK, QRadar)
- Security orchestration and automated response (SOAR) workflows
- Threat hunting and incident response automation
- Security data lake processing pipelines
Out-of-Scope Use
- General-purpose text generation
- Non-security related field extraction
- Real-time processing without proper input validation
- Decision-making for critical security responses without human oversight
Bias, Risks, and Limitations
Technical Limitations
- Optimized specifically for security log formats seen during training
- May struggle with completely novel log formats or schemas
- Performance may degrade on logs with unusual encoding or formatting
- Quantized version (Q4_K_M) shows 5% accuracy reduction vs FP16
Security Considerations
- Model outputs should be validated before use in automated security workflows
- Not suitable for real-time critical security decisions without human oversight
- Training data may contain biases from specific security environments
- Should not be the sole source of truth for security incident classification
Recommendations
- Always validate model outputs in production security environments
- Implement fallback mechanisms for handling novel or malformed inputs
- Regular retraining recommended as new log formats emerge
- Use FP16 version for maximum accuracy, Q4_K_M for resource-constrained deployments
How to Get Started with the Model
from transformers import AutoTokenizer, AutoModelForCausalLM
import torch
# Load the model and tokenizer
tokenizer = AutoTokenizer.from_pretrained("LLMSIEM/logem")
model = AutoModelForCausalLM.from_pretrained("LLMSIEM/logem")
# Example usage
prompt = "Extract security fields from the following log: [your log here]"
inputs = tokenizer(prompt, return_tensors="pt")
with torch.no_grad():
outputs = model.generate(
inputs.input_ids,
max_length=512,
temperature=0.1,
do_sample=False,
pad_token_id=tokenizer.eos_token_id
)
result = tokenizer.decode(outputs[0], skip_special_tokens=True)
print(result)
Using with Ollama (Recommended for Production)
# Pull the quantized version
ollama pull LLMSIEM/logem
# Run inference
ollama run LLMSIEM/logem "Extract fields from: SSH login from 10.0.0.5 by root"
Training Details
Training Data
The model was fine-tuned on a curated dataset of security logs and corresponding structured field extractions, including:
- Network security events (firewall, IDS/IPS)
- Authentication logs (successful/failed logins)
- System security events (file access, process execution)
- Application security logs (web servers, databases)
Dataset characteristics:
- 21 standardized test cases for evaluation
- Diverse log formats and security event types
- JSON-formatted target outputs for structured field extraction
Training Procedure
Training Hyperparameters
- Base model: Qwen3-0.6B
- Training regime: Mixed precision (fp16)
- Fine-tuning approach: Supervised fine-tuning on field extraction tasks
- Optimization: Task-specific training for SIEM applications
Model Variants
- FP16 Version: 1.2 GB, maximum accuracy (0.833 F1)
- Q4_K_M Quantized: 396 MB, production-optimized (0.800 F1)
Evaluation
Testing Data, Factors & Metrics
Testing Data
- 21 standardized security log parsing test cases
- Diverse log formats from multiple security tools
- Ground truth structured outputs for comparison
Metrics
- Perfect Match Rate: Percentage of test cases with 100% accurate field extraction
- F1 Score: Harmonic mean of precision and recall for field detection
- Precision: Accuracy of extracted fields
- Response Time: Average inference latency
Results
Model | Perfect Matches | Avg F1 | Precision | Speed | Size |
---|---|---|---|---|---|
LLMSIEM/logem (FP16) | 14/21 (66.7%) | 0.833 | 0.848 | 1.00s | 1.2 GB |
LLMSIEM/logem (Q4_K_M) | 13/21 (61.9%) | 0.800 | 0.819 | 1.00s | 396 MB |
Gemma:12B | 15/21 (71.4%) | 0.790 | 0.788 | 3.06s | 5.0 GB |
Qwen3:0.6B (base) | 9/21 (42.9%) | 0.651 | 0.636 | 1.57s | 522 MB |
Key Findings
- +28% F1 improvement over base Qwen3-0.6B model
- Outperforms 12B models in F1 score despite being 20x smaller
- 3x faster than comparable accuracy models
- 12.6x smaller than Gemma while maintaining superior performance
Environmental Impact
Training a specialized 0.6B parameter model requires significantly less computational resources compared to training larger models from scratch:
- Hardware Type: NVIDIA GPU (RTX3060)
- Training approach: Fine-tuning (more efficient than training from scratch)
- Base model efficiency: Starting from pre-trained Qwen3-0.6B reduces carbon footprint
- Production efficiency: Smaller model size reduces inference energy consumption
Technical Specifications
Model Architecture
- Architecture: Transformer decoder (Qwen3 family)
- Parameters: 0.6 billion
- Context length: [Inherited from Qwen3-0.6B]
- Vocabulary size: [Inherited from Qwen3-0.6B]
Compute Infrastructure
- Training: Fine-tuning on security-specific datasets
- Inference: Optimized for CPU and GPU deployment
- Quantization: GGML Q4_K_M for edge deployment
Citation
If you use this model in your research or applications, please cite:
@misc{llmsiem-logem-2025,
title={LLMSIEM/logem: A Fine-tuned Language Model for Security Log Analysis},
author={[Hassan Shehata]},
year={2025},
url={https://huggingface.co/LLMSIEM/logem},
note={Fine-tuned from Qwen3-0.6B for SIEM applications}
}
Model Card Authors
[Hassan Shehata/LLMSIEM]
Model Card Contact
For questions about this model, please contact:
- Email: [[email protected]]
- LinkedIn: [https://www.linkedin.com/in/hassan-shehata-503272172/]
- GitHub: [Your GitHub Profile]
This model is part of the LLMSIEM research series exploring the application of Large Language Models in cybersecurity and SIEM workflows.
- Downloads last month
- 11