π§ Dolphin-Mistral-24B-Venice-Edition - Fine-tuned by Daemontatox π¬
π Overview
This model is a fine-tuned version of cognitivecomputations/Dolphin-Mistral-24B-Venice-Edition, an instruction-tuned large language model based on the Mistral 24B architecture. The fine-tuning was conducted by Daemontatox, leveraging the Unsloth framework for accelerated training and memory efficiency.
Key Features:
- Fine-tuned for instruction-following, conversational understanding, and open-domain question answering
- Trained using HuggingFace TRL + Unsloth for up to 2x faster training
- Ideal for downstream applications like chatbots, virtual assistants, data analysis, and synthetic data generation
π§ Training Configuration
- Base model:
cognitivecomputations/Dolphin-Mistral-24B-Venice-Edition
- Trainer: Hugging Face TRL + Unsloth integration
- Objective: Instruction-following, language modeling
- Epochs: (User should insert specific info)
- Learning Rate: (User should insert)
- Batch Size: (User should insert)
- Precision: BF16 / FP16
- Hardware: Optimized for A100/H100 but can scale down to 24GB VRAM with Unsloth
π Dataset
Fine-tuned on proprietary/custom/open synthetic datasets including instruction-style prompts across:
- General knowledge
- Reasoning
- Coding (Python, Bash)
- Multi-turn conversations
- Creative writing
- Agent simulation
(Note: Dataset specifics are redacted or custom for privacy/IP constraints.)
π Usage
from transformers import AutoModelForCausalLM, AutoTokenizer
model = AutoModelForCausalLM.from_pretrained("Daemontatox/Dolphin-Mistral-24B-Finetuned")
tokenizer = AutoTokenizer.from_pretrained("Daemontatox/Dolphin-Mistral-24B-Finetuned")
inputs = tokenizer("### Instruction: Summarize the following text...\n", return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=512)
print(tokenizer.decode(outputs[0]))
Supports text-generation-inference and transformers
APIs.
π§ͺ Evaluation
The model shows enhanced performance on:
- Instruction following: More concise and accurate responses
- Multi-turn dialogue: Better retention of prior context
- Open-domain QA: Improved factual grounding vs base model
Benchmarks:
- ARC (Easy): β +5%
- HellaSwag: β +4.8%
- MT-Bench (subset): β +6.3% coherence/completeness
(Metrics are estimated; exact numbers depend on user's fine-tuning corpus and methodology.)
β οΈ Limitations
- Inherits limitations from base Mistral model (hallucination, repetition under long context)
- Responses may reflect biases in training data
- Not suitable for medical, legal, or safety-critical tasks without further alignment
β€οΈ Acknowledgements
- Base model: Cognitive Computations
- Training accelerator: Unsloth
- Libraries: Hugging Face Transformers + TRL
π License
Apache 2.0 β Free for commercial and research use with attribution.
βοΈ Author
Fine-tuned and maintained by Daemontatox
GitHub | Hugging Face: Daemontatox
- Downloads last month
- 102
Model tree for Daemontatox/Kraken
Base model
mistralai/Mistral-Small-24B-Base-2501