File size: 3,234 Bytes
0133ba7 765ca6f a265d36 0133ba7 765ca6f 0133ba7 765ca6f 0133ba7 765ca6f 0133ba7 765ca6f 0133ba7 765ca6f 0133ba7 765ca6f 0133ba7 765ca6f 0133ba7 765ca6f 0133ba7 765ca6f |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 |
---
license: llama3.3
base_model: SicariusSicariiStuff/Negative_LLAMA_70B
tags:
- llama-3.3
- finetune
- roleplay
- chat
- wings-of-fire
datasets:
- Darkhn/WOF_QA_V2
- Darkhn/WOF_Pretraining
- Darkhn/WOF_V2_All_Good_Stories
- Darkhn/WOF_V1_All_Incredible_Stories
---
# Model Name - L3.3-70B-Animus-V1
<img src="special-edition-box-set-poster-v0-e7a67voh505f1.webp" alt="Wings_of_Fire" width="700"/>
## Model Description
This is a fine-tuned version of `SicariusSicariiStuff/Negative_LLAMA_70B` specialized for roleplaying and instruction-following within the *Wings of Fire* universe.
The model was trained on a 3-million-token dataset extracted from the *Wings of Fire* book series to build a strong foundation of domain knowledge. It was fine-tuned for 2 epochs on a curated dataset of conversational examples.
The goal of this model is to provide a high-quality, immersive, and lore-accurate conversational experience for fans of the series. It can adopt character personas, answer questions about the world, and engage in creative storytelling.
## Training Details
### Training Hardware
The model was trained on a single NVIDIA H200 GPU.
### Training Procedure
A QLoRA (Quantized Low-Rank Adaptation) approach was used for efficient fine-tuning.
### Training Data
The training process involved two main stages:
1. **Domain training:** The base model was adapted to the *Wings of Fire* universe using a custom dataset of **3 million tokens** compiled directly from the book series. This step saturated the model with the specific lore, characters, and writing style of the source material.
2. **Instruction & Chat Fine-tuning:** The model was fine-tuned on a mixed dataset of **2,200 examples**:
* **1,400 Roleplay Conversations:** Multi-turn conversational examples designed to teach the model how to adopt and maintain character personas from the series.
* **800 Assistant Examples:** Instruction-response pairs focused on answering lore questions and following commands within the context of the *Wings of Fire* world.
## Intended Use & Limitations
* **Intended Use:** This model is intended for creative and roleplaying purposes within the *Wings of Fire* universe. It is designed for fans of the series and is not a general-purpose chatbot.
* **Limitations & Quirks:**
* Performance on tasks outside of its training domain (general knowledge, coding, etc.) is not guaranteed and will likely be poor.
* The model may "hallucinate" or generate plausible but non-canonical information.
* **Content:** The base model (`Negative_LLAMA_70B`) has a slight NSFW inclination. While this finetune was **not** trained for NSFW content, some of that behavior may bleed through. It is not the intended purpose of this model.
* **Formatting:** The model has a strong tendency to use asterisks (`*...*`) for actions and emphasis. This is a learned behavior from its training data.
* **Safety:** This model has not undergone additional safety alignment beyond what was included in its base. Standard responsible AI practices should be followed.
## Acknowledgements
* Credit to Meta for the powerful Llama 3 architecture.
* Credit to `SicariusSicariiStuff` for the `Negative_LLAMA_70B` base model. |