Text Generation
Transformers
Safetensors
GGUF
mistral
roblox
luau
code
sft
trl
unsloth
conversational
text-generation-inference
boatbomber commited on
Commit
e581e7c
·
verified ·
1 Parent(s): faa562a

Add token counts to training data info

Browse files
Files changed (1) hide show
  1. README.md +91 -91
README.md CHANGED
@@ -1,91 +1,91 @@
1
- ---
2
- license: apache-2.0
3
- datasets:
4
- - TorpedoSoftware/the-luau-stack
5
- - TorpedoSoftware/roblox-info-dump
6
- language:
7
- - en
8
- - fr
9
- - de
10
- - es
11
- - pt
12
- - it
13
- base_model:
14
- - unsloth/Devstral-Small-2507
15
- tags:
16
- - roblox
17
- - luau
18
- - code
19
- - sft
20
- - transformers
21
- - trl
22
- - unsloth
23
- ---
24
-
25
- # Luau Devstral 24B Instruct v0.1
26
-
27
- A Roblox Luau focused finetune of [Devstral Small 2507](https://huggingface.co/mistralai/Devstral-Small-2507).
28
-
29
- ## Model Details
30
-
31
- ### Model Description
32
-
33
- Devstral Small 2507 is a powerful choice for local inference, achieving SOTA open source results at just 24B parameters. However, Roblox gamedev and Luau programming are generally not well represented in LLM training data. This model fine tunes Devstral on a corpus of permissively licensed Luau code and Roblox documentation, improving the model's Luau programming capabilities. Additionally, the jinja chat template contains a default system prompt that steers the model's Luau capabilities even further.
34
-
35
- - **Developed by:** Zack Williams ([boatbomber](https://huggingface.co/boatbomber))
36
- - **Funded by:** [Torpedo Software LLC](https://huggingface.co/TorpedoSoftware)
37
- - **License:** [Apache 2.0](https://www.tldrlegal.com/license/apache-license-2-0-apache-2-0)
38
- - **Finetuned from model:** [unsloth/Devstral-Small-2507](https://huggingface.co/unsloth/Devstral-Small-2507)
39
-
40
- ### Model Sources
41
-
42
- - **Repository:** https://huggingface.co/mistralai/Devstral-Small-2507
43
- - **Blog:** https://mistral.ai/news/devstral-2507
44
-
45
-
46
- ## Training Details
47
-
48
- ### Training Data
49
-
50
- 1. https://huggingface.co/datasets/TorpedoSoftware/the-luau-stack
51
-
52
- Format:
53
- ```md
54
- Repository: {repo_name}
55
- Repository Description: {repo_description}
56
-
57
- File Path: `{file_path}`
58
- File Content:
59
- ```Lua
60
- {file_content}
61
- ```\
62
- ```
63
- 2. https://huggingface.co/datasets/TorpedoSoftware/roblox-info-dump
64
-
65
- Format:
66
- ```md
67
- Roblox Creator Docs: {url}
68
- ```md
69
- {content}
70
- ```\
71
- ```
72
-
73
- ### Training Process
74
-
75
- Trained a LoRA adapter (r=64) at full precision on two epochs of the dataset for a total of 54,630 steps and 43.40 E FLOPs. Then merged the final adapter checkpoint into a BF16 model.
76
-
77
- ### Training Loss Curve
78
-
79
- ![Training Loss Curve](assets/training_loss.png)
80
-
81
- ### Imatrix Calibration
82
-
83
- The imatrix for the GGUF quantizations was computed using 5.73MB of text containing a combination of [technical.txt](https://huggingface.co/datasets/froggeric/imatrix/blob/main/technical.txt), [groups_merged.txt](huggingface.co/datasets/froggeric/imatrix/blob/main/groups_merged.txt), and content from [the-luau-stack](https://huggingface.co/datasets/TorpedoSoftware/the-luau-stack) & [roblox-info-dump](https://huggingface.co/datasets/TorpedoSoftware/roblox-info-dump). This created an imatrix that is well suited to the specialized tasks this model is designed for while still maintaining broader intelligence as well. While we do provide several quantizations already, the `imatrix.gguf` is included in this repository should you want to create other quants yourself.
84
-
85
- ## Environmental Impact
86
-
87
- Carbon emissions estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
88
-
89
- - **Hardware Type:** A100 80GB PCIe
90
- - **Hours used:** 60
91
- - **Carbon Emitted:** ~4.5 kg CO2eq (equivalent to ~10.1 miles driven by an average ICE car)
 
1
+ ---
2
+ license: apache-2.0
3
+ datasets:
4
+ - TorpedoSoftware/the-luau-stack
5
+ - TorpedoSoftware/roblox-info-dump
6
+ language:
7
+ - en
8
+ - fr
9
+ - de
10
+ - es
11
+ - pt
12
+ - it
13
+ base_model:
14
+ - unsloth/Devstral-Small-2507
15
+ tags:
16
+ - roblox
17
+ - luau
18
+ - code
19
+ - sft
20
+ - transformers
21
+ - trl
22
+ - unsloth
23
+ ---
24
+
25
+ # Luau Devstral 24B Instruct v0.1
26
+
27
+ A Roblox Luau focused finetune of [Devstral Small 2507](https://huggingface.co/mistralai/Devstral-Small-2507).
28
+
29
+ ## Model Details
30
+
31
+ ### Model Description
32
+
33
+ Devstral Small 2507 is a powerful choice for local inference, achieving SOTA open source results at just 24B parameters. However, Roblox gamedev and Luau programming are generally not well represented in LLM training data. This model fine tunes Devstral on a corpus of permissively licensed Luau code and Roblox documentation, improving the model's Luau programming capabilities. Additionally, the jinja chat template contains a default system prompt that steers the model's Luau capabilities even further.
34
+
35
+ - **Developed by:** Zack Williams ([boatbomber](https://huggingface.co/boatbomber))
36
+ - **Funded by:** [Torpedo Software LLC](https://huggingface.co/TorpedoSoftware)
37
+ - **License:** [Apache 2.0](https://www.tldrlegal.com/license/apache-license-2-0-apache-2-0)
38
+ - **Finetuned from model:** [unsloth/Devstral-Small-2507](https://huggingface.co/unsloth/Devstral-Small-2507)
39
+
40
+ ### Model Sources
41
+
42
+ - **Repository:** https://huggingface.co/mistralai/Devstral-Small-2507
43
+ - **Blog:** https://mistral.ai/news/devstral-2507
44
+
45
+
46
+ ## Training Details
47
+
48
+ ### Training Data
49
+
50
+ 1. https://huggingface.co/datasets/TorpedoSoftware/the-luau-stack
51
+
52
+ 25.917M lines of real Luau code, 0.452B tokens. Format:
53
+ ```txt
54
+ Repository: {repo_name}
55
+ Repository Description: {repo_description}
56
+
57
+ File Path: `{file_path}`
58
+ File Content:
59
+ ```Lua
60
+ {file_content}
61
+ ```\
62
+ ```
63
+ 1. https://huggingface.co/datasets/TorpedoSoftware/roblox-info-dump
64
+
65
+ 19.6K pages of multilingual Roblox documentation, 0.149B tokens. Format:
66
+ ```txt
67
+ Roblox Creator Docs: {url}
68
+ ```md
69
+ {content}
70
+ ```\
71
+ ```
72
+
73
+ ### Training Process
74
+
75
+ Trained a LoRA adapter (r=64) at full precision on two epochs of the dataset for a total of 54,630 steps and 43.40 E FLOPs. Then merged the final adapter checkpoint into a BF16 model.
76
+
77
+ ### Training Loss Curve
78
+
79
+ ![Training Loss Curve](assets/training_loss.png)
80
+
81
+ ### Imatrix Calibration
82
+
83
+ The imatrix for the GGUF quantizations was computed using 5.73MB of text containing a combination of [technical.txt](https://huggingface.co/datasets/froggeric/imatrix/blob/main/technical.txt), [groups_merged.txt](huggingface.co/datasets/froggeric/imatrix/blob/main/groups_merged.txt), and content from [the-luau-stack](https://huggingface.co/datasets/TorpedoSoftware/the-luau-stack) & [roblox-info-dump](https://huggingface.co/datasets/TorpedoSoftware/roblox-info-dump). This created an imatrix that is well suited to the specialized tasks this model is designed for while still maintaining broader intelligence as well. While we do provide several quantizations already, the `imatrix.gguf` is included in this repository should you want to create other quants yourself.
84
+
85
+ ## Environmental Impact
86
+
87
+ Carbon emissions estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
88
+
89
+ - **Hardware Type:** A100 80GB PCIe
90
+ - **Hours used:** 60
91
+ - **Carbon Emitted:** ~4.5 kg CO2eq (equivalent to ~10.1 miles driven by an average ICE car)