boatbomber commited on
Commit
005d8b3
·
1 Parent(s): e35a404

Add GGUF quants

Browse files
Gemma-3-27B-Roblox-Luau-Q3_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f407c6a06851b66cc4deb718c52d640387c78ed8b89062788cb1bcbb46f7874d
3
+ size 14043341568
Gemma-3-27B-Roblox-Luau-Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:431dc15aad46a47e9f7fe9b3945096c12489b8194c54df4ad34d8623c99bb1d3
3
+ size 17339606400
Gemma-3-27B-Roblox-Luau-Q5_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5e2fbd799e6ca97a78a96209fbcf00b82f593e06f23b7e99bf69a1d9af0373b1
3
+ size 20240797056
Gemma-3-27B-Roblox-Luau-Q6_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1eedb94c6134dbc9a60760c1930cafda228b4520007d69244508c12e0efdef9e
3
+ size 23323312128
Gemma-3-27B-Roblox-Luau-Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f7dd57b5376e98d53ff78614e4cb84f5fd1d89d7028a8d9764baf7aa36ef3c40
3
+ size 30205704960
README.md CHANGED
@@ -18,3 +18,13 @@ datasets:
18
  # Gemma-3-27B-Roblox-Luau
19
 
20
  A fine tune of [google/gemma-3-27b-it](google/gemma-3-27b-it) using [boatbomber/roblox-info-dump](https://huggingface.co/datasets/boatbomber/roblox-info-dump) and [boatbomber/the-luau-stack](https://huggingface.co/datasets/boatbomber/the-luau-stack) for Roblox domain knowledge.
 
 
 
 
 
 
 
 
 
 
 
18
  # Gemma-3-27B-Roblox-Luau
19
 
20
  A fine tune of [google/gemma-3-27b-it](google/gemma-3-27b-it) using [boatbomber/roblox-info-dump](https://huggingface.co/datasets/boatbomber/roblox-info-dump) and [boatbomber/the-luau-stack](https://huggingface.co/datasets/boatbomber/the-luau-stack) for Roblox domain knowledge.
21
+
22
+ Available quants:
23
+
24
+ | Quant | Size | Notes |
25
+ | ------ | ------- | ----- |
26
+ | Q8_O | 30.21GB | High resource use, but generally acceptable. Use only when accuracy is crucial. |
27
+ | Q6_K | 23.32GB | Uses Q6_K for all tensors. Good for high end GPUs. |
28
+ | Q5_K_M | 20.24GB | Uses Q6_K for half of the attention.wv and feed_forward.w2 tensors, else Q5_K |
29
+ | Q4_K_M | 17.34GB | **Recommended.** Uses Q6_K for half of the attention.wv and feed_forward.w2 tensors, else Q4_K |
30
+ | Q3_K_M | 14.04GB | Uses Q4_K for the attention.wv, attention.wo, and feed_forward.w2 tensors, else Q3_K. Quality is noticeably degraded. |