apepkuss79 commited on
Commit
21e0aef
·
verified ·
1 Parent(s): f6c47c4

Update models

Browse files
.gitattributes CHANGED
@@ -33,3 +33,16 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ Tessa-T1-14B-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
37
+ Tessa-T1-14B-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
38
+ Tessa-T1-14B-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
39
+ Tessa-T1-14B-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
40
+ Tessa-T1-14B-Q4_0.gguf filter=lfs diff=lfs merge=lfs -text
41
+ Tessa-T1-14B-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
42
+ Tessa-T1-14B-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
43
+ Tessa-T1-14B-Q5_0.gguf filter=lfs diff=lfs merge=lfs -text
44
+ Tessa-T1-14B-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
45
+ Tessa-T1-14B-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
46
+ Tessa-T1-14B-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
47
+ Tessa-T1-14B-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
48
+ Tessa-T1-14B-f16.gguf filter=lfs diff=lfs merge=lfs -text
Tessa-T1-14B-Q2_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:573365cca000a8358936a058e6ea16015ac5747e518cc3cbae48ffadf1351eee
3
+ size 5770498048
Tessa-T1-14B-Q3_K_L.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0ee26ab11ee7d188ea2eac9626bd0b54df9fd5905e85b02d88906897fbc49ecc
3
+ size 7924768768
Tessa-T1-14B-Q3_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2f03e9cfb89f465a0694adc72aab44b98577ed67c2aa8d3521c677b298e5f8bb
3
+ size 7339204608
Tessa-T1-14B-Q3_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:98223345fa00090129c4662db5302c2bb24120aa3c934cc7fd0dc2665f7894c4
3
+ size 6659596288
Tessa-T1-14B-Q4_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d9f0453522f6d31c9a9fa60903a0dd625d45a1f1a69863adb9d9c5ca334f7cf0
3
+ size 8517726208
Tessa-T1-14B-Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f7a94d9030bfea072c6edb8fd0b958bd59003d00f163219a3fe7f8dedbeafde8
3
+ size 8988110848
Tessa-T1-14B-Q4_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fd80b0cd487f279859b5002dac3a860daafd70ec47f6e52d87849dbbee061611
3
+ size 8573431808
Tessa-T1-14B-Q5_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b4061a54c99d97f6dd4451faf3425cb811353dabf6e3472cbacea176115a0d1
3
+ size 10266554368
Tessa-T1-14B-Q5_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e62a1d50588d9f38f3f4a5327977cb134c384ee9bce4c2b79939b493b5dd627d
3
+ size 10508873728
Tessa-T1-14B-Q5_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c0f1d7c9fe7c0070e22b59ef26ef59ac3dcad9e30588cf6b30a1b96a48242721
3
+ size 10266554368
Tessa-T1-14B-Q6_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:51ce9bb15a11003f9c17e45fc403b198d39d26f2440106e3ba7780cb81e0e75c
3
+ size 12124684288
Tessa-T1-14B-Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a45ac1c59719d0018a4b860a2f2fea2ca08c24e95a292297c230d21dc812073c
3
+ size 15701598208
Tessa-T1-14B-f16.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:260ba313da0f90cadc99676b81d1ee2030e30e67522fb04a130309a3cae92735
3
+ size 29547716608
config.json ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "Qwen2ForCausalLM"
4
+ ],
5
+ "attention_dropout": 0.0,
6
+ "bos_token_id": 151643,
7
+ "eos_token_id": 151645,
8
+ "hidden_act": "silu",
9
+ "hidden_size": 5120,
10
+ "initializer_range": 0.02,
11
+ "intermediate_size": 13824,
12
+ "max_position_embeddings": 120000,
13
+ "max_window_layers": 70,
14
+ "model_type": "qwen2",
15
+ "num_attention_heads": 40,
16
+ "num_hidden_layers": 48,
17
+ "num_key_value_heads": 8,
18
+ "pad_token_id": 151665,
19
+ "rms_norm_eps": 1e-06,
20
+ "rope_scaling": {
21
+ "factor": 3.662109375,
22
+ "rope_type": "linear",
23
+ "type": "linear"
24
+ },
25
+ "rope_theta": 1000000.0,
26
+ "sliding_window": null,
27
+ "tie_word_embeddings": false,
28
+ "torch_dtype": "bfloat16",
29
+ "transformers_version": "4.50.0.dev0",
30
+ "unsloth_fixed": true,
31
+ "unsloth_version": "2025.3.18",
32
+ "use_cache": true,
33
+ "use_sliding_window": false,
34
+ "vocab_size": 152064
35
+ }