Io2007 commited on
Commit
da8ebb4
·
verified ·
1 Parent(s): a074b70

Upload folder using huggingface_hub

Browse files
README.md CHANGED
@@ -1,37 +1,43 @@
1
  ---
2
  base_model:
3
- - google/gemma-3-1b-pt
4
- - huihui-ai/gemma-3-1b-it-abliterated
5
- - prithivMLmods/gemma-3-1b-it-abliterated
6
  - lunahr/gemma-3-1b-it-abliterated
 
7
  - google/gemma-3-1b-pt
 
 
8
  tags:
9
- - merge
10
  - mergekit
11
- - lazymergekit
12
- - google/gemma-3-1b-pt
13
- - huihui-ai/gemma-3-1b-it-abliterated
14
- - prithivMLmods/gemma-3-1b-it-abliterated
15
- - lunahr/gemma-3-1b-it-abliterated
16
  ---
 
17
 
18
- # gemma-3-1b-big
19
 
20
- gemma-3-1b-big is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing):
21
- * [google/gemma-3-1b-pt](https://huggingface.co/google/gemma-3-1b-pt)
22
- * [huihui-ai/gemma-3-1b-it-abliterated](https://huggingface.co/huihui-ai/gemma-3-1b-it-abliterated)
23
- * [prithivMLmods/gemma-3-1b-it-abliterated](https://huggingface.co/prithivMLmods/gemma-3-1b-it-abliterated)
 
 
 
 
24
  * [lunahr/gemma-3-1b-it-abliterated](https://huggingface.co/lunahr/gemma-3-1b-it-abliterated)
 
25
  * [google/gemma-3-1b-pt](https://huggingface.co/google/gemma-3-1b-pt)
 
 
 
26
 
27
- ## 🧩 Configuration
28
 
29
  ```yaml
 
30
  dtype: bfloat16
31
  merge_method: passthrough
32
  slices:
33
  - sources:
34
- - layer_range: [0, 5]
35
  model: google/gemma-3-1b-pt
36
  - sources:
37
  - layer_range: [5, 10]
@@ -43,31 +49,7 @@ slices:
43
  - layer_range: [15, 20]
44
  model: lunahr/gemma-3-1b-it-abliterated
45
  - sources:
46
- - layer_range: [20, 25]
47
  model: google/gemma-3-1b-pt
48
- ```
49
-
50
- ## 💻 Usage
51
-
52
- ```python
53
- !pip install -qU transformers accelerate
54
-
55
- from transformers import AutoTokenizer
56
- import transformers
57
- import torch
58
 
59
- model = "Io2007/gemma-3-1b-big"
60
- messages = [{"role": "user", "content": "What is a large language model?"}]
61
-
62
- tokenizer = AutoTokenizer.from_pretrained(model)
63
- prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
64
- pipeline = transformers.pipeline(
65
- "text-generation",
66
- model=model,
67
- torch_dtype=torch.float16,
68
- device_map="auto",
69
- )
70
-
71
- outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
72
- print(outputs[0]["generated_text"])
73
- ```
 
1
  ---
2
  base_model:
 
 
 
3
  - lunahr/gemma-3-1b-it-abliterated
4
+ - prithivMLmods/gemma-3-1b-it-abliterated
5
  - google/gemma-3-1b-pt
6
+ - huihui-ai/gemma-3-1b-it-abliterated
7
+ library_name: transformers
8
  tags:
 
9
  - mergekit
10
+ - merge
11
+
 
 
 
12
  ---
13
+ # merge
14
 
15
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
16
 
17
+ ## Merge Details
18
+ ### Merge Method
19
+
20
+ This model was merged using the Passthrough merge method.
21
+
22
+ ### Models Merged
23
+
24
+ The following models were included in the merge:
25
  * [lunahr/gemma-3-1b-it-abliterated](https://huggingface.co/lunahr/gemma-3-1b-it-abliterated)
26
+ * [prithivMLmods/gemma-3-1b-it-abliterated](https://huggingface.co/prithivMLmods/gemma-3-1b-it-abliterated)
27
  * [google/gemma-3-1b-pt](https://huggingface.co/google/gemma-3-1b-pt)
28
+ * [huihui-ai/gemma-3-1b-it-abliterated](https://huggingface.co/huihui-ai/gemma-3-1b-it-abliterated)
29
+
30
+ ### Configuration
31
 
32
+ The following YAML configuration was used to produce this model:
33
 
34
  ```yaml
35
+
36
  dtype: bfloat16
37
  merge_method: passthrough
38
  slices:
39
  - sources:
40
+ - layer_range: [0, 7]
41
  model: google/gemma-3-1b-pt
42
  - sources:
43
  - layer_range: [5, 10]
 
49
  - layer_range: [15, 20]
50
  model: lunahr/gemma-3-1b-it-abliterated
51
  - sources:
52
+ - layer_range: [19, 26]
53
  model: google/gemma-3-1b-pt
 
 
 
 
 
 
 
 
 
 
54
 
55
+ ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
added_tokens.json CHANGED
@@ -1,3 +1,3 @@
1
- {
2
- "<image_soft_token>": 262144
3
- }
 
1
+ {
2
+ "<image_soft_token>": 262144
3
+ }
config.json CHANGED
@@ -7,10 +7,7 @@
7
  "attn_logit_softcapping": null,
8
  "bos_token_id": 2,
9
  "cache_implementation": "hybrid",
10
- "eos_token_id": [
11
- 1,
12
- 106
13
- ],
14
  "final_logit_softcapping": null,
15
  "head_dim": 256,
16
  "hidden_activation": "gelu_pytorch_tanh",
@@ -20,7 +17,7 @@
20
  "max_position_embeddings": 32768,
21
  "model_type": "gemma3_text",
22
  "num_attention_heads": 4,
23
- "num_hidden_layers": 25,
24
  "num_key_value_heads": 1,
25
  "pad_token_id": 0,
26
  "query_pre_attn_scalar": 256,
 
7
  "attn_logit_softcapping": null,
8
  "bos_token_id": 2,
9
  "cache_implementation": "hybrid",
10
+ "eos_token_id": 1,
 
 
 
11
  "final_logit_softcapping": null,
12
  "head_dim": 256,
13
  "hidden_activation": "gelu_pytorch_tanh",
 
17
  "max_position_embeddings": 32768,
18
  "model_type": "gemma3_text",
19
  "num_attention_heads": 4,
20
+ "num_hidden_layers": 29,
21
  "num_key_value_heads": 1,
22
  "pad_token_id": 0,
23
  "query_pre_attn_scalar": 256,
mergekit_config.yml CHANGED
@@ -3,7 +3,7 @@ dtype: bfloat16
3
  merge_method: passthrough
4
  slices:
5
  - sources:
6
- - layer_range: [0, 5]
7
  model: google/gemma-3-1b-pt
8
  - sources:
9
  - layer_range: [5, 10]
@@ -15,5 +15,5 @@ slices:
15
  - layer_range: [15, 20]
16
  model: lunahr/gemma-3-1b-it-abliterated
17
  - sources:
18
- - layer_range: [20, 25]
19
  model: google/gemma-3-1b-pt
 
3
  merge_method: passthrough
4
  slices:
5
  - sources:
6
+ - layer_range: [0, 7]
7
  model: google/gemma-3-1b-pt
8
  - sources:
9
  - layer_range: [5, 10]
 
15
  - layer_range: [15, 20]
16
  model: lunahr/gemma-3-1b-it-abliterated
17
  - sources:
18
+ - layer_range: [19, 26]
19
  model: google/gemma-3-1b-pt
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a6c7b710fba668105304f0a7faf1148bce16cdbf8eeab6b5d0253ab127147527
3
- size 1946125472
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7818ace264d1dceb7013007417abceaa79e8ad8cb10d28d24614de79337d19ee
3
+ size 2160868424
special_tokens_map.json CHANGED
@@ -1,27 +1,33 @@
1
- {
2
- "boi_token": "<start_of_image>",
3
- "bos_token": {
4
- "content": "<bos>",
5
- "lstrip": false,
6
- "normalized": false,
7
- "rstrip": false,
8
- "single_word": false
9
- },
10
- "eoi_token": "<end_of_image>",
11
- "eos_token": {
12
- "content": "<eos>",
13
- "lstrip": false,
14
- "normalized": false,
15
- "rstrip": false,
16
- "single_word": false
17
- },
18
- "image_token": "<image_soft_token>",
19
- "pad_token": "<eos>",
20
- "unk_token": {
21
- "content": "<unk>",
22
- "lstrip": false,
23
- "normalized": false,
24
- "rstrip": false,
25
- "single_word": false
26
- }
27
- }
 
 
 
 
 
 
 
1
+ {
2
+ "boi_token": "<start_of_image>",
3
+ "bos_token": {
4
+ "content": "<bos>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false
9
+ },
10
+ "eoi_token": "<end_of_image>",
11
+ "eos_token": {
12
+ "content": "<eos>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false
17
+ },
18
+ "image_token": "<image_soft_token>",
19
+ "pad_token": {
20
+ "content": "<pad>",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false
25
+ },
26
+ "unk_token": {
27
+ "content": "<unk>",
28
+ "lstrip": false,
29
+ "normalized": false,
30
+ "rstrip": false,
31
+ "single_word": false
32
+ }
33
+ }
tokenizer_config.json CHANGED
The diff for this file is too large to render. See raw diff