TroyDoesAI commited on
Commit
397897f
·
verified ·
1 Parent(s): 4b6ddb3

New blend, then trained on top with my blacksheep dataset

Browse files
README.md CHANGED
@@ -1,8 +1,48 @@
1
- ---
2
- license: apache-2.0
3
- ---
4
-
5
- Please Dont Download This Yet, It has only seen 1 Epoch of my experimental Dataset
6
- - This is not ready for use yet for RP, context length is only 8k currently.
7
-
8
- I am fine tuning on a 32k context length dataset so it will take me a couple days before its ready to be used.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model:
3
+ - mlabonne/NeuralDaredevil-7B
4
+ - fhai50032/RolePlayLake-7B
5
+ library_name: transformers
6
+ tags:
7
+ - mergekit
8
+ - merge
9
+
10
+ ---
11
+ # BlackSheep-RP
12
+
13
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
14
+
15
+ ## Merge Details
16
+ ### Merge Method
17
+
18
+ This model was merged using the SLERP merge method.
19
+
20
+ ### Models Merged
21
+
22
+ The following models were included in the merge:
23
+ * [mlabonne/NeuralDaredevil-7B](https://huggingface.co/mlabonne/NeuralDaredevil-7B)
24
+ * [fhai50032/RolePlayLake-7B](https://huggingface.co/fhai50032/RolePlayLake-7B)
25
+
26
+ ### Configuration
27
+
28
+ The following YAML configuration was used to produce this model:
29
+
30
+ ```yaml
31
+ slices:
32
+ - sources:
33
+ - model: mlabonne/NeuralDaredevil-7B
34
+ layer_range: [0, 32]
35
+ - model: fhai50032/RolePlayLake-7B
36
+ layer_range: [0, 32]
37
+ merge_method: slerp
38
+ base_model: mlabonne/NeuralDaredevil-7B
39
+ parameters:
40
+ t:
41
+ - filter: self_attn
42
+ value: [0, 0.5, 0.3, 0.7, 1]
43
+ - filter: mlp
44
+ value: [1, 0.5, 0.7, 0.3, 0]
45
+ - value: 0.5
46
+ dtype: bfloat16
47
+
48
+ ```
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "mlabonne/NeuralDaredevil-7B",
3
  "architectures": [
4
  "MistralForCausalLM"
5
  ],
 
1
  {
2
+ "_name_or_path": "BlackSheep-RP",
3
  "architectures": [
4
  "MistralForCausalLM"
5
  ],
mergekit_config.yml CHANGED
@@ -2,7 +2,7 @@ slices:
2
  - sources:
3
  - model: mlabonne/NeuralDaredevil-7B
4
  layer_range: [0, 32]
5
- - model: SanjiWatsuki/Kunoichi-DPO-v2-7B
6
  layer_range: [0, 32]
7
  merge_method: slerp
8
  base_model: mlabonne/NeuralDaredevil-7B
 
2
  - sources:
3
  - model: mlabonne/NeuralDaredevil-7B
4
  layer_range: [0, 32]
5
+ - model: fhai50032/RolePlayLake-7B
6
  layer_range: [0, 32]
7
  merge_method: slerp
8
  base_model: mlabonne/NeuralDaredevil-7B
model-00001-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bf0f1ffd685c36842849ae7fa27c8e4d9bc4d4095ec3dcab3a6790525ff93d6f
3
  size 4886547008
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bd3a55614bf5acdfe42c903599e5885fe70c1535943c5b417c09237be29d0581
3
  size 4886547008
model-00002-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1b5188777153ac6b930f738c15ad166b2fe3bd2adb2709bf10bb3852b74fc83e
3
  size 4915916176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dcc4dfc50f888b2cdb184f72d99fe4b8493248d6745caf020cc5f03d3c09204e
3
  size 4915916176
model-00003-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b75c0a61e826ed3a389572c8f3c86ae24d43bd5dfd1e5148a8f65e624bf31e6a
3
  size 4681034848
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9ad941b6ef97df0f20367f3f2b69b64657b093e72cfe77ff9de1e4cc4564c567
3
  size 4681034848