Blackroot commited on
Commit
28797a8
·
verified ·
1 Parent(s): 49dfdc1

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -0
README.md CHANGED
@@ -32,6 +32,8 @@ Stock for the "True Merge" -- This was a TIES Merge, the reasoning is explained
32
  - Sao10K/L3.3-70B-Euryale-v2.3
33
  - (Custom Base Model-Stock Soup -- Recipe Below)
34
 
 
 
35
 
36
  # Why a different approach?
37
  As some users had noted, particularly thanks to |GodZio| and The-Istar, the previous Mirai's instruct format was very unclear. Infact, when testing Llama-3 instruct format it seemed just broken, and, it was. Why? Well, the issue was with merging multiple models with different stopping tokens. I'll leave a tecnical explanation below for my assumption about why this happened. The long story short, I changed strategies for this model. It's very different, and expects the Llama-3 format to be used.
 
32
  - Sao10K/L3.3-70B-Euryale-v2.3
33
  - (Custom Base Model-Stock Soup -- Recipe Below)
34
 
35
+ One note here, I wasn't really sure how to state this in the huggingface tags. This model is actually THREE different merges. There's a base history merge, which was rolled into a base model merge, and you can see we merged the bases with our instruct models. Whew. I tried to give a thorough overview of model contributions, but not all of them contribute to the most direct merge.
36
+
37
 
38
  # Why a different approach?
39
  As some users had noted, particularly thanks to |GodZio| and The-Istar, the previous Mirai's instruct format was very unclear. Infact, when testing Llama-3 instruct format it seemed just broken, and, it was. Why? Well, the issue was with merging multiple models with different stopping tokens. I'll leave a tecnical explanation below for my assumption about why this happened. The long story short, I changed strategies for this model. It's very different, and expects the Llama-3 format to be used.