sororicide-12B-Farer-Mell-Unslop

This is a merge of pre-trained language models created using mergekit.

Wayfarer is an interesting model, so I decided to merge it with Mag-Mell and Unslop-Nemo. That's all, just seemed like a good combination.

Eleventh model.

Merge Details

Merge Method

This model was merged using the NuSLERP merge method using TheDrummer/UnslopNemo-12B-v4 as a base.

Models Merged

The following models were included in the merge:

Configuration

The following YAML configuration was used to produce this model:

models:
  - model: LatitudeGames/Wayfarer-12B
    parameters:
      weight:
        - filter: self_attn
          value: [0.4, 0.5, 0.4]
        - filter: mlp
          value: [0.7, 0.5, 0.6, 0.7, 0.6]
        - value: [0.5]
  - model: inflatebot/MN-12B-Mag-Mell-R1
    parameters:
      weight:
        - filter: self_attn
          value: [0.6, 0.5, 0.6]
        - filter: mlp
          value: [0.3, 0.5, 0.4, 0.3, 0.4]
        - value: [0.5]
base_model: TheDrummer/UnslopNemo-12B-v4
merge_method: nuslerp
dtype: bfloat16
chat_template: "chatml"
tokenizer:
  source: union
parameters:
  normalize: true
  int8_mask: true

Downloads last month
43
Safetensors
Model size
12.2B params
Tensor type
BF16
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for redrix/sororicide-12B-Farer-Mell-Unslop

Collection including redrix/sororicide-12B-Farer-Mell-Unslop