Phi-4 SLERP Merge Model

Model Description

This is a merged language model created using the Spherical Linear Interpolation (SLERP) merge method, allowing for a smooth blend of features from both parent models across different layers. The merge optimizes reasoning, general knowledge, and task-specific performance by strategically interpolating attention and MLP components.


Merge Details

Merge Method:
The model was merged using SLERP (Spherical Linear Interpolation) rather than a traditional linear merge, ensuring a well-balanced combination of both source models while maintaining coherent weight transitions.

Base Model:

  • bunnycore/Phi-4-RR-Shoup (used as the primary base)

Models Merged

The following models were included in this merge:

  1. bunnycore/Phi-4-RR-Shoup (Primary base)
  2. bunnycore/Phi-4-Model-Stock-v4

Configuration

The following YAML configuration was used to produce this merged model:

slices:
- sources:
  - model: bunnycore/Phi-4-RR-Shoup
    layer_range:
    - 0
    - 32
  - model: bunnycore/Phi-4-Model-Stock-v4
    layer_range:
    - 0
    - 32
merge_method: slerp
base_model: bunnycore/Phi-4-RR-Shoup
parameters:
  t:
  - filter: self_attn
    value:
    - 0
    - 0.5
    - 0.3
    - 0.7
    - 1
  - filter: mlp
    value:
    - 1
    - 0.5
    - 0.7
    - 0.3
    - 0
  - value: 0.5
dtype: bfloat16
Downloads last month
93
Safetensors
Model size
6.65B params
Tensor type
F32
·
FP16
·
U8
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support