Marcoroni-7b-DPO-Merge

Marcoroni-7b-DPO-Merge is a merge of the following models using mergekit and inspired by Maxime Labonne's work:

🧩 Configuration

models:
  - model: madatnlp/marcoroni-7b-v3-safetensor
    # no parameters necessary for base model
  - model: fblgit/UNA-TheBeagle-7b-v1
    parameters:
      density: 0.3
      weight: 0.5
  - model: udkai/Turdus
    parameters:
      density: 0.7
      weight: 0.3
merge_method: ties
base_model: madatnlp/marcoroni-7b-v3-safetensor
parameters:
  normalize: true
dtype: float16

πŸ’» Example Python Code

from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline

model_name_or_path = "nfaheem/Marcoroni-7b-DPO-Merge"
model = AutoModelForCausalLM.from_pretrained(model_name_or_path,
                                             device_map="auto",
                                             revision="main")

tokenizer = AutoTokenizer.from_pretrained(model_name_or_path, use_fast=True)

prompt = "Write a story about llamas"
system_message = "You are a story writing assistant"
prompt_template=f'''{prompt}
'''

print("\n\n*** Generate:")

input_ids = tokenizer(prompt_template, return_tensors='pt').input_ids.cuda()
output = model.generate(inputs=input_ids, temperature=0.7, do_sample=True, top_p=0.95, top_k=40, max_new_tokens=512)
print(tokenizer.decode(output[0]))

# Inference can also be done using transformers' pipeline

print("*** Pipeline:")
pipe = pipeline(
    "text-generation",
    model=model,
    tokenizer=tokenizer,
    max_new_tokens=512,
    do_sample=True,
    temperature=0.7,
    top_p=0.95,
    top_k=40,
    repetition_penalty=1.1
)

print(pipe(prompt_template)[0]['generated_text'])

πŸ“‹ Summary Eval:

Average ARC HellaSwag MMLU TruthfulQA Winogrande GSM8K
74.9 73.04 88.8 64.24 70.47 85.24 67.63

πŸ“ˆ Huggingface Leaderboard

It's Ranked # 1 on HuggingFace Leaderboard among around 13B parameters (01/15/2024)

Model Average ARC HellaSwag MMLU Truthful QA Winogrande GSM8K
nfaheem/Marcoroni-7b-DPO-Merge 74.9 73.04 88.8 64.24 70.47 85.24 67.63
mlabonne/Beagle14-7b 74.76 72.95 87.95 64.7 68.38 82.64 71.42
udkai/Turdus 74.66 73.38 88.56 64.52 67.11 86.66 67.7
CultriX/MergeTrix-7B 74.33 72.24 87.84 64.88 66.27 83.5 71.19
fblgit/UNA-TheBeagle-7b-v1 73.87 73.04 88 63.48 69.85 82.16 66.72

image/png

Downloads last month
1,754
Safetensors
Model size
7.24B params
Tensor type
FP16
Β·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Model tree for nfaheem/Marcoroni-7b-DPO-Merge

Adapters
1 model
Quantizations
1 model

Spaces using nfaheem/Marcoroni-7b-DPO-Merge 17