PEFT
Safetensors
German
llama
trl
sft
Generated from Trainer

LLäMmlein 7B Chat

This is an early preview of our instruction-tuned 7B model, trained using limited German-language resources. Please note that it is not the final version - we are actively working on improvements!

Find more details on our page and our preprint!

Example Usage

from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained("LSX-UniWue/LLaMmlein_7B_chat")
tokenizer = AutoTokenizer.from_pretrained("LSX-UniWue/LLaMmlein_7B_chat")
model = model.to("mps")

messages = [
    {
        "role": "user",
        "content": "Was sind die wichtigsten Sehenswürdigkeiten von Berlin?",
    },
]

chat = tokenizer.apply_chat_template(
    messages,
    return_tensors="pt",
    add_generation_prompt=True,
).to("mps")


print(
    tokenizer.decode(
        model.generate(
            chat,
            max_new_tokens=100,
            pad_token_id=tokenizer.pad_token_id,
            eos_token_id=tokenizer.eos_token_id,
            repetition_penalty=1.1,
        )[0],
        skip_special_tokens=False,
    )
)
Downloads last month
0
Safetensors
Model size
6.74B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for LSX-UniWue/LLaMmlein_7B_chat

Adapter
(1)
this model
Quantizations
2 models

Datasets used to train LSX-UniWue/LLaMmlein_7B_chat

Collection including LSX-UniWue/LLaMmlein_7B_chat