You need to agree to share your contact information to access this model

This repository is publicly accessible, but you have to accept the conditions to access its files and content.

Log in or Sign Up to review the conditions and access this model content.

Update @ 2025.03.14: The updated version, T3Q-qwen2.5-14b-v1.2-e2 release

Update @ 2025.03.14: The updated version, T3Q-qwen2.5-32b-v1.2-e2 release

Model Summary

T3Q-qwen2.5-14b-v1.0-e3 is a post-trained version of the Qwen/Qwen2.5-14B-Instruct-1M model.
(LoRA-8-4-0.0001-cosine-32-16 with train_data_v1.0)
image/png

Global Open LLM Leaderboard Performance

This model achieved 1st place in performance among models under 32b in the Global Open LLM Leaderboard.
image/png

Page Capture Page Capture

Quick Start

Here provides a code snippet with apply_chat_template to show you how to load the tokenizer and model and how to generate contents.

from transformers import AutoModelForCausalLM, AutoTokenizer

model_name = "JungZoona/T3Q-qwen2.5-14b-v1.0-e3"

model = AutoModelForCausalLM.from_pretrained(
    model_name,
    torch_dtype="auto",
    device_map="auto"
)
tokenizer = AutoTokenizer.from_pretrained(model_name)

prompt = "Give me a short introduction to large language model."
messages = [
    {"role": "user", "content": prompt}
]
text = tokenizer.apply_chat_template(
    messages,
    tokenize=False,
    add_generation_prompt=True
)
model_inputs = tokenizer([text], return_tensors="pt").to(model.device)

generated_ids = model.generate(
    **model_inputs,
    max_new_tokens=512
)
generated_ids = [
    output_ids[len(input_ids):] for input_ids, output_ids in zip(model_inputs.input_ids, generated_ids)
]

response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]
Downloads last month
0
Safetensors
Model size
14.8B params
Tensor type
BF16
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Model tree for xi0v/T3Q-qwen2.5-14b-v1.0-e3-archive

Base model

Qwen/Qwen2.5-14B
Finetuned
(22)
this model