Model Card for Model ID
kakaocorp/kanana-nano-2.1b-instruct
๋ชจ๋ธ์ kuotient/gsm8k-ko
๋ฐ์ดํฐ์
์ผ๋ก COT GRPO๋ฅผ ํ์ต์ํจ LoRA ์ด๋ํฐ์
๋๋ค.
How to Get Started with the Model
import torch
from transformers import AutoModelForCausalLM, AutoTokenizer, TextStreamer
from peft import PeftModel
model_name = "kakaocorp/kanana-nano-2.1b-instruct"
peft_model_id = "rycont/kanana-2.1b-lora-reasoning"
base_model = AutoModelForCausalLM.from_pretrained(
model_name,
torch_dtype=torch.bfloat16,
trust_remote_code=True,
)
model = PeftModel.from_pretrained(base_model, peft_model_id).to("cuda")
tokenizer = AutoTokenizer.from_pretrained(model_name)
streamer = TextStreamer(tokenizer)
SYSTEM_PROMPT = """
You are a helpful AI assistant developed by Kakao. Respond in the following format:
<reasoning>
...
</reasoning>
<answer>
...
</answer>
"""
messages = [
{"role" : "system", "content" : SYSTEM_PROMPT},
{"role" : "user", "content" : f"์ฒ ์๊ฐ ํ ๋ณ์ ๊ธธ์ด๊ฐ 5km์ธ ์ ์ฌ๊ฐํ ๋ชจ์์ ๊ณต์์์ ๋๋ ๋ฅผ ๋ฐ๋ผ ๋๋ฌด๋ฅผ ์ฌ์ผ๋ ค๊ณ ํด. ๋๋ฌด ์ฌ์ด ๊ฐ๊ฒฉ์ 500m์ผ. ํ ๋๋ฌด๋ฅผ ์ฌ์ ๋ 17๋ฒ์ ์ฝ์ง์ด ํ์ํ๋ฐ, ๊ทผ๋ก๊ธฐ์ค๋ฒ์ ํ ์ฌ๋์ ์ธ์์์ 31๋ฒ์ ์ฝ์ง๋ฐ์ ๋ชปํด. ๊ทธ๋ ๋ค๋ฉด ์ฒ ์๊ฐ ๋๋ฌด์ฌ๊ธฐ๋ฅผ ์๋ฃํ๊ธฐ ์ํด์๋ ๋ช๋ช
์ ์ธ๋ถ๋ฅผ ์ถ๊ฐ๋ก ๊ณ ์ฉํด์ผ ํ ๊น?"},
{"role" : "system", "content" : "<reason> ์ฌํธํก ํ๊ณ , ์ฐจ๊ทผ์ฐจ๊ทผ ์๊ฐํด๋ณด์. ์ผ๋จ, "},
]
input_ids = tokenizer.apply_chat_template(
messages,
tokenize=True,
continue_final_message=True,
return_tensors="pt"
).to("cuda")
_ = model.eval()
# with model.disable_adapter():
with torch.no_grad():
output = model.generate(
input_ids,
max_new_tokens=1024,
streamer=streamer,
tokenizer=tokenizer,
stop_strings="</answer>"
)
print(tokenizer.decode(output[0]))
Framework versions
- PEFT 0.14.0
- Downloads last month
- 4
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support
HF Inference deployability: The model has no pipeline_tag.
Model tree for rycont/kanana-2.1b-lora-reasoning
Base model
kakaocorp/kanana-nano-2.1b-instruct