Input Models input text only.
Output Models generate text only.
Base Model beomi/Yi-Ko-6B
Training Dataset
Implementation Code
from transformers import AutoModelForCausalLM, AutoTokenizer
import torch
repo = "ifuseok/yi-ko-playtus-instruct-v0.2"
OpenOrca = AutoModelForCausalLM.from_pretrained(
repo,
return_dict=True,
torch_dtype=torch.float16,
device_map='auto'
)
OpenOrca_tokenizer = AutoTokenizer.from_pretrained(repo)
Prompt Example
<|system|>
μμ€ν
λ©μμ§ μ
λλ€. <|endoftext|>
<|user|>
μ μ μ
λλ€.<|endoftext|>
<|assistant|>
μ΄μμ€ν΄νΈ μ
λλ€.<|endoftext|>
- Downloads last month
- 2,029
Inference Providers
NEW
This model is not currently available via any of the supported Inference Providers.