You need to agree to share your contact information to access this model

This repository is publicly accessible, but you have to accept the conditions to access its files and content.

Log in or Sign Up to review the conditions and access this model content.

Model Card for Model TwinDoc/RedWhale-2-3B

meta-llama/Llama-3.2-3B λͺ¨λΈλ‘œλΆ€ν„° μ‚¬μ „ν•™μŠ΅ν•œ λͺ¨λΈμž…λ‹ˆλ‹€. μ‚¬μ „ν•™μŠ΅μ€ ν•œκ΅­μ–΄ Corpus둜 μ§„ν–‰ν•˜μ˜€μŠ΅λ‹ˆλ‹€.

Model Details

Model Description

  • Developed by: AgileSoda
  • Model type: Llama
  • Language(s) (NLP): ν•œκ΅­μ–΄
  • License: [More Information Needed]
  • Finetuned from model [optional]: TwinDoc/RedWhale-2-3B-Instruct
  • Foundation Model: meta-llama/Llama-3.2-3B

Model Sources [optional]

  • Repository: [More Information Needed]
  • Paper [optional]: [More Information Needed]
  • Demo [optional]: [More Information Needed]

Uses

RedWhale-2-3B λͺ¨λΈ μ‚¬μš© 방법은 meta-llama/Llama-3.2-3B λͺ¨λΈ μ‚¬μš© 방법과 λ™μΌν•©λ‹ˆλ‹€. μ‚¬μš©ν•˜κ³ μž ν•˜λŠ” μ„œλΉ™ μ—”μ§„μ˜ 곡식 λ¬Έμ„œλ₯Ό μ°Έκ³ ν•˜μ„Έμš”. λ‹€μŒμ€ μ˜ˆμ‹œμž…λ‹ˆλ‹€.

Direct Use

usage with Transformers μ˜ˆμ‹œ μ½”λ“œλŠ” transformers == 4.48.1μ—μ„œ μž‘μ„±λ˜μ—ˆμŠ΅λ‹ˆλ‹€.

from transformers import AutoModelForCausalLM,AutoTokenizer
import torch

loading_args = {"torch_dtype": torch.bfloat16, "device_map": "auto"} ## for multi gpu loading
model = AutoModelForCausalLM.from_pretrained("TwinDoc/RedWhale-2-3B",**loading_args)
tokenizer = AutoTokenizer.from_pretrained("TwinDoc/RedWhale-2-3B")

text = "λŒ€ν•œλ―Όκ΅­μ˜ μˆ˜λ„λŠ” "
inputs = tokenizer(text,return_tensors="pt")
outputs = model.generate(**inputs,max_new_tokens = 100)
>>> print(tokenizer.decode(outputs[0]))
"<|begin_of_text|>λŒ€ν•œλ―Όκ΅­μ˜ μˆ˜λ„λŠ” 4개의 μ‹œλ„ μ€‘μ—μ„œ κ°€μž₯ μž‘μ€ λ„μ‹œλ‘œ μ„œμšΈμ‹œκ°€ 605.2㎒λ₯Ό μ°¨μ§€ν•œλ‹€. μ„œμšΈμ‹œμ˜ 면적은 605.2㎒이며, κ·Έ μ€‘μ—μ„œ 222.2γŽ’κ°€ μ„œμšΈμ˜ 쀑심뢀인 μ’…λ‘œκ΅¬μ— 속해 μžˆλ‹€. μ„œμšΈμ‹œμ˜ 면적은 605.2㎒이며, κ·Έ μ€‘μ—μ„œ 222.2γŽ’κ°€ μ„œμšΈμ˜ 쀑심뢀인 μ’…λ‘œκ΅¬μ— 속해 μžˆλ‹€. μ„œμšΈμ‹œ"

Out-of-Scope Use

μ‚¬μ „ν•™μŠ΅λ§Œ μ§„ν–‰ν•œ λͺ¨λΈμ΄κΈ° λ•Œλ¬Έμ— Instruction을 λ”°λ₯΄λŠ” λŠ₯λ ₯은 μ—†μŠ΅λ‹ˆλ‹€. νŠΉμ • Task에 λ°”λ‘œ μ‚¬μš©ν•˜κΈ° λ³΄λ‹€λŠ” Fine-Tuning을 μœ„ν•œ Baseλͺ¨λΈλ‘œ μ‚¬μš©ν•˜λŠ” 것을 ꢌμž₯ν•©λ‹ˆλ‹€.

Training Details

Training Data

Compute Infrastructure

Hardware

  • H100 80GB * 1EA
Downloads last month
0
Safetensors
Model size
3.21B params
Tensor type
F32
Β·
BF16
Β·
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Collection including TwinDoc/RedWhale-2-3B