AnIma / Ocelot f_1

Update @ 2025.08.04: First release of malpyung_korean_culture_qna_sota

This model card corresponds to the 10.8B Instruct version of the Yanolja EEVE model.

Resources and Technical Documentation:

Citation

@misc {ai-AnIma/malpyung_korean_culture_qna_sota,
    author       = { {frcp, nebchi, DaKu00, philosokey-M} },
    title        = { malpyung_rag },
    year         = 2025,
    url          = { https://huggingface.co/ai-AnIma/malpyung_korean_culture_qna_sota },
    publisher    = { Hugging Face }
}

Model Developers: frcp, nebchi, DaKu00, philosokey-M

๐Ÿ“ ๊ณผ์ œ ๊ฐœ์š”: ๊ตญ๋ฆฝ๊ตญ์–ด์› ํ•œ๊ตญ๋ฌธํ™” ์งˆ์˜์‘๋‹ต ๋Œ€ํšŒ

์ด ๊ณผ์ œ๋Š” ํ•œ๊ตญ ์ „ํ†ต๋ฌธํ™”, ์—ญ์‚ฌ, ์‚ฌํšŒ, ๊ณผํ•™๊ธฐ์ˆ  ๋“ฑ ๋‹ค์–‘ํ•œ ๋ถ„์•ผ์— ๋Œ€ํ•œ ์ธ๊ณต์ง€๋Šฅ์˜ ํ•œ๊ตญ๋ฌธํ™” ์ง€์‹์„ ์ข…ํ•ฉ์ ์œผ๋กœ ํ‰๊ฐ€ํ•˜๊ธฐ ์œ„ํ•œ ๋ชฉ์ ์„ ๊ฐ€์ง‘๋‹ˆ๋‹ค.

  • ๋ฌธํ•ญ ์œ ํ˜•:
    • ์„ ๋‹คํ˜•: ์ •๋‹ต ๋ฒˆํ˜ธ(์ •์ˆ˜)๋ฅผ ์ถœ๋ ฅ
    • ๋‹จ๋‹ตํ˜•: ๋‹จ์–ด, ๊ตฌ, ๊ธฐํ˜ธ ๋“ฑ 5์–ด์ ˆ ์ดํ•˜์˜ ๋‹จ๋‹ต ์ถœ๋ ฅ
    • ์„œ์ˆ ํ˜•: 300~500์ž ๋‚ด์™ธ์˜ ๋ฌธ์žฅ ์ƒ์„ฑ---

๐Ÿง  Model Details

  • ํ•™์Šต ๋ฐฉ์‹: ๊ณ ํ’ˆ์งˆ ํ•œ๊ตญ์–ด ์ธ์ŠคํŠธ๋Ÿญ์…˜ ๋ฐ์ดํ„ฐ์…‹์„ ํ™œ์šฉํ•œ Full fine-tuning

  • Alignment Tech:

    • WizardLM์˜ Evol-Instruct ํ”„๋ ˆ์ž„์›Œํฌ๋ฅผ ๋„์ž…ํ•˜์—ฌ seed data ๊ธฐ๋ฐ˜, Depth/Breadth๊ฐ€ ํ’๋ถ€ํ•œ ๋‹ค์–‘ํ•œ ์œ ํ˜•์˜ instruction dataset์„ ์ƒ์„ฑํ•˜์˜€์Šต๋‹ˆ๋‹ค.
    • ๋ฐ์ดํ„ฐ ์ „์ฒ˜๋ฆฌ ๋‹จ๊ณ„์—์„œ SimHash ๊ธฐ๋ฐ˜ ์ค‘๋ณต ์ œ๊ฑฐ๋ฅผ ์ˆ˜ํ–‰ํ•˜์—ฌ, ํ•™์Šต ๋ฐ์ดํ„ฐ์˜ ๋‹ค์–‘์„ฑ์„ ํ™•๋ณดํ•˜๊ณ  ๊ณผ์ ํ•ฉ์„ ๋ฐฉ์ง€ํ•˜์˜€์Šต๋‹ˆ๋‹ค.
    • FineWeb ๊ธฐ๋ฐ˜ ํ’ˆ์งˆ ํ•„ํ„ฐ๋ง์„ ์ ์šฉํ•˜์—ฌ, ๋…ธ์ด์ฆˆ๊ฐ€ ๋งŽ๊ฑฐ๋‚˜ ์ •๋ณด์„ฑ์ด ๋–จ์–ด์ง€๋Š” ๋ฐ์ดํ„ฐ๋ฅผ ์ œ๊ฑฐํ•จ์œผ๋กœ์จ ์‘๋‹ต์˜ ๋ช…ํ™•์„ฑ๊ณผ ์ •๋ฐ€๋„๋ฅผ ๋†’์˜€์Šต๋‹ˆ๋‹ค.
    • Curriculum Learning ์ „๋žต์„ ํ™œ์šฉํ•˜์—ฌ, ์‰ฌ์šด ๋ฌธ์ œ๋ถ€ํ„ฐ ์ ์ง„์ ์œผ๋กœ ๋ณต์žกํ•œ ๋ฌธ์ œ๋กœ ํ•™์Šต ๋‚œ์ด๋„๋ฅผ ์กฐ์ ˆํ•˜๋ฉฐ ๋ชจ๋ธ์˜ ์ผ๋ฐ˜ํ™” ์„ฑ๋Šฅ์„ ํ–ฅ์ƒ์‹œ์ผฐ์Šต๋‹ˆ๋‹ค.

๐Ÿ› ๏ธ ๋ชจ๋ธ ์‚ฌ์šฉ ์˜ˆ์‹œ

from transformers import AutoTokenizer, AutoModelForCausalLM, pipeline

tokenizer = AutoTokenizer.from_pretrained("ai-AnIma/malpyung_korean_language_rag_sota")
model = AutoModelForCausalLM.from_pretrained("ai-AnIma/malpyung_korean_language_rag_sota, device_map="auto")

pipe = pipeline("text-generation", model=model, tokenizer=tokenizer, max_new_tokens=4096) 

messages = [
    {"role": "user", "content": "2024๋…„ ๊ธฐ์ค€ ํ•œ๊ตญ์˜ ์ผ๋ฐ˜ ๊ฐ€์ •์—์„œ ํƒœ์–‘๊ด‘ ๋ฐœ์ „๊ธฐ๋ฅผ ์„ค์น˜ํ•˜๊ณ  ์‹ถ์œผ๋ฉด ์–ด๋–ป๊ฒŒ ํ•ด์•ผ ํ•˜๋‚˜์š”?"}]

prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)

outputs = pipe(prompt, temperature=0.2)
print(outputs[0]["generated_text"][len(prompt):])


results

์ผ๋ฐ˜ ๊ฐ€์ •์—์„œ ํƒœ์–‘๊ด‘ ๋ฐœ์ „๊ธฐ๋ฅผ ์„ค์น˜ํ•˜๊ธฐ ์œ„ํ•ด์„œ๋Š” ํ•œ๊ตญ์—๋„ˆ์ง€๊ณต๋‹จ '๊ทธ๋ฆฐํ™ˆ'์—์„œ ์‹œ๊ณต์—…์ฒด๋ฅผ ์„ ํƒํ•œ ๋’ค ์—…์ฒด๊ฐ€ ๊ฐ€์ •์„ ๋ฐฉ๋ฌธํ•˜์—ฌ ์„ค์น˜ ๊ฐ€๋Šฅ ์—ฌ๋ถ€ ๋ฐ ์ ์ ˆ์„ฑ์„ ํ‰๊ฐ€ํ•˜๋Š” ๊ณผ์ •์ด ๋จผ์ € ์ด๋ฃจ์–ด์ ธ์•ผ ํ•œ๋‹ค. ์ ์ ˆ์„ฑ ํ‰๊ฐ€์— ํ†ต๊ณผ๋˜๋ฉด ์—…์ฒด๊ฐ€ ์‚ฌ์—…์‹ ์ฒญ์„œ๋ฅผ ๊ทธ๋ฆฐํ™ˆ์— ์ œ์ถœํ•˜๊ณ  ๊ณต๋‹จ์˜ ๊ฒ€ํ† ๋ฅผ ๋ฐ›๊ฒŒ ๋œ๋‹ค. ๊ณต๋‹จ์˜ ์Šน์ธ์ด ๋‚œ ํ›„ ์‹ ์ฒญ์ž๊ฐ€ ์˜ˆ์น˜๊ธˆ์„ ๋‚ฉ๋ถ€ํ•˜๊ฒŒ ๋˜๋ฉฐ ๊ทธ ํ›„ ์„ค๋น„๋ฅผ ์‹œ์ž‘ํ•œ๋‹ค. ์„ค๋น„๋Š” ์„ ์ •ํ•œ ์—…์ฒด์— ์˜ํ•ด ์ด๋ฃจ์–ด์ง„๋‹ค. ์„ค์น˜๊ฐ€ ์™„๋ฃŒ๋˜๋ฉด ์„ค์น˜ ํ™•์ธ ์ ˆ์ฐจ๋ฅผ ๊ฑฐ์ณ ์‹ ์ฒญ์ž์—๊ฒŒ ๋ณด์กฐ๊ธˆ์ด ์ง€๊ธ‰๋œ๋‹ค. ์ดˆ๊ธฐ ์„ค์น˜ ๋น„์šฉ์ด ๋ถ€๋‹ด๋˜๋Š” ๊ฒฝ์šฐ ๋Œ€์—ฌ ์‚ฌ์—…์ž๋ฅผ ํ†ตํ•ด ํƒœ์–‘๊ด‘ ๋ฐœ์ „๊ธฐ๋ฅผ ๋Œ€์—ฌํ•˜๋Š” ๋ฐฉ๋ฒ•๋„ ์žˆ๋‹ค.

Evaluation Results - ๊ตญ๋ฆฝ๊ตญ์–ด์› ํ•œ๊ตญ๋ฌธํ™” ์งˆ์˜์‘๋‹ต ๋Œ€ํšŒ

์ด ๋ชจ๋ธ์€ ๊ตญ๋ฆฝ๊ตญ์–ด์› ์ฃผ๊ด€ ํ•œ๊ตญ๋ฌธํ™” ์งˆ์˜์‘๋‹ต ๋Œ€ํšŒ์— ์ฐธ์—ฌํ•˜์—ฌ ๋‹ค์Œ๊ณผ ๊ฐ™์€ ์„ฑ๋Šฅ์„ ๊ธฐ๋กํ•˜์˜€์Šต๋‹ˆ๋‹ค.

๋ชจ๋ธ๋ช… ํ‰๊ฐ€ ์ ์ˆ˜ Accuracy Exact Match ROUGE-1 BERTScore BLEURT ์ œ์ถœ์ผ์‹œ
n_try01 79.1005 92.75 86.27 45.10 73.45 56.29 2025.07.17 15:24
Downloads last month
12
Safetensors
Model size
11B params
Tensor type
BF16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support