gemma-2-9b-HangulFixer

  • Base Model: unsloth/gemma-2-9b-bnb-4bit
  • Developed by: SeongeonKim
  • Base Model License: Apache 2.0
  • Finetuned Model License: CC BY-NC 4.0 (๋น„์˜๋ฆฌ์  ์‚ฌ์šฉ๋งŒ ํ—ˆ์šฉ)
  • Language: Korean (ko)
  • Tags:
    • text-generation-inference
    • transformers
    • unsloth
    • gemma2
    • trl
    • sft

๋ชจ๋ธ ์„ค๋ช…

gemma-2-9b-HangulFixer๋Š” Unsloth์™€ Hugging Face์˜ TRL ๋ผ์ด๋ธŒ๋Ÿฌ๋ฆฌ๋ฅผ ์‚ฌ์šฉํ•ด fine-tuningํ•œ ํ…์ŠคํŠธ ์ƒ์„ฑ ๋ชจ๋ธ์ž…๋‹ˆ๋‹ค.
์ด ๋ชจ๋ธ์€ ๋‚œ๋…ํ™”๋œ ํ•œ๊ธ€ ํ˜ธํ…” ๋ฆฌ๋ทฐ๋ฅผ ์›๋ž˜์˜ ๋ช…ํ™•ํ•˜๊ณ  ์ž์—ฐ์Šค๋Ÿฌ์šด ๋ฆฌ๋ทฐ๋กœ ๋ณต์›ํ•˜๊ธฐ ์œ„ํ•ด ์„ค๊ณ„๋˜์—ˆ์Šต๋‹ˆ๋‹ค.


๋ฐ์ดํ„ฐ ๋ฐฐ๊ฒฝ

ํ•œ๊ตญ ์ˆ™์†Œ ์˜ˆ์•ฝ ์‚ฌ์ดํŠธ์—์„œ๋Š” ๋ถ€์ •์ ์ธ ๋ฆฌ๋ทฐ๊ฐ€ ์‚ญ์ œ๋  ์ˆ˜ ์žˆ๋Š” ์šฐ๋ ค๊ฐ€ ์žˆ์Šต๋‹ˆ๋‹ค. ์ด๋ฅผ ํ”ผํ•˜๋ฉด์„œ ํ•œ๊ตญ ์‚ฌ์šฉ์ž๋“ค์—๊ฒŒ๋งŒ ์œ ์šฉํ•œ ์ •๋ณด๋ฅผ ์ „๋‹ฌํ•˜๊ธฐ ์œ„ํ•ด ์ผ๋ถ€ ์‚ฌ์šฉ์ž๋“ค์€ ๋ฆฌ๋ทฐ๋ฅผ ๋‚œ๋…ํ™”ํ•˜๋Š” ๋ฐฉ์‹์„ ์‚ฌ์šฉํ•ด ์†”์งํ•œ ํ›„๊ธฐ๋ฅผ ๋‚จ๊ธฐ๊ณ ์ž ํ•ฉ๋‹ˆ๋‹ค.
์ด ๋ฐฉ์‹์€ ํ•œ๊ตญ ์‚ฌ์šฉ์ž๋“ค์—๊ฒŒ๋Š” ์†”์งํ•œ ์ •๋ณด๋ฅผ ์ œ๊ณตํ•˜๋Š” ์žฅ์ ์ด ์žˆ์ง€๋งŒ, ์ˆ™์†Œ ์šด์˜์ž์™€ ๋ฐฉ๋ฌธ๊ฐ ๊ฐ„ ํ”ผ๋“œ๋ฐฑ ์ „๋‹ฌ์„ ์–ด๋ ต๊ฒŒ ๋งŒ๋“ค์–ด ์„œ๋น„์Šค ํ’ˆ์งˆ ํ–ฅ์ƒ์— ์žฅ์• ๊ฐ€ ๋  ์ˆ˜ ์žˆ์Šต๋‹ˆ๋‹ค.


ํ•™์Šต ๋ฐ์ดํ„ฐ

์ด ๋ชจ๋ธ์€ SeongeonKim/ko-scrambled_v0.1 ๋ฐ์ดํ„ฐ์…‹์˜ ํ•™์Šต ๋ฐ์ดํ„ฐ๋ฅผ ์‚ฌ์šฉํ•ด ํ•™์Šต๋˜์—ˆ์Šต๋‹ˆ๋‹ค.
์ด 11,263๊ฐœ์˜ ๋‚œ๋…ํ™”๋œ ํ•œ๊ธ€ ํ˜ธํ…” ๋ฆฌ๋ทฐ์™€ ๊ทธ์— ์ƒ์‘ํ•˜๋Š” ๋ณต์›๋œ ์›๋ณธ ๋ฆฌ๋ทฐ๋ฅผ ์‚ฌ์šฉํ•ด fine-tuning์ด ์ง„ํ–‰๋˜์—ˆ์Šต๋‹ˆ๋‹ค.
์ด ๋ฐ์ดํ„ฐ์…‹์€ ๋ฆฌ๋ทฐ์˜ ๋‚œ๋…ํ™”๋œ ํ˜•ํƒœ(input)์™€ ๋ณต์›๋œ ํ˜•ํƒœ(output)๋กœ ๊ตฌ์„ฑ๋˜์–ด ์žˆ์Šต๋‹ˆ๋‹ค.

์˜ˆ์ œ ๋ฐ์ดํ„ฐ:

  • Input: ์‰ญ๋”ฑ ์œ• ํ„บ๋ ค์œค ๊ตผ๋ฉ•์›จ ๋ท”ํ–… ์™„์ฉ ๋ณ„๋ฃŒ. 2๋นก 3์œŒ์—ฃ 88๋งŽ ์–ธ ์ผค์ณ…ํ–‡๋ˆˆ๋Žจ...
  • Output: ์‹ํƒ ์œ„ ๋”๋Ÿฌ์šด ๊ธˆ์•ก์— ๋น„ํ•ด ์™„์ „ ๋ณ„๋กœ. 2๋ฐ• 3์ผ์— 88๋งŒ ์› ๊ฒฐ์ œํ–ˆ๋Š”๋ฐ...

๋ชจ๋ธ ์„ฑ๋Šฅ

  • Base Model: unsloth/gemma-2-9b-bnb-4bit๋ฅผ ๊ธฐ๋ฐ˜์œผ๋กœ ํ•™์Šต๋˜์—ˆ์Šต๋‹ˆ๋‹ค.
  • Training Efficiency: Unsloth๋ฅผ ์‚ฌ์šฉํ•ด 2๋ฐฐ ๋” ๋น ๋ฅธ ์†๋„๋กœ ํ•™์Šต ์™„๋ฃŒ.
  • Target Task: ํ…์ŠคํŠธ ๋ณต์›
  • Language: ํ•œ๊ธ€

๊ฐœ์„  ์‚ฌํ•ญ (v0.1)

  • ์–ผ๋ฆฌ ์Šคํ† ํ•‘(Early Stopping) ๋„์ž…: EarlyStoppingCallback์„ ์ถ”๊ฐ€ํ•˜์—ฌ ํ‰๊ฐ€ ์†์‹ค(eval_loss) ๊ธฐ์ค€์œผ๋กœ ํ•™์Šต ๊ณผ์ •์„ ์ž๋™ ์ค‘๋‹จ, ๊ณผ์ ํ•ฉ ๋ฐฉ์ง€.
  • ํ•™์Šต ํšจ์œจ์„ฑ ํ–ฅ์ƒ: ๋ฐ์ดํ„ฐ ๋ณ‘๋ ฌ ์ฒ˜๋ฆฌ(dataset_num_proc=2) ๋ฐ ๋ช…์‹œ์  ํŒŒ๋ผ๋ฏธํ„ฐ ์„ค์ •์„ ํ†ตํ•ด ํ•™์Šต ์•ˆ์ •์„ฑ๊ณผ ์†๋„ ๊ฐœ์„ .
  • ํ‰๊ฐ€ ์ง€ํ‘œ ๋ชจ๋‹ˆํ„ฐ๋ง: ํ•™์Šต ํ’ˆ์งˆ์„ ๋†’์ด๊ธฐ ์œ„ํ•ด eval_loss ๋ชจ๋‹ˆํ„ฐ๋ง ์ถ”๊ฐ€.

์‚ฌ์šฉ ๋ฐฉ๋ฒ•

์•„๋ž˜๋Š” ๋ชจ๋ธ์„ ์‚ฌ์šฉํ•˜๋Š” ๊ธฐ๋ณธ์ ์ธ ์˜ˆ์ œ์ž…๋‹ˆ๋‹ค:

from transformers import AutoModelForCausalLM, AutoTokenizer

# ๋ชจ๋ธ ๋กœ๋“œ
model_name = "SeongeonKim/gemma-2-9b-HangulFixer"
model = AutoModelForCausalLM.from_pretrained(model_name).to("cuda")
tokenizer = AutoTokenizer.from_pretrained(model_name)

# ํ…Œ์ŠคํŠธ ์ž…๋ ฅ
input_text = "๋…€๋ฎจ๋„’๋ญ… ๋งŒ์ฃก์ˆญ๋Ÿฌ์œค ํšจํ…œ๋คผ์—์˜ค. ํ‘ธ์‹ธ๋ˆผ ์˜ฐ๋ฉด ์ฝ• ์ธ„์ฉํ•™๊ผฌ ์‹ถ์€ ์ฝง์‰ฐ์›จ์˜ค."
inputs = tokenizer(input_text, return_tensors="pt").to("cuda")
outputs = model.generate(**inputs, max_new_tokens=128)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))

์ƒ์—…์  ์ด์šฉ ๊ด€๋ จ ์ œํ•œ ์‚ฌํ•ญ

์ด ๋ชจ๋ธ์€ ํ•™์Šต ๋ฐ์ดํ„ฐ์˜ ๋ผ์ด์„ ์Šค์— ๋”ฐ๋ผ CC BY-NC 4.0 ๋ผ์ด์„ ์Šค๋ฅผ ๋”ฐ๋ฆ…๋‹ˆ๋‹ค.
์ด ๋ผ์ด์„ ์Šค๋Š” ๋น„์˜๋ฆฌ์  ์‚ฌ์šฉ๋งŒ ํ—ˆ์šฉํ•˜๋ฉฐ, ์ƒ์—…์  ์šฉ๋„๋กœ ์‚ฌ์šฉํ•˜๋ ค๋ฉด ๋ณ„๋„์˜ ํ—ˆ๊ฐ€๊ฐ€ ํ•„์š”ํ•ฉ๋‹ˆ๋‹ค.
๋ฐ์ดํ„ฐ๋ฅผ ์ด์šฉํ•˜๋Š” ๊ฒฝ์šฐ, ์—ฐ๊ตฌ ๋ชฉ์ ์œผ๋กœ ์‚ฌ์šฉํ•˜๊ฑฐ๋‚˜ ํ•™์Šต ๋ฐ์ดํ„ฐ์˜ ์ถœ์ฒ˜์™€ ๋ผ์ด์„ ์Šค๋ฅผ ๋ช…ํ™•ํžˆ ํ‘œ๊ธฐํ•ด์•ผ ํ•ฉ๋‹ˆ๋‹ค.

  • ๋ผ์ด์„ ์Šค ์„ธ๋ถ€์‚ฌํ•ญ: CC BY-NC 4.0

๋ชจ๋ธ ๊ฐœ๋ฐœ ๊ด€๋ จ ์ฐธ๊ณ  ์‚ฌํ•ญ

  • ์‚ฌ์šฉ๋œ ๊ธฐ์ˆ : Unsloth์™€ Hugging Face์˜ TRL(SFT) ์‚ฌ์šฉ
  • ๋ฐ์ดํ„ฐ ์ถœ์ฒ˜: SeongeonKim/ko-scrambled_v0.1 ๋ฐ์ดํ„ฐ์…‹
  • ์‚ฌ์šฉ ์ œํ•œ: ๋ฐ์ดํ„ฐ์˜ ๋ผ์ด์„ ์Šค๋ฅผ ์ค€์ˆ˜ํ•˜๋ฉฐ, ์ƒ์—…์  ์šฉ๋„๋กœ ์‚ฌ์šฉ ๋ถˆ๊ฐ€

This gemma2 model was trained 2x faster with Unsloth and Huggingface's TRL library.

Downloads last month
19
Safetensors
Model size
9.24B params
Tensor type
BF16
ยท
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Model tree for SeongeonKim/gemma-2-9b-HangulFixer_v0.1

Base model

google/gemma-2-9b
Finetuned
(358)
this model