RoBERTa Korean Hanja Extended - MLM ํŒŒ์ธํŠœ๋‹ ๋ชจ๋ธ

hwp0725/roberta-korean-hanja-stdict ๋ฒ ์ด์Šค ๋ชจ๋ธ์„ ํ•œ์ž ์ฝ”ํผ์Šค๋กœ MLM(Masked Language Modeling) ํŒŒ์ธํŠœ๋‹ํ•œ ๋ชจ๋ธ์ž…๋‹ˆ๋‹ค.

๋ชจ๋ธ ์ •๋ณด

ํ•ญ๋ชฉ ๊ฐ’
๋ฒ ์ด์Šค ๋ชจ๋ธ hwp0725/roberta-korean-hanja-stdict
ํŒŒ๋ผ๋ฏธํ„ฐ ์ˆ˜ 345.1M
ํ•™์Šต ๋ฐฉ์‹ MLM (Masked Language Modeling)
์—ํฌํฌ 3

๊ด€๋ จ ๋ชจ๋ธ

๋ชจ๋ธ ์„ค๋ช…
hwp0725/roberta-korean-hanja-stdict vocab ํ™•์žฅ ๋ฒ ์ด์Šค ๋ชจ๋ธ (MLM ํ•™์Šต ์ „)
hwp0725/roberta-korean-hanja-stdict-mlm MLM ํŒŒ์ธํŠœ๋‹ ์™„๋ฃŒ ๋ชจ๋ธ (ํ˜„์žฌ ๋ชจ๋ธ)

์‚ฌ์šฉ๋ฒ•

from transformers import AutoTokenizer, AutoModelForMaskedLM, pipeline

tokenizer = AutoTokenizer.from_pretrained("hwp0725/roberta-korean-hanja-stdict-mlm")
model = AutoModelForMaskedLM.from_pretrained("hwp0725/roberta-korean-hanja-stdict-mlm")

fill_mask = pipeline("fill-mask", model=model, tokenizer=tokenizer)
result = fill_mask("ๅญ”ๅญๆ›ฐ๏ผšๅญธ่€Œๆ™‚็ฟ’ไน‹๏ผŒไธไบฆ[MASK]ไนŽ")
print(result)

ํ•™์Šต ํ•˜์ดํผํŒŒ๋ผ๋ฏธํ„ฐ

ํ•ญ๋ชฉ ๊ฐ’
Learning rate 2e-05
Train batch size 16
Seed 42
Optimizer AdamW (betas=(0.9,0.999), epsilon=1e-08)
LR scheduler linear (warmup 500 steps)
Epochs 3
Mixed precision Native AMP

ํ”„๋ ˆ์ž„์›Œํฌ ๋ฒ„์ „

  • Transformers 5.0.0
  • PyTorch 2.4.1+cu124
  • Datasets 4.5.0
  • Tokenizers 0.22.2

๋ผ์ด์„ ์Šค

CC BY-SA 4.0

Downloads last month
-
Safetensors
Model size
0.3B params
Tensor type
F32
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for hwp0725/roberta-korean-hanja-stdict-mlm