RoBERTa Korean Hanja Extended - MLM ํ์ธํ๋ ๋ชจ๋ธ
hwp0725/roberta-korean-hanja-stdict ๋ฒ ์ด์ค ๋ชจ๋ธ์ ํ์ ์ฝํผ์ค๋ก MLM(Masked Language Modeling) ํ์ธํ๋ํ ๋ชจ๋ธ์ ๋๋ค.
๋ชจ๋ธ ์ ๋ณด
| ํญ๋ชฉ | ๊ฐ |
|---|---|
| ๋ฒ ์ด์ค ๋ชจ๋ธ | hwp0725/roberta-korean-hanja-stdict |
| ํ๋ผ๋ฏธํฐ ์ | 345.1M |
| ํ์ต ๋ฐฉ์ | MLM (Masked Language Modeling) |
| ์ํฌํฌ | 3 |
๊ด๋ จ ๋ชจ๋ธ
| ๋ชจ๋ธ | ์ค๋ช |
|---|---|
| hwp0725/roberta-korean-hanja-stdict | vocab ํ์ฅ ๋ฒ ์ด์ค ๋ชจ๋ธ (MLM ํ์ต ์ ) |
| hwp0725/roberta-korean-hanja-stdict-mlm | MLM ํ์ธํ๋ ์๋ฃ ๋ชจ๋ธ (ํ์ฌ ๋ชจ๋ธ) |
์ฌ์ฉ๋ฒ
from transformers import AutoTokenizer, AutoModelForMaskedLM, pipeline
tokenizer = AutoTokenizer.from_pretrained("hwp0725/roberta-korean-hanja-stdict-mlm")
model = AutoModelForMaskedLM.from_pretrained("hwp0725/roberta-korean-hanja-stdict-mlm")
fill_mask = pipeline("fill-mask", model=model, tokenizer=tokenizer)
result = fill_mask("ๅญๅญๆฐ๏ผๅญธ่ๆ็ฟไน๏ผไธไบฆ[MASK]ไน")
print(result)
ํ์ต ํ์ดํผํ๋ผ๋ฏธํฐ
| ํญ๋ชฉ | ๊ฐ |
|---|---|
| Learning rate | 2e-05 |
| Train batch size | 16 |
| Seed | 42 |
| Optimizer | AdamW (betas=(0.9,0.999), epsilon=1e-08) |
| LR scheduler | linear (warmup 500 steps) |
| Epochs | 3 |
| Mixed precision | Native AMP |
ํ๋ ์์ํฌ ๋ฒ์
- Transformers 5.0.0
- PyTorch 2.4.1+cu124
- Datasets 4.5.0
- Tokenizers 0.22.2
๋ผ์ด์ ์ค
CC BY-SA 4.0
- Downloads last month
- -
Model tree for hwp0725/roberta-korean-hanja-stdict-mlm
Base model
klue/roberta-large
Finetuned
KoichiYasuoka/roberta-large-korean-hanja
Finetuned
hwp0725/roberta-korean-hanja-stdict