KAIdol Qwen3-14B Korean Role-Playing
K-pop AI μμ΄λ μ±λ΄μ μν΄ νμΈνλλ Qwen3-14B λͺ¨λΈμ λλ€.
Model Description
- Base Model: Qwen/Qwen3-14B
- Fine-tuning Dataset: KREW/korean-role-playing
- Training Method: Full Fine-tuning with DeepSpeed ZeRO-3
- Training Hardware: AWS ml.p4d.24xlarge (8x NVIDIA A100 40GB)
- Training Time: ~13.5 hours
Training Details
Hyperparameters
| Parameter | Value |
|---|---|
| Learning Rate | 2e-5 |
| Epochs | 3 |
| Batch Size (effective) | 128 |
| Max Sequence Length | 4096 |
| Optimizer | AdamW |
| LR Scheduler | Cosine |
| Warmup Ratio | 0.03 |
| Weight Decay | 0.01 |
| Precision | bfloat16 |
Training Configuration
- Distributed Strategy: DeepSpeed ZeRO-3
- Gradient Checkpointing: Enabled
- Flash Attention 2: Enabled
Usage
With Transformers
from transformers import AutoModelForCausalLM, AutoTokenizer
model_id = "developer-lunark/kaidol-qwen3-14b-korean-rp"
tokenizer = AutoTokenizer.from_pretrained(model_id, trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained(
model_id,
torch_dtype="auto",
device_map="auto",
trust_remote_code=True
)
messages = [
{"role": "system", "content": "λΉμ μ μΉκ·Όνκ³ λ€μ ν AI μμ΄λμ
λλ€."},
{"role": "user", "content": "μλ
νμΈμ! μ€λ κΈ°λΆμ΄ μ΄λ μΈμ?"}
]
text = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
inputs = tokenizer(text, return_tensors="pt").to(model.device)
outputs = model.generate(**inputs, max_new_tokens=512, do_sample=True, temperature=0.7)
response = tokenizer.decode(outputs[0], skip_special_tokens=True)
print(response)
With vLLM
from vllm import LLM, SamplingParams
llm = LLM(
model="developer-lunark/kaidol-qwen3-14b-korean-rp",
trust_remote_code=True,
max_model_len=4096
)
sampling_params = SamplingParams(temperature=0.7, max_tokens=512)
outputs = llm.generate(["μλ
νμΈμ!"], sampling_params)
Intended Use
μ΄ λͺ¨λΈμ K-pop AI μμ΄λ μΊλ¦ν°μμ λ‘€νλ μ΄ λνλ₯Ό μν΄ μ€κ³λμμ΅λλ€:
- μΉκ·Όνκ³ λ€μ ν λν
- μΊλ¦ν° μΌκ΄μ± μ μ§
- νκ΅μ΄ μμ°μ€λ¬μ΄ νν
Limitations
- λ‘€νλ μ΄/λν λͺ©μ μΌλ‘ μ΅μ νλμ΄ μμ΄ μΌλ°μ μΈ QAλ μΆλ‘ μμ μλ μ ν©νμ§ μμ μ μμ΅λλ€
- 14B λͺ¨λΈμ΄λ―λ‘ μΆλ‘ μ μ΅μ 28GB VRAMμ΄ νμν©λλ€
License
This model is released under the Apache 2.0 license.
Citation
@misc{kaidol-qwen3-14b-korean-rp,
author = {developer-lunark},
title = {KAIdol Qwen3-14B Korean Role-Playing},
year = {2024},
publisher = {Hugging Face},
url = {https://huggingface.co/developer-lunark/kaidol-qwen3-14b-korean-rp}
}
- Downloads last month
- 562