KAIdol Qwen3-14B Korean Role-Playing

K-pop AI μ•„μ΄λŒ 챗봇을 μœ„ν•΄ νŒŒμΈνŠœλ‹λœ Qwen3-14B λͺ¨λΈμž…λ‹ˆλ‹€.

Model Description

  • Base Model: Qwen/Qwen3-14B
  • Fine-tuning Dataset: KREW/korean-role-playing
  • Training Method: Full Fine-tuning with DeepSpeed ZeRO-3
  • Training Hardware: AWS ml.p4d.24xlarge (8x NVIDIA A100 40GB)
  • Training Time: ~13.5 hours

Training Details

Hyperparameters

Parameter Value
Learning Rate 2e-5
Epochs 3
Batch Size (effective) 128
Max Sequence Length 4096
Optimizer AdamW
LR Scheduler Cosine
Warmup Ratio 0.03
Weight Decay 0.01
Precision bfloat16

Training Configuration

  • Distributed Strategy: DeepSpeed ZeRO-3
  • Gradient Checkpointing: Enabled
  • Flash Attention 2: Enabled

Usage

With Transformers

from transformers import AutoModelForCausalLM, AutoTokenizer

model_id = "developer-lunark/kaidol-qwen3-14b-korean-rp"

tokenizer = AutoTokenizer.from_pretrained(model_id, trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained(
    model_id,
    torch_dtype="auto",
    device_map="auto",
    trust_remote_code=True
)

messages = [
    {"role": "system", "content": "당신은 μΉœκ·Όν•˜κ³  λ‹€μ •ν•œ AI μ•„μ΄λŒμž…λ‹ˆλ‹€."},
    {"role": "user", "content": "μ•ˆλ…•ν•˜μ„Έμš”! 였늘 기뢄이 μ–΄λ– μ„Έμš”?"}
]

text = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
inputs = tokenizer(text, return_tensors="pt").to(model.device)

outputs = model.generate(**inputs, max_new_tokens=512, do_sample=True, temperature=0.7)
response = tokenizer.decode(outputs[0], skip_special_tokens=True)
print(response)

With vLLM

from vllm import LLM, SamplingParams

llm = LLM(
    model="developer-lunark/kaidol-qwen3-14b-korean-rp",
    trust_remote_code=True,
    max_model_len=4096
)

sampling_params = SamplingParams(temperature=0.7, max_tokens=512)
outputs = llm.generate(["μ•ˆλ…•ν•˜μ„Έμš”!"], sampling_params)

Intended Use

이 λͺ¨λΈμ€ K-pop AI μ•„μ΄λŒ μΊλ¦­ν„°μ™€μ˜ λ‘€ν”Œλ ˆμ΄ λŒ€ν™”λ₯Ό μœ„ν•΄ μ„€κ³„λ˜μ—ˆμŠ΅λ‹ˆλ‹€:

  • μΉœκ·Όν•˜κ³  λ‹€μ •ν•œ λŒ€ν™”
  • 캐릭터 일관성 μœ μ§€
  • ν•œκ΅­μ–΄ μžμ—°μŠ€λŸ¬μš΄ ν‘œν˜„

Limitations

  • λ‘€ν”Œλ ˆμ΄/λŒ€ν™” λͺ©μ μœΌλ‘œ μ΅œμ ν™”λ˜μ–΄ μžˆμ–΄ 일반적인 QAλ‚˜ μΆ”λ‘  μž‘μ—…μ—λŠ” μ ν•©ν•˜μ§€ μ•Šμ„ 수 μžˆμŠ΅λ‹ˆλ‹€
  • 14B λͺ¨λΈμ΄λ―€λ‘œ 좔둠에 μ΅œμ†Œ 28GB VRAM이 ν•„μš”ν•©λ‹ˆλ‹€

License

This model is released under the Apache 2.0 license.

Citation

@misc{kaidol-qwen3-14b-korean-rp,
  author = {developer-lunark},
  title = {KAIdol Qwen3-14B Korean Role-Playing},
  year = {2024},
  publisher = {Hugging Face},
  url = {https://huggingface.co/developer-lunark/kaidol-qwen3-14b-korean-rp}
}
Downloads last month
562
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for developer-lunark/kaidol-qwen3-14b-korean-rp

Finetuned
Qwen/Qwen3-14B
Finetuned
(156)
this model