YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

DeBERTaV3-Large Reward Model โ€” formatguard & preference distillation (ckpt-799)

  • Base backbone: microsoft/deberta-v3-large
  • Init RM: yungshun317/deberta-v3-large-format-guard
  • This snapshot adds anti-format-spam preference tuning.

Usage

from transformers import AutoTokenizer, AutoModelForSequenceClassification
tok = AutoTokenizer.from_pretrained("yungshun317/deberta-v3-large-format-guard-preference-distillation")
rm  = AutoModelForSequenceClassification.from_pretrained("yungshun317/deberta-v3-large-format-guard-preference-distillation")
Downloads last month
-
Safetensors
Model size
0.4B params
Tensor type
BF16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support