YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)
DeBERTaV3-Large Reward Model โ formatguard & preference distillation (ckpt-799)
- Base backbone:
microsoft/deberta-v3-large - Init RM:
yungshun317/deberta-v3-large-format-guard - This snapshot adds anti-format-spam preference tuning.
Usage
from transformers import AutoTokenizer, AutoModelForSequenceClassification
tok = AutoTokenizer.from_pretrained("yungshun317/deberta-v3-large-format-guard-preference-distillation")
rm = AutoModelForSequenceClassification.from_pretrained("yungshun317/deberta-v3-large-format-guard-preference-distillation")
- Downloads last month
- -
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐ Ask for provider support