--- datasets: - ner_distillbert language: - en library_name: pytorch license: apache-2.0 tags: - ner - pytorch - mlflow task: token-classification --- # ner_distillbert ## Model Description This is a Named Entity Recognition (NER) model based on DistilBERT, a distilled version of BERT that retains 97% of BERT's performance while being 60% smaller and faster. The model identifies and classifies named entities in text such as persons, organizations, locations, and other predefined categories. **Model Details:** - **Model Name:** ner_distillbert - **Version:** 1 - **Task:** Ner - **Framework:** pytorch - **Language(s):** en - **License:** apache-2.0 ## Intended Uses This model is designed for ner tasks. Please evaluate on your specific use case before production deployment. ## Training Details ### Training Data - **Dataset:** ner_distillbert - **Dataset Size:** Not specified ### Training Configuration ## Usage ### Installation ```bash pip install transformers torch # For transformers models # OR pip install -r requirements.txt # For other frameworks ``` ### Basic Usage ```python # Load and use the model from inference import Model # See inference.py in the repository model = Model(openchs/ner_distillbert_v1) predictions = model.predict([President Biden met with Chancellor Merkel at the White House to discuss NATO policies.]) print(predictions) ``` ## Performance Metrics ### Evaluation Results | Metric | Value | |--------|-------| | Epoch | 3.0000 | | Eval Accuracy | 0.9608 | | Eval F1 | 0.9433 | | Eval Loss | 0.1171 | | Eval Precision | 0.9324 | | Eval Recall | 0.9608 | | Eval Runtime | 0.1943 | | Eval Samples Per Second | 82.3410 | | Eval Steps Per Second | 10.2930 | ## MLflow Tracking - **Experiment:** N/A - **Run ID:** `N/A` - **Training Date:** N/A ## Citation ```bibtex @misc{ner_distillbert_1, title={ner_distillbert}, author={BITZ-AI TEAM}, year={2025}, publisher={Hugging Face}, url={https://huggingface.co/marlonbino/ner_distillbert} } ```