nerT / README.md
TerminatorPower's picture
Upload BertForSequenceClassification
fce9a20 verified
---
language:
- tr
library_name: transformers
license: mit
metrics:
- f1
- accuracy
- recall
tags:
- ner
- token-classification
- turkish
---
# Model Card for Turkish Named Entity Recognition Model
<!-- Provide a quick summary of what the model is/does. -->
This model performs Named Entity Recognition (NER) for Turkish text, identifying and classifying entities such as person names, locations, and organizations. Model got 0.9599 F1 on validation set.
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is a fine-tuned BERT model for Turkish Named Entity Recognition (NER). It is based on the `dbmdz/bert-base-turkish-uncased` model and has been trained on a custom Turkish NER dataset.
- **Developed by:** Ezel Bayraktar (ai@bayraktarlar.dev)
- **Model type:** Token Classification (Named Entity Recognition)
- **Language(s) (NLP):** Turkish
- **License:** MIT
- **Finetuned from model:** dbmdz/bert-base-turkish-uncased
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
This model can be used directly for Named Entity Recognition tasks in Turkish text. It identifies and labels entities such as person names (PER), locations (LOC), and organizations (ORG).
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
The model can be integrated into larger natural language processing pipelines for Turkish, such as information extraction systems, question answering, or text summarization.
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
This model should not be used for languages other than Turkish or for tasks beyond Named Entity Recognition. It may not perform well on domain-specific text or newly emerging named entities not present in the training data.
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
The model may inherit biases present in the training data or the pre-trained BERT model it was fine-tuned from. It may not perform consistently across different domains or types of Turkish text.
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users should evaluate the model's performance on their specific domain and use case. For critical applications, human review of the model's outputs is recommended.
## How to Get Started with the Model
Use the code below to get started with the model.
```python
from transformers import pipeline
nert = pipeline('ner', model='TerminatorPower/nerT', tokenizer='TerminatorPower/nerT')
answer = nert("Mustafa Kemal Atatürk, 19 Mayıs 1919'da Samsun'a çıktı.")
print(answer)