|
---
|
|
library_name: transformers
|
|
license: mit
|
|
base_model: base-uncased
|
|
tags:
|
|
- bert
|
|
- fine-tuning
|
|
- text-classification
|
|
model-index:
|
|
- name: NLP_with_Disaster_Tweets
|
|
results:
|
|
- task:
|
|
type: text-classification
|
|
name: Text Classification
|
|
metrics:
|
|
- name: Accuracy
|
|
type: accuracy
|
|
value: 0.835
|
|
language:
|
|
- en
|
|
---
|
|
|
|
# Disaster Tweets Classification
|
|
|
|
This model is fine-tuned BERT for classifying whether a tweet is about a real disaster or not.
|
|
|
|
## Model Description
|
|
|
|
- Based on `bert-base-uncased`
|
|
- Fine-tuned for binary classification task
|
|
- Achieves 83.5% accuracy on validation set
|
|
- Trained on Kaggle's "Natural Language Processing with Disaster Tweets" competition dataset
|
|
|
|
## How to Use
|
|
|
|
```python
|
|
from transformers import AutoConfig, AutoTokenizer, AutoModelForSequenceClassification
|
|
|
|
# Load model and tokenizer
|
|
tokenizer = AutoTokenizer.from_pretrained("real-jiakai/NLP_with_Disaster_Tweets")
|
|
model = AutoModelForSequenceClassification.from_pretrained("real-jiakai/NLP_with_Disaster_Tweets")
|
|
|
|
# Example usage
|
|
text = "There was a major earthquake in California"
|
|
inputs = tokenizer(text, return_tensors="pt", padding=True, truncation=True, max_length=512)
|
|
outputs = model(**inputs)
|
|
predicted_class = outputs.logits.argmax(-1).item()
|
|
```
|
|
|
|
## License
|
|
|
|
This model is licensed under the [MIT](https://opensource.org/license/mit) License.
|
|
|
|
## Citation
|
|
|
|
If you use this model in your work, please cite:
|
|
|
|
```
|
|
@misc{NLP_with_Disaster_Tweets,
|
|
author = {real-jiakai},
|
|
title = {NLP_with_Disaster_Tweets},
|
|
year = {2024},
|
|
url = {https://huggingface.co/real-jiakai/NLP_with_Disaster_Tweets},
|
|
publisher = {Hugging Face}
|
|
}
|
|
```
|
|
|
|
|
|
|