File size: 1,702 Bytes
19e71ef c613d29 2bb7d6b c613d29 2bb7d6b c613d29 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 |
---
license: cc-by-2.0
---
# ColBERT_Humor
## Dataset Description
- **Paper:** [Colbert: Using bert sentence embedding for humor detection](https://arxiv.org/abs/2004.12765)
## Dataset Summary
ColBERT Humor contains 200,000 labeled short texts, equally distributed between humorous and non-humorous content. The dataset was created to overcome the limitations of prior humor detection datasets, which were characterized by inconsistencies in text length, word count, and formality, making them easy to predict with simple models without truly understanding the nuances of humor. The two sources for this dataset are the News Category dataset, featuring 200k news headlines from the Huffington Post (2012-2018), and a collection of 231,657 Reddit jokes. The texts have been rigorously preprocessed to ensure syntactic similarity, requiring models to delve into the linguistic intricacies to distinguish humor, effectively providing a more complex and substantial platform for humor detection research.
For the details of this dataset, we refer you to the original [paper](https://arxiv.org/abs/2004.12765).
Metadata in Creative Language Toolkit ([CLTK](https://github.com/liyucheng09/cltk))
- CL Type: Humor
- Task Type: detection
- Size: 200k
- Created time: 2020
### Citation Information
If you find this dataset helpful, please cite:
```
@article{annamoradnejad2020colbert,
title={Colbert: Using bert sentence embedding for humor detection},
author={Annamoradnejad, Issa and Zoghi, Gohar},
journal={arXiv preprint arXiv:2004.12765},
year={2020}
}
```
### Contributions
If you have any queries, please open an issue or direct your queries to [mail](mailto:yucheng.li@surrey.ac.uk). |