Datasets:

Modalities:
Text
Formats:
csv
ArXiv:
Libraries:
Datasets
pandas
License:
File size: 1,702 Bytes
19e71ef
 
 
c613d29
 
 
 
 
 
 
2bb7d6b
c613d29
2bb7d6b
 
 
 
 
c613d29
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
---
license: cc-by-2.0
---

# ColBERT_Humor

## Dataset Description

- **Paper:** [Colbert: Using bert sentence embedding for humor detection](https://arxiv.org/abs/2004.12765)

## Dataset Summary

ColBERT Humor contains 200,000 labeled short texts, equally distributed between humorous and non-humorous content. The dataset was created to overcome the limitations of prior humor detection datasets, which were characterized by inconsistencies in text length, word count, and formality, making them easy to predict with simple models without truly understanding the nuances of humor. The two sources for this dataset are the News Category dataset, featuring 200k news headlines from the Huffington Post (2012-2018), and a collection of 231,657 Reddit jokes. The texts have been rigorously preprocessed to ensure syntactic similarity, requiring models to delve into the linguistic intricacies to distinguish humor, effectively providing a more complex and substantial platform for humor detection research.

For the details of this dataset, we refer you to the original [paper](https://arxiv.org/abs/2004.12765).

Metadata in Creative Language Toolkit ([CLTK](https://github.com/liyucheng09/cltk))
- CL Type: Humor
- Task Type: detection
- Size: 200k
- Created time: 2020

### Citation Information

If you find this dataset helpful, please cite:

```
@article{annamoradnejad2020colbert,
  title={Colbert: Using bert sentence embedding for humor detection},
  author={Annamoradnejad, Issa and Zoghi, Gohar},
  journal={arXiv preprint arXiv:2004.12765},
  year={2020}
}
```

### Contributions

If you have any queries, please open an issue or direct your queries to [mail](mailto:yucheng.li@surrey.ac.uk).