Edit model card

Hungarian Sentence-level Sentiment Analysis Model with XLM-RoBERTa

For further models, scripts and details, see our repository or our demo site.

  • Pretrained model used: XLM-RoBERTa base
  • Finetuned on Hungarian Twitter Sentiment (HTS) Corpus
  • Labels: 0 (negative), 1 (positive)

Limitations

  • max_seq_length = 128

Results

Model HTS2 HTS5
huBERT 85.56 68.99
XLM-RoBERTa 85.56 66.50

Citation

If you use this model, please cite the following paper:

@article {laki-yang-sentiment,
      author = {Laki, László János and Yang, Zijian Győző},
      title = {Sentiment Analysis with Neural Models for Hungarian},
      journal = {Acta Polytechnica Hungarica},
      year = {2023},
      publisher = {Obuda University},
      volume = {20},
      number = {5},
      doi = {10.12700/APH.20.5.2023.5.8},
      pages=      {109--128},
      url = {https://acta.uni-obuda.hu/Laki_Yang_134.pdf}
}
Downloads last month
38
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.