Edit model card

xlm-roberta-base-finetuned-panx-all-langs

This model is a fine-tuned version of xlm-roberta-base on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.3089
  • F1 Score: 0.8140

Model description

"트랜슀 포머λ₯Ό ν™œμš©ν•œ μžμ—°μ–΄ 처리" O'Reilly 책을 ν™œμš©ν•˜μ˜€μŠ΅λ‹ˆλ‹€.

xlm-roberta-base λ₯Ό μ΄μš©ν•˜μ—¬ 파인 νŠœλ‹μ„ ν•œ 객체λͺ… 인식 μž…λ‹ˆλ‹€.

각 언어별 f1_score
ko : f1_score = 0.8611821192789028
en : f1_score = 0.7868391074180795
ja : f1_score = 0.6440401846320934
es : f1_score = 0.8533862565120316

Intended uses & limitations

μ½”λ“œλŠ” 책을 ν™œμš© 및 μ°Έκ³  ν•˜μ˜€μŠ΅λ‹ˆλ‹€.

Training and evaluation data

XTREME 데이터 셋을 μ΄μš©ν•˜μ˜€μŠ΅λ‹ˆλ‹€. 특히 PAN-X 데이터 셋을 μ΄μš©ν–ˆμŠ΅λ‹ˆλ‹€.

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 5e-05
  • train_batch_size: 24
  • eval_batch_size: 24
  • seed: 42
  • optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
  • lr_scheduler_type: linear
  • num_epochs: 3

Training results

Training Loss Epoch Step Validation Loss F1 Score
0.5093 1.0 917 0.3530 0.7679
0.309 2.0 1834 0.3101 0.8029
0.2176 3.0 2751 0.3089 0.8140

Framework versions

  • Transformers 4.46.2
  • Pytorch 2.5.1+cu124
  • Datasets 3.1.0
  • Tokenizers 0.20.3
Downloads last month
5
Safetensors
Model size
277M params
Tensor type
F32
Β·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for tommyjin/xlm-roberta-base-finetuned-panx-all-langs

Finetuned
(2623)
this model