Edit model card

xlm-roberta-large-ctebmsp

This model is a finetuned version of xlm-roberta-large for the CT-EBM-SP (Clinical Trials for Evidence-based Medicine in Spanish) dataset used in a benchmark in the paper A comparative analysis of Spanish Clinical encoder-based models on NER and classification tasks. The model has a F1 of 0.906

Please refer to the original publication for more information.

Parameters used

parameter Value
batch size 64
learning rate 2e-05
classifier dropout 0.1
warmup ratio 0
warmup steps 0
weight decay 0
optimizer AdamW
epochs 10
early stopping patience 3

BibTeX entry and citation info

@article{10.1093/jamia/ocae054,
    author = {García Subies, Guillem and Barbero Jiménez, Álvaro and Martínez Fernández, Paloma},
    title = {A comparative analysis of Spanish Clinical encoder-based models on NER and classification tasks},
    journal = {Journal of the American Medical Informatics Association},
    volume = {31},
    number = {9},
    pages = {2137-2146},
    year = {2024},
    month = {03},
    issn = {1527-974X},
    doi = {10.1093/jamia/ocae054},
    url = {https://doi.org/10.1093/jamia/ocae054},
}
Downloads last month
28
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Dataset used to train IIC/xlm-roberta-large-ctebmsp

Collection including IIC/xlm-roberta-large-ctebmsp

Evaluation results