Model Card for Model ID
This is a XLM-RoBERTa-base fine-tuned model on 6.5K (premise, hypothesis) sentence pairs from the ASSIN2 (Avaliação de Similaridade Semântica e Inferência Textual) corpus. The original references are: Unsupervised Cross-Lingual Representation Learning At Scale, ASSIN2: Avaliação de Similaridade Semântica e Inferência Textual, respectivelly. This model is suitable for Brazilian Portuguese.
Model Details
Model Description
- Developed by: Giovani Tavares and Felipe Ribas Serras
- Oriented By: Felipe Ribas Serras, Renata Wassermann and Marcelo Finger
- Model type: Transformer-based text classifier
- Language(s) (NLP): Portuguese
- License: mit
- Finetuned from model XLM-RoBERTa-base
Model Sources
- Repository: Natural-Portuguese-Language-Inference
- Paper: This is an ongoing research. We are currently writing a paper where we fully describe our experiments.
Uses
Direct Use
This fine-tuned version of XLM-RoBERTa-base performs Natural Language Inference (NLI), which is a text classification task.
The (premise, hypothesis) entailment definition used is the same as the one found in Salvatore's paper [1].
Therefore, this fine-tuned version of XLM-RoBERTa-base classifies pairs of sentences in the form (premise, hypothesis) into the classes ENTAILMENT or NONE.
Demo
from transformers import AutoModelForSequenceClassification, AutoTokenizer
import torch
model_path = "giotvr/xlm_roberta_base_assin2_fine_tuned"
premise = "As mudanças climáticas são uma ameaça séria para a biodiversidade do planeta."
hypothesis ="A biodiversidade do planeta é seriamente ameaçada pelas mudanças climáticas."
tokenizer = XLMRobertaTokenizer.from_pretrained(model_path, use_auth_token=True)
input_pair = tokenizer(premise, hypothesis, return_tensors="pt",padding=True, truncation=True)
model = AutoModelForSequenceClassification.from_pretrained(model_path, use_auth_token=True)
with torch.no_grad():
logits = model(**input_pair).logits
probs = torch.nn.functional.softmax(logits, dim=-1)
probs, sorted_indices = torch.sort(probs, descending=True)
for i, score in enumerate(probs[0]):
print(f"Class {sorted_indices[0][i]}: {score.item():.4f}")
Recommendations
This model should be used for scientific purposes only. It was not tested for production environments.
Fine-Tuning Details
Fine-Tuning Data
Train Dataset: ASSIN
Evaluation Dataset used for Hyperparameter Tuning: ASSIN2's validation split
Test Datasets:
This is a fine tuned version of XLM-RoBERTa-base using the ASSIN2 (Avaliação de Similaridade Semântica e Inferência textual) dataset. ASSIN2 is a corpus annotated with hypothesis/premise Portuguese sentence pairs suitable for detecting textual entailment or neutral relationship between the members of such pairs. Such corpus is balanced with 7k ptbr (Brazilian Portuguese) sentence pairs.
Fine-Tuning Procedure
The model's fine-tuning procedure can be summarized in three major subsequent tasks:
- Data Processing: ASSIN2's validation and train splits were loaded from the Hugging Face Hub and processed afterwards;
- Hyperparameter Tuning: XLM-RoBERTa-base's hyperparameters were chosen with the help of the [Weights & Biases] API to track the results and upload the fine-tuned models;
- Final Model Loading and Testing: using the cross-tests approach described in the this section, the models' performance were measured using different datasets and metrics.
Hyperparameter Tuning
The following hyperparameters were tested in order to maximize the evaluation accuracy.
- Number of Training Epochs: $(1,2,3)$
- Per Device Train Batch Size: $(8,16,32)$
- Learning Rate: $(1e−5, 2e−5, 5e−5)$
The hyperaparemeter tuning experiments were run and tracked using the Weights & Biases' API and can be found at this link.
Training Hyperparameters
The hyperparameter tuning performed yelded the following values:
- Number of Training Epochs: $3$
- Per Device Train Batch Size: $8$
- Learning Rate: $5e-5$
Evaluation
ASSIN
Testing this model in ASSIN's test split was straightforward because this model was fine tuned using ASSIN2's training set which contains the same labels as ASSIN. Hence, it can predict the same labels as the ones found in ASSIN's test set.
ASSIN2
Testing this model in ASSIN2's test split is straightforward because this model was fine tuned using ASSIN2's training set and therefore can predict the same labels as the ones found in its test set.
PLUE/MNLI
Testing this model in PLUE/MNLI was only possible by considering PLUE/MNLI's contradiction and neutral labels as NONE and PLUE/MNLI's entailment label as equivalent to the ENTAILMENT predicted by the model.
More information on how such mapping is performed can be found in Modelos para Inferência em Linguagem Natural que entendem a Língua Portuguesa.
Metrics
The model's performance metrics for each test dataset are presented separately. Accuracy, f1 score, precision and recall were the metrics used to every evaluation performed. Such metrics are reported below. More information on such metrics them will be available in our ongoing research paper.
Results
test set | accuracy | f1 score | precision | recall |
---|---|---|---|---|
assin | 0.78 | 0.78 | 0.80 | 0.80 |
assin2 | 0.86 | 0.86 | 0.87 | 0.86 |
plue/mnli | 0.71 | 0.67 | 0.71 | 0.71 |
Model Examination
Some interpretability work is being done in order to understand the model's behavior. Such details will be available in the previoulsy referred paper.
References
- Downloads last month
- 12