activations: Tanh batch_size: 1 class_identifier: unified_metric dropout: 0.1 encoder_learning_rate: 1.0e-06 encoder_model: XLM-RoBERTa final_activation: null hidden_sizes: - 3072 - 1024 input_segments: - mt - src keep_embeddings_frozen: true layer: mix layer_norm: false layer_transformation: sparsemax layerwise_decay: 0.95 learning_rate: 1.5e-05 loss: mse loss_lambda: 0.65 nr_frozen_epochs: 0.3 optimizer: AdamW pool: avg pretrained_model: Davlan/afro-xlmr-large sent_layer: mix train_data: - data/1720-da.mlqe-src.csv validation_data: - data/wmt-ende-newstest2021.csv - data/wmt-enru-newstest2021.csv - data/wmt-zhen-newstest2021.csv word_layer: 24 word_level_training: false word_weights: - 0.15 - 0.85