Indonesian T5 Language Models
Collection
Indonesian T5 models pre-trained with nanoT5 and fine-tuned on IndoNLG tasks. GitHub: https://github.com/LazarusNLP/IndoT5/
•
5 items
•
Updated
IndoNanoT5 Base is an Indonesian sequence-to-sequence language model based on the T5 architecture. We conducted pre-training on an open-source Indonesian corpus of uonlp/CulturaX. On a held-out subset of the corpus, our model achieved an evaluation loss of 2.082 or a perplexity of about 8.02.
This model was trained using the nanoT5 PyTorch framework. All training was done on an NVIDIA H100 GPU. LazarusNLP/IndoNanoT5-base is released under Apache 2.0 license.
from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
model_checkpoint = "LazarusNLP/IndoNanoT5-base"
tokenizer = AutoTokenizer.from_pretrained(model_checkpoint)
model = AutoModelForSeq2SeqLM.from_pretrained(model_checkpoint)
Around 4B tokens from the following corpora were used during pre-training.
The following hyperparameters were used during training:
total_steps
: 65536input_length
: 512batch_size
: 128grad_acc
: 1base_lr
: 5e-3optimizer
: AdamWScaled with betas=(0.9,0.999)
and epsilon=1e-08
weight_decay
: 0.0lr_scheduler
: cosinewarmup_steps
: 10000final_cosine
: 1e-5grad_clip
: 1.0precision
: bf16
We would like to acknowledge nanoT5 for inspiring this project.
BhinnekaLM is developed with love by: