Indonesian T5 Language Models
Collection
Indonesian T5 models pre-trained with nanoT5 and fine-tuned on IndoNLG tasks. GitHub: https://github.com/LazarusNLP/IndoT5/
•
5 items
•
Updated
IndoNanoT5 Base is an Indonesian sequence-to-sequence language model based on the T5 architecture. We conducted pre-training on an open-source Indonesian corpus of uonlp/CulturaX. On a held-out subset of the corpus, our model achieved an evaluation loss of 2.082 or a perplexity of about 8.02.
This model was trained using the nanoT5 PyTorch framework. All training was done on an NVIDIA H100 GPU. LazarusNLP/IndoNanoT5-base is released under Apache 2.0 license.
from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
model_checkpoint = "LazarusNLP/IndoNanoT5-base"
tokenizer = AutoTokenizer.from_pretrained(model_checkpoint)
model = AutoModelForSeq2SeqLM.from_pretrained(model_checkpoint)
Around 4B tokens from the following corpora were used during pre-training.
The following hyperparameters were used during training:
total_steps: 65536input_length: 512batch_size: 128grad_acc: 1base_lr: 5e-3optimizer: AdamWScaled with betas=(0.9,0.999) and epsilon=1e-08weight_decay: 0.0lr_scheduler: cosinewarmup_steps: 10000final_cosine: 1e-5grad_clip: 1.0precision: bf16We would like to acknowledge nanoT5 for inspiring this project.
BhinnekaLM is developed with love by: