IndoBERT Lite Base IndoNLI Distil mDeBERTa

IndoBERT Lite Base IndoNLI Distil mDeBERTa is a natural language inference (NLI) model based on the ALBERT model. The model was originally the pre-trained indobenchmark/indobert-lite-base-p1 model, which is then fine-tuned on IndoNLI's dataset consisting of Indonesian Wikipedia, news, and Web articles [1], whilst being distilled from MoritzLaurer/mDeBERTa-v3-base-xnli-multilingual-nli-2mil7.

Evaluation Results

dev Acc. test_lay Acc. test_expert Acc.
IndoNLI 77.19 74.42 61.22

Model

Model #params Arch. Training/Validation data (text)
indobert-lite-base-p1-indonli-distil-mdeberta 11.7M ALBERT Base IndoNLI

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 16
  • eval_batch_size: 16
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 5

Training results

Training Loss Epoch Step Validation Loss Accuracy F1 Precision Recall
0.5053 1.0 646 0.4511 0.7506 0.7462 0.7530 0.7445
0.4516 2.0 1292 0.4458 0.7692 0.7683 0.7684 0.7697
0.4192 3.0 1938 0.4433 0.7701 0.7677 0.7685 0.7673
0.3647 4.0 2584 0.4497 0.7720 0.7699 0.7697 0.7701
0.3502 5.0 3230 0.4530 0.7679 0.7661 0.7658 0.7668

Framework versions

  • Transformers 4.36.2
  • Pytorch 2.1.2+cu121
  • Datasets 2.16.1
  • Tokenizers 0.15.0

References

[1] Mahendra, R., Aji, A. F., Louvan, S., Rahman, F., & Vania, C. (2021, November). IndoNLI: A Natural Language Inference Dataset for Indonesian. Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics.

Downloads last month
94
Safetensors
Model size
11.7M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for LazarusNLP/indobert-lite-base-p1-indonli-distil-mdeberta

Finetuned
(4)
this model

Dataset used to train LazarusNLP/indobert-lite-base-p1-indonli-distil-mdeberta

Collection including LazarusNLP/indobert-lite-base-p1-indonli-distil-mdeberta