A Fine-Tuned BART Pre-trained Language Model for the Indonesian Question-Answering Task

Authors

  • Alfonso Darren Vincentio Informatics Department, Universitas Multimedia Nusantara, Tangerang, Indonesia
  • Seng Hansun Informatics Department, Universitas Multimedia Nusantara, Tangerang, Indonesia
Volume: 15 | Issue: 2 | Pages: 21398-21403 | April 2025 | https://doi.org/10.48084/etasr.9828

Abstract

The information extraction process from a given context can be time consuming and a Pre-trained Language Model (PLM) based on the transformer architecture could reduce the time needed to obtain the information. Moreover, PLM is easily fine-tuned to accomplish certain tasks, one of which is the Question-Answering (QA) task. In literature, QA tasks are generally fine-tuned using encoder-based PLMs, such as the Bidirectional Encoder Representations from Transformers (BERT), where the generated answers come from the extraction process of the context. In order to be able to return more abstract answers, a PLM with Natural Language Generation (NLG) capability, such as the Bidirectional and Auto-Regressive Transformer (BART), is needed. In this study, we aim to fine-tune the NLG PLM using BART to build a more abstractive generative QA task. Based on the experimental results, the fine-tuned BART model performs well with an 85.84 F1 score and a 59.42 Exact Match (EM) score.

Keywords:

BART, BERT, NLG, PLM, QA task

Downloads

Download data is not yet available.

References

A. Bouziane, D. Bouchiha, N. Doumi, and M. Malki, "Question Answering Systems: Survey and Trends," Procedia Computer Science, vol. 73, pp. 366–375, Jan. 2015.

J. Devlin, M.-W. Chang, K. Lee, and K. Toutanova, "BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding," in Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Minneapolis, MN, USA, Mar. 2019, pp. 4171–4186.

A. Radford, J. Wu, R. Child, D. Luan, D. Amodei, and I. Sutskever, "Language Models are Unsupervised Multitask Learners," 2019. Available: https://cdn.openai.com/better-language-models/language_

models_are_unsupervised_multitask_learners.pdf

S. Giddaluru, S. M. Maturi, O. Ooruchintala, and M. Munirathnam, "Stance Detection in Hinglish Data using the BART-large-MNLI Integration Model," Engineering, Technology & Applied Science Research, vol. 14, no. 4, pp. 15477–15481, Aug. 2024.

M. Lewis et al., "BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension," in 58th Annual Meeting of the Association for Computational Linguistics, Jul. 2020, pp. 7871–7880.

A. Vaswani et al., "Attention is all you need," in 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, Dec. 2017, pp. 6000–6010.

T. B. Brown et al., "Language models are few-shot learners," in 34th International Conference on Neural Information Processing Systems, Vancouver, BC, Canada, Dec. 2020, pp. 1877–1901.

Y. Liu et al., "Multilingual Denoising Pre-training for Neural Machine Translation," Transactions of the Association for Computational Linguistics, vol. 8, pp. 726–742, Dec. 2020.

Y. Tang et al., "Multilingual Translation from Denoising Pre-Training," in Findings of the Association for Computational Linguistics: ACL-IJCNLP, Aug. 2021, pp. 3450–3466.

J. H. Clark et al., "TyDi QA: A Benchmark for Information-Seeking Question Answering in Typologically Diverse Languages," Transactions of the Association for Computational Linguistics, vol. 8, pp. 454–470, Jul. 2020.

T. Wolf et al., "Transformers: State-of-the-Art Natural Language Processing," in Conference on Empirical Methods in Natural Language Processing: System Demonstrations, Jul. 2020, pp. 38–45.

K. Papineni, S. Roukos, T. Ward, and W.-J. Zhu, "BLEU: a method for automatic evaluation of machine translation," in 40th Annual Meeting on Association for Computational Linguistics, Philadelphia, PA, USA, Jul. 2012, pp. 311–318.

C.-Y. Lin, "ROUGE: A Package for Automatic Evaluation of Summaries," in Text Summarization Branches Out, Barcelona, Spain, Apr. 2004, pp. 74–81.

R. A. Putri and A. Oh, "IDK-MRC: Unanswerable Questions for Indonesian Machine Reading Comprehension," in Conference on Empirical Methods in Natural Language Processing, Abu Dhabi, United Arab Emirates, Dec. 2022, pp. 6918–6933.

M. Fuadi and A. D. Wibawa, "Automatic Question Generation from Indonesian Texts Using Text-to-Text Transformers," in International Conference on Electrical and Information Technology, Malang, Indonesia, Sep. 2022, pp. 84–89.

B. Richardson and A. Wicaksana, "Comparison of indobert-lite and roberta in text mining for Indonesian language question answering application," International Journal of Innovative Computing, Information and Control, vol. 18, no. 6, pp. 1719–1734, Dec. 2022.

M. M. Henry, G. N. Elwirehardja, and B. Pardamean, "Automatic question generation for bahasa indonesia examination using copynet," Procedia Computer Science, vol. 245, pp. 953–962, Jan. 2024.

P. Rajpurkar, J. Zhang, K. Lopyrev, and P. Liang, "SQuAD: 100,000+ Questions for Machine Comprehension of Text," in Conference on Empirical Methods in Natural Language Processing, Austin, TX, USA, Nov. 2016, pp. 2383–2392.

P. Rajpurkar, R. Jia, and P. Liang, "Know What You Don`t Know: Unanswerable Questions for SQuAD," in 56th Annual Meeting of the Association for Computational Linguistics, Melbourne, Australia, Jul. 2018, pp. 784–789.

M. Artetxe, S. Ruder, and D. Yogatama, "On the Cross-lingual Transferability of Monolingual Representations," in 58th Annual Meeting of the Association for Computational Linguistics, Jul. 2020, pp. 4623–4637.

L. Xue et al., "mT5: A Massively Multilingual Pre-trained Text-to-Text Transformer," in Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Jun. 2021, pp. 483–498.

Y. Liu et al., "RoBERTa: A Robustly Optimized BERT Pretraining Approach." arXiv, Jul. 26, 2019.

V. Sanh, L. Debut, J. Chaumond, and T. Wolf, "DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter," in 33rd Conference on Neural Information Processing Systems, Vancouver, BC, Canada, Dec. 2019, pp. 1–5.

C. Raffel et al., "Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer," Journal of Machine Learning Research, vol. 21, no. 140, pp. 1–67, 2020.

S. Hansun, A. Suryadibrata, R. Nurhasanah, and J. Fitra, "Tweets sentiment on ppkm policy as a covid-19 response in indonesia," Indian Journal of Computer Science and Engineering, vol. 13, no. 1, pp. 51–58, Feb. 2022.

S. Hansun, A. Wicaksana, and A. Q. M. Khaliq, "Multivariate cryptocurrency prediction: comparative analysis of three recurrent neural networks approaches," Journal of Big Data, vol. 9, no. 1, Apr. 2022, Art. no. 50.

T. A. Wotaifi and B. N. Dhannoon, "An Effective Hybrid Deep Neural Network for Arabic Fake News Detection," Baghdad Science Journal, vol. 20, no. 4, pp. 1392–1392, Aug. 2023.

A. K. Jassim, M. J. Hamzah, and A. H. Aliwy, "Using Graph Mining Method in Analyzing Turkish Loanwords Derived from Arabic Language," Baghdad Science Journal, vol. 19, no. 6, pp. 1369–1369, Dec. 2022.

B. Wilie et al., "IndoNLU: Benchmark and Resources for Evaluating Indonesian Natural Language Understanding," in 1st Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics and the 10th International Joint Conference on Natural Language Processing, Suzhou, China, Dec. 2020, pp. 843–857.

S. Cahyawijaya et al., "IndoNLG: Benchmark and Resources for Evaluating Indonesian Natural Language Generation," in Conference on Empirical Methods in Natural Language Processing, Nov. 2021, pp. 8875–8898.

A. Wang, A. Singh, J. Michael, F. Hill, O. Levy, and S. Bowman, "GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding," in EMNLP Workshop BlackboxNLP: Analyzing and Interpreting Neural Networks for NLP, Brussels, Belgium, Nov. 2018, pp. 353–355.

S. Gehrmann et al., "The GEM Benchmark: Natural Language Generation, its Evaluation and Metrics," in 1st Workshop on Natural Language Generation, Evaluation, and Metrics, Bangkok, Thailand, Aug. 2021, pp. 96–120.

Downloads

How to Cite

[1]
Vincentio, A.D. and Hansun, S. 2025. A Fine-Tuned BART Pre-trained Language Model for the Indonesian Question-Answering Task. Engineering, Technology & Applied Science Research. 15, 2 (Apr. 2025), 21398–21403. DOI:https://doi.org/10.48084/etasr.9828.

Metrics

Abstract Views: 1
PDF Downloads: 2

Metrics Information