[LREC-COLING 2024] Benchmarking Large Language Models for Persian: A Preliminary Study Focusing on ChatGPT
This repo covers the implementation of the following paper: Benchmarking Large Language Models for Persian: A Preliminary Study Focusing on ChatGPT by Amirhossein Abaskohi, Sara Baruni, Mostafa Masoudi, Nesa Abbasi, Mohammad Hadi Babalou, Ali Edalat, Sepehr Kamahi, Samin Mahdizadeh Sani, Nikoo Naghavian, Danial Namazifard, Pouya Sadeghi and Yadollah Yaghoobzadeh , accepted to LREC-COLING 2024.
This paper explores the efficacy of large language models (LLMs) for Persian. While ChatGPT and consequent LLMs have shown remarkable performance in English, their efficiency for more low-resource languages remains an open question. We present the first comprehensive benchmarking study of LLMs across diverse Persian language tasks. Our primary focus is on GPT-3.5-turbo, but we also include GPT-4 and OpenChat-3.5 to provide a more holistic evaluation. Our assessment encompasses a diverse set of tasks categorized into classic, reasoning, and knowledge-based domains. To enable a thorough comparison, we evaluate LLMs against existing task-specific fine-tuned models. Given the limited availability of Persian datasets for reasoning tasks, we introduce two new benchmarks: one based on elementary school math questions and another derived from the entrance exams for 7th and 10th grades. Our findings reveal that while LLMs, especially GPT-4, excel in tasks requiring reasoning abilities and a broad understanding of general knowledge, they often lag behind smaller pre-trained models fine-tuned specifically for particular tasks. Additionally, we observe improved performance when test sets are translated to English before inputting them into GPT-3.5. These results highlight the significant potential for enhancing LLM performance in the Persian language. This is particularly noteworthy due to the unique attributes of Persian, including its distinct alphabet and writing styles.
The benchmarks
and prompts
used in our paper can be found in Benchmarks
directory.
In it for each task there are three files available:
.jsonl
file which includes the test samplesprompt.py
file which includes our prompts both in English and Farsi(Persian)sample.ipynb
which is a sample notebook for getting the evaluation results.
For the new benchmarks introduced in the paper you find them using the following links:
For the other datasets, checkout the paper for the used datasets.
We evaluated GPT-3.5, GPT-4, and OpenChat 13 tasks. The results are as followed:
GPT-3.5 | GPT-4 | ||||
---|---|---|---|---|---|
Persian Prompt (N-shot) | English Prompt (N-shot) | ||||
Category | Task | Metric | 0 | 1 | 3 |
Classic | Sentiment | Macro F1 | .725 | .804 | .791 |
Classic | Emotion | Macro F1 | .492 | .513 | .537 |
Classic | NER | Macro F1 | .578 | .472 | .589 |
Classic | MT (En → Fa) | Bleu | 7.5 | 6.9 | 7.3 |
Classic | MT (Fa → En) | Bleu | 10.5 | 10.8 | 11.0 |
Classic | Reading | F1 | .535 | .643 | .642 |
Reasoning | Textual | Macro F1 | .375 | .318 | .320 |
Reasoning | Textual | Macro F1 | .348 | .356 | .368 |
Reasoning | Multi-choice QA (math & logic) | Acc | .450 | .450 | .435 |
Reasoning | Elementary | Acc | .535 | .435 | .565 |
Reasoning | Math | Math | .209 | .375 | .503 |
Know | Multi-choice QA (literature) | Acc | .280 | .295 | .275 |
Know | Multi-choice QA (common) | Acc | .385 | .395 | .445 |
OpenChat | ||||
---|---|---|---|---|
Persian Prompt (N-shot) | English Prompt (N-shot) | |||
Category | Task | Metric | 0 | 1 |
Classic | Sentiment | Macro F1 | .460 | .484 |
Classic | Emotion | Macro F1 | .186 | .327 |
Classic | NER | Macro F1 | .241 | .603 |
Classic | MT (En → Fa) | Bleu | 5.7 | 6.3 |
Classic | MT (Fa → En) | Bleu | 9.1 | 9.1 |
Classic | Reading | F1 | .506 | .528 |
Reasoning | Textual | Macro F1 | .338 | .468 |
Reasoning | Textual | Macro F1 | .370 | .415 |
Reasoning | Multi-choice QA (math & logic) | Acc | .180 | .260 |
Reasoning | Elementary | Acc | .555 | .455 |
Reasoning | Math | Math | .128 | .229 |
Know | Multi-choice QA (literature) | Acc | .215 | .275 |
Know | Multi-choice QA (common) | Acc | .345 | .310 |
The SOTA models used as a baseline in the paper are as follows:
For each experiment, there as notebook in the paper where you can follow them step by step. Remember to replace you API-KEY
for the models. In addition, for ChatGPT
experiments, as mentioned in the paper we used two different versions. For GPT-3.5 we used gpt-3.5-turbo-0125
and for GPT-4 we used gpt-4-0125-preview
.
@misc{abaskohi2024benchmarking,
title={Benchmarking Large Language Models for Persian: A Preliminary Study Focusing on ChatGPT},
author={Amirhossein Abaskohi and Sara Baruni and Mostafa Masoudi and Nesa Abbasi and Mohammad Hadi Babalou and Ali Edalat and Sepehr Kamahi and Samin Mahdizadeh Sani and Nikoo Naghavian and Danial Namazifard and Pouya Sadeghi and Yadollah Yaghoobzadeh},
year={2024},
eprint={2404.02403},
archivePrefix={arXiv},
primaryClass={cs.CL}
}