ThaiOCRBench is the first comprehensive benchmark for evaluating vision-language models (VLMs) on Thai text-rich visual understanding tasks. Inspired by OCRBench v2, it includes 2,808 human-annotated samples across 13 tasks such as table parsing, chart reading, OCR, key information extraction, and visual question answering. The benchmark provides standardized zero-shot evaluation for both proprietary and open-source models, revealing performance gaps and advancing document understanding for low-resource languages.
2025.10.25🚀 Our paper ThaiOCRBench has been accepted to the IJCNLP-AACL 2025 Main Conference!
👉 📄 Read the Paper
👉 💻 Huggingface dataset
All Python dependencies required for the evaluation process are specified in the requirements.txt. To set up the environment, simply run the following commands in the project directory:
conda create -n thai_ocrbench python==3.10 -y
conda activate thai_ocrbench
pip install -r requirements.txtTo evaluate the model's performance on ThaiOCRBench, please run the following command.
CUDA_VISIBLE_DEVICES=0 python ./eval_scripts/run_inference.py \
--model_name qwen3b \
--output_path "./pred_folder/qwen3b.json" \
--hf_token "YOUR_TOKEN" \
--max_samples 10After obtaining the inference results from the model, you can use the following scripts to calculate the final score for ThaiOCRBench.
python ./eval_scripts/eval.py --input_path ./pred_folder/qwen3b.json --output_path ./res_folder/qwen3b.json
We did not benchmark Typhoon OCR because:
- The response format is different.
- Typhoon OCR only supports a single task — "Document Parsing"
If you use ThaiOCRBench in your research or applications, please cite our work:
@misc{nonesung2025thaiocrbenchtaskdiversebenchmarkvisionlanguage,
title={ThaiOCRBench: A Task-Diverse Benchmark for Vision-Language Understanding in Thai},
author={Surapon Nonesung and Teetouch Jaknamon and Sirinya Chaiophat and Natapong Nitarach and Chanakan Wittayasakpan and Warit Sirichotedumrong and Adisai Na-Thalang and Kunat Pipatanakul},
year={2025},
eprint={2511.04479},
archivePrefix={arXiv},
primaryClass={cs.CL},
url={https://arxiv.org/abs/2511.04479},
}
