ChartQAPro: A More Diverse and Challenging Benchmark for Chart Question Answering
January 17, 2026 ยท View on GitHub
๐คDataset | ๐ฅ๏ธCode | ๐Paper
Dataset
You can find our dataset on huggingface: ๐คChartQAPro Dataset
Evaluation Results
โ Evaluation Instructions
ChartQAPro is supported in VLMEvalKit, the standard evaluation toolkit for VisionโLanguage Models.
We strongly recommend using VLMEvalKit for consistent and reproducible evaluation.
๐ https://github.com/open-compass/VLMEvalKit
To evaluate your model on ChartQAPro using our repo codebase (NOT RECOMMENDED), follow the steps below:
1. Format Your Predictions
Save your model's predictions in a .json file that contains a list of dictionaries.
Each dictionary should include the following keys (first three keys taken from the original huggingface dataset):
"Answer": the ground truth answer"Question Type": the type of the question (e.g., Factoid, MCQ, etc.)"Year": useful for evaluating year-based answers"prediction": your modelโs predicted answer
๐ Example Format
[
{
"Answer": ["2016"]
"Question Type": "Factoid",
"Year": ["YES"]
"prediction": "2016"
},
...
]
2. Install Required Dependencies
pip install anls pandas
3. Run the Evaluation Script
python evaluate_predictions.py --predictions-file path/to/your/predictions.json
This will print your modelโs performance across different question types and the overall score, following the official evaluation metrics used in the paper. ๐
๐ฌ Contact
If you have any questions about this work, please contact Ahmed Masry using the following email addresses: amasry17@ku.edu.tr, ahmed.elmasry24653@gmail.com, or masry20@yorku.ca.
๐ Citation
If you use ChartQAPro in your research, please cite:
@misc{masry2025chartqaprodiversechallengingbenchmark,
title={ChartQAPro: A More Diverse and Challenging Benchmark for Chart Question Answering},
author={Ahmed Masry and Mohammed Saidul Islam and Mahir Ahmed and Aayush Bajaj and Firoz Kabir and Aaryaman Kartha and Md Tahmid Rahman Laskar and Mizanur Rahman and Shadikur Rahman and Mehrad Shahmohammadi and Megh Thakkar and Md Rizwan Parvez and Enamul Hoque and Shafiq Joty},
year={2025},
eprint={2504.05506},
archivePrefix={arXiv},
primaryClass={cs.CL},
url={https://arxiv.org/abs/2504.05506},
}