- Authors: Shankar Kantharaj, Xuan Long Do, Rixie Tiffany Ko Leong, Jia Qing Tan, Enamul Hoque, Shafiq Joty
- Paper Link: OpenCQA
The OpenCQA images are available in the chart_images folder. The full annotations are available in the etc/data(full_summary_article) folder. They are saved as (train/val/test)_extended.json files for the train/val/test split.The dataset has the following structure:
{
image_no: [ image_file,
title,
article,
summary,
question,
abstractive_answer,
extractive_answer
],
....
....
....
}
Please refer to VL-T5
Please refer to T5
Please refer to BART
Please refer to BERTQA
Please refer to DOC2GRND
Please refer to ELECTRA
Please refer to GPT2
If you have any questions about this work, please contact Enamul Hoque using the following email address: [email protected].
Please cite our paper if you use our models or dataset in your research.
@inproceedings{kantharaj-etal-2022-opencqa,
title = "{O}pen{CQA}: Open-ended Question Answering with Charts",
author = "Kantharaj, Shankar and
Do, Xuan Long and
Leong, Rixie Tiffany and
Tan, Jia Qing and
Hoque, Enamul and
Joty, Shafiq",
booktitle = "Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing",
month = dec,
year = "2022",
address = "Abu Dhabi, United Arab Emirates",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2022.emnlp-main.811",
pages = "11817--11837",
abstract = "Charts are very popular to analyze data and convey important insights. People often analyze visualizations to answer open-ended questions that require explanatory answers. Answering such questions are often difficult and time-consuming as it requires a lot of cognitive and perceptual efforts. To address this challenge, we introduce a new task called OpenCQA, where the goal is to answer an open-ended question about a chart with descriptive texts. We present the annotation process and an in-depth analysis of our dataset. We implement and evaluate a set of baselines under three practical settings. In the first setting, a chart and the accompanying article is provided as input to the model. The second setting provides only the relevant paragraph(s) to the chart instead of the entire article, whereas the third setting requires the model to generate an answer solely based on the chart. Our analysis of the results show that the top performing models generally produce fluent and coherent text while they struggle to perform complex logical and arithmetic reasoning.",
}