Skip to content

Latest commit

 

History

History
50 lines (33 loc) · 1.87 KB

File metadata and controls

50 lines (33 loc) · 1.87 KB

Task-name

Paper

Title: ChartQA: A Benchmark for Question Answering about Charts with Visual and Logical Reasoning

Abstract: In this work, we present a large-scale benchmark covering 9.6K human-written questions as well as 23.1K questions generated from human-written chart summaries.

Short description of paper / benchmark goes here:

Homepage: https://github.com/vis-nlp/ChartQA

Citation

@misc{masry2022chartqabenchmarkquestionanswering,
      title={ChartQA: A Benchmark for Question Answering about Charts with Visual and Logical Reasoning},
      author={Ahmed Masry and Do Xuan Long and Jia Qing Tan and Shafiq Joty and Enamul Hoque},
      year={2022},
      eprint={2203.10244},
      archivePrefix={arXiv},
      primaryClass={cs.CL},
      url={https://arxiv.org/abs/2203.10244},
}

Groups, Tags, and Tasks

Tasks

  • chartqa: Prompt taken from on mistral-evals: https://github.com/mistralai/mistral-evals/blob/main/eval/tasks/chartqa.py
  • chartqa_llama: variant as implemented in https://github.com/meta-llama/llama-models/blob/main/models/llama3_3/eval_details.md
  • chartqa_llama_90: similar to chartqa_llama but specific to the 90B models of llama 3.2

Checklist

For adding novel benchmarks/datasets to the library:

  • Is the task an existing benchmark in the literature?
    • Have you referenced the original paper that introduced the task?
    • If yes, does the original paper provide a reference implementation? If so, have you checked against the reference implementation and documented how to run such a test?

If other tasks on this dataset are already supported:

  • Is the "Main" variant of this task clearly denoted?
  • Have you provided a short sentence in a README on what each new variant adds / evaluates?
  • Have you noted which, if any, published evaluation setups are matched by this variant?

Changelog