Datasets:

Modalities:
Image
Text
Languages:
English
Size:
< 1K
ArXiv:
Libraries:
Datasets
License:
BenchLMM / README.md
Ziruibest's picture
Update README.md
4902993
metadata
license: apache-2.0
task_categories:
  - visual-question-answering
language:
  - en
pretty_name: BenchLMM
size_categories:
  - n<1K

Dataset Card for BenchLMM

BenchLMM is a benchmarking dataset focusing on the cross-style visual capability of large multimodal models. It evaluates these models' performance in various visual contexts.

Dataset Details

Dataset Description

  • Curated by: Rizhao Cai, Zirui Song, Dayan Guan, Zhenhao Chen, Xing Luo, Chenyu Yi, and Alex Kot.
  • Funded by : Supported in part by the Rapid-Rich Object Search (ROSE) Lab of Nanyang Technological University and the NTU-PKU Joint Research Institute.
  • Shared by : AIFEG.
  • Language(s) (NLP): English.
  • License: Apache-2.0.

Dataset Sources

  • Repository: GitHub - AIFEG/BenchLMM
  • Paper : Cai, R., Song, Z., Guan, D., et al. (2023). BenchLMM: Benchmarking Cross-style Visual Capability of Large Multimodal Models. arXiv:2312.02896.

Uses

Direct Use

The dataset can be used to benchmark large multimodal models, especially focusing on their capability to interpret and respond to different visual styles.

Dataset Structure

  • Directory Structure:
    • baseline/: Baseline code for LLaVA and InstructBLIP.
    • evaluate/: Python code for model evaluation.
    • evaluate_results/: Evaluation results of baseline models.
    • jsonl/: JSONL files with questions, image locations, and answers.

Dataset Creation

Curation Rationale

Developed to assess large multimodal models' performance in diverse visual contexts, helping to understand their capabilities and limitations.

Source Data

Data Collection and Processing

The dataset consists of various visual questions and corresponding answers, structured to evaluate multimodal model performance.

Bias, Risks, and Limitations

Users should consider the specific visual contexts and question types included in the dataset when interpreting model performance.

Citation

BibTeX: @misc{cai2023benchlmm, title={BenchLMM: Benchmarking Cross-style Visual Capability of Large Multimodal Models}, author={Rizhao Cai and Zirui Song and Dayan Guan and Zhenhao Chen and Xing Luo and Chenyu Yi and Alex Kot}, year={2023}, eprint={2312.02896}, archivePrefix={arXiv}, primaryClass={cs.CV} }

APA: Cai, R., Song, Z., Guan, D., Chen, Z., Luo, X., Yi, C., & Kot, A. (2023). BenchLMM: Benchmarking Cross-style Visual Capability of Large Multimodal Models. arXiv preprint arXiv:2312.02896.

Acknowledgements

This research is supported in part by the Rapid-Rich Object Search (ROSE) Lab of Nanyang Technological University and the NTU-PKU Joint Research Institute.