--- license: apache-2.0 task_categories: - visual-question-answering language: - en pretty_name: BenchLMM size_categories: - n<1K --- # Dataset Card for BenchLMM BenchLMM is a benchmarking dataset focusing on the cross-style visual capability of large multimodal models. It evaluates these models' performance in various visual contexts. ## Dataset Details ### Dataset Description - **Curated by:** Rizhao Cai, Zirui Song, Dayan Guan, Zhenhao Chen, Xing Luo, Chenyu Yi, and Alex Kot. - **Funded by :** Supported in part by the Rapid-Rich Object Search (ROSE) Lab of Nanyang Technological University and the NTU-PKU Joint Research Institute. - **Shared by :** AIFEG. - **Language(s) (NLP):** English. - **License:** Apache-2.0. ### Dataset Sources - **Repository:** [GitHub - AIFEG/BenchLMM](https://github.com/AIFEG/BenchLMM) - **Paper :** Cai, R., Song, Z., Guan, D., et al. (2023). BenchLMM: Benchmarking Cross-style Visual Capability of Large Multimodal Models. arXiv:2312.02896. ## Uses ### Direct Use The dataset can be used to benchmark large multimodal models, especially focusing on their capability to interpret and respond to different visual styles. ## Dataset Structure - **Directory Structure:** - `baseline/`: Baseline code for LLaVA and InstructBLIP. - `evaluate/`: Python code for model evaluation. - `evaluate_results/`: Evaluation results of baseline models. - `jsonl/`: JSONL files with questions, image locations, and answers. ## Dataset Creation ### Curation Rationale Developed to assess large multimodal models' performance in diverse visual contexts, helping to understand their capabilities and limitations. ### Source Data #### Data Collection and Processing The dataset consists of various visual questions and corresponding answers, structured to evaluate multimodal model performance. ## Bias, Risks, and Limitations Users should consider the specific visual contexts and question types included in the dataset when interpreting model performance. ## Citation **BibTeX:** @misc{cai2023benchlmm, title={BenchLMM: Benchmarking Cross-style Visual Capability of Large Multimodal Models}, author={Rizhao Cai and Zirui Song and Dayan Guan and Zhenhao Chen and Xing Luo and Chenyu Yi and Alex Kot}, year={2023}, eprint={2312.02896}, archivePrefix={arXiv}, primaryClass={cs.CV} } **APA:** Cai, R., Song, Z., Guan, D., Chen, Z., Luo, X., Yi, C., & Kot, A. (2023). BenchLMM: Benchmarking Cross-style Visual Capability of Large Multimodal Models. arXiv preprint arXiv:2312.02896. ## Acknowledgements This research is supported in part by the Rapid-Rich Object Search (ROSE) Lab of Nanyang Technological University and the NTU-PKU Joint Research Institute.