| dataset_info: | |
| features: | |
| - name: question_id | |
| dtype: string | |
| - name: image | |
| dtype: image | |
| - name: question | |
| dtype: string | |
| - name: answer | |
| dtype: string | |
| - name: category | |
| dtype: string | |
| configs: | |
| - config_name: default | |
| data_files: | |
| - split: test | |
| path: data/test-*.parquet | |
| # MME | |
| MME (MultiModal Evaluation) is a comprehensive benchmark for evaluating multimodal large language models across diverse perception and cognition tasks. | |