--- license: cc-by-nc-4.0 language: - en task_categories: - question-answering - visual-question-answering task_ids: - multiple-choice-qa pretty_name: EgoMemReason size_categories: - n<1K tags: - egocentric-video - long-video-understanding - memory - multimodal - benchmark - video-qa configs: - config_name: default data_files: - split: test path: annotations_public.jsonl --- # EgoMemReason **A Memory-driven Reasoning Benchmark for Long-Horizon Egocentric Video Understanding.** 500 multiple-choice questions over **week-long egocentric video** (built on [EgoLife](https://egolife-ai.github.io/)) that evaluate three complementary kinds of memory: - **Entity memory** — track how object states evolve across days - **Event memory** — recall and order activities separated by hours or days - **Behavior memory** — abstract recurring patterns from sparse, repeated observations Average **5.1 evidence segments per question** and **25.9 hours of memory backtracking** — 2× both metrics over the strongest prior week-long benchmark. ## Links - 🧠 **Leaderboard (HF Space):** - 💻 **Code & reference eval scripts:** - 🌐 **Project page:** - 🎬 **EgoLife video frames (separate license):** - 📄 **Paper:** ## Composition | Memory type | Capability (`query_type`) | # Qs | |---|---|---:| | Entity | Cumulative State Tracking | 100 | | Entity | Temporal Counting | 100 | | Event | Event Ordering | 100 | | Event | Event Linking | 100 | | Behavior | Spatial Preference | 50 | | Behavior | Activity Pattern | 50 | | **Total** | | **500** | ## Schema This dataset releases the **public** version — questions and options only, no answer keys (the held-out answer key lives in a private dataset, and submissions are scored against it by the leaderboard Space). ```json { "example_id": 1, "p_id": "A1_JAKE_DAY7_19_00_00_q001", "identity": "A1_JAKE", "query_time": "DAY7, 19:00:00", "question": "What do I most often eat for breakfast?", "options": { "A": "Pancake", "B": "Rice", "C": "Burger", "D": "Dumplings" }, "query_type": "Activity Pattern" } ``` Note that **questions have 4-10 options** (letters A-J). The valid answer set for any given question is the keys of its `options` dict; Event Ordering questions tend to have the most options. ## How to evaluate 1. Get this dataset: ```python from datasets import load_dataset ds = load_dataset("Ted412/EgoMemReason")["test"] ``` 2. Get the underlying EgoLife video frames (separate license, see ) — we don't redistribute video here. 3. For each item, sample frames from `(identity, query_time)` backwards in time and run your model to pick one letter from `options.keys()`. 4. Format the predictions as a JSON list: ```json [ {"example_id": 1, "predicted_answer": "A"}, ... ] ``` 5. Submit it on the leaderboard Space: . Per-split + overall accuracy are computed automatically. The reference inference scripts for 12 MLLMs and 5 agentic frameworks (Gemini, GPT-5, Qwen3-VL, InternVL3.5, Molmo2, VideoLLaMA3, InternVideo2.5, LongVA, AVP, Ego-R1, SiLVR, WorldMM, …) live in the [GitHub repo](https://github.com/Ziyang412/EgoMemReason). ## License - **EgoMemReason annotations** (this dataset): [CC BY-NC 4.0](https://creativecommons.org/licenses/by-nc/4.0/) — academic research and benchmarking are permitted; commercial use requires written permission. - **EgoLife video frames** (not redistributed here): governed by the [EgoLife data license](https://egolife-ai.github.io/) — you must accept their terms separately. ## Citation ```bibtex @misc{wang2026egomemreasonmemorydrivenreasoningbenchmark, title={EgoMemReason: A Memory-Driven Reasoning Benchmark for Long-Horizon Egocentric Video Understanding}, author={Ziyang Wang and Yue Zhang and Shoubin Yu and Ce Zhang and Zengqi Zhao and Jaehong Yoon and Hyunji Lee and Gedas Bertasius and Mohit Bansal}, year={2026}, eprint={2605.09874}, archivePrefix={arXiv}, primaryClass={cs.CV}, url={https://arxiv.org/abs/2605.09874}, } ```