| | --- |
| | pretty_name: ENACT |
| | language: |
| | - en |
| | task_categories: |
| | - visual-question-answering |
| | configs: |
| | - config_name: default |
| | data_files: |
| | - QA.zip |
| | dataset_info: |
| | features: |
| | - name: id |
| | dtype: string |
| | - name: type |
| | dtype: string |
| | - name: task_name |
| | dtype: string |
| | - name: key_frame_ids |
| | sequence: string |
| | - name: images |
| | sequence: string |
| | - name: question |
| | dtype: string |
| | - name: options |
| | sequence: string |
| | - name: gt_answer |
| | sequence: int32 |
| | license: mit |
| | tags: |
| | - agent |
| | size_categories: |
| | - 1K<n<10K |
| | --- |
| | |
| | # ENACT: Evaluating Embodied Cognition with World Modeling of Egocentric Interaction |
| |
|
| | ENACT is a benchmark dataset for evaluating **embodied cognition** in vision–language models via **egocentric world modeling**. It probes whether models can reason about how the world changes under sequences of actions, using long-horizon household activities in a mobile manipulation setting. |
| |
|
| | - **Project page:** https://enact-embodied-cognition.github.io/ |
| | - **Code & evaluation:** https://github.com/mll-lab-nu/ENACT |
| | - **Paper** https://arxiv.org/abs/2511.20937 |
| |
|
| |
|
| | ## Dataset Summary |
| |
|
| | Each ENACT example is a **multi-image, multi-step reasoning problem** built from robot trajectories: |
| |
|
| | - **Forward world modeling** |
| | - Input: one **current state image**, several **future state images** (shuffled), and a list of **actions in correct order**. |
| | - Task: output a Python list of integers giving the **correct chronological order of future images** (e.g., `[1, 3, 2]`). |
| |
|
| | - **Inverse world modeling** |
| | - Input: an **ordered sequence of images** showing state changes, plus **actions in shuffled order**. |
| | - Task: output a Python list of integers giving the **correct chronological order of actions** (e.g., `[2, 3, 1]`). |
| |
|
| | All images are egocentric RGB observations rendered from long-horizon household tasks (e.g., assembling gift baskets, bringing water, preparing lunch boxes, cleaning up a desk). |
| |
|
| |
|
| | ## File Structure |
| |
|
| | After unpacking, the dataset has the following structure: |
| |
|
| | ```text |
| | . |
| | ├── enact_ordering.jsonl # All QA examples (one JSON per line) |
| | └── images/ |
| | ├── forward_world_modeling_3_steps/ |
| | ├── forward_world_modeling_4_steps/ |
| | ├── ... |
| | ├── forward_world_modeling_10_steps/ |
| | ├── inverse_world_modeling_3_steps/ |
| | ├── ... |
| | └── inverse_world_modeling_10_steps/ |
| | ```` |
| |
|
| | Each task folder (e.g., `forward_world_modeling_3_steps/`) contains one subfolder per activity, such as: |
| |
|
| | ```text |
| | images/forward_world_modeling_3_steps/ |
| | ├── assembling_gift_baskets_1749468508582193/ |
| | ├── bringing_water_1750844141719178/ |
| | ├── ... |
| | ``` |
| |
|
| | Inside each activity folder are the PNGs for that trajectory (current state and future states, or ordered states in the inverse setting). |
| |
|
| |
|
| | ## JSONL Format |
| |
|
| | Each line in `enact_ordering.jsonl` is a JSON object: |
| |
|
| | ```json |
| | { |
| | "id": "assembling_gift_baskets_1749468508582193_forward_world_modeling_3_steps_cfbcc15c", |
| | "type": "forward_world_modeling_3_steps", |
| | "task_name": "assembling_gift_baskets_1749468508582193", |
| | "key_frame_ids": ["4150", "11360", "11834"], |
| | "images": [ |
| | "QA/images/forward_world_modeling_3_steps/..._cur_state.png", |
| | "QA/images/forward_world_modeling_3_steps/..._next_state_1.png", |
| | "QA/images/forward_world_modeling_3_steps/..._next_state_2.png" |
| | ], |
| | "question": "...natural language instructions and actions...", |
| | "options": [], |
| | "gt_answer": [1, 2] |
| | } |
| | ``` |
| |
|
| | * **`id`** – unique identifier for this QA instance. |
| | * **`type`** – question type and horizon, e.g. `forward_world_modeling_3_steps` or `inverse_world_modeling_4_steps`. |
| | * **`task_name`** – underlying household task instance. |
| | * **`key_frame_ids`** – frame indices of selected key frames in the trajectory. |
| | * **`images`** – relative paths to PNG images: |
| | |
| | * index 0 is the **current state**; |
| | * subsequent entries are **future states** (forward) or later states (inverse). |
| | * **`question`** – natural language prompt specifying the task setup, actions, and the required output as a Python list of integers. |
| | * **`gt_answer`** – ground-truth ordering of image or action labels (list of integers, e.g. `[1, 3, 2]`). |
| |
|
| |
|
| | ## Usage |
| | To evaluate, follow the scripts in the code repository: [https://github.com/mll-lab-nu/ENACT](https://github.com/mll-lab-nu/ENACT) |
| |
|
| |
|
| | ## Citation |
| |
|
| | If you use ENACT, please cite the paper: |
| | ``` |
| | @article{wang2025enact, |
| | title={ENACT: Evaluating Embodied Cognition with World Modeling of Egocentric Interaction}, |
| | author={Wang, Qineng and Huang, Wenlong and Zhou, Yu and Yin, Hang |
| | and Bao, Tianwei and Lyu, Jianwen and Liu, Weiyu and Zhang, Ruohan |
| | and Wu, Jiajun and Li, Fei-Fei and Li, Manling}, |
| | journal={arXiv preprint arXiv:2511.20937}, |
| | year={2025} |
| | } |
| | ``` |