A Large-scale Embodied Question Answering (EQA) benchmark and method
Embodied Question Answering (EQA) is a challenging task in embodied intelligence that requires agents to dynamically explore 3D environments, actively gather visual information, and perform multi-step reasoning to answer questions. However, current EQA approaches suffer from critical limitations in exploration efficiency, dataset design, and evaluation metrics. Moreover, existing datasets often introduce biases or prior knowledge, leading to disembodied reasoning, while frontier-based exploration strategies struggle in cluttered environments and fail to ensure fine-grained exploration of task-relevant areas. To address these challenges, we construct the EXPloration-awaRe Embodied queStion anSwering Benchmark (EXPRESS-Bench), the largest dataset designed specifically to evaluate both exploration and reasoning capabilities. EXPRESS-Bench consists of 777 exploration trajectories and 2,044 question-trajectory pairs. To improve exploration efficiency, we propose Fine-EQA, a hybrid exploration model that integrates frontier-based and goal-oriented navigation to guide agents toward task-relevant regions more effectively. Additionally, we introduce a novel evaluation metric, Exploration-Answer Consistency (EAC), which ensures faithful assessment by measuring the alignment between answer grounding and exploration reliability. Extensive experimental comparisons with state-of-the-art EQA models demonstrate the effectiveness of our EXPRESS-Bench in advancing embodied exploration and question reasoning.
Set up the conda environment (Linux, Python 3.9):
conda env create -f environment.yml
conda activate fine-eqa
pip install -e .
Install the latest version of Habitat-Sim on headless machines:
conda install habitat-sim headless -c conda-forge -c aihabitat
Install Prismatic VLM:
git https://github.com/TRI-ML/prismatic-vlms.git
cd prismatic-vlms
pip install -e .
EXPRESS-Bench comprises 777 exploration trajectories and 2,044 question-trajectory pairs. The corresponding question-answer pairs are stored in express-bench.json, while the full set of episodes for EXPRESS-Bench can be accessed from [Google Drive], [Baidu] and [ModelScope].
To obtain the train and val splits of the HM3D dataset, please download them here. Note that semantic annotations are required, and access must be requested in advance.
Afterward, your data directory structure should be:
|→ data
|→ episode
|→ 0000-00006-HkseAnWCgqk
|→ ...
|→ hm3d
|→ train
|→ 00000-kfPV7w3FaU5
|→ ...
|→ val
|→ 00800-TEEsavR23oF
|→ ...
|→ express-bench.json
To run the Fine-EQA model, you can use the following command:
python main.py -cf fine_eqa.yaml
This project is built upon the explore-eqa. We sincerely thank the authors for their excellent work and open-source contribution, which served as a solid foundation for our development.
If you use this code for your research, please cite our paper.
@inproceedings{EXPRESSBench,
title={Beyond the Destination: A Novel Benchmark for Exploration-Aware Embodied Question Answering},
author={Jiang, Kaixuan and Liu, Yang and Chen, Weixing and Luo, Jingzhou and Chen, Ziliang and Pan, Ling and Li, Guanbin and Lin, Liang},
year={2025}
booktitle={IEEE/CVF International Conference on Computer Vision (ICCV)}
}
If you have any question about this code, feel free to reach ([email protected] or [email protected]).