Weakly Supervised Multimodal Affordance Grounding for Egocentric Images(AAAI 2024)
Appendix
Link: Appendix.pdf
Abstract:
To enhance the interaction between intelligent systems and the environment, locating the affordance regions of objects is crucial. These regions correspond to specific areas that provide distinct functionalities. Humans often acquire the ability to identify these regions through action demonstrations and verbal instructions. In this paper, we present a novel multimodal framework that extracts affordance knowledge from exocentric images, which depict human-object interactions, as well as from accompanying textual descriptions that describe the performed actions. The extracted knowledge is then transferred to egocentric images. To achieve this goal, we propose the HOI-Transfer Module, which utilizes local perception to disentangle individual actions within exocentric images. This module effectively captures localized features and correlations between actions, leading to valuable affordance knowledge. Additionally, we introduce the Pixel-Text Fusion Module, which fuses affordance knowledge by identifying regions in egocentric images that bear resemblances to the textual features defining affordances. We employ a Weakly Supervised Multimodal Affordance (WSMA) learning approach, utilizing image-level labels for training. Through extensive experiments, we demonstrate the superiority of our proposed method in terms of evaluation metrics and visual results when compared to existing affordance grounding models. Furthermore, ablation experiments confirm the effectiveness of our approach.
We run in the following environment:
- A NVIDIA GeForce RTX 3090
- Python(3.8)
- Pytorch(1.10.0)
- model for Dino_vit(No need to download separately, the code is already included)
- model for text_enconder(clip): You can find it here
git clone https://github.com/xulingjing88/WSMA.git
cd WSMABefore training, you need to preprocess the data
python preprocessing.pySet 'data_root' to the path of the dataset, 'divide' to the dataset name (Seen or Unseen or HICO-IIF), and then you can start training by running train.py.
python train.pyWe would like to express our gratitude to the following repositories for their contributions and inspirations: Cross-View-AG, LOCATE, Dino, CLIP.
