In-Person presentation / poster accept
A Simple Approach for Visual Room Rearrangement: 3D Mapping and Semantic Search
Brandon Trabucco · Gunnar Sigurdsson · Robinson Piramuthu · Gaurav Sukhatme · Ruslan Salakhutdinov
MH1-2-3-4 #148
Keywords: [ Object Rearrangement ] [ Embodied AI ] [ deep learning ] [ Reinforcement Learning ]
Physically rearranging objects is an important capability for embodied agents. Visual room rearrangement evaluates an agent's ability to rearrange objects in a room to a desired goal based solely on visual input. We propose a simple yet effective method for this problem: (1) search for and map which objects need to be rearranged, and (2) rearrange each object until the task is complete. Our approach consists of an off-the-shelf semantic segmentation model, voxel-based semantic map, and semantic search policy to efficiently find objects that need to be rearranged. Our method was the winning submission to the AI2-THOR Rearrangement Challenge in the 2022 Embodied AI Workshop at CVPR 2022, and improves on current state-of-the-art end-to-end reinforcement learning-based methods that learn visual room rearrangement policies from 0.53% correct rearrangement to 16.56%, using only 2.7% as many samples from the environment.