Known object pose estimation is essential for a robot to interact with the real world. It is the first and fundamental task if the robot wants to manipulate the object. This problem is particularly challenging when the environment is complicated with clutters or the object itself is occluded. Changing in lightings and difficult orientations of the objects also bring challenges to the pose estimation algorithm. Most of the modern approaches need to obtain a large number of training data with accurate ground truth annotations to find the correspondence and output predictions. An alternative is to use a search-based algorithm that finds a pose best explains the scene in all possible rendered poses, which does not require prior knowledge or training except the model of the targeting object. PERCH(PErception Via SeaRCH) is an example that uses depth data to converge to a globally optimal solution by searching over a specific space.
In this work, we propose PERCH color-only version, a pose estimation algorithm that only needs a single RGB image and the mesh model of the target object. The experiment results from publicly available datasets show that our algorithm achieves high accuracy, especially in high occlusion scenes without the need for any annotation and training.
Maxim Likhachev (Advisor)
Zoom Participation. See announcement.