The motivation behind this research is that standard LIDAR sensors used for LIDAR-Inertial SLAM, or just mapping in some cases, are prohibitively expensive. We hope to reproduce results from $10,000 LIDAR sensors with a depth camera aboard a pan-tilt gimbal, which can cost less than $1,000 in total. In order to do this, however, we must develop smart and efficient algorithms to control the pose of the camera to best aid in localization and map creation, given characteristics of the environment.
As a first attempt at maximizing camera views, we used Deep-Q Learning to learn a control policy maximizing the number of new features seen per view. This policy is essentially open-loop and doesn't exploit environmental information, so we are currently extending this work to be environmentally aware.