According to this paper
RTAB-Map as an Open-Source Lidar and Visual SLAM Library for Large-Scale and Long-Term Online Operation,
the motion estimation is done by PnP RANSAC.
but according to this answer
http://official-rtab-map-forum.206.s1.nabble.com/How-does-RGBD-Slam-works-td191.htmlit is done by 3D features feature RANSAC
I'm kind of confused here. So how it is done?
To my knowledge, PnpRANSAC is more like a mono camera case. Match 2D points to the known 3D features.
With RGBD and Stereo we have the 3D inforamtion, why do we still need PnP RANSAC in the paper mentioned
Thank you for your repaly