-
Notifications
You must be signed in to change notification settings - Fork 559
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Multiple camera SLAM and localization #1235
Comments
Can you give example/screenshot of what you mean by "two cameras detect the lines"? For the masked images, what are they? |
There are some lines on the floor surface, I would like to segment those lines out and use those segmented lines for the localization and mapping. |
Masked images are images only contains the lines |
Based on those pictures, it seems that the textured floor has actually good visual features to track. Removing them would make localization/visual odometry worst. To use a mask like this to limit features extraction, you can apply it to the depth image. By default rtabmap won't extract features when there is no valid depth.
I feel that if the environment is structured with lines and you can detect them quite accurately, I would suggest to try 2D lidar-based SLAM/localization approaches by converting the lines seen by he camera into a fake laser scan. A similar option would be to use icp_odometry using a 2D PointCloud2 |
Hi,
I really admire your work and currently I am working on the multiple cameras setup for SLAM and localization.
The difficult thing of this is that two cameras detect the lines and the masked images are used to generate point clouds and use those point clouds for localization and map building. Could you give me some advice?
Thank you so much.
The text was updated successfully, but these errors were encountered: