4.6 Article

Multi-Objective Location and Mapping Based on Deep Learning and Visual Slam

Journal

SENSORS
Volume 22, Issue 19, Pages -

Publisher

MDPI
DOI: 10.3390/s22197576

Keywords

deep learning; target tracking; visual SLAM; multi-objective location; semantic mapping

Ask authors/readers for more resources

SLAM technology is used for locating and mapping in unknown environments, but the constructed maps often lack readability and interactivity, making it difficult to grasp the information accurately. To enable intelligent robots to interact meaningfully with their environment, it is necessary to understand both the geometric and semantic properties of the scene. The proposed method reduces absolute positional errors, constructs dense semantic point cloud maps, and segments point cloud models of objects in the environment with high accuracy.
Simultaneous localization and mapping (SLAM) technology can be used to locate and build maps in unknown environments, but the constructed maps often suffer from poor readability and interactivity, and the primary and secondary information in the map cannot be accurately grasped. For intelligent robots to interact in meaningful ways with their environment, they must understand both the geometric and semantic properties of the scene surrounding them. Our proposed method can not only reduce the absolute positional errors (APE) and improve the positioning performance of the system but also construct the object-oriented dense semantic point cloud map and output point cloud model of each object to reconstruct each object in the indoor scene. In fact, eight categories of objects are used for detection and semantic mapping using coco weights in our experiments, and most objects in the actual scene can be reconstructed in theory. Experiments show that the number of points in the point cloud is significantly reduced. The average positioning error of the eight categories of objects in Technical University of Munich (TUM) datasets is very small. The absolute positional error of the camera is also reduced with the introduction of semantic constraints, and the positioning performance of the system is improved. At the same time, our algorithm can segment the point cloud model of objects in the environment with high accuracy.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.6
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available