In dynamic scenarios,visual simultaneous localization and mapping(SLAM)algorithms often incorrectly incorporate dynamic points during camera pose computation,leading to reduced accuracy and robustness.This paper prese...In dynamic scenarios,visual simultaneous localization and mapping(SLAM)algorithms often incorrectly incorporate dynamic points during camera pose computation,leading to reduced accuracy and robustness.This paper presents a dynamic SLAM algorithm that leverages object detection and regional dynamic probability.Firstly,a parallel thread employs the YOLOX object detectionmodel to gather 2D semantic information and compensate for missed detections.Next,an improved K-means++clustering algorithm clusters bounding box regions,adaptively determining the threshold for extracting dynamic object contours as dynamic points change.This process divides the image into low dynamic,suspicious dynamic,and high dynamic regions.In the tracking thread,the dynamic point removal module assigns dynamic probability weights to the feature points in these regions.Combined with geometric methods,it detects and removes the dynamic points.The final evaluation on the public TUM RGB-D dataset shows that the proposed dynamic SLAM algorithm surpasses most existing SLAM algorithms,providing better pose estimation accuracy and robustness in dynamic environments.展开更多
In dynamic scenes,the pose estimation and map consistency of visual simultaneous localisation and mapping(visual SLAM)are affected by intermittent changes in object motion states.An adaptive motion-state estimation an...In dynamic scenes,the pose estimation and map consistency of visual simultaneous localisation and mapping(visual SLAM)are affected by intermittent changes in object motion states.An adaptive motion-state estimation and feature-reuse mechanism is proposed which restores features once objects become stationary.Camera ego-motion is com-pensated via projection-based point-to-point red-green-blue-depth(RGB-D)Iterative Closest Point;the alignment residual yields a short-term jitter score.An Extended Kalman Filter fuses the centre-pixel trajectory and depth of the object,using depth innovation as strong evidence to suppress false triggers.Applied adaptive decision thresholds involve resolution,ego-motion intensity,jitter,and reference depth,and are combined with dual/single triggering and hysteresis to achieve robust switching.When an object is considered static,its feature points are reused.On the Bonn RGB-D Dynamic Dataset(BONN)and TUM RGB-D SLAM Dataset and Benchmark(TUM),the proposed method matches or exceeds baselines:In intermittent-motion-dominated BONN sequences Placing_non_box,it re-duces the root-mean-square of the absolute trajectory error(ATE-RMSE)by 27%relative to the baseline,remains comparable to Ellipsoid-SLAM on TUM,and consistently outperforms ORB-SLAM3 in dynamic scenes.The hysteresis counter reading on Placing_non_box2 shows that the proposed method can reduce the motion-state misclassification rate by nearly 40%.From the ablation experiment results,we confirm that adaptive thresholds yield the most significant optimisation effect.The approach improves robustness and map completeness in dynamic environments without degrading performance in low-dynamic settings.展开更多
Background As visual simultaneous localization and mapping(SLAM)is primarily based on the assumption of a static scene,the presence of dynamic objects in the frame causes problems such as a deterioration of system rob...Background As visual simultaneous localization and mapping(SLAM)is primarily based on the assumption of a static scene,the presence of dynamic objects in the frame causes problems such as a deterioration of system robustness and inaccurate position estimation.In this study,we propose a YGC-SLAM for indoor dynamic environments based on the ORB-SLAM2 framework combined with semantic and geometric constraints to improve the positioning accuracy and robustness of the system.Methods First,the recognition accuracy of YOLOv5 was improved by introducing the convolution block attention model and the improved EIOU loss function,whereby the prediction frame converges quickly for better detection.The improved YOLOv5 was then added to the tracking thread for dynamic target detection to eliminate dynamic points.Subsequently,multi-view geometric constraints were used for re-judging to further eliminate dynamic points while enabling more useful feature points to be retained and preventing the semantic approach from over-eliminating feature points,causing a failure of map building.The K-means clustering algorithm was used to accelerate this process and quickly calculate and determine the motion state of each cluster of pixel points.Finally,a strategy for drawing keyframes with de-redundancy was implemented to construct a clear 3D dense static point-cloud map.Results Through testing on TUM dataset and a real environment,the experimental results show that our algorithm reduces the absolute trajectory error by 98.22%and the relative trajectory error by 97.98%compared with the original ORBSLAM2,which is more accurate and has better real-time performance than similar algorithms,such as DynaSLAM and DS-SLAM.Conclusions The YGC-SLAM proposed in this study can effectively eliminate the adverse effects of dynamic objects,and the system can better complete positioning and map building tasks in complex environments.展开更多
Dynamic visual SLAM (Simultaneous Localization and Mapping) is an important research area, but existing methods struggle to balance real-time performance and accuracy in removing dynamic feature points, especially whe...Dynamic visual SLAM (Simultaneous Localization and Mapping) is an important research area, but existing methods struggle to balance real-time performance and accuracy in removing dynamic feature points, especially when semantic information is missing. This paper presents a novel dynamic SLAM system that uses optical flow tracking and epipolar geometry to identify dynamic feature points and applies a regional dynamic probability method to improve removal accuracy. We developed two innovative algorithms for precise pruning of dynamic regions: first, using optical flow and epipolar geometry to identify and prune dynamic areas while preserving static regions on stationary dynamic objects to optimize tracking performance;second, propagating dynamic probabilities across frames to mitigate the impact of semantic information loss in some frames. Experiments show that our system significantly reduces trajectory and pose errors in dynamic scenes, achieving dynamic feature point removal accuracy close to that of semantic segmentation methods, while maintaining high real-time performance. Our system performs exceptionally well in highly dynamic environments, especially where complex dynamic objects are present, demonstrating its advantage in handling dynamic scenarios. The experiments also show that while traditional methods may fail in tracking when semantic information is lost, our approach effectively reduces the misidentification of dynamic regions caused by such loss, thus improving system robustness and accuracy.展开更多
A great number of visual simultaneous localization and mapping(VSLAM)systems need to assume static features in the environment.However,moving objects can vastly impair the performance of a VSLAM system which relies on...A great number of visual simultaneous localization and mapping(VSLAM)systems need to assume static features in the environment.However,moving objects can vastly impair the performance of a VSLAM system which relies on the static-world assumption.To cope with this challenging topic,a real-time and robust VSLAM system based on ORB-SLAM2 for dynamic environments was proposed.To reduce the influence of dynamic content,we incorporate the deep-learning-based object detection method in the visual odometry,then the dynamic object probability model is added to raise the efficiency of object detection deep neural network and enhance the real-time performance of our system.Experiment with both on the TUM and KITTI benchmark dataset,as well as in a real-world environment,the results clarify that our method can significantly reduce the tracking error or drift,enhance the robustness,accuracy and stability of the VSLAM system in dynamic scenes.展开更多
Visual simultaneous localization and mapping(SLAM)is crucial in robotics and autonomous driving.However,traditional visual SLAM faces challenges in dynamic environments.To address this issue,researchers have proposed ...Visual simultaneous localization and mapping(SLAM)is crucial in robotics and autonomous driving.However,traditional visual SLAM faces challenges in dynamic environments.To address this issue,researchers have proposed semantic SLAM,which combines object detection,semantic segmentation,instance segmentation,and visual SLAM.Despite the growing body of literature on semantic SLAM,there is currently a lack of comprehensive research on the integration of object detection and visual SLAM.Therefore,this study aims to gather information from multiple databases and review relevant literature using specific keywords.It focuses on visual SLAM based on object detection,covering different aspects.Firstly,it discusses the current research status and challenges in this field,highlighting methods for incorporating semantic information from object detection networks into mileage measurement,closed-loop detection,and map construction.It also compares the characteristics and performance of various visual SLAM object detection algorithms.Lastly,it provides an outlook on future research directions and emerging trends in visual SLAM.Research has shown that visual SLAM based on object detection has significant improvements compared to traditional SLAM in dynamic point removal,data association,point cloud segmentation,and other technologies.It can improve the robustness and accuracy of the entire SLAM system and can run in real time.With the continuous optimization of algorithms and the improvement of hardware level,object visual SLAM has great potential for development.展开更多
This article presents a brief survey to visual simultaneous localization and mapping (SLAM) systems applied to multiple independently moving agents, such as a team of ground or aerial vehicles, a group of users holdin...This article presents a brief survey to visual simultaneous localization and mapping (SLAM) systems applied to multiple independently moving agents, such as a team of ground or aerial vehicles, a group of users holding augmented or virtual reality devices. Such visual SLAM system, name as collaborative visual SLAM, is different from a typical visual SLAM deployed on a single agent in that information is exchanged or shared among different agents to achieve better robustness, efficiency, and accuracy. We review the representative works on this topic proposed in the past ten years and describe the key components involved in designing such a system including collaborative pose estimation and mapping tasks, as well as the emerging topic of decentralized architecture. We believe this brief survey could be helpful to someone who are working on this topic or developing multi-agent applications, particularly micro-aerial vehicle swarm or collaborative augmented/virtual reality.展开更多
Feature selection is always an important issue in the visual SLAM (simultaneous location and mapping) literature. Considering that the location estimation can be improved by tracking features with larger value of vi...Feature selection is always an important issue in the visual SLAM (simultaneous location and mapping) literature. Considering that the location estimation can be improved by tracking features with larger value of visible time, a new feature selection method based on motion estimation is proposed. First, a k-step iteration algorithm is presented for visible time estimation using an affme motion model; then a delayed feature detection method is introduced for efficiently detecting features with the maximum visible time. As a means of validation for the proposed method, both simulation and real data experiments are carded out. Results show that the proposed method can improve both the estimation performance and the computational performance compared with the existing random feature selection method.展开更多
This paper presents a hierarchical simultaneous localization and mapping(SLAM) system for a small unmanned aerial vehicle(UAV) using the output of an inertial measurement unit(IMU) and the bearing-only observati...This paper presents a hierarchical simultaneous localization and mapping(SLAM) system for a small unmanned aerial vehicle(UAV) using the output of an inertial measurement unit(IMU) and the bearing-only observations from an onboard monocular camera.A homography based approach is used to calculate the motion of the vehicle in 6 degrees of freedom by image feature match.This visual measurement is fused with the inertial outputs by an indirect extended Kalman filter(EKF) for attitude and velocity estimation.Then,another EKF is employed to estimate the position of the vehicle and the locations of the features in the map.Both simulations and experiments are carried out to test the performance of the proposed system.The result of the comparison with the referential global positioning system/inertial navigation system(GPS/INS) navigation indicates that the proposed SLAM can provide reliable and stable state estimation for small UAVs in GPS-denied environments.展开更多
A dynamic visual Simultaneous Localization and Mapping(SLAM)algorithm is proposed in this paper,which combines the YOLACT network with the geometric method to design a dynamic point detection module for eliminating dy...A dynamic visual Simultaneous Localization and Mapping(SLAM)algorithm is proposed in this paper,which combines the YOLACT network with the geometric method to design a dynamic point detection module for eliminating dynamic points.The dense optical flow-based dynamic point detection scheme is adopted to make up for the problem that the elimination algorithm based on the instance segmentation network overrelies on object prior information.Aiming at the low accuracy of the original output mask of YOLACT,a mask post-processing method based on image processing and morphology is proposed to repair the dynamic point mask output by the YOLACT network.Finally,this module is integrated into the Oriented FAST and Rotated BRIEF SLAM 2(ORB-SLAM2)framework to construct a visual SLAM system adapted to dynamic scenes.The proposed algorithm is tested and verified on the public TUM dataset,which proves the effectiveness of the proposed module.Compared with the ORB-SLAM2 system,the localization accuracy of the proposed algorithm is improved by 93.4%in indoor dynamic scenes.展开更多
针对动态场景下视觉SLAM(Simultaneous Localization and Mapping)系统中深度学习分割网络实时性不足,以及相机非期望运动导致位姿估计偏差的问题,提出一种基于跨域掩膜分割的视觉SLAM算法.该算法采用轻量化YOLO-fastest网络结合背景减...针对动态场景下视觉SLAM(Simultaneous Localization and Mapping)系统中深度学习分割网络实时性不足,以及相机非期望运动导致位姿估计偏差的问题,提出一种基于跨域掩膜分割的视觉SLAM算法.该算法采用轻量化YOLO-fastest网络结合背景减除法实现运动物体检测,利用深度图结合深度阈值分割构建跨域掩膜分割机制,并设计相机运动几何校正策略补偿检测框坐标误差,在实现运动物体分割的同时提升处理速度.为优化特征点利用率,采用金字塔光流对动态特征点进行帧间连续跟踪与更新,同时确保仅由静态特征点参与位姿估计过程.在TUM数据集上进行系统性评估,实验结果表明,相比于ORB-SLAM3算法,该算法的绝对位姿误差平均降幅达97.1%,与使用深度学习分割网络的DynaSLAM和DS-SLAM的动态SLAM算法相比,其单帧跟踪时间大幅减少,在精度与效率之间实现了更好的平衡.展开更多
针对弱纹理和变光照环境下基于点特征的视觉SLAM(simultaneous localization and mapping)算法轨迹漂移的问题,提出了一种基于改进自适应阈值ELSED算法(Adaptive-ELSED)的快速点线融合双目视觉SLAM算法。通过在ELSED算法中添加自适应阈...针对弱纹理和变光照环境下基于点特征的视觉SLAM(simultaneous localization and mapping)算法轨迹漂移的问题,提出了一种基于改进自适应阈值ELSED算法(Adaptive-ELSED)的快速点线融合双目视觉SLAM算法。通过在ELSED算法中添加自适应阈值矩阵,动态调整不同光照条件下梯度阈值,并使用长度抑制和短线合并策略,提高线特征的质量。利用基于双目几何约束和图像结构相似性(SSIM)进行快速线段特征三角化。基于历史位姿及误差分析获取初始位姿,通过自适应因子实现光束法平差过程中点线特征的更有效融合。实验结果表明,所提算法在提高线特征质量的同时,耗时仅为LSD算法的50%,线特征匹配速度较传统LBD算法提升67%,挑战性场景下轨迹误差较ORB-SLAM3降低62.2%,系统的平均跟踪帧率为27帧/s,在保证系统实时性的同时,显著提升了系统在弱纹理、变光照环境下的精度和鲁棒性。展开更多
针对传统开源的激光惯性里程计(LIO,lidar-inertial odometry)和即时定位与地图构建(SLAM,simultaneous localization and mapping)结合的LIO-SLAM在室内复杂环境中受激光特征稀疏与动态遮挡影响、定位精度下降等问题,提出一种融合视觉...针对传统开源的激光惯性里程计(LIO,lidar-inertial odometry)和即时定位与地图构建(SLAM,simultaneous localization and mapping)结合的LIO-SLAM在室内复杂环境中受激光特征稀疏与动态遮挡影响、定位精度下降等问题,提出一种融合视觉里程计的改进方法。在保持LIO-SLAM激光惯性紧耦合框架的基础上,引入基于ORB特征的三维定位与地图构建算法(ORB-SLAM)作为独立的视觉里程计模块,为系统提供高频率、丰富纹理的视觉约束信息。通过自适应权重融合策略,实现激光、惯性与视觉观测的多源优化,增强了在弱几何约束、纹理丰富但结构复杂环境中的鲁棒性。在多种典型室内场景(走廊、开放大厅及动态人群环境)中开展了实验验证。结果表明,相较于原始LIO-SLAM,整体轨迹误差降低至原始系统的70%。研究验证了视觉-激光-惯性多模态融合在室内复杂环境下的可行性与有效性,为高精度室内自主定位与地图构建提供了新的思路。展开更多
近年来,神经辐射场在三维重建任务中展现出卓越性能。然而,应用在视觉同时定位与地图构建(simultaneous localization and mapping,SLAM)中因缺乏全局优化机制容易导致系统定位精度不足以及重建失败。针对该问题,本文提出一种融合深度...近年来,神经辐射场在三维重建任务中展现出卓越性能。然而,应用在视觉同时定位与地图构建(simultaneous localization and mapping,SLAM)中因缺乏全局优化机制容易导致系统定位精度不足以及重建失败。针对该问题,本文提出一种融合深度学习位姿估计与神经隐式表征的视觉SLAM系统。通过稠密束调整层以及高效的全局优化机制对相机位姿和深度进行像素级的循环迭代,并基于神经辐射场方法更新全局一致的隐式重建表面,使得系统在精准定位的同时能够重建高保真场景,并且在此基础上引入语言查询机制,增强系统的交互能力。在EuRoC和Replica数据集上进行大量实验,在不同的输入条件下,分别与3类基准方法进行对比,结果表明该系统在跟踪鲁棒性和重建精度方面相较于现有方法表现更优。本方法可为后续基于神经辐射场的视觉SLAM方法提供参考。展开更多
即时定位与地图构建(Simultaneous Localization and Mapping,SLAM)是一种重要的定位与地图构建技术。传统的视觉SLAM假设环境完全静止,但在动态场景中,SLAM识别的特征点很可能位于运动物体上,这会导致定位误差和地图不准确。为解决该问...即时定位与地图构建(Simultaneous Localization and Mapping,SLAM)是一种重要的定位与地图构建技术。传统的视觉SLAM假设环境完全静止,但在动态场景中,SLAM识别的特征点很可能位于运动物体上,这会导致定位误差和地图不准确。为解决该问题,提出了一种基于深度学习的视觉SLAM系统,利用嵌入MobileNetV3轻量化改进的YOLOv8模型识别运动物体,并将其范围内的特征点剔除,从而提高系统在动态场景下的鲁棒性、稳定性和追踪精度。实验结果表明,改进后的YOLOv8在计算速度上加快了5%以上,满足视觉SLAM系统的需求。最终,基于KITTI数据集00序列的实验结果显示,改进后的SLAM系统在动态场景下相较于ORB-SLAM2,轨迹精度提高了12.53%以上。展开更多
基金the National Natural Science Foundation of China(No.62063006)to the Guangxi Natural Science Foundation under Grant(Nos.2023GXNSFAA026025,AA24010001)+3 种基金to the Innovation Fund of Chinese Universities Industry-University-Research(ID:2023RY018)to the Special Guangxi Industry and Information Technology Department,Textile and Pharmaceutical Division(ID:2021 No.231)to the Special Research Project of Hechi University(ID:2021GCC028)to the Key Laboratory of AI and Information Processing,Education Department of Guangxi Zhuang Autonomous Region(Hechi University),No.2024GXZDSY009。
文摘In dynamic scenarios,visual simultaneous localization and mapping(SLAM)algorithms often incorrectly incorporate dynamic points during camera pose computation,leading to reduced accuracy and robustness.This paper presents a dynamic SLAM algorithm that leverages object detection and regional dynamic probability.Firstly,a parallel thread employs the YOLOX object detectionmodel to gather 2D semantic information and compensate for missed detections.Next,an improved K-means++clustering algorithm clusters bounding box regions,adaptively determining the threshold for extracting dynamic object contours as dynamic points change.This process divides the image into low dynamic,suspicious dynamic,and high dynamic regions.In the tracking thread,the dynamic point removal module assigns dynamic probability weights to the feature points in these regions.Combined with geometric methods,it detects and removes the dynamic points.The final evaluation on the public TUM RGB-D dataset shows that the proposed dynamic SLAM algorithm surpasses most existing SLAM algorithms,providing better pose estimation accuracy and robustness in dynamic environments.
文摘In dynamic scenes,the pose estimation and map consistency of visual simultaneous localisation and mapping(visual SLAM)are affected by intermittent changes in object motion states.An adaptive motion-state estimation and feature-reuse mechanism is proposed which restores features once objects become stationary.Camera ego-motion is com-pensated via projection-based point-to-point red-green-blue-depth(RGB-D)Iterative Closest Point;the alignment residual yields a short-term jitter score.An Extended Kalman Filter fuses the centre-pixel trajectory and depth of the object,using depth innovation as strong evidence to suppress false triggers.Applied adaptive decision thresholds involve resolution,ego-motion intensity,jitter,and reference depth,and are combined with dual/single triggering and hysteresis to achieve robust switching.When an object is considered static,its feature points are reused.On the Bonn RGB-D Dynamic Dataset(BONN)and TUM RGB-D SLAM Dataset and Benchmark(TUM),the proposed method matches or exceeds baselines:In intermittent-motion-dominated BONN sequences Placing_non_box,it re-duces the root-mean-square of the absolute trajectory error(ATE-RMSE)by 27%relative to the baseline,remains comparable to Ellipsoid-SLAM on TUM,and consistently outperforms ORB-SLAM3 in dynamic scenes.The hysteresis counter reading on Placing_non_box2 shows that the proposed method can reduce the motion-state misclassification rate by nearly 40%.From the ablation experiment results,we confirm that adaptive thresholds yield the most significant optimisation effect.The approach improves robustness and map completeness in dynamic environments without degrading performance in low-dynamic settings.
基金Supported by Jiangsu Key R&D Program(BE2021622)Jiangsu Postgraduate Practice and Innovation Program(SJCX23_0395).
文摘Background As visual simultaneous localization and mapping(SLAM)is primarily based on the assumption of a static scene,the presence of dynamic objects in the frame causes problems such as a deterioration of system robustness and inaccurate position estimation.In this study,we propose a YGC-SLAM for indoor dynamic environments based on the ORB-SLAM2 framework combined with semantic and geometric constraints to improve the positioning accuracy and robustness of the system.Methods First,the recognition accuracy of YOLOv5 was improved by introducing the convolution block attention model and the improved EIOU loss function,whereby the prediction frame converges quickly for better detection.The improved YOLOv5 was then added to the tracking thread for dynamic target detection to eliminate dynamic points.Subsequently,multi-view geometric constraints were used for re-judging to further eliminate dynamic points while enabling more useful feature points to be retained and preventing the semantic approach from over-eliminating feature points,causing a failure of map building.The K-means clustering algorithm was used to accelerate this process and quickly calculate and determine the motion state of each cluster of pixel points.Finally,a strategy for drawing keyframes with de-redundancy was implemented to construct a clear 3D dense static point-cloud map.Results Through testing on TUM dataset and a real environment,the experimental results show that our algorithm reduces the absolute trajectory error by 98.22%and the relative trajectory error by 97.98%compared with the original ORBSLAM2,which is more accurate and has better real-time performance than similar algorithms,such as DynaSLAM and DS-SLAM.Conclusions The YGC-SLAM proposed in this study can effectively eliminate the adverse effects of dynamic objects,and the system can better complete positioning and map building tasks in complex environments.
基金the National Natural Science Foundation of China(No.62063006)to the Guangxi Natural Science Foundation under Grant(Nos.2023GXNSFAA026025,AA24010001)+4 种基金to the Innovation Fund of Chinese Universities Industry-University-Research(ID:2023RY018)to the Special Guangxi Industry and Information Technology Department,Textile and Pharmaceutical Division(ID:2021 No.231)to the Special Research Project of Hechi University(ID:2021GCC028)to the Key Laboratory of AI and Information Processing,Education Department of Guangxi Zhuang Autonomous Region(Hechi University)No.2024GXZDSY009.
文摘Dynamic visual SLAM (Simultaneous Localization and Mapping) is an important research area, but existing methods struggle to balance real-time performance and accuracy in removing dynamic feature points, especially when semantic information is missing. This paper presents a novel dynamic SLAM system that uses optical flow tracking and epipolar geometry to identify dynamic feature points and applies a regional dynamic probability method to improve removal accuracy. We developed two innovative algorithms for precise pruning of dynamic regions: first, using optical flow and epipolar geometry to identify and prune dynamic areas while preserving static regions on stationary dynamic objects to optimize tracking performance;second, propagating dynamic probabilities across frames to mitigate the impact of semantic information loss in some frames. Experiments show that our system significantly reduces trajectory and pose errors in dynamic scenes, achieving dynamic feature point removal accuracy close to that of semantic segmentation methods, while maintaining high real-time performance. Our system performs exceptionally well in highly dynamic environments, especially where complex dynamic objects are present, demonstrating its advantage in handling dynamic scenarios. The experiments also show that while traditional methods may fail in tracking when semantic information is lost, our approach effectively reduces the misidentification of dynamic regions caused by such loss, thus improving system robustness and accuracy.
基金the National Natural Science Foundation of China(No.61671470).
文摘A great number of visual simultaneous localization and mapping(VSLAM)systems need to assume static features in the environment.However,moving objects can vastly impair the performance of a VSLAM system which relies on the static-world assumption.To cope with this challenging topic,a real-time and robust VSLAM system based on ORB-SLAM2 for dynamic environments was proposed.To reduce the influence of dynamic content,we incorporate the deep-learning-based object detection method in the visual odometry,then the dynamic object probability model is added to raise the efficiency of object detection deep neural network and enhance the real-time performance of our system.Experiment with both on the TUM and KITTI benchmark dataset,as well as in a real-world environment,the results clarify that our method can significantly reduce the tracking error or drift,enhance the robustness,accuracy and stability of the VSLAM system in dynamic scenes.
基金the National Natural Science Foundation of China(No.62063006)to the Natural Science Foundation of Guangxi Province(No.2023GXNS-FAA026025)+3 种基金to the Innovation Fund of Chinese Universities Industry-University-Research(ID:2021RYC06005)to the Research Project for Young and Middle-aged Teachers in Guangxi Universities(ID:2020KY15013)to the Special Research Project of Hechi University(ID:2021GCC028)supported by the Project of Outstanding Thousand Young Teachers’Training in Higher Education Institutions of Guangxi,Guangxi Colleges and Universities Key Laboratory of AI and Information Processing(Hechi University),Education Department of Guangxi Zhuang Autonomous Region.
文摘Visual simultaneous localization and mapping(SLAM)is crucial in robotics and autonomous driving.However,traditional visual SLAM faces challenges in dynamic environments.To address this issue,researchers have proposed semantic SLAM,which combines object detection,semantic segmentation,instance segmentation,and visual SLAM.Despite the growing body of literature on semantic SLAM,there is currently a lack of comprehensive research on the integration of object detection and visual SLAM.Therefore,this study aims to gather information from multiple databases and review relevant literature using specific keywords.It focuses on visual SLAM based on object detection,covering different aspects.Firstly,it discusses the current research status and challenges in this field,highlighting methods for incorporating semantic information from object detection networks into mileage measurement,closed-loop detection,and map construction.It also compares the characteristics and performance of various visual SLAM object detection algorithms.Lastly,it provides an outlook on future research directions and emerging trends in visual SLAM.Research has shown that visual SLAM based on object detection has significant improvements compared to traditional SLAM in dynamic point removal,data association,point cloud segmentation,and other technologies.It can improve the robustness and accuracy of the entire SLAM system and can run in real time.With the continuous optimization of algorithms and the improvement of hardware level,object visual SLAM has great potential for development.
基金Project Grant JZX7Y2-0190258055601National Natural Science Foundation of China(61402283).
文摘This article presents a brief survey to visual simultaneous localization and mapping (SLAM) systems applied to multiple independently moving agents, such as a team of ground or aerial vehicles, a group of users holding augmented or virtual reality devices. Such visual SLAM system, name as collaborative visual SLAM, is different from a typical visual SLAM deployed on a single agent in that information is exchanged or shared among different agents to achieve better robustness, efficiency, and accuracy. We review the representative works on this topic proposed in the past ten years and describe the key components involved in designing such a system including collaborative pose estimation and mapping tasks, as well as the emerging topic of decentralized architecture. We believe this brief survey could be helpful to someone who are working on this topic or developing multi-agent applications, particularly micro-aerial vehicle swarm or collaborative augmented/virtual reality.
文摘Feature selection is always an important issue in the visual SLAM (simultaneous location and mapping) literature. Considering that the location estimation can be improved by tracking features with larger value of visible time, a new feature selection method based on motion estimation is proposed. First, a k-step iteration algorithm is presented for visible time estimation using an affme motion model; then a delayed feature detection method is introduced for efficiently detecting features with the maximum visible time. As a means of validation for the proposed method, both simulation and real data experiments are carded out. Results show that the proposed method can improve both the estimation performance and the computational performance compared with the existing random feature selection method.
基金supported by National High Technology Research Development Program of China (863 Program) (No.2011AA040202)National Science Foundation of China (No.51005008)
文摘This paper presents a hierarchical simultaneous localization and mapping(SLAM) system for a small unmanned aerial vehicle(UAV) using the output of an inertial measurement unit(IMU) and the bearing-only observations from an onboard monocular camera.A homography based approach is used to calculate the motion of the vehicle in 6 degrees of freedom by image feature match.This visual measurement is fused with the inertial outputs by an indirect extended Kalman filter(EKF) for attitude and velocity estimation.Then,another EKF is employed to estimate the position of the vehicle and the locations of the features in the map.Both simulations and experiments are carried out to test the performance of the proposed system.The result of the comparison with the referential global positioning system/inertial navigation system(GPS/INS) navigation indicates that the proposed SLAM can provide reliable and stable state estimation for small UAVs in GPS-denied environments.
文摘A dynamic visual Simultaneous Localization and Mapping(SLAM)algorithm is proposed in this paper,which combines the YOLACT network with the geometric method to design a dynamic point detection module for eliminating dynamic points.The dense optical flow-based dynamic point detection scheme is adopted to make up for the problem that the elimination algorithm based on the instance segmentation network overrelies on object prior information.Aiming at the low accuracy of the original output mask of YOLACT,a mask post-processing method based on image processing and morphology is proposed to repair the dynamic point mask output by the YOLACT network.Finally,this module is integrated into the Oriented FAST and Rotated BRIEF SLAM 2(ORB-SLAM2)framework to construct a visual SLAM system adapted to dynamic scenes.The proposed algorithm is tested and verified on the public TUM dataset,which proves the effectiveness of the proposed module.Compared with the ORB-SLAM2 system,the localization accuracy of the proposed algorithm is improved by 93.4%in indoor dynamic scenes.
文摘针对动态场景下视觉SLAM(Simultaneous Localization and Mapping)系统中深度学习分割网络实时性不足,以及相机非期望运动导致位姿估计偏差的问题,提出一种基于跨域掩膜分割的视觉SLAM算法.该算法采用轻量化YOLO-fastest网络结合背景减除法实现运动物体检测,利用深度图结合深度阈值分割构建跨域掩膜分割机制,并设计相机运动几何校正策略补偿检测框坐标误差,在实现运动物体分割的同时提升处理速度.为优化特征点利用率,采用金字塔光流对动态特征点进行帧间连续跟踪与更新,同时确保仅由静态特征点参与位姿估计过程.在TUM数据集上进行系统性评估,实验结果表明,相比于ORB-SLAM3算法,该算法的绝对位姿误差平均降幅达97.1%,与使用深度学习分割网络的DynaSLAM和DS-SLAM的动态SLAM算法相比,其单帧跟踪时间大幅减少,在精度与效率之间实现了更好的平衡.
文摘针对弱纹理和变光照环境下基于点特征的视觉SLAM(simultaneous localization and mapping)算法轨迹漂移的问题,提出了一种基于改进自适应阈值ELSED算法(Adaptive-ELSED)的快速点线融合双目视觉SLAM算法。通过在ELSED算法中添加自适应阈值矩阵,动态调整不同光照条件下梯度阈值,并使用长度抑制和短线合并策略,提高线特征的质量。利用基于双目几何约束和图像结构相似性(SSIM)进行快速线段特征三角化。基于历史位姿及误差分析获取初始位姿,通过自适应因子实现光束法平差过程中点线特征的更有效融合。实验结果表明,所提算法在提高线特征质量的同时,耗时仅为LSD算法的50%,线特征匹配速度较传统LBD算法提升67%,挑战性场景下轨迹误差较ORB-SLAM3降低62.2%,系统的平均跟踪帧率为27帧/s,在保证系统实时性的同时,显著提升了系统在弱纹理、变光照环境下的精度和鲁棒性。
文摘针对传统开源的激光惯性里程计(LIO,lidar-inertial odometry)和即时定位与地图构建(SLAM,simultaneous localization and mapping)结合的LIO-SLAM在室内复杂环境中受激光特征稀疏与动态遮挡影响、定位精度下降等问题,提出一种融合视觉里程计的改进方法。在保持LIO-SLAM激光惯性紧耦合框架的基础上,引入基于ORB特征的三维定位与地图构建算法(ORB-SLAM)作为独立的视觉里程计模块,为系统提供高频率、丰富纹理的视觉约束信息。通过自适应权重融合策略,实现激光、惯性与视觉观测的多源优化,增强了在弱几何约束、纹理丰富但结构复杂环境中的鲁棒性。在多种典型室内场景(走廊、开放大厅及动态人群环境)中开展了实验验证。结果表明,相较于原始LIO-SLAM,整体轨迹误差降低至原始系统的70%。研究验证了视觉-激光-惯性多模态融合在室内复杂环境下的可行性与有效性,为高精度室内自主定位与地图构建提供了新的思路。
文摘近年来,神经辐射场在三维重建任务中展现出卓越性能。然而,应用在视觉同时定位与地图构建(simultaneous localization and mapping,SLAM)中因缺乏全局优化机制容易导致系统定位精度不足以及重建失败。针对该问题,本文提出一种融合深度学习位姿估计与神经隐式表征的视觉SLAM系统。通过稠密束调整层以及高效的全局优化机制对相机位姿和深度进行像素级的循环迭代,并基于神经辐射场方法更新全局一致的隐式重建表面,使得系统在精准定位的同时能够重建高保真场景,并且在此基础上引入语言查询机制,增强系统的交互能力。在EuRoC和Replica数据集上进行大量实验,在不同的输入条件下,分别与3类基准方法进行对比,结果表明该系统在跟踪鲁棒性和重建精度方面相较于现有方法表现更优。本方法可为后续基于神经辐射场的视觉SLAM方法提供参考。
文摘即时定位与地图构建(Simultaneous Localization and Mapping,SLAM)是一种重要的定位与地图构建技术。传统的视觉SLAM假设环境完全静止,但在动态场景中,SLAM识别的特征点很可能位于运动物体上,这会导致定位误差和地图不准确。为解决该问题,提出了一种基于深度学习的视觉SLAM系统,利用嵌入MobileNetV3轻量化改进的YOLOv8模型识别运动物体,并将其范围内的特征点剔除,从而提高系统在动态场景下的鲁棒性、稳定性和追踪精度。实验结果表明,改进后的YOLOv8在计算速度上加快了5%以上,满足视觉SLAM系统的需求。最终,基于KITTI数据集00序列的实验结果显示,改进后的SLAM系统在动态场景下相较于ORB-SLAM2,轨迹精度提高了12.53%以上。