期刊文献+
共找到15篇文章
< 1 >
每页显示 20 50 100
Unsupervised Monocular Depth Estimation with Edge Enhancement for Dynamic Scenes
1
作者 Peicheng Shi Yueyue Tang +3 位作者 Yi Li Xinlong Dong Yu Sun Aixi Yang 《Computers, Materials & Continua》 2025年第8期3321-3343,共23页
In the dynamic scene of autonomous vehicles,the depth estimation of monocular cameras often faces the problem of inaccurate edge depth estimation.To solve this problem,we propose an unsupervised monocular depth estima... In the dynamic scene of autonomous vehicles,the depth estimation of monocular cameras often faces the problem of inaccurate edge depth estimation.To solve this problem,we propose an unsupervised monocular depth estimation model based on edge enhancement,which is specifically aimed at the depth perception challenge in dynamic scenes.The model consists of two core networks:a deep prediction network and a motion estimation network,both of which adopt an encoder-decoder architecture.The depth prediction network is based on the U-Net structure of ResNet18,which is responsible for generating the depth map of the scene.The motion estimation network is based on the U-Net structure of Flow-Net,focusing on the motion estimation of dynamic targets.In the decoding stage of the motion estimation network,we innovatively introduce an edge-enhanced decoder,which integrates a convolutional block attention module(CBAM)in the decoding process to enhance the recognition ability of the edge features of moving objects.In addition,we also designed a strip convolution module to improve the model’s capture efficiency of discrete moving targets.To further improve the performance of the model,we propose a novel edge regularization method based on the Laplace operator,which effectively accelerates the convergence process of themodel.Experimental results on the KITTI and Cityscapes datasets show that compared with the current advanced dynamic unsupervised monocular model,the proposed model has a significant improvement in depth estimation accuracy and convergence speed.Specifically,the rootmean square error(RMSE)is reduced by 4.8%compared with the DepthMotion algorithm,while the training convergence speed is increased by 36%,which shows the superior performance of the model in the depth estimation task in dynamic scenes. 展开更多
关键词 dynamic scenes unsupervised learning monocular depth edge enhancement
在线阅读 下载PDF
Fast speedometer identification in dynamic scene based on phase correlation 被引量:1
2
作者 王昱棠 付梦印 杨毅 《Journal of Beijing Institute of Technology》 EI CAS 2012年第3期394-399,共6页
Speedometer identification has been researched for many years.The common approaches to that problem are usually based on image subtraction,which does not adapt to image offsets caused by camera vibration.To cope with ... Speedometer identification has been researched for many years.The common approaches to that problem are usually based on image subtraction,which does not adapt to image offsets caused by camera vibration.To cope with the rapidity,robust and accurate requirements of this kind of work in dynamic scene,a fast speedometer identification algorithm is proposed,it utilizes phase correlation method based on regional entire template translation to estimate the offset between images.In order to effectively reduce unnecessary computation and false detection rate,an improved linear Hough transform method with two optimization strategies is presented for pointer line detection.Based on VC++ 6.0 software platform with OpenCV library,the algorithm performance under experiments has shown that it celerity and precision. 展开更多
关键词 speedometer dynamic scene image sequence phase correlation improved linear Hough transform
在线阅读 下载PDF
Introduction to Visual Surveillance of Dynamic Scenes
3
作者 Steve Maybank 《自动化学报》 EI CSCD 北大核心 2003年第3期319-320,共2页
关键词 of Introduction to Visual Surveillance of dynamic scenes
全文增补中
DKP-SLAM:A Visual SLAM for Dynamic Indoor Scenes Based on Object Detection and Region Probability
4
作者 Menglin Yin Yong Qin Jiansheng Peng 《Computers, Materials & Continua》 SCIE EI 2025年第1期1329-1347,共19页
In dynamic scenarios,visual simultaneous localization and mapping(SLAM)algorithms often incorrectly incorporate dynamic points during camera pose computation,leading to reduced accuracy and robustness.This paper prese... In dynamic scenarios,visual simultaneous localization and mapping(SLAM)algorithms often incorrectly incorporate dynamic points during camera pose computation,leading to reduced accuracy and robustness.This paper presents a dynamic SLAM algorithm that leverages object detection and regional dynamic probability.Firstly,a parallel thread employs the YOLOX object detectionmodel to gather 2D semantic information and compensate for missed detections.Next,an improved K-means++clustering algorithm clusters bounding box regions,adaptively determining the threshold for extracting dynamic object contours as dynamic points change.This process divides the image into low dynamic,suspicious dynamic,and high dynamic regions.In the tracking thread,the dynamic point removal module assigns dynamic probability weights to the feature points in these regions.Combined with geometric methods,it detects and removes the dynamic points.The final evaluation on the public TUM RGB-D dataset shows that the proposed dynamic SLAM algorithm surpasses most existing SLAM algorithms,providing better pose estimation accuracy and robustness in dynamic environments. 展开更多
关键词 Visual SLAM dynamic scene YOLOX K-means++clustering dynamic probability
在线阅读 下载PDF
Self-Supervised Monocular Depth Estimation with Scene Dynamic Pose
5
作者 Jing He Haonan Zhu +1 位作者 Chenhao Zhao Minrui Zhao 《Computers, Materials & Continua》 2025年第6期4551-4573,共23页
Self-supervised monocular depth estimation has emerged as a major research focus in recent years,primarily due to the elimination of ground-truth depth dependence.However,the prevailing architectures in this domain su... Self-supervised monocular depth estimation has emerged as a major research focus in recent years,primarily due to the elimination of ground-truth depth dependence.However,the prevailing architectures in this domain suffer from inherent limitations:existing pose network branches infer camera ego-motion exclusively under static-scene and Lambertian-surface assumptions.These assumptions are often violated in real-world scenarios due to dynamic objects,non-Lambertian reflectance,and unstructured background elements,leading to pervasive artifacts such as depth discontinuities(“holes”),structural collapse,and ambiguous reconstruction.To address these challenges,we propose a novel framework that integrates scene dynamic pose estimation into the conventional self-supervised depth network,enhancing its ability to model complex scene dynamics.Our contributions are threefold:(1)a pixel-wise dynamic pose estimation module that jointly resolves the pose transformations of moving objects and localized scene perturbations;(2)a physically-informed loss function that couples dynamic pose and depth predictions,designed to mitigate depth errors arising from high-speed distant objects and geometrically inconsistent motion profiles;(3)an efficient SE(3)transformation parameterization that streamlines network complexity and temporal pre-processing.Extensive experiments on the KITTI and NYU-V2 benchmarks show that our framework achieves state-of-the-art performance in both quantitative metrics and qualitative visual fidelity,significantly improving the robustness and generalization of monocular depth estimation under dynamic conditions. 展开更多
关键词 Monocular depth estimation self-supervised learning scene dynamic pose estimation dynamic-depth constraint pixel-wise dynamic pose
在线阅读 下载PDF
MG-SLAM: RGB-D SLAM Based on Semantic Segmentation for Dynamic Environment in the Internet of Vehicles 被引量:1
6
作者 Fengju Zhang Kai Zhu 《Computers, Materials & Continua》 2025年第2期2353-2372,共20页
The Internet of Vehicles (IoV) has become an important direction in the field of intelligent transportation, in which vehicle positioning is a crucial part. SLAM (Simultaneous Localization and Mapping) technology play... The Internet of Vehicles (IoV) has become an important direction in the field of intelligent transportation, in which vehicle positioning is a crucial part. SLAM (Simultaneous Localization and Mapping) technology plays a crucial role in vehicle localization and navigation. Traditional Simultaneous Localization and Mapping (SLAM) systems are designed for use in static environments, and they can result in poor performance in terms of accuracy and robustness when used in dynamic environments where objects are in constant movement. To address this issue, a new real-time visual SLAM system called MG-SLAM has been developed. Based on ORB-SLAM2, MG-SLAM incorporates a dynamic target detection process that enables the detection of both known and unknown moving objects. In this process, a separate semantic segmentation thread is required to segment dynamic target instances, and the Mask R-CNN algorithm is applied on the Graphics Processing Unit (GPU) to accelerate segmentation. To reduce computational cost, only key frames are segmented to identify known dynamic objects. Additionally, a multi-view geometry method is adopted to detect unknown moving objects. The results demonstrate that MG-SLAM achieves higher precision, with an improvement from 0.2730 m to 0.0135 m in precision. Moreover, the processing time required by MG-SLAM is significantly reduced compared to other dynamic scene SLAM algorithms, which illustrates its efficacy in locating objects in dynamic scenes. 展开更多
关键词 Visual SLAM dynamic scene semantic segmentation GPU acceleration key segmentation frame
在线阅读 下载PDF
Key Techniques and Challenges in NeRF-Based Dynamic 3D Reconstruction
7
作者 LU Ping FENG Daquan +2 位作者 SHI Wenzhe LI Wan LIN Jiaxin 《ZTE Communications》 2025年第3期71-80,共10页
This paper explores the key techniques and challenges in dynamic scene reconstruction with neural radiance fields(NeRF).As an emerging computer vision method,the NeRF has wide application potential,especially in excel... This paper explores the key techniques and challenges in dynamic scene reconstruction with neural radiance fields(NeRF).As an emerging computer vision method,the NeRF has wide application potential,especially in excelling at 3D reconstruction.We first introduce the basic principles and working mechanisms of NeRFs,followed by an in-depth discussion of the technical challenges faced by 3D reconstruction in dynamic scenes,including problems in perspective and illumination changes of moving objects,recognition and modeling of dynamic objects,real-time requirements,data acquisition and calibration,motion estimation,and evaluation mechanisms.We also summarize current state-of-theart approaches to address these challenges,as well as future research trends.The goal is to provide researchers with an in-depth understanding of the application of NeRFs in dynamic scene reconstruction,as well as insights into the key issues faced and future directions. 展开更多
关键词 neural radiance fields 3D computer vision dynamic scene reconstruction
在线阅读 下载PDF
Progressive edge-sensing dynamic scene deblurring
8
作者 Tianlin Zhang Jinjiang Li Hui Fan 《Computational Visual Media》 SCIE EI CSCD 2022年第3期495-508,共14页
Deblurring images of dynamic scenes is a challenging task because blurring occurs due to a combination of many factors.In recent years,the use of multi-scale pyramid methods to recover high-resolution sharp images has... Deblurring images of dynamic scenes is a challenging task because blurring occurs due to a combination of many factors.In recent years,the use of multi-scale pyramid methods to recover high-resolution sharp images has been extensively studied.We have made improvements to the lack of detail recovery in the cascade structure through a network using progressive integration of data streams.Our new multi-scale structure and edge feature perception design deals with changes in blurring at different spatial scales and enhances the sensitivity of the network to blurred edges.The coarse-to-fine architecture restores the image structure,first performing global adjustments,and then performing local refinement.In this way,not only is global correlation considered,but also residual information is used to significantly improve image restoration and enhance texture details.Experimental results show quantitative and qualitative improvements over existing methods. 展开更多
关键词 image deblurring dynamic scenes multiscale edge features
原文传递
A temporal-spatial background modeling of dynamic scenes
9
作者 Jiuyue HAO Chao LI +1 位作者 Zhang XIONG Ejaz HUSSAIN 《Frontiers of Materials Science》 SCIE CSCD 2011年第3期290-299,共10页
Moving object detection in dynamic scenes is a basic task in a surveillance system for sensor data collection. In this paper, we present a powerful back- ground subtraction algorithm called Gaussian-kernel density est... Moving object detection in dynamic scenes is a basic task in a surveillance system for sensor data collection. In this paper, we present a powerful back- ground subtraction algorithm called Gaussian-kernel density estimator (G-KDE) that improves the accuracy and reduces the computational load. The main innovation is that we divide the changes of background into continuous and stable changes to deal with dynamic scenes and moving objects that first merge into the background, and separately model background using both KDE model and Gaussian models. To get a temporal- spatial background model, the sample selection is based on the concept of region average at the update stage. In the detection stage, neighborhood information content (NIC) is implemented which suppresses the false detection due to small and un-modeled movements in the scene. The experimental results which are generated on three separate sequences indicate that this method is well suited for precise detection of moving objects in complex scenes and it can be efficiently used in various detection systems. 展开更多
关键词 temporal-spatial background model Gaus-sian-kemel density estimator (G-KDE) dynamic scenes neighborhood information content (NIC) moving objectdetection
原文传递
HDR-Net-Fusion:Real-time 3D dynamic scene reconstruction with a hierarchical deep reinforcement network 被引量:1
10
作者 Hao-Xuan Song Jiahui Huang +1 位作者 Yan-Pei Cao Tai-Jiang Mu 《Computational Visual Media》 EI CSCD 2021年第4期419-435,共17页
Reconstructing dynamic scenes with commodity depth cameras has many applications in computer graphics,computer vision,and robotics.However,due to the presence of noise and erroneous observations from data capturing de... Reconstructing dynamic scenes with commodity depth cameras has many applications in computer graphics,computer vision,and robotics.However,due to the presence of noise and erroneous observations from data capturing devices and the inherently ill-posed nature of non-rigid registration with insufficient information,traditional approaches often produce low-quality geometry with holes,bumps,and misalignments.We propose a novel 3D dynamic reconstruction system,named HDR-Net-Fusion,which learns to simultaneously reconstruct and refine the geometry on the fly with a sparse embedded deformation graph of surfels,using a hierarchical deep reinforcement(HDR)network.The latter comprises two parts:a global HDR-Net which rapidly detects local regions with large geometric errors,and a local HDR-Net serving as a local patch refinement operator to promptly complete and enhance such regions.Training the global HDR-Net is formulated as a novel reinforcement learning problem to implicitly learn the region selection strategy with the goal of improving the overall reconstruction quality.The applicability and efficiency of our approach are demonstrated using a large-scale dynamic reconstruction dataset.Our method can reconstruct geometry with higher quality than traditional methods. 展开更多
关键词 dynamic 3D scene reconstruction deep reinforcement learning point cloud completion deep neural networks
原文传递
Multi-exposure fusion for high dynamic range scene
11
作者 Shen Xiaohe Liu Jinghong 《High Technology Letters》 EI CAS 2017年第4期343-349,共7页
Due to the existing limited dynamic range a camera cannot reveal all the details in a high-dynamic range scene. In order to solve this problem,this paper presents a multi-exposure fusion method for getting high qualit... Due to the existing limited dynamic range a camera cannot reveal all the details in a high-dynamic range scene. In order to solve this problem,this paper presents a multi-exposure fusion method for getting high quality images in high dynamic range scene. First,a set of multi-exposure images is obtained by multiple exposures in a same scene and their brightness condition is analyzed. Then,multi-exposure images under the same scene are decomposed using dual-tree complex wavelet transform( DT-CWT),and their low and high frequency components are obtained. Weight maps according to the brightness condition are assigned to the low components for fusion. Maximizing the region Sum Modified-Laplacian( SML) is adopted for high-frequency components fusing. Finally,the fused image is acquired by subjecting the low and high frequency coefficients to inverse DT-CWT.Experimental results show that the proposed approach generates high quality results with uniform distributed brightness and rich details. The proposed method is efficient and robust in varies scenes. 展开更多
关键词 multi-exposure fusion high dynamic range scene dual-tree complex wavelet transform(DT-CWT) brightness analysis
在线阅读 下载PDF
Hybrid Dynamic Point Removal and Ellipsoid Modelling of Object-Based Semantic SLAM
12
作者 Qingyang Xu Siwei Huang +2 位作者 Yong Song Bao Pang Chengjin Zhang 《IET Cyber-Systems and Robotics》 2025年第3期25-47,共23页
For the issue of low positioning accuracy in dynamic environments with traditional simultaneous localisation and mapping(SLAM),a dynamic point removal strategy combining object detection and optical flow tracking has ... For the issue of low positioning accuracy in dynamic environments with traditional simultaneous localisation and mapping(SLAM),a dynamic point removal strategy combining object detection and optical flow tracking has been proposed.To fully utilise the semantic information,an ellipsoid model of the detected semantic objects was first constructed based on the plane and point cloud constraints,which assists in loop closure detection.Bilateral semantic map matching was achieved through the Kuhn-Munkres(KM)algorithm maximum weight assignment,and the pose transformation between local and global maps was determined by the random sample consensus(RANSAC)algorithm.Finally,a stable semantic SLAM system suitable for dy-namic environments was constructed.The effectiveness of achieving the system's positioning accuracy under dynamic inter-ference and large visual-inertial loop closure was verified by the experiment. 展开更多
关键词 dynamic scene ellipsoid modeling hybrid dynamic feature point object detection semantic object simultaneous localization and mopping(SLAM)
原文传递
A robust RGB-D visual odometry with moving object detection in dynamic indoor scenes
13
作者 Xianglong Zhang Haiyang Yu Yan Zhuang 《IET Cyber-Systems and Robotics》 EI 2023年第1期79-88,共10页
Simultaneous localisation and mapping(SLAM)are the basis for many robotic applications.As the front end of SLAM,visual odometry is mainly used to estimate camera pose.In dynamic scenes,classical methods are deteriorat... Simultaneous localisation and mapping(SLAM)are the basis for many robotic applications.As the front end of SLAM,visual odometry is mainly used to estimate camera pose.In dynamic scenes,classical methods are deteriorated by dynamic objects and cannot achieve satisfactory results.In order to improve the robustness of visual odometry in dynamic scenes,this paper proposed a dynamic region detection method based on RGBD images.Firstly,all feature points on the RGB image are classified as dynamic and static using a triangle constraint and the epipolar geometric constraint successively.Meanwhile,the depth image is clustered using the K-Means method.The classified feature points are mapped to the clustered depth image,and a dynamic or static label is assigned to each cluster according to the number of dynamic feature points.Subsequently,a dynamic region mask for the RGB image is generated based on the dynamic clusters in the depth image,and the feature points covered by the mask are all removed.The remaining static feature points are applied to estimate the camera pose.Finally,some experimental results are provided to demonstrate the feasibility and performance. 展开更多
关键词 dynamic indoor scenes moving object detection RGB-D SLAM visual odometry
原文传递
Dynamic background modeling using tensor representation and ant colony optimization 被引量:1
14
作者 PENG LiZhong ZHANG Fan ZHOU BingYin 《Science China Mathematics》 SCIE CSCD 2017年第11期2287-2302,共16页
Background modeling and subtraction is a fundamental problem in video analysis. Many algorithms have been developed to date, but there are still some challenges in complex environments, especially dynamic scenes in wh... Background modeling and subtraction is a fundamental problem in video analysis. Many algorithms have been developed to date, but there are still some challenges in complex environments, especially dynamic scenes in which backgrounds are themselves moving, such as rippling water and swaying trees. In this paper, a novel background modeling method is proposed for dynamic scenes by combining both tensor representation and swarm intelligence. We maintain several video patches, which are naturally represented as higher order tensors,to represent the patterns of background, and utilize tensor low-rank approximation to capture the dynamic nature. Furthermore, we introduce an ant colony algorithm to improve the performance. Experimental results show that the proposed method is robust and adaptive in dynamic environments, and moving objects can be perfectly separated from the complex dynamic background. 展开更多
关键词 background modeling dynamic scenes tensor representation ant colony optimization
原文传递
DFPC-SLAM:A Dynamic Feature Point Constraints-Based SLAM Using Stereo Vision for Dynamic Environment
15
作者 Bo Zeng Chengqun Song +1 位作者 Cheng Jun Yuhang Kang 《Guidance, Navigation and Control》 2023年第1期46-60,共15页
Visual SLAM methods usually presuppose that the scene is static, so the SLAM algorithm formobile robots in dynamic scenes often results in a signicant decrease in accuracy due to thein°uence of dynamic objects. I... Visual SLAM methods usually presuppose that the scene is static, so the SLAM algorithm formobile robots in dynamic scenes often results in a signicant decrease in accuracy due to thein°uence of dynamic objects. In this paper, feature points are divided into dynamic and staticfrom semantic information and multi-view geometry information, and then static region featurepoints are added to the pose-optimization, and static scene maps are established for dynamicscenes. Finally, experiments are conducted in dynamic scenes using the KITTI dataset, and theresults show that the proposed algorithm has higher accuracy in highly dynamic scenes comparedto the visual SLAM baseline. 展开更多
关键词 SLAM stereo vision semantic segmentation multi-view geometry dynamic scenes.
在线阅读 下载PDF
上一页 1 下一页 到第
使用帮助 返回顶部