期刊文献+
共找到190篇文章
< 1 2 10 >
每页显示 20 50 100
Self-Supervised Monocular Depth Estimation with Scene Dynamic Pose
1
作者 Jing He Haonan Zhu +1 位作者 Chenhao Zhao Minrui Zhao 《Computers, Materials & Continua》 2025年第6期4551-4573,共23页
Self-supervised monocular depth estimation has emerged as a major research focus in recent years,primarily due to the elimination of ground-truth depth dependence.However,the prevailing architectures in this domain su... Self-supervised monocular depth estimation has emerged as a major research focus in recent years,primarily due to the elimination of ground-truth depth dependence.However,the prevailing architectures in this domain suffer from inherent limitations:existing pose network branches infer camera ego-motion exclusively under static-scene and Lambertian-surface assumptions.These assumptions are often violated in real-world scenarios due to dynamic objects,non-Lambertian reflectance,and unstructured background elements,leading to pervasive artifacts such as depth discontinuities(“holes”),structural collapse,and ambiguous reconstruction.To address these challenges,we propose a novel framework that integrates scene dynamic pose estimation into the conventional self-supervised depth network,enhancing its ability to model complex scene dynamics.Our contributions are threefold:(1)a pixel-wise dynamic pose estimation module that jointly resolves the pose transformations of moving objects and localized scene perturbations;(2)a physically-informed loss function that couples dynamic pose and depth predictions,designed to mitigate depth errors arising from high-speed distant objects and geometrically inconsistent motion profiles;(3)an efficient SE(3)transformation parameterization that streamlines network complexity and temporal pre-processing.Extensive experiments on the KITTI and NYU-V2 benchmarks show that our framework achieves state-of-the-art performance in both quantitative metrics and qualitative visual fidelity,significantly improving the robustness and generalization of monocular depth estimation under dynamic conditions. 展开更多
关键词 monocular depth estimation self-supervised learning scene dynamic pose estimation dynamic-depth constraint pixel-wise dynamic pose
在线阅读 下载PDF
Enhancing Underwater Monocular Depth Estimation with Lpg-Lap Unet for Target Tracking Mission
2
作者 YAO Peng WANG Yalu 《Journal of Ocean University of China》 2026年第1期161-170,共10页
Accurately estimating depth from underwater monocular images is essential for the target tracking task of unmanned underwater vehicles.This work proposes a method based on the Lpg-Lap Unet architecture.First,the Unet ... Accurately estimating depth from underwater monocular images is essential for the target tracking task of unmanned underwater vehicles.This work proposes a method based on the Lpg-Lap Unet architecture.First,the Unet architecture integrates Laplacian pyramid depth residuals and Sobel operators to improve the boundary details in depth images,which may suffer from the feature loss caused by upsampling and the blurriness of underwater images.Multiscale local planar guidance layers then fully exploit the intermediate depth features,and a comprehensive loss function ensures robustness and accuracy.Experimental results on benchmarks demonstrate the effectiveness of Lpg-Lap Unet and its superior performance over state-of-the-art models.An underwater target tracking system is then designed to further validate its real-time capabilities in the AirSim simulation platform. 展开更多
关键词 underwater monocular depth estimation Laplacian pyramid multiscale local planar guidance underwater target tracking
在线阅读 下载PDF
Self-Supervised Monocular Depth Estimation by Digging into Uncertainty Quantification
3
作者 李远珍 郑圣杰 +3 位作者 谭梓欣 曹拓 罗飞 肖春霞 《Journal of Computer Science & Technology》 SCIE EI CSCD 2023年第3期510-525,共16页
Based on well-designed network architectures and objective functions,self-supervised monocular depth estimation has made great progress.However,lacking a specific mechanism to make the network learn more about the reg... Based on well-designed network architectures and objective functions,self-supervised monocular depth estimation has made great progress.However,lacking a specific mechanism to make the network learn more about the regions containing moving objects or occlusion scenarios,existing depth estimation methods likely produce poor results for them.Therefore,we propose an uncertainty quantification method to improve the performance of existing depth estimation networks without changing their architectures.Our uncertainty quantification method consists of uncertainty measurement,the learning guidance by uncertainty,and the ultimate adaptive determination.Firstly,with Snapshot and Siam learning strategies,we measure the uncertainty degree by calculating the variance of pre-converged epochs or twins during training.Secondly,we use the uncertainty to guide the network to strengthen learning about those regions with more uncertainty.Finally,we use the uncertainty to adaptively produce the final depth estimation results with a balance of accuracy and robustness.To demonstrate the effectiveness of our uncertainty quantification method,we apply it to two state-of-the-art models,Monodepth2 and Hints.Experimental results show that our method has improved the depth estimation performance in seven evaluation metrics compared with two baseline models and exceeded the existing uncertainty method. 展开更多
关键词 self-supervised monocular depth estimation uncertainty quantification variance
原文传递
On Robust Cross-view Consistency in Self-supervised Monocular Depth Estimation
4
作者 Haimei Zhao Jing Zhang +2 位作者 Zhuo Chen Bo Yuan Dacheng Tao 《Machine Intelligence Research》 EI CSCD 2024年第3期495-513,共19页
Remarkable progress has been made in self-supervised monocular depth estimation (SS-MDE) by exploring cross-view consistency, e.g., photometric consistency and 3D point cloud consistency. However, they are very vulner... Remarkable progress has been made in self-supervised monocular depth estimation (SS-MDE) by exploring cross-view consistency, e.g., photometric consistency and 3D point cloud consistency. However, they are very vulnerable to illumination variance, occlusions, texture-less regions, as well as moving objects, making them not robust enough to deal with various scenes. To address this challenge, we study two kinds of robust cross-view consistency in this paper. Firstly, the spatial offset field between adjacent frames is obtained by reconstructing the reference frame from its neighbors via deformable alignment, which is used to align the temporal depth features via a depth feature alignment (DFA) loss. Secondly, the 3D point clouds of each reference frame and its nearby frames are calculated and transformed into voxel space, where the point density in each voxel is calculated and aligned via a voxel density alignment (VDA) loss. In this way, we exploit the temporal coherence in both depth feature space and 3D voxel space for SS-MDE, shifting the “point-to-point” alignment paradigm to the “region-to-region” one. Compared with the photometric consistency loss as well as the rigid point cloud alignment loss, the proposed DFA and VDA losses are more robust owing to the strong representation power of deep features as well as the high tolerance of voxel density to the aforementioned challenges. Experimental results on several outdoor benchmarks show that our method outperforms current state-of-the-art techniques. Extensive ablation study and analysis validate the effectiveness of the proposed losses, especially in challenging scenes. The code and models are available at https://github.com/sunnyHelen/RCVC-depth. 展开更多
关键词 3D vision depth estimation cross-view consistency self-supervised learning monocular perception
原文传递
High Quality Monocular Video Depth Estimation Based on Mask Guided Refinement
5
作者 Huixiao Pan Qiang Zhao 《Journal of Beijing Institute of Technology》 2025年第1期18-27,共10页
Depth maps play a crucial role in various practical applications such as computer vision,augmented reality,and autonomous driving.How to obtain clear and accurate depth information in video depth estimation is a signi... Depth maps play a crucial role in various practical applications such as computer vision,augmented reality,and autonomous driving.How to obtain clear and accurate depth information in video depth estimation is a significant challenge faced in the field of computer vision.However,existing monocular video depth estimation models tend to produce blurred or inaccurate depth information in regions with object edges and low texture.To address this issue,we propose a monocular depth estimation model architecture guided by semantic segmentation masks,which introduces semantic information into the model to correct the ambiguous depth regions.We have evaluated the proposed method,and experimental results show that our method improves the accuracy of edge depth,demonstrating the effectiveness of our approach. 展开更多
关键词 monocular video depth estimation depth refinement edge depth accuracy semantic segmentation
在线阅读 下载PDF
Bridging 2D and 3D Object Detection:Advances in Occlusion Handling through Depth Estimation
6
作者 Zainab Ouardirhi Mostapha Zbakh Sidi Ahmed Mahmoudi 《Computer Modeling in Engineering & Sciences》 2025年第6期2509-2571,共63页
Object detection in occluded environments remains a core challenge in computer vision(CV),especially in domains such as autonomous driving and robotics.While Convolutional Neural Network(CNN)-based twodimensional(2D)a... Object detection in occluded environments remains a core challenge in computer vision(CV),especially in domains such as autonomous driving and robotics.While Convolutional Neural Network(CNN)-based twodimensional(2D)and three-dimensional(3D)object detection methods havemade significant progress,they often fall short under severe occlusion due to depth ambiguities in 2D imagery and the high cost and deployment limitations of 3D sensors such as Light Detection and Ranging(LiDAR).This paper presents a comparative review of recent 2D and 3D detection models,focusing on their occlusion-handling capabilities and the impact of sensor modalities such as stereo vision,Time-of-Flight(ToF)cameras,and LiDAR.In this context,we introduce FuDensityNet,our multimodal occlusion-aware detection framework that combines Red-Green-Blue(RGB)images and LiDAR data to enhance detection performance.As a forward-looking direction,we propose a monocular depth-estimation extension to FuDensityNet,aimed at replacing expensive 3D sensors with a more scalable CNN-based pipeline.Although this enhancement is not experimentally evaluated in this manuscript,we describe its conceptual design and potential for future implementation. 展开更多
关键词 Object detection occlusion handling multimodal fusion monocular 3D sensors depth estimation
在线阅读 下载PDF
ELDE-Net:Efficient Light-Weight Depth Estimation Network for Deep Reinforcement Learning-Based Mobile Robot Path Planning
7
作者 Thai-Viet Dang Dinh-Manh-Cuong Tran +1 位作者 Nhu-Nghia Bui Phan Xuan Tan 《Computers, Materials & Continua》 2025年第11期2651-2680,共30页
Precise and robust three-dimensional object detection(3DOD)presents a promising opportunity in the field of mobile robot(MR)navigation.Monocular 3DOD techniques typically involve extending existing twodimensional obje... Precise and robust three-dimensional object detection(3DOD)presents a promising opportunity in the field of mobile robot(MR)navigation.Monocular 3DOD techniques typically involve extending existing twodimensional object detection(2DOD)frameworks to predict the three-dimensional bounding box(3DBB)of objects captured in 2D RGB images.However,these methods often require multiple images,making them less feasible for various real-time scenarios.To address these challenges,the emergence of agile convolutional neural networks(CNNs)capable of inferring depth froma single image opens a new avenue for investigation.The paper proposes a novel ELDENet network designed to produce cost-effective 3DBounding Box Estimation(3D-BBE)froma single image.This novel framework comprises the PP-LCNet as the encoder and a fast convolutional decoder.Additionally,this integration includes a Squeeze-Exploit(SE)module utilizing the Math Kernel Library for Deep Neural Networks(MKLDNN)optimizer to enhance convolutional efficiency and streamline model size during effective training.Meanwhile,the proposed multi-scale sub-pixel decoder generates high-quality depth maps while maintaining a compact structure.Furthermore,the generated depthmaps provide a clear perspective with distance details of objects in the environment.These depth insights are combined with 2DOD for precise evaluation of 3D Bounding Boxes(3DBB),facilitating scene understanding and optimal route planning for mobile robots.Based on the estimated object center of the 3DBB,the Deep Reinforcement Learning(DRL)-based obstacle avoidance strategy for MRs is developed.Experimental results demonstrate that our model achieves state-of-the-art performance across three datasets:NYU-V2,KITTI,and Cityscapes.Overall,this framework shows significant potential for adaptation in intelligent mechatronic systems,particularly in developing knowledge-driven systems for mobile robot navigation. 展开更多
关键词 3D bounding box estimation depth estimation mobile robot navigation monocular camera object detection
在线阅读 下载PDF
Monocular Depth Estimation with Sharp Boundary
8
作者 Xin Yang Qingling Chang +2 位作者 Shiting Xu Xinlin Liu Yan Cui 《Computer Modeling in Engineering & Sciences》 SCIE EI 2023年第7期573-592,共20页
Monocular depth estimation is the basic task in computer vision.Its accuracy has tremendous improvement in the decade with the development of deep learning.However,the blurry boundary in the depth map is a serious pro... Monocular depth estimation is the basic task in computer vision.Its accuracy has tremendous improvement in the decade with the development of deep learning.However,the blurry boundary in the depth map is a serious problem.Researchers find that the blurry boundary is mainly caused by two factors.First,the low-level features,containing boundary and structure information,may be lost in deep networks during the convolution process.Second,themodel ignores the errors introduced by the boundary area due to the few portions of the boundary area in the whole area,during the backpropagation.Focusing on the factors mentioned above.Two countermeasures are proposed to mitigate the boundary blur problem.Firstly,we design a scene understanding module and scale transformmodule to build a lightweight fuse feature pyramid,which can deal with low-level feature loss effectively.Secondly,we propose a boundary-aware depth loss function to pay attention to the effects of the boundary’s depth value.Extensive experiments show that our method can predict the depth maps with clearer boundaries,and the performance of the depth accuracy based on NYU-Depth V2,SUN RGB-D,and iBims-1 are competitive. 展开更多
关键词 monocular depth estimation object boundary blurry boundary scene global information feature fusion scale transform boundary aware
在线阅读 下载PDF
Boosting Unsupervised Monocular Depth Estimation with Auxiliary Semantic Information
9
作者 Hui Ren Nan Gao Jia Li 《China Communications》 SCIE CSCD 2021年第6期228-243,共16页
Learning-based multi-task models have been widely used in various scene understanding tasks,and complement each other,i.e.,they allow us to consider prior semantic information to better infer depth.We boost the unsupe... Learning-based multi-task models have been widely used in various scene understanding tasks,and complement each other,i.e.,they allow us to consider prior semantic information to better infer depth.We boost the unsupervised monocular depth estimation using semantic segmentation as an auxiliary task.To address the lack of cross-domain datasets and catastrophic forgetting problems encountered in multi-task training,we utilize existing methodology to obtain redundant segmentation maps to build our cross-domain dataset,which not only provides a new way to conduct multi-task training,but also helps us to evaluate results compared with those of other algorithms.In addition,in order to comprehensively use the extracted features of the two tasks in the early perception stage,we use a strategy of sharing weights in the network to fuse cross-domain features,and introduce a novel multi-task loss function to further smooth the depth values.Extensive experiments on KITTI and Cityscapes datasets show that our method has achieved state-of-the-art performance in the depth estimation task,as well improved semantic segmentation. 展开更多
关键词 unsupervised monocular depth estimation semantic segmentation multi-task model
在线阅读 下载PDF
RADepthNet:Reflectance-aware monocular depth estimation
10
作者 Chuxuan LI Ran YI +5 位作者 Saba Ghazanfar ALI Lizhuang MA Enhua WU Jihong WANG Lijuan MAO Bin SHENG 《Virtual Reality & Intelligent Hardware》 2022年第5期418-431,共14页
Background Monocular depth estimation aims to predict a dense depth map from a single RGB image,and has important applications in 3D reconstruction,automatic driving,and augmented reality.However,existing methods dire... Background Monocular depth estimation aims to predict a dense depth map from a single RGB image,and has important applications in 3D reconstruction,automatic driving,and augmented reality.However,existing methods directly feed the original RGB image into the model to extract depth features without avoiding the interference of depth-irrelevant information on depth-estimation accuracy,which leads to inferior performance.Methods To remove the influence of depth-irrelevant information and improve the depth-prediction accuracy,we propose RADepthNet,a novel reflectance-guided network that fuses boundary features.Specifically,our method predicts depth maps using the following three steps:(1)Intrinsic Image Decomposition.We propose a reflectance extraction module consisting of an encoder-decoder structure to extract the depth-related reflectance.Through an ablation study,we demonstrate that the module can reduce the influence of illumination on depth estimation.(2)Boundary Detection.A boundary extraction module,consisting of an encoder,refinement block,and upsample block,was proposed to better predict the depth at object boundaries utilizing gradient constraints.(3)Depth Prediction Module.We use an encoder different from(2)to obtain depth features from the reflectance map and fuse boundary features to predict depth.In addition,we proposed FIFADataset,a depth-estimation dataset applied in soccer scenarios.Results Extensive experiments on a public dataset and our proposed FIFADataset show that our method achieves state-of-the-art performance. 展开更多
关键词 monocular depth estimation Deep learning Intrinsic image decomposition
在线阅读 下载PDF
Self-supervised coarse-to-fine monocular depth estimation using a lightweight attention module
11
作者 Yuanzhen Li Fei Luo Chunxia Xiao 《Computational Visual Media》 SCIE EI CSCD 2022年第4期631-647,共17页
Self-supervised monocular depth estimation has been widely investigated and applied in previous works.However,existing methods suffer from texture-copy,depth drift,and incomplete structure.It is difficult for normal C... Self-supervised monocular depth estimation has been widely investigated and applied in previous works.However,existing methods suffer from texture-copy,depth drift,and incomplete structure.It is difficult for normal CNN networks to completely understand the relationship between the object and its surrounding environment.Moreover,it is hard to design the depth smoothness loss to balance depth smoothness and sharpness.To address these issues,we propose a coarse-to-fine method with a normalized convolutional block attention module(NCBAM).In the coarse estimation stage,we incorporate the NCBAM into depth and pose networks to overcome the texture-copy and depth drift problems.Then,we use a new network to refine the coarse depth guided by the color image and produce a structure-preserving depth result in the refinement stage.Our method can produce results competitive with state-of-the-art methods.Comprehensive experiments prove the effectiveness of our two-stage method using the NCBAM. 展开更多
关键词 monocular depth estimation texture copy depth drift attention module
原文传递
TalentDepth:基于多尺度注意力机制的复杂天气场景单目深度估计模型
12
作者 张航 卫守林 殷继彬 《计算机科学》 北大核心 2025年第S1期442-448,共7页
对于复杂天气场景图像模糊、低对比度和颜色失真所导致的深度信息预测不准的问题,以往的研究均以标准场景的深度图作为先验信息来对该类场景进行深度估计。然而,这一方式存在先验信息精度较低等问题。对此,提出一个基于多尺度注意力机... 对于复杂天气场景图像模糊、低对比度和颜色失真所导致的深度信息预测不准的问题,以往的研究均以标准场景的深度图作为先验信息来对该类场景进行深度估计。然而,这一方式存在先验信息精度较低等问题。对此,提出一个基于多尺度注意力机制的单目深度估计模型TalentDepth,以实现对复杂天气场景的预测。首先,在编码器中融合多尺度注意力机制,在减少计算成本的同时,保留每个通道的信息,提高特征提取的效率和能力。其次,针对图像深度不清晰的问题,基于几何一致性,提出深度区域细化(Depth Region Refinement,DSR)模块,过滤不准确的像素点,以提高深度信息的可靠性。最后,输入图像翻译模型所生成的复杂样本,并计算相应原始图像上的标准损失来指导模型的自监督训练。在NuScence,KITTI和KITTI-C这3个数据集上,相比于基线模型,所提模型对误差和精度均有优化。 展开更多
关键词 单目深度估计 自监督学习 多尺度注意力 知识提炼 深度学习
在线阅读 下载PDF
轻量化的低成本海洋机器人深度估计方法EDepth
13
作者 陈东烁 柴春来 +1 位作者 叶航 张思赟 《计算机应用》 北大核心 2025年第S1期106-113,共8页
针对传统单目深度估计方法在海洋环境中存在的精度低、鲁棒性差、运行速度慢和难以部署等问题,提出一种轻量化的海洋机器人深度估计方法,命名为EDepth(EfficientDepth)。该方法旨在提升低成本海洋机器人的三维(3D)感知能力。首先,利用... 针对传统单目深度估计方法在海洋环境中存在的精度低、鲁棒性差、运行速度慢和难以部署等问题,提出一种轻量化的海洋机器人深度估计方法,命名为EDepth(EfficientDepth)。该方法旨在提升低成本海洋机器人的三维(3D)感知能力。首先,利用水下光衰减先验,通过空间转换将输入数据从原始RGB(Red-Green-Blue)图像空间映射到RBI(Red-BlueIntensity)输入域,从而提高深度估计的准确性;其次,采用高效的EfficientFormerV2作为特征提取模块,并结合视觉注意力机制MiniViT(Mini Vision Transformer)和光衰减模块实现深度信息的有效提取和处理;此外,通过自适应分区的设计,MiniViT模块能够动态调整深度区间,从而提高深度估计的精度;最后,优化网络结构,从而在不牺牲性能的前提下,实现高效的计算。实验结果表明,EDepth在RGB-D(Red-Green-Blue Depth)数据集USOD10K上的深度估计性能显著优于传统方法。具体来说,EDepth在平均绝对相对误差(Abs Rel)上达到了0.587,而DenseDepth为0.519,尽管DenseDepth在某些指标上表现更佳,但相较于DenseDepth的4 461万参数和171.44 MB的内存占用,EDepth仅有461万参数,减少了89.67%的参数量,而内存占用减少至23.56 MB,且在单个CPU上EDepth的每秒帧数(FPS)达到了14.11,明显优于DenseDepth的2.45。可见,EDepth在深度估计性能和计算效率之间取得了良好的平衡。 展开更多
关键词 三维感知 自适应分区 计算效率 EfficientFormerV2 海洋机器人 单目深度估计
在线阅读 下载PDF
LpDepth:基于拉普拉斯金字塔的自监督单目深度估计
14
作者 曹明伟 邢景杰 +1 位作者 程宜风 赵海锋 《计算机科学》 北大核心 2025年第3期33-40,共8页
自监督单目深度估计受到了国内外研究人员的广泛关注。现有基于深度学习的自监督单目深度估计方法主要采用编码器-解码器结构。然而,这些方法在编码过程中对输入图像进行下采样操作,导致部分图像信息,尤其是图像的边界信息丢失,进而影... 自监督单目深度估计受到了国内外研究人员的广泛关注。现有基于深度学习的自监督单目深度估计方法主要采用编码器-解码器结构。然而,这些方法在编码过程中对输入图像进行下采样操作,导致部分图像信息,尤其是图像的边界信息丢失,进而影响深度图的精度。针对上述问题,提出一种基于拉普拉斯金字塔的自监督单目深度估计方法(Self-supervised Monocular Depth Estimation Based on the Laplace Pyramid,LpDepth)。此方法的核心思想是:首先,使用拉普拉斯残差图丰富编码特征,以弥补在下采样过程中丢失的特征信息;其次,在下采样过程中使用最大池化层突显和放大特征信息,使编码器在特征提取过程中更容易地提取到训练模型所需要的特征信息;最后,使用残差模块解决过拟合问题,提高解码器对特征的利用效率。在KITTI和Make3D等数据集上对所提方法进行了测试,同时将其与现有经典方法进行了比较。实验结果证明了所提方法的有效性。 展开更多
关键词 单目深度估计 拉普拉斯金字塔 残差网络 深度图
在线阅读 下载PDF
DepthMamba:多尺度VisionMamba架构的单目深度估计
15
作者 徐志斌 张孙杰 《计算机应用研究》 北大核心 2025年第3期944-948,共5页
在单目深度估计领域,虽然基于CNN和Transformer的模型已经得到了广泛的研究,但是CNN全局特征提取不足,Transformer则具有二次计算复杂性。为了克服这些限制,提出了一种用于单目深度估计的端到端模型,命名为DepthMamba。该模型能够高效... 在单目深度估计领域,虽然基于CNN和Transformer的模型已经得到了广泛的研究,但是CNN全局特征提取不足,Transformer则具有二次计算复杂性。为了克服这些限制,提出了一种用于单目深度估计的端到端模型,命名为DepthMamba。该模型能够高效地捕捉全局信息并减少计算负担。具体地,该方法引入了视觉状态空间(VSS)模块构建编码器-解码器架构,以提高模型提取多尺度信息和全局信息的能力。此外,还设计了MLPBins深度预测模块,旨在优化深度图的平滑性和整洁性。最后在室内场景NYU_Depth V2数据集和室外场景KITTI数据集上进行了综合实验,实验结果表明:与基于视觉Transformer架构的Depthformer相比,该方法网络参数量减少了27.75%,RMSE分别减少了6.09%和2.63%,验证了算法的高效性和优越性。 展开更多
关键词 单目深度估计 Vmamba Bins深度预测 状态空间模型
在线阅读 下载PDF
Monocular depth estimation based on deep learning for intraoperative guidance using surface-enhanced Raman scattering imaging
16
作者 ANIWAT JUHONG BO LI +12 位作者 YIFAN LIU CHENG-YOU YAO CHIA-WEI YANG A.K.M.ATIQUE ULLAH KUNLI LIU RYAN P.LEWANDOWSKI JACK R.HARKEMA DALEN W.AGNEW YU LEO LEI GARY D.LUKER XUEFEI HUANG WIBOOL PIYAWATTANAMETHA ZHEN QIU 《Photonics Research》 2025年第2期550-560,共11页
Imaging of surface-enhanced Raman scattering(SERS) nanoparticles(NPs) has been intensively studied for cancer detection due to its high sensitivity, unconstrained low signal-to-noise ratios, and multiplexing detection... Imaging of surface-enhanced Raman scattering(SERS) nanoparticles(NPs) has been intensively studied for cancer detection due to its high sensitivity, unconstrained low signal-to-noise ratios, and multiplexing detection capability. Furthermore, conjugating SERS NPs with various biomarkers is straightforward, resulting in numerous successful studies on cancer detection and diagnosis. However, Raman spectroscopy only provides spectral data from an imaging area without co-registered anatomic context. 展开更多
关键词 raman spectroscopy cancer detection surface enhanced raman scattering imaging intraoperative guidance monocular depth estimation anatomic context deep learning sers nanoparticles
原文传递
局部特征引导的室内自监督单目深度估计方法的改进
17
作者 艾浩军 张锋 +2 位作者 吕鹏飞 唐雪华 王中元 《计算机研究与发展》 北大核心 2026年第2期338-351,共14页
近年来,自监督单目深度估计方法取得了显著的性能提升,但在复杂的室内场景生成结构化深度图时性能明显下降,为此,提出局部特征引导知识蒸馏的自监督单目深度估计方法LoFtDepth改进训练过程。首先,使用预训练的深度估计网络预测结构化的... 近年来,自监督单目深度估计方法取得了显著的性能提升,但在复杂的室内场景生成结构化深度图时性能明显下降,为此,提出局部特征引导知识蒸馏的自监督单目深度估计方法LoFtDepth改进训练过程。首先,使用预训练的深度估计网络预测结构化的相对深度图作为深度先验,从中提取局部特征作为边界点引导局部深度估计细化,减少深度无关特征的干扰,将深度先验中的边界知识传递到自监督深度估计网络中。同时,引入逆自动掩模加权的表面法线损失,通过对齐自监督网络预测的深度图和深度先验在无纹理区域的法线方向来提升深度估计精度。最后,根据相机运动的连续性,对相机位姿残差估计施加位姿一致性约束以适应室内场景相机位姿的频繁变化来减小训练误差和提升模型性能。主要的室内公开数据集上的实验结果表明,LoFtDepth性能提升显著,将相对误差降至0.121,且生成的深度图具有更高的全局准确度和良好的结构特征。 展开更多
关键词 单目深度估计 自监督学习 局部特征 知识蒸馏 表面法线约束
在线阅读 下载PDF
特征偏置与注意力聚合的自监督单目深度估计网络
18
作者 贾瑞明 余俊 +2 位作者 徐霄 王涵 郑瀚 《计算机工程与应用》 北大核心 2026年第1期253-263,共11页
提出一种特征偏置与注意力聚合的网络模型,解决单目深度估计任务中存在的深度不连续、局部偏差大等问题。特征偏置增强模块通过在网络内部特征通道维度上添加随机偏置,强化网络对位置信息的处理,增强模型的泛化能力。此外,重新设计卷积... 提出一种特征偏置与注意力聚合的网络模型,解决单目深度估计任务中存在的深度不连续、局部偏差大等问题。特征偏置增强模块通过在网络内部特征通道维度上添加随机偏置,强化网络对位置信息的处理,增强模型的泛化能力。此外,重新设计卷积与自注意力的双路聚合结构,使用深度卷积与非压缩门控结构模块,更加准确地推理像点的深度信息。在公开数据集KITTI和Make3D上分别进行了对比实验和泛化验证实验,对比实验的精度提升了2.01%,跨数据集泛化实验的精度提高了5.23%。 展开更多
关键词 单目深度估计 自监督 聚合注意力 特征偏置增强
在线阅读 下载PDF
基于单目深度估计的冬小麦株高提取方法
19
作者 张辰阳 吴门新 +5 位作者 李峰 张继波 张承明 邱炳文 崔兆韵 孔开昕 《农业工程学报》 北大核心 2026年第2期177-184,共8页
为了满足利用图像技术测量冬小麦株高的需要,该研究提出了一种基于单目深度估计的冬小麦株高提取方法,该方法以相机采集的冬小麦图像作为输入,通过目标区域定位获取有效作物信息,生成像素级深度信息,再将深度信息转换为作物真实株高;在... 为了满足利用图像技术测量冬小麦株高的需要,该研究提出了一种基于单目深度估计的冬小麦株高提取方法,该方法以相机采集的冬小麦图像作为输入,通过目标区域定位获取有效作物信息,生成像素级深度信息,再将深度信息转换为作物真实株高;在训练过程中,使用像素级约束与尺度一致性约束进行联合监督,提高了深度估计精度与株高提取结果的可靠性。在山东泰安农业气象试验站采集冬小麦图像数据用于开展试验,选取BTS、FCRN、DORN和DPT作为对比模型。试验结果表明,深度生成网络在均方根误差(2.759)、对数均方根误差(0.157)、相对误差(0.152)和平方相对误差(0.907)等指标上均优于对比模型。进一步将深度估计结果转换为株高,并与实测值进行对比分析,该方法准确率达到98.74%,优于BTS(92.68%)、FCRN(97.17%)、DORN(97.44%)和DPT(98.40%),证明了该方法在冬小麦长势监测中的有效性和可靠性,能够为生产实践提供理论指导。 展开更多
关键词 冬小麦 株高 单目深度估计 注意力机制
在线阅读 下载PDF
基于深度图像渲染的裸眼3D图像合成研究
20
作者 苏杭 许英朝 +3 位作者 王素彬 邱骏毅 刘博 洪荣辉 《激光杂志》 北大核心 2026年第1期162-170,共9页
针对裸眼3D图像合成中存在的硬件成本高、遮挡区域有空洞及视点间串扰问题,提出了一种基于深度图像渲染的裸眼3D图像合成方法。引入单目深度估计模型Depth Anything V2,通过自监督优化与多任务联合训练实现了高精度深度信息提取,显著降... 针对裸眼3D图像合成中存在的硬件成本高、遮挡区域有空洞及视点间串扰问题,提出了一种基于深度图像渲染的裸眼3D图像合成方法。引入单目深度估计模型Depth Anything V2,通过自监督优化与多任务联合训练实现了高精度深度信息提取,显著降低对深度相机的依赖,降低成本;其次,提出了深度跳变检测与纹理外插联合的空洞填充算法,有效解决遮挡区域像素缺失问题;最后优化了像素重分配策略,通过交错式列像素分布抑制视点间光线交叠,提升了裸眼3D显示的适配性与清晰度。在多种高分辨率图像上,联合空洞填充算法的PSNR与SSIM值较于现有先进算法,分别提升0.34 dB和0.007,视差像素交错映射策略实现视点间光线交叠区域减少47%~49%,光线交叠角度平均增加2.2°~2.3°。 展开更多
关键词 单目深度估计 深度跳变检测 三维映射 视差像素交错映射
原文传递
上一页 1 2 10 下一页 到第
使用帮助 返回顶部