期刊文献+
共找到294篇文章
< 1 2 15 >
每页显示 20 50 100
DDFNet:real-time salient object detection with dual-branch decoding fusion for steel plate surface defects
1
作者 Tao Wang Wang-zhe Du +5 位作者 Xu-wei Li Hua-xin Liu Yuan-ming Liu Xiao-miao Niu Ya-xing Liu Tao Wang 《Journal of Iron and Steel Research International》 2025年第8期2421-2433,共13页
A novel dual-branch decoding fusion convolutional neural network model(DDFNet)specifically designed for real-time salient object detection(SOD)on steel surfaces is proposed.DDFNet is based on a standard encoder–decod... A novel dual-branch decoding fusion convolutional neural network model(DDFNet)specifically designed for real-time salient object detection(SOD)on steel surfaces is proposed.DDFNet is based on a standard encoder–decoder architecture.DDFNet integrates three key innovations:first,we introduce a novel,lightweight multi-scale progressive aggregation residual network that effectively suppresses background interference and refines defect details,enabling efficient salient feature extraction.Then,we propose an innovative dual-branch decoding fusion structure,comprising the refined defect representation branch and the enhanced defect representation branch,which enhance accuracy in defect region identification and feature representation.Additionally,to further improve the detection of small and complex defects,we incorporate a multi-scale attention fusion module.Experimental results on the public ESDIs-SOD dataset show that DDFNet,with only 3.69 million parameters,achieves detection performance comparable to current state-of-the-art models,demonstrating its potential for real-time industrial applications.Furthermore,our DDFNet-L variant consistently outperforms leading methods in detection performance.The code is available at https://github.com/13140W/DDFNet. 展开更多
关键词 Steel plate surface defect Real-time detection salient object detection Dual-branch decoder Multi-scale attention fusion Multi-scale residual fusion
原文传递
Salient Object Detection Based on Multi-Strategy Feature Optimization
2
作者 Libo Han Sha Tao +3 位作者 Wen Xia Weixin Sun Li Yan Wanlin Gao 《Computers, Materials & Continua》 2025年第2期2431-2449,共19页
At present, salient object detection (SOD) has achieved considerable progress. However, the methods that perform well still face the issue of inadequate detection accuracy. For example, sometimes there are problems of... At present, salient object detection (SOD) has achieved considerable progress. However, the methods that perform well still face the issue of inadequate detection accuracy. For example, sometimes there are problems of missed and false detections. Effectively optimizing features to capture key information and better integrating different levels of features to enhance their complementarity are two significant challenges in the domain of SOD. In response to these challenges, this study proposes a novel SOD method based on multi-strategy feature optimization. We propose the multi-size feature extraction module (MSFEM), which uses the attention mechanism, the multi-level feature fusion, and the residual block to obtain finer features. This module provides robust support for the subsequent accurate detection of the salient object. In addition, we use two rounds of feature fusion and the feedback mechanism to optimize the features obtained by the MSFEM to improve detection accuracy. The first round of feature fusion is applied to integrate the features extracted by the MSFEM to obtain more refined features. Subsequently, the feedback mechanism and the second round of feature fusion are applied to refine the features, thereby providing a stronger foundation for accurately detecting salient objects. To improve the fusion effect, we propose the feature enhancement module (FEM) and the feature optimization module (FOM). The FEM integrates the upper and lower features with the optimized features obtained by the FOM to enhance feature complementarity. The FOM uses different receptive fields, the attention mechanism, and the residual block to more effectively capture key information. Experimental results demonstrate that our method outperforms 10 state-of-the-art SOD methods. 展开更多
关键词 salient object detection multi-strategy feature optimization feedback mechanism
在线阅读 下载PDF
Multi-Stream Temporally Enhanced Network for Video Salient Object Detection
3
作者 Dan Xu Jiale Ru Jinlong Shi 《Computers, Materials & Continua》 SCIE EI 2024年第1期85-104,共20页
Video salient object detection(VSOD)aims at locating the most attractive objects in a video by exploring the spatial and temporal features.VSOD poses a challenging task in computer vision,as it involves processing com... Video salient object detection(VSOD)aims at locating the most attractive objects in a video by exploring the spatial and temporal features.VSOD poses a challenging task in computer vision,as it involves processing complex spatial data that is also influenced by temporal dynamics.Despite the progress made in existing VSOD models,they still struggle in scenes of great background diversity within and between frames.Additionally,they encounter difficulties related to accumulated noise and high time consumption during the extraction of temporal features over a long-term duration.We propose a multi-stream temporal enhanced network(MSTENet)to address these problems.It investigates saliency cues collaboration in the spatial domain with a multi-stream structure to deal with the great background diversity challenge.A straightforward,yet efficient approach for temporal feature extraction is developed to avoid the accumulative noises and reduce time consumption.The distinction between MSTENet and other VSOD methods stems from its incorporation of both foreground supervision and background supervision,facilitating enhanced extraction of collaborative saliency cues.Another notable differentiation is the innovative integration of spatial and temporal features,wherein the temporal module is integrated into the multi-stream structure,enabling comprehensive spatial-temporal interactions within an end-to-end framework.Extensive experimental results demonstrate that the proposed method achieves state-of-the-art performance on five benchmark datasets while maintaining a real-time speed of 27 fps(Titan XP).Our code and models are available at https://github.com/RuJiaLe/MSTENet. 展开更多
关键词 Video salient object detection deep learning temporally enhanced foreground-background collaboration
在线阅读 下载PDF
Local saliency consistency-based label inference for weakly supervised salient object detection using scribble annotations
4
作者 Shuo Zhao Peng Cui +1 位作者 Jing Shen Haibo Liu 《CAAI Transactions on Intelligence Technology》 SCIE EI 2024年第1期239-249,共11页
Recently,weak supervision has received growing attention in the field of salient object detection due to the convenience of labelling.However,there is a large performance gap between weakly supervised and fully superv... Recently,weak supervision has received growing attention in the field of salient object detection due to the convenience of labelling.However,there is a large performance gap between weakly supervised and fully supervised salient object detectors because the scribble annotation can only provide very limited foreground/background information.Therefore,an intuitive idea is to infer annotations that cover more complete object and background regions for training.To this end,a label inference strategy is proposed based on the assumption that pixels with similar colours and close positions should have consistent labels.Specifically,k-means clustering algorithm was first performed on both colours and coordinates of original annotations,and then assigned the same labels to points having similar colours with colour cluster centres and near coordinate cluster centres.Next,the same annotations for pixels with similar colours within each kernel neighbourhood was set further.Extensive experiments on six benchmarks demonstrate that our method can significantly improve the performance and achieve the state-of-the-art results. 展开更多
关键词 label inference salient object detection weak supervision
在线阅读 下载PDF
Automatic salient object segmentation using saliency map and color segmentation 被引量:1
5
作者 HAN Sung-ho JUNG Gye-dong +2 位作者 LEE Sangh-yuk HONG Yeong-pyo LEE Sang-hun 《Journal of Central South University》 SCIE EI CAS 2013年第9期2407-2413,共7页
A new method for automatic salient object segmentation is presented.Salient object segmentation is an important research area in the field of object recognition,image retrieval,image editing,scene reconstruction,and 2... A new method for automatic salient object segmentation is presented.Salient object segmentation is an important research area in the field of object recognition,image retrieval,image editing,scene reconstruction,and 2D/3D conversion.In this work,salient object segmentation is performed using saliency map and color segmentation.Edge,color and intensity feature are extracted from mean shift segmentation(MSS)image,and saliency map is created using these features.First average saliency per segment image is calculated using the color information from MSS image and generated saliency map.Then,second average saliency per segment image is calculated by applying same procedure for the first image to the thresholding,labeling,and hole-filling applied image.Thresholding,labeling and hole-filling are applied to the mean image of the generated two images to get the final salient object segmentation.The effectiveness of proposed method is proved by showing 80%,89%and 80%of precision,recall and F-measure values from the generated salient object segmentation image and ground truth image. 展开更多
关键词 salient object visual attention saliency map color segmentation
在线阅读 下载PDF
A Multiscale Superpixel-Level Salient Object Detection Model Using Local-Global Contrast Cue
6
作者 穆楠 徐新 +1 位作者 王英林 张晓龙 《Journal of Shanghai Jiaotong university(Science)》 EI 2017年第1期121-128,共8页
The goal of salient object detection is to estimate the regions which are most likely to attract human's visual attention. As an important image preprocessing procedure to reduce the computational complexity, sali... The goal of salient object detection is to estimate the regions which are most likely to attract human's visual attention. As an important image preprocessing procedure to reduce the computational complexity, salient object detection is still a challenging problem in computer vision. In this paper, we proposed a salient object detection model by integrating local and global superpixel contrast at multiple scales. Three features are computed to estimate the saliency of superpixel. Two optimization measures are utilized to refine the resulting saliency map. Extensive experiments with the state-of-the-art saliency models on four public datasets demonstrate the effectiveness of the proposed model. 展开更多
关键词 salient object detection superpixel multiple scales local contrast global contrast TP 391.4 A
原文传递
Salient Object Detection Based on a Novel Combination Framework Using the Perceptual Matching and Subjective-Objective Mapping Technologies
7
作者 Jian Han Jialu Li +3 位作者 Meng Liu Zhe Ren Zhimin Cao Xingbin Liu 《Journal of Beijing Institute of Technology》 EI CAS 2023年第1期95-106,共12页
The integrity and fineness characterization of non-connected regions and contours is a major challenge for existing salient object detection.The key to address is how to make full use of the subjective and objective s... The integrity and fineness characterization of non-connected regions and contours is a major challenge for existing salient object detection.The key to address is how to make full use of the subjective and objective structural information obtained in different steps.Therefore,by simulating the human visual mechanism,this paper proposes a novel multi-decoder matching correction network and subjective structural loss.Specifically,the loss pays different attentions to the foreground,boundary,and background of ground truth map in a top-down structure.And the perceived saliency is mapped to the corresponding objective structure of the prediction map,which is extracted in a bottom-up manner.Thus,multi-level salient features can be effectively detected with the loss as constraint.And then,through the mapping of improved binary cross entropy loss,the differences between salient regions and objects are checked to pay attention to the error prone region to achieve excellent error sensitivity.Finally,through tracking the identifying feature horizontally and vertically,the subjective and objective interaction is maximized.Extensive experiments on five benchmark datasets demonstrate that compared with 12 state-of-the-art methods,the algorithm has higher recall and precision,less error and strong robustness and generalization ability,and can predict complete and refined saliency maps. 展开更多
关键词 salient object detection subjective-objective mapping perceptional separation and matching error sensitivity non-connected region detection
在线阅读 下载PDF
Cross-modal attention and reinforcement for RGB-T salient object detection
8
作者 Bi Hongbo Sun Weihan +3 位作者 Zhang Jiayuan Xia Bingjie Guo Yingwei Zhang Cong 《The Journal of China Universities of Posts and Telecommunications》 2025年第2期44-55,共12页
Exploring the interaction between red,green,blue(RGB)and thermal infrared modalities is critical to the success of RGB-thermal(RGB-T)salient object detection(RGB-T SOD).In this paper,a cross-modal attention and reinfo... Exploring the interaction between red,green,blue(RGB)and thermal infrared modalities is critical to the success of RGB-thermal(RGB-T)salient object detection(RGB-T SOD).In this paper,a cross-modal attention and reinforcement network(CAR-Net)was proposed to explore the implicit relationship between the two modalities,which fully leverages the beneficial expression and complementary fusion of the two modalities.Specifically,CAR-Net has a cross-modal attention module(CAM)that enables efficient interaction and key information extraction through joint attention.It also includes a feature strengthener module(FSM)for improved representation using channel rank and loop methods.A large number of experiments show that the CAR-Net achieves the best performance on three publicly available datasets. 展开更多
关键词 RGB-thermal(RGB-T)salient object detection(RGB-T SOD) ATTENTION feature strengthener multi-modal fusion
原文传递
A Novel Divide and Conquer Solution for Long-term Video Salient Object Detection
9
作者 Yun-Xiao Li Cheng-Li-Zhao Chen +2 位作者 Shuai Li Ai-Min Hao Hong Qin 《Machine Intelligence Research》 EI CSCD 2024年第4期684-703,共20页
Recently,a new research trend in our video salient object detection(VSOD)research community has focused on enhancing the detection results via model self-fine-tuning using sparsely mined high-quality keyframes from th... Recently,a new research trend in our video salient object detection(VSOD)research community has focused on enhancing the detection results via model self-fine-tuning using sparsely mined high-quality keyframes from the given sequence.Although such a learning scheme is generally effective,it has a critical limitation,i.e.,the model learned on sparse frames only possesses weak generalization ability.This situation could become worse on“long”videos since they tend to have intensive scene variations.Moreover,in such videos,the keyframe information from a longer time span is less relevant to the previous,which could also cause learning conflict and deteriorate the model performance.Thus,the learning scheme is usually incapable of handling complex pattern modeling.To solve this problem,we propose a divide-and-conquer framework,which can convert a complex problem domain into multiple simple ones.First,we devise a novel background consistency analysis(BCA)which effectively divides the mined frames into disjoint groups.Then for each group,we assign an individual deep model on it to capture its key attribute during the fine-tuning phase.During the testing phase,we design a model-matching strategy,which could dynamically select the best-matched model from those fine-tuned ones to handle the given testing frame.Comprehensive experiments show that our method can adapt severe background appearance variation coupling with object movement and obtain robust saliency detection compared with the previous scheme and the state-of-the-art methods. 展开更多
关键词 Video salient object detection background consistency analysis weakly supervised learning long-term information background shift.
原文传递
Saliency Rank:Two-stage manifold ranking for salient object detection 被引量:5
10
作者 Wei Qi Ming-Ming Cheng +2 位作者 Ali Borji Huchuan Lu Lian-Fa Bai 《Computational Visual Media》 2015年第4期309-320,共12页
Salient object detection remains one of the most important and active research topics in computer vision,with wide-ranging applications to object recognition,scene understanding,image retrieval,context aware image edi... Salient object detection remains one of the most important and active research topics in computer vision,with wide-ranging applications to object recognition,scene understanding,image retrieval,context aware image editing,image compression,etc. Most existing methods directly determine salient objects by exploring various salient object features.Here,we propose a novel graph based ranking method to detect and segment the most salient object in a scene according to its relationship to image border(background) regions,i.e.,the background feature.Firstly,we use regions/super-pixels as graph nodes,which are fully connected to enable both long range and short range relations to be modeled. The relationship of each region to the image border(background) is evaluated in two stages:(i) ranking with hard background queries,and(ii) ranking with soft foreground queries. We experimentally show how this two-stage ranking based salient object detection method is complementary to traditional methods,and that integrated results outperform both. Our method allows the exploitation of intrinsic image structure to achieve high quality salient object determination using a quadratic optimization framework,with a closed form solution which can be easily computed.Extensive method evaluation and comparison using three challenging saliency datasets demonstrate that our method consistently outperforms 10 state-of-theart models by a big margin. 展开更多
关键词 salient object detection manifold ranking visual attention SALIENCY
原文传递
Light field salient object detection:A review and benchmark 被引量:2
11
作者 Keren Fu Yao Jiang +3 位作者 Ge-Peng Ji Tao Zhou Qijun Zhao Deng-Ping Fan 《Computational Visual Media》 SCIE EI CSCD 2022年第4期509-534,共26页
Salient object detection(SOD)is a long-standing research topic in computer vision with increasing interest in the past decade.Since light fields record comprehensive information of natural scenes that benefit SOD in a... Salient object detection(SOD)is a long-standing research topic in computer vision with increasing interest in the past decade.Since light fields record comprehensive information of natural scenes that benefit SOD in a number of ways,using light field inputs to improve saliency detection over conventional RGB inputs is an emerging trend.This paper provides the first comprehensive review and a benchmark for light field SOD,which has long been lacking in the saliency community.Firstly,we introduce light fields,including theory and data forms,and then review existing studies on light field SOD,covering ten traditional models,seven deep learning-based models,a comparative study,and a brief review.Existing datasets for light field SOD are also summarized.Secondly,we benchmark nine representative light field SOD models together with several cutting-edge RGB-D SOD models on four widely used light field datasets,providing insightful discussions and analyses,including a comparison between light field SOD and RGB-D SOD models.Due to the inconsistency of current datasets,we further generate complete data and supplement focal stacks,depth maps,and multi-view images for them,making them consistent and uniform.Our supplemental data make a universal benchmark possible.Lastly,light field SOD is a specialised problem,because of its diverse data representations and high dependency on acquisition hardware,so it differs greatly from other saliency detection tasks.We provide nine observations on challenges and future directions,and outline several open issues.All the materials including models,datasets,benchmarking results,and supplemented light field datasets are publicly available at https://github.com/kerenfu/LFSOD-Survey. 展开更多
关键词 light field salient object detection(SOD) deep learning BENCHMARKING
原文传递
WGI-Net:A weighted group integration network for RGB-D salient object detection
12
作者 Yanliang Ge Cong Zhang +2 位作者 Kang Wang Ziqi Liu Hongbo Bi 《Computational Visual Media》 EI CSCD 2021年第1期115-125,共11页
Salient object detection is used as a preprocess in many computer vision tasks(such as salient object segmentation,video salient object detection,etc.).When performing salient object detection,depth information can pr... Salient object detection is used as a preprocess in many computer vision tasks(such as salient object segmentation,video salient object detection,etc.).When performing salient object detection,depth information can provide clues to the location of target objects,so effective fusion of RGB and depth feature information is important.In this paper,we propose a new feature information aggregation approach,weighted group integration(WGI),to effectively integrate RGB and depth feature information.We use a dual-branch structure to slice the input RGB image and depth map separately and then merge the results separately by concatenation.As grouped features may lose global information about the target object,we also make use of the idea of residual learning,taking the features captured by the original fusion method as supplementary information to ensure both accuracy and completeness of the fused information.Experiments on five datasets show that our model performs better than typical existing approaches for four evaluation metrics. 展开更多
关键词 weighted group depth information RGBD information salient object detection deep learning
原文传递
Salient object extraction for user-targeted video content association
13
作者 Jia LI Han-nan YU +2 位作者 Yong-hong TIAN Tie-jun HUANG Wen GAO 《Journal of Zhejiang University-Science C(Computers and Electronics)》 SCIE EI 2010年第11期850-859,共10页
The increasing amount of videos on the Internet and digital libraries highlights the necessity and importance of interactive video services such as automatically associating additional materials(e.g.,advertising logos... The increasing amount of videos on the Internet and digital libraries highlights the necessity and importance of interactive video services such as automatically associating additional materials(e.g.,advertising logos and relevant selling information) with the video content so as to enrich the viewing experience.Toward this end,this paper presents a novel approach for user-targeted video content association(VCA) .In this approach,the salient objects are extracted automatically from the video stream using complementary saliency maps.According to these salient objects,the VCA system can push the related logo images to the users.Since the salient objects often correspond to important video content,the associated images can be considered as content-related.Our VCA system also allows users to associate images to the preferred video content through simple interactions by the mouse and an infrared pen.Moreover,by learning the preference of each user through collecting feedbacks on the pulled or pushed images,the VCA system can provide user-targeted services.Experimental results show that our approach can effectively and efficiently extract the salient objects.Moreover,subjective evaluations show that our system can provide content-related and user-targeted VCA services in a less intrusive way. 展开更多
关键词 salient object extraction User-targeted video content association Complementary saliency maps
原文传递
Salient object detection: A survey 被引量:53
14
作者 Ali Borji Ming-Ming Cheng +2 位作者 Qibin Hou Huaizu Jiang Jia Li 《Computational Visual Media》 CSCD 2019年第2期117-150,共34页
Detecting and segmenting salient objects from natural scenes, often referred to as salient object detection, has attracted great interest in computer vision. While many models have been proposed and several applicatio... Detecting and segmenting salient objects from natural scenes, often referred to as salient object detection, has attracted great interest in computer vision. While many models have been proposed and several applications have emerged, a deep understanding of achievements and issues remains lacking. We aim to provide a comprehensive review of recent progress in salient object detection and situate this field among other closely related areas such as generic scene segmentation, object proposal generation, and saliency for fixation prediction. Covering 228 publications, we survey i) roots, key concepts, and tasks, ii) core techniques and main modeling trends, and iii) datasets and evaluation metrics for salient object detection. We also discuss open problems such as evaluation metrics and dataset bias in model performance, and suggest future research directions. 展开更多
关键词 salient object detection SALIENCY visual ATTENTION REGIONS of INTEREST
原文传递
Detection of salient objects with focused attention based on spatial and temporal coherence 被引量:4
15
作者 WU Yang ZHENG NanNing +2 位作者 YUAN ZeJian JIANG HuaiZu LIU Tie 《Chinese Science Bulletin》 SCIE EI CAS 2011年第10期1055-1062,共8页
The understanding and analysis of video content are fundamentally important for numerous applications,including video summarization,retrieval,navigation,and editing.An important part of this process is to detect salie... The understanding and analysis of video content are fundamentally important for numerous applications,including video summarization,retrieval,navigation,and editing.An important part of this process is to detect salient (which usually means important and interesting) objects in video segments.Unlike existing approaches,we propose a method that combines the saliency measurement with spatial and temporal coherence.The integration of spatial and temporal coherence is inspired by the focused attention in human vision.In the proposed method,the spatial coherence of low-level visual grouping cues (e.g.appearance and motion) helps per-frame object-background separation,while the temporal coherence of the object properties (e.g.shape and appearance) ensures consistent object localization over time,and thus the method is robust to unexpected environment changes and camera vibrations.Having developed an efficient optimization strategy based on coarse-to-fine multi-scale dynamic programming,we evaluate our method using a challenging dataset that is freely available together with this paper.We show the effectiveness and complementariness of the two types of coherence,and demonstrate that they can significantly improve the performance of salient object detection in videos. 展开更多
关键词 时间相干性 空间相干性 连贯性 检测 突出 视频内容 组成部分 人类视觉
在线阅读 下载PDF
多特征聚合的边界引导视频图像显著目标检测
16
作者 张荣国 郑晓鸽 +2 位作者 王丽芳 胡静 刘小君 《中国图象图形学报》 北大核心 2025年第4期1141-1154,共14页
目的视频显著目标检测的目的是识别和突出显示视频中的重要对象或区域。现有的方法在挖掘边界线索和时空特征之间的相关性方面存在不足,并且在特征聚合过程中未能充分考虑相关的上下文信息,导致检测结果不够精确。因此提出了多特征聚合... 目的视频显著目标检测的目的是识别和突出显示视频中的重要对象或区域。现有的方法在挖掘边界线索和时空特征之间的相关性方面存在不足,并且在特征聚合过程中未能充分考虑相关的上下文信息,导致检测结果不够精确。因此提出了多特征聚合的边界引导网络,进行显著目标边界信息和显著目标时空信息之间的互补协作。方法首先,提取视频帧显著目标的空间和运动特征,在不同分辨率下将显著目标边界特征与显著目标时空特征耦合,突出运动目标边界的特征,更准确地定位视频显著目标;其次,采用了多层特征注意聚合模块以提高不同特征的表征能力,使得各相异特征得以充分利用;同时在训练阶段采用混合损失来帮助网络学习,以更加准确地分割出运动目标显著的边界区域,获得期望的显著目标。结果实验在4个数据集上与现有的5种方法进行了比较,所提方法在4个数据集上的F-measure值均优于对比方法。在DAVIS(densely annotated video segmentation)数据集上,与性能次优的模型相比,F-measure值提高了0.2%,S-measure值略低于最优值0.7%;在FBMS(Freiburg-Berkeley motion segmentation)数据集上,F-measure值比次优值提高了0.9%;在ViSal数据集上,平均绝对误差(mean absolute error,MAE)值仅低于最优方法STVS(spatial temporal video salient)0.1%,F-measure值比STVS提高了0.2%;在MCL据集上,所提方法实现了最优的MAE值2.2%,S-measure值和F-measure值比次优方法SSAV(saliency-shift aware VSOD)分别提高了1.6%和0.6%。结论提出的方法能够有效提升检测出的视频显著目标的边界质量。 展开更多
关键词 视频图像 显著性目标检测 边界引导 多尺度特征 特征聚合
原文传递
畸变自适应与位置感知的360°全景图像显著目标检测网络
17
作者 陈晓雷 杜泽龙 +1 位作者 张学功 王兴 《中国图象图形学报》 北大核心 2025年第8期2758-2774,共17页
目的现有360°全景图像显著目标检测方法一定程度上解决了360°全景图像投影后的几何畸变问题,但是这些方法面对复杂场景或是前景与背景对比度较低的场景时,容易受到背景干扰,导致检测效果不佳。为了同时解决几何畸变和背景干扰... 目的现有360°全景图像显著目标检测方法一定程度上解决了360°全景图像投影后的几何畸变问题,但是这些方法面对复杂场景或是前景与背景对比度较低的场景时,容易受到背景干扰,导致检测效果不佳。为了同时解决几何畸变和背景干扰,提出一种畸变自适应与位置感知网络(distortion-adaptive and position-aware network,DPNet)。方法提出两个对畸变和位置敏感的自适应检测模块:畸变自适应模块(distortion-adaptive module,DAM)和位置感知模块(position-aware module,PAM)。它们可以帮助模型根据等矩形投影的特点和具体图像决定应该关注图像的哪些区域。在此基础上,进一步提出一个显著信息增强模块(salient information enhancement module,SIEM),该模块用高级特征指导低级特征,过滤其中的非显著信息,防止背景干扰对360°显著目标检测效果的影响。结果实验在2个公开数据集(360-SOD,360-SSOD)上与13种新颖方法进行了客观指标和主观结果的比较,在8个评价指标上的综合性能优于13种对比方法。本文还设置了泛化性实验,采用交叉验证的方式表明了本文模型优秀的泛化性能。结论本文所提出的360°全景图像显著目标检测模型DPNet,同时考虑了360°全景图像投影后的几何畸变问题和复杂场景下的背景干扰问题,能够有效地、完全自适应地检测显著目标。 展开更多
关键词 360°全景图像 显著目标检测(SOD) 畸变自适应 位置感知 抗背景干扰
原文传递
面向360°全景图像显著目标检测的畸变语义聚合网络
18
作者 陈晓雷 张学功 +1 位作者 杜泽龙 王兴 《中国图象图形学报》 北大核心 2025年第7期2451-2467,共17页
目的为了有效应对360°全景图像的几何畸变和大视野特性带来的挑战,提出一种畸变自适应语义聚合网络(distortion semantic aggregation network,DSANet)。该网络能够提升360°全景图像显著目标检测性能。方法DSANet由3个模块组... 目的为了有效应对360°全景图像的几何畸变和大视野特性带来的挑战,提出一种畸变自适应语义聚合网络(distortion semantic aggregation network,DSANet)。该网络能够提升360°全景图像显著目标检测性能。方法DSANet由3个模块组成:畸变自适应校正模块(distortion aware calibration module,DACM)、多尺度语义注意力聚合模块(multiscale semantic attention aggregation module,MSAAM)以及渐进式细化模块(progressive refinement module,PRM)。DACM模块利用不同扩张率的可变形卷积来学习自适应权重矩阵,校正360°全景图像中的几何畸变;MSAAM模块结合注意力机制和可变形卷积,提取并融合全局语义特征与局部细节特征,生成多尺度语义特征;PRM模块逐层融合多尺度语义特征,进一步提升检测精度。MSAAM模块与PRM模块相配合,解决360°全景图像的大视野问题。结果在两个公开数据集360-SOD和360-SSOD(共计1605幅图像)上进行的实验表明,DSANet在6种主流评价指标(Max F-measure、Mean F-measure、MAE(mean absolute error)、Max E-measure、Mean E-measure、Structure-measure)上均优于其他方法。结论本文方法在多个客观评价指标上表现突出,同时生成的显著目标图像在边缘轮廓性和空间结构细节信息上更为清晰。 展开更多
关键词 深度学习 显著目标检测(SOD) 360°全景图像 几何畸变 大视野
原文传递
多尺度特征信息融合的显著性目标检测算法
19
作者 陈爽 王润豪 +1 位作者 罗林棋 何毅 《科学技术与工程》 北大核心 2025年第23期9919-9926,共8页
显著性目标检测是计算机视觉领域热门研究之一,而显著性目标检测在处理背景复杂的图像时存在检测目标边缘模糊的问题。提出了一种多尺度特征信息融合的网络结构U^(2)Net_MFF(U^(2)Net_multi-scale feature fusion)。该网络通过对每层解... 显著性目标检测是计算机视觉领域热门研究之一,而显著性目标检测在处理背景复杂的图像时存在检测目标边缘模糊的问题。提出了一种多尺度特征信息融合的网络结构U^(2)Net_MFF(U^(2)Net_multi-scale feature fusion)。该网络通过对每层解码器增加跳跃连接,使每层解码器不仅能够获得上一级解码器的特征还能获得本层以上所有更浅层次的特征信息,以此来获取目标区域更多的细节特征,提高网络的检测精度;将动态上采样DySample融入到残差U型块中来提高上采样后特征图的质量,并结合多维度协作注意力机制抑制图像复杂背景干扰的同时提升对显著性目标区域的敏感程度。实验表明,改进后的算法在准确率、精确率、召回率、均交并比和F 1分数上分别提升了1%、0.3%、0.4%、0.7%和0.8%,证明该算法具有更强的抗干扰能力和更清晰的目标边界检测效果。 展开更多
关键词 显著性目标 多尺度特征融合 注意力机制 U^(2)Net
在线阅读 下载PDF
基于深度学习的RGB-D图像显著性目标检测前沿进展 被引量:2
20
作者 黄年昌 杨阳 +1 位作者 张强 韩军功 《计算机学报》 北大核心 2025年第2期284-316,共33页
显著性目标检测是计算机视觉领域的基础问题之一,旨在对图像中最吸引人注意的目标进行检测和分割。随着深度学习技术的发展,基于RGB(Red-Green-Blue)图像的显著性目标检测算法取得了巨大进步,在简单场景下已经取得较为满意的结果。然而... 显著性目标检测是计算机视觉领域的基础问题之一,旨在对图像中最吸引人注意的目标进行检测和分割。随着深度学习技术的发展,基于RGB(Red-Green-Blue)图像的显著性目标检测算法取得了巨大进步,在简单场景下已经取得较为满意的结果。然而,局限于可见光相机的成像能力,RGB图像易受到光照条件的影响,且无法捕捉场景的三维空间信息。相应地,基于RGB图像的显著性目标检测算法通常难以在一些复杂场景下取得较好的检测效果。近年来,随着深度成像技术不断发展和硬件成本不断降低,深度相机得到了广泛应用。其捕获的场景空间信息,与可见光图像获取的场景细节信息相互补充,有助于提升复杂场景下显著性目标检测性能。因此,RGB-深度(RGB-Depth,RGB-D)图像显著性目标检测引起了学者广泛研究。本文对近期基于深度学习的RGB-D图像显著性目标检测算法进行了整理和分析。首先,分析了多模态RGB-D图像显著性目标检测所面临的关键问题,并以此对现有算法解决这些关键问题的主要思路和方法进行了总结和梳理。然后,介绍了用于RGB-D图像显著性目标检测算法研究的主流数据集和常用性能评价指标,并对各类主流模型进行了定量比较和定性分析。最后,本文进一步分析了RGB-D图像显著性目标检测领域有待解决的问题,同时对今后可能的研究趋势进行了展望。 展开更多
关键词 显著性目标检测 RGB图像 深度图像 深度学习 多模态图像处理
在线阅读 下载PDF
上一页 1 2 15 下一页 到第
使用帮助 返回顶部