期刊文献+
共找到287篇文章
< 1 2 15 >
每页显示 20 50 100
Fast recognition using convolutional neural network for the coal particle density range based on images captured under multiple light sources 被引量:7
1
作者 Feiyan Bai Minqiang Fan +1 位作者 Hongli Yang Lianping Dong 《International Journal of Mining Science and Technology》 SCIE EI CAS CSCD 2021年第6期1053-1061,共9页
A method based on multiple images captured under different light sources at different incident angles was developed to recognize the coal density range in this study.The innovation is that two new images were construc... A method based on multiple images captured under different light sources at different incident angles was developed to recognize the coal density range in this study.The innovation is that two new images were constructed based on images captured under four single light sources.Reconstruction image 1 was constructed by fusing greyscale versions of the original images into one image,and Reconstruction image2 was constructed based on the differences between the images captured under the different light sources.Subsequently,the four original images and two reconstructed images were input into the convolutional neural network AlexNet to recognize the density range in three cases:-1.5(clean coal) and+1.5 g/cm^(3)(non-clean coal);-1.8(non-gangue) and+1.8 g/cm^(3)(gangue);-1.5(clean coal),1.5-1.8(middlings),and+1.8 g/cm^(3)(gangue).The results show the following:(1) The reconstructed images,especially Reconstruction image 2,can effectively improve the recognition accuracy for the coal density range compared with images captured under single light source.(2) The recognition accuracies for gangue and non-gangue,clean coal and non-clean coal,and clean coal,middlings,and gangue reached88.44%,86.72% and 77.08%,respectively.(3) The recognition accuracy increases as the density moves further away from the boundary density. 展开更多
关键词 COAL Density range Image Multiple light sources convolutional neural network
在线阅读 下载PDF
Intelligent Fusion of Infrared and Visible Image Data Based on Convolutional Sparse Representation and Improved Pulse-Coupled Neural Network 被引量:3
2
作者 Jingming Xia Yi Lu +1 位作者 Ling Tan Ping Jiang 《Computers, Materials & Continua》 SCIE EI 2021年第4期613-624,共12页
Multi-source information can be obtained through the fusion of infrared images and visible light images,which have the characteristics of complementary information.However,the existing acquisition methods of fusion im... Multi-source information can be obtained through the fusion of infrared images and visible light images,which have the characteristics of complementary information.However,the existing acquisition methods of fusion images have disadvantages such as blurred edges,low contrast,and loss of details.Based on convolution sparse representation and improved pulse-coupled neural network this paper proposes an image fusion algorithm that decompose the source images into high-frequency and low-frequency subbands by non-subsampled Shearlet Transform(NSST).Furthermore,the low-frequency subbands were fused by convolutional sparse representation(CSR),and the high-frequency subbands were fused by an improved pulse coupled neural network(IPCNN)algorithm,which can effectively solve the problem of difficulty in setting parameters of the traditional PCNN algorithm,improving the performance of sparse representation with details injection.The result reveals that the proposed method in this paper has more advantages than the existing mainstream fusion algorithms in terms of visual effects and objective indicators. 展开更多
关键词 Image fusion infrared image visible light image non-downsampling shear wave transform improved PCNN convolutional sparse representation
在线阅读 下载PDF
M2ANet:Multi-branch and multi-scale attention network for medical image segmentation 被引量:1
3
作者 Wei Xue Chuanghui Chen +3 位作者 Xuan Qi Jian Qin Zhen Tang Yongsheng He 《Chinese Physics B》 2025年第8期547-559,共13页
Convolutional neural networks(CNNs)-based medical image segmentation technologies have been widely used in medical image segmentation because of their strong representation and generalization abilities.However,due to ... Convolutional neural networks(CNNs)-based medical image segmentation technologies have been widely used in medical image segmentation because of their strong representation and generalization abilities.However,due to the inability to effectively capture global information from images,CNNs can easily lead to loss of contours and textures in segmentation results.Notice that the transformer model can effectively capture the properties of long-range dependencies in the image,and furthermore,combining the CNN and the transformer can effectively extract local details and global contextual features of the image.Motivated by this,we propose a multi-branch and multi-scale attention network(M2ANet)for medical image segmentation,whose architecture consists of three components.Specifically,in the first component,we construct an adaptive multi-branch patch module for parallel extraction of image features to reduce information loss caused by downsampling.In the second component,we apply residual block to the well-known convolutional block attention module to enhance the network’s ability to recognize important features of images and alleviate the phenomenon of gradient vanishing.In the third component,we design a multi-scale feature fusion module,in which we adopt adaptive average pooling and position encoding to enhance contextual features,and then multi-head attention is introduced to further enrich feature representation.Finally,we validate the effectiveness and feasibility of the proposed M2ANet method through comparative experiments on four benchmark medical image segmentation datasets,particularly in the context of preserving contours and textures. 展开更多
关键词 medical image segmentation convolutional neural network multi-branch attention multi-scale feature fusion
原文传递
Explainable Diabetic Retinopathy Detection Using a Distributed CNN and LightGBM Framework
4
作者 Pooja Bidwai Shilpa Gite +1 位作者 Biswajeet Pradhan Abdullah Almari 《Computers, Materials & Continua》 2025年第8期2645-2676,共32页
Diabetic Retinopathy(DR)is a critical disorder that affects the retina due to the constant rise in diabetics and remains the major cause of blindness across the world.Early detection and timely treatment are essential... Diabetic Retinopathy(DR)is a critical disorder that affects the retina due to the constant rise in diabetics and remains the major cause of blindness across the world.Early detection and timely treatment are essential to mitigate the effects of DR,such as retinal damage and vision impairment.Several conventional approaches have been proposed to detect DR early and accurately,but they are limited by data imbalance,interpretability,overfitting,convergence time,and other issues.To address these drawbacks and improve DR detection accurately,a distributed Explainable Convolutional Neural network-enabled Light Gradient Boosting Machine(DE-ExLNN)is proposed in this research.The model combines an explainable Convolutional Neural Network(CNN)and Light Gradient Boosting Machine(LightGBM),achieving highly accurate outcomes in DR detection.LightGBM serves as the detection model,and the inclusion of an explainable CNN addresses issues that conventional CNN classifiers could not resolve.A custom dataset was created for this research,containing both fundus and OCTA images collected from a realtime environment,providing more accurate results compared to standard conventional DR datasets.The custom dataset demonstrates notable accuracy,sensitivity,specificity,and Matthews Correlation Coefficient(MCC)scores,underscoring the effectiveness of this approach.Evaluations against other standard datasets achieved an accuracy of 93.94%,sensitivity of 93.90%,specificity of 93.99%,and MCC of 93.88%for fundus images.For OCTA images,the results obtained an accuracy of 95.30%,sensitivity of 95.50%,specificity of 95.09%,andMCC of 95%.Results prove that the combination of explainable CNN and LightGBMoutperforms othermethods.The inclusion of distributed learning enhances the model’s efficiency by reducing time consumption and complexity while facilitating feature extraction. 展开更多
关键词 Diabetic retinopathy explainable convolutional neural network light gradient boosting machine fundus image custom dataset
在线阅读 下载PDF
基于DGA的LightGBM-ICOA-CNN变压器故障诊断方法
5
作者 孙涛 陈鑫 +3 位作者 杜民生 郭文凯 张佳 李亚兴 《电气工程学报》 北大核心 2025年第6期459-468,共10页
为提高基于深度学习的变压器故障诊断精度,提出了基于油中溶解气体分析(Dissolved gas analysis,DGA)的LightGBM-ICOA-CNN变压器故障诊断方法。首先,基于变压器油中溶解气体含量对变压器特征变量进行丰富,利用轻量梯度提升机算法(Light ... 为提高基于深度学习的变压器故障诊断精度,提出了基于油中溶解气体分析(Dissolved gas analysis,DGA)的LightGBM-ICOA-CNN变压器故障诊断方法。首先,基于变压器油中溶解气体含量对变压器特征变量进行丰富,利用轻量梯度提升机算法(Light gradient boosting machine,LightGBM)量化其重要性,实现特征变量优选;其次,引入改进浣熊优化算法(Improved coati optimization algorithm,ICOA)对卷积神经网络(Convolutional neural network,CNN)的学习率、卷积核大小与数量、全连接层神经元数量等超参数实现优化,提高模型诊断结果的准确率;最后,通过算例分析对建立的LightGBM-ICOA-CNN方法性能进行评估,验证了所提方法对变压器故障诊断的有效性,且收敛性较好,精度较高。 展开更多
关键词 变压器 故障诊断 轻量梯度提升机 特征变量 改进浣熊优化算法 卷积神经网络
在线阅读 下载PDF
面向还原的结构光投影三维物理轮廓测量技术
6
作者 杨彦雄 田苗 张磊 《激光杂志》 北大核心 2026年第2期77-82,共6页
为了减少三维物理轮廓测量误差,研究一种面向还原的结构光投影三维物理轮廓测量技术。利用相机捕捉变形的光场信息,采用N步相移法选取双频时间相位展开方法实施相位解包裹,获取准确的三维物理轮廓相位信息,采用卷积神经网络拟合光学过... 为了减少三维物理轮廓测量误差,研究一种面向还原的结构光投影三维物理轮廓测量技术。利用相机捕捉变形的光场信息,采用N步相移法选取双频时间相位展开方法实施相位解包裹,获取准确的三维物理轮廓相位信息,采用卷积神经网络拟合光学过程与结构光投影三维物理轮廓相位间复杂的非线性关系,实现对受光学干扰的结构光图案的准确还原。采用损失函数训练卷积神经网络,衡量还原结果与真实标签之间的差异,输出精准的三维物理轮廓测量结果。本研究采用6层卷积神经网络,基于稀疏描述子损失函数与全卷积交叉熵检测损失函数组合训练,在10帧图像测试中实现低于1.6%的误还原比例及0.01 mm以下的三维物理轮廓测量不确定度。 展开更多
关键词 结构光投影 三维物理轮廓 非线性映射 卷积神经网络 轮廓还原 双频时间相位
原文传递
LFSepNet:融合Transformer的照明和面部特征解耦人脸识别方法
7
作者 黎克迅 高治军 《计算机工程与应用》 北大核心 2026年第4期201-209,共9页
在低光环境下,人脸识别面临图像质量低、特征模糊等诸多挑战,导致现有方法难以提取鲁棒且辨识度高的特征,从而严重影响识别性能。为应对这一问题,提出了一种新颖的非成对低光人脸识别模型LFSepNet(low-light face separation network)... 在低光环境下,人脸识别面临图像质量低、特征模糊等诸多挑战,导致现有方法难以提取鲁棒且辨识度高的特征,从而严重影响识别性能。为应对这一问题,提出了一种新颖的非成对低光人脸识别模型LFSepNet(low-light face separation network)。与传统基于卷积神经网络(convolutional neural network,CNN)架构的训练方法不同,LFSepNet采用Transformer架构,更有效地捕捉长距离依赖关系,从而克服卷积神经网络在局部感受野上的限制。由于低光环境下的人脸图像往往整体偏暗,仅有少数区域可能包含较丰富的照明信息,传统CNN在特征提取时容易受限于局部区域,难以充分利用这些关键信息。相比之下,Transformer通过自注意力机制实现全局信息建模,使网络能够更全面地整合亮度不均的人脸图像信息,从而提升特征解耦的效果和低光人脸识别的准确性。LFSepNet模型包含自适应亮度分离模块和自适应照明间隙损失,通过动态分离人脸与光照特征,减少光照干扰,同时进一步优化特征分离效果,使模型能够提取更加精确和鲁棒的特征。实验结果表明,LFSepNet在多个低光人脸数据集上的性能均优于现有方法,特别是在极端低光条件下,其识别精度显著提升。该研究为低光人脸识别提供了基于非成对设置的有效解决方案,并在实际应用中展现了良好的潜力。 展开更多
关键词 低光人脸识别 深度学习 TRANSFORMER 特征解耦 卷积神经网络(CNN) LFSepNet
在线阅读 下载PDF
基于GCNet网络的玻璃瓶缺陷检测算法
8
作者 傅莉 房志磊 +1 位作者 席剑辉 任艳 《沈阳航空航天大学学报》 2026年第1期48-55,共8页
针对当前玻璃瓶缺陷检测深度学习模型计算量大、参数量多、难以部署的问题,探索一种高效的轻量化解决方案。针对此问题,结合YOLOv8的网络结构设计特征提取网络GCNet。首先,采用GhostConv代替标准卷积;其次,为了降低YOLOv8瓶颈层的参数... 针对当前玻璃瓶缺陷检测深度学习模型计算量大、参数量多、难以部署的问题,探索一种高效的轻量化解决方案。针对此问题,结合YOLOv8的网络结构设计特征提取网络GCNet。首先,采用GhostConv代替标准卷积;其次,为了降低YOLOv8瓶颈层的参数量和计算量,设计瓶颈层卷积模块,并重新搭建瓶颈层;最后,结合C2f模块的结构设计构建新的CM模块。新的特征提取网络较原YOLOv8网络有着更低的参数量。在特征融合部分采用重新构建的重复加权双向特征融合金字塔结构,解决随着网络层数的加深导致特征信息丢失的问题。同时针对边界框的回归问题上,结合WIoU与Inner-ShapeIoU,提高模型的回归收敛速度。结果表明,相较于YOLOv8算法,由上述方法组成的YOLOv8-DB参数量降低了45.8%,计算量下降了11.9%,精度提升了0.4%。改进后的模型能够有效地降低占用的计算资源,更好地适用于特定工业检测环境。 展开更多
关键词 深度学习 缺陷检测 多尺度卷积 WIoU 轻量化
在线阅读 下载PDF
基于Multi-Light模型的奶牛个体识别研究 被引量:3
9
作者 付丽丽 李士军 +2 位作者 孔朔琳 宫鹤 李思函 《黑龙江畜牧兽医》 CAS 北大核心 2023年第3期41-45,51,132,133,共8页
为了解决大规模智能化奶牛养殖场对奶牛个体识别存在模型大、识别速度慢的问题,试验构建了一种用于识别奶牛个体的多尺度轻量化卷积神经网络(Multi-Light)模型,将拍摄的奶牛图像经过标注后利用DeepLab V3模型从复杂背景中分割出单头奶... 为了解决大规模智能化奶牛养殖场对奶牛个体识别存在模型大、识别速度慢的问题,试验构建了一种用于识别奶牛个体的多尺度轻量化卷积神经网络(Multi-Light)模型,将拍摄的奶牛图像经过标注后利用DeepLab V3模型从复杂背景中分割出单头奶牛图像;在Multi-Light模型中引入空洞卷积,保证该模型参数量不变的同时增强提取图像全局信息的能力;加入多尺度卷积模块增强该模型对不同尺度特征点的检测能力,在该模型中使用短路连接以保证特征不丢失,提升模型的识别精度;此外,利用通道注意力机制提高了该模型识别精度,同时使该模型具有更多的非线性;最后将分割得到的奶牛图像数据集输入Multi-Light模型进行训练。结果表明:Multi-Light模型对奶牛个体识别的精度达98.51%,高于其他经典模型对奶牛个体的识别率;与轻量级模型对比,Multi-Light模型的大小为5.86 MB,在具备高识别精度的前提下参数量较少。说明试验所搭建的Multi-Light模型克服了传统方法中需要对特征进行人为提取、提取特征方法不够鲁棒、识别模型参数量大及识别速度慢的缺点,为奶牛个体轻量化识别提供了参考。 展开更多
关键词 空洞卷积 多尺度 轻量化 奶牛识别 跳跃连接
原文传递
Development of an automatic monitoring system for rice light-trap pests based on machine vision 被引量:18
10
作者 YAO Qing FENG Jin +9 位作者 TANG Jian XU Wei-gen ZHU Xu-hua YANG Bao-jun LU Jun XIE Yi-ze YAO Bo WU Shu-zhen KUAI Nai-yang WANG Li-jun 《Journal of Integrative Agriculture》 SCIE CAS CSCD 2020年第10期2500-2513,共14页
Monitring pest populations in paddy fields is important to effectively implement integrated pest management.Light traps are widely used to monitor field pests all over the world.Most conventional light traps still inv... Monitring pest populations in paddy fields is important to effectively implement integrated pest management.Light traps are widely used to monitor field pests all over the world.Most conventional light traps still involve manual identification of target pests from lots of trapped insects,which is time-consuming,labor-intensive and error-prone,especially in pest peak periods.In this paper,we developed an automatic monitoring system for rice light-trap pests based on machine vision.This system is composed of an itelligent light trap,a computer or mobile phone client platform and a cloud server.The light trap firstly traps,kills and disperses insects,then collects images of trapped insects and sends each image to the cloud server.Five target pests in images are automatically identifed and counted by pest identification models loaded in the server.To avoid light-trap insects piling up,a vibration plate and a moving rotation conveyor belt are adopted to disperse these trapped insects.There was a close correlation(r=0.92)between our automatic and manual identification methods based on the daily pest number of one-year images from one light trap.Field experiments demonstrated the effectiveness and accuracy of our automatic light trap monitoring system. 展开更多
关键词 automatic monitoring system light trap rice pest machine vision image processing convolutional neural network
在线阅读 下载PDF
基于CNN-LSTM-lightGBM组合的超短期风电功率预测方法 被引量:27
11
作者 王愈轩 刘尔佳 黄永章 《科学技术与工程》 北大核心 2022年第36期16067-16074,共8页
近年来,风电装机规模逐年增加,风电数据采集量呈现规模化增长,面对海量多维、强波动的风电数据,风电功率预测精度仍面临一定的挑战。为提高风电功率预测精度,提出了基于卷积神经网络(convolutional neural networks,CNN)-长短期记忆网络... 近年来,风电装机规模逐年增加,风电数据采集量呈现规模化增长,面对海量多维、强波动的风电数据,风电功率预测精度仍面临一定的挑战。为提高风电功率预测精度,提出了基于卷积神经网络(convolutional neural networks,CNN)-长短期记忆网络(long short-term memory,LSTM)和梯度提升学习(light gradient boosting machine,lightGBM)组合的超短期风电功率预测方法。首先,分别建立CNN-LSTM和lightGBM的风电功率超短期预测模型。其中,CNN-LSTM模型采用CNN对风电数据集进行特征处理,并将其作为LSTM模型的数据输入,从而建立CNN-LSTM融合的预测模型;然后,采用误差倒数法对CNN-LSTM和lightGBM的预测数据进行加权组合,建立CNN-LSTM-lightGBM组合的预测模型;最后,采用张北曹碾沟风电场的风电数据集,以未来4 h风电功率为预测目标,验证了组合模型的有效性。预测结果表明:相较于其他3种单一模型,组合模型具有更高的预测精度。 展开更多
关键词 卷积神经网络(CNN) 长短期记忆网络(LSTM) 梯度提升学习(lightGBM) 组合模型 风电功率预测
在线阅读 下载PDF
DLA+: A Light Aggregation Network for Object Classification and Detection 被引量:1
12
作者 Fu-Tian Wang Li Yang +2 位作者 Jin Tang Si-Bao Chen Xin Wang 《International Journal of Automation and computing》 EI CSCD 2021年第6期963-972,共10页
An efficient convolution neural network(CNN) plays a crucial role in various visual tasks like object classification or detection, etc. The most common way to construct a CNN is stacking the same convolution block or ... An efficient convolution neural network(CNN) plays a crucial role in various visual tasks like object classification or detection, etc. The most common way to construct a CNN is stacking the same convolution block or complex connection. These approaches may be efficient but the parameter size and computation(Comp) have explosive growth. So we present a novel architecture called"DLA+", which could obtain the feature from the different stages, and by the newly designed convolution block, could achieve better accuracy, while also dropping the computation six times compared to the baseline. We design some experiments about classification and object detection. On the CIFAR10 and VOC data-sets, we get better precision and faster speed than other architecture. The lightweight network even allows us to deploy to some low-performance device like drone, laptop, etc. 展开更多
关键词 light weight image classification channel attention efficient convolution object detection
原文传递
Deep learning-assisted common temperature measurement based on visible light imaging
13
作者 朱佳仪 何志民 +8 位作者 黄成 曾峻 林惠川 陈福昌 余超群 李燕 张永涛 陈焕庭 蒲继雄 《Chinese Physics B》 SCIE EI CAS CSCD 2024年第8期230-236,共7页
Real-time,contact-free temperature monitoring of low to medium range(30℃-150℃)has been extensively used in industry and agriculture,which is usually realized by costly infrared temperature detection methods.This pap... Real-time,contact-free temperature monitoring of low to medium range(30℃-150℃)has been extensively used in industry and agriculture,which is usually realized by costly infrared temperature detection methods.This paper proposes an alternative approach of extracting temperature information in real time from the visible light images of the monitoring target using a convolutional neural network(CNN).A mean-square error of<1.119℃was reached in the temperature measurements of low to medium range using the CNN and the visible light images.Imaging angle and imaging distance do not affect the temperature detection using visible optical images by the CNN.Moreover,the CNN has a certain illuminance generalization ability capable of detection temperature information from the images which were collected under different illuminance and were not used for training.Compared to the conventional machine learning algorithms mentioned in the recent literatures,this real-time,contact-free temperature measurement approach that does not require any further image processing operations facilitates temperature monitoring applications in the industrial and civil fields. 展开更多
关键词 convolutional neural network visible light image temperature measurement low-to-medium-range temperatures
原文传递
Mapping local-scale working population and daytime population densities using points-of-interest and nighttime light satellite imageries
14
作者 Yeran Sun Jing Xie +2 位作者 Yu Wang Ting On Chan Zhao-Yong Sun 《Geo-Spatial Information Science》 CSCD 2024年第6期1852-1867,共16页
In this study,we proposed a multi-source approach for mapping local-scale population density of England.Specifically,we mapped both the working and daytime population densities by integrating the multi-source data suc... In this study,we proposed a multi-source approach for mapping local-scale population density of England.Specifically,we mapped both the working and daytime population densities by integrating the multi-source data such as residential population density,point-of-interest density,point-of-interest category mix,and nighttime light intensity.It is demonstrated that combining remote sensing and social sensing data provides a plausible way to map annual working or daytime population densities.In this paper,we trained models with England-wide data and subsequently tested these models with Wales-wide data.In addition,we further tested the models with England-wide data at a higher level of spatial granularity.Particularly,the random forest and convolutional neural network models were adopted to map population density.The estimated results and validation suggest that the three built models have high prediction accuracies at the local authority district level.It is shown that the convolutional neural network models have the greatest prediction accuracies at the local authority district level though they are most time-consuming.The models trained with the data at the local authority district level are less appropriately applicable to test data at a higher level of spatial granularity.The proposed multi-source approach performs well in mapping local-scale population density.It indicates that combining remote sensing and social sensing data is advantageous to mapping socioeconomic variables. 展开更多
关键词 Nighttime light imagery point-of-interest working population daytime population convolutional neural network
原文传递
采用卷积神经网络的室内可见光定位方法 被引量:2
15
作者 王亮 孙海燕 《导航定位学报》 北大核心 2025年第1期128-136,共9页
针对多径反射与系统噪声导致室内可见光定位精度下降的问题,提出一种基于扩张卷积网络的室内可见光三维定位方法:基于皮尔森相关性系数对采集的接收信号强度向量进行过滤,删除系统噪声引起的非线性失真接收信号强度向量,以提高训练的神... 针对多径反射与系统噪声导致室内可见光定位精度下降的问题,提出一种基于扩张卷积网络的室内可见光三维定位方法:基于皮尔森相关性系数对采集的接收信号强度向量进行过滤,删除系统噪声引起的非线性失真接收信号强度向量,以提高训练的神经网络精度;然后,将接收信号强度向量集建立的指纹库传入神经网络进行训练,利用神经网络较强的三维空间结构表达能力拟合多径反射和系统噪声下的非线性指纹库。仿真结果表明,在7 m×7 m×3 m的室内环境下,所提方法的平均定位误差可达0.91 cm,其中90%样本的定位误差小于1.17 cm;此外,所提方法的平均定位误差较全连接神经网络和卷积神经网络可分别降低0.82 cm和0.56 cm,证明所提方法在多径反射与系统噪声环境下具有较好的定位性能。 展开更多
关键词 可见光通信系统 室内定位 物联网 卷积神经网络(CNN) 可见光定位
在线阅读 下载PDF
基于卷积神经网络的线结构光高精度三维测量方法 被引量:2
16
作者 叶涛 何威燃 +2 位作者 刘国鹏 欧阳煜 王斌 《仪器仪表学报》 北大核心 2025年第2期183-195,共13页
线结构光视觉三维测量技术因其高精度和非接触的三维重建优势而被广泛应用。然而,现有的线结构光三维测量方法在标定过程中往往面临较高的耦合性问题,且在复杂环境下,背景噪声和光照变化会严重干扰条纹的提取,导致结构光条纹中心定位精... 线结构光视觉三维测量技术因其高精度和非接触的三维重建优势而被广泛应用。然而,现有的线结构光三维测量方法在标定过程中往往面临较高的耦合性问题,且在复杂环境下,背景噪声和光照变化会严重干扰条纹的提取,导致结构光条纹中心定位精度下降,进而影响整体三维测量的精度和鲁棒性。针对上述问题,提出了一种基于卷积神经网络的鲁棒三维测量方法。首先,设计了一种创新性的残差U型块特征金字塔网络(RSU-FPN),旨在实现背景噪声的干扰抑制和结构光条纹区域中心的高精度鲁棒提取。其次,构建了一种新型的线结构光视觉传感器,并提出了一种分离式测量模型,成功将摄像机标定与光平面标定解耦,极大地提高了系统的灵活性与扩展性。通过这种解耦的标定方式,避免了传统标定方法中存在的耦合问题,使得整个测量系统更加高效且易于调整。实验结果表明,所提出的基于卷积神经网络的鲁棒三维测量方法,在复杂背景下能够实现结构光条纹中心的高精度提取,利用提取出的光条纹中心进行标定,其均方根误差分别为x方向0.005 mm、y方向0.009 mm以及z方向0.097 mm。并且,该方法在不同表面类型(如漫反射表面和光滑反射表面)上均能实现高精度的三维重建,验证了其在实际应用中的优越性和强大的鲁棒性。 展开更多
关键词 线结构光 三维测量 卷积神经网络 残差U型块特征金字塔网络 背景噪声抑制
原文传递
基于跳跃连接神经网络的无监督弱光图像增强算法 被引量:2
17
作者 刘洋 刘思瑞 +1 位作者 徐晓淼 王竹筠 《电子测量与仪器学报》 北大核心 2025年第5期208-216,共9页
针对Zero-DCE网络存在细节丢失和不同亮度区域处理结果出现差异等问题,设计了一种基于增强深度曲线估计网络(EnDCE-Net)的无监督弱光图像增强算法。通过探索弱光图像与未配对的正常光照图像之间的潜在映射关系,实现了对低光照场景下图... 针对Zero-DCE网络存在细节丢失和不同亮度区域处理结果出现差异等问题,设计了一种基于增强深度曲线估计网络(EnDCE-Net)的无监督弱光图像增强算法。通过探索弱光图像与未配对的正常光照图像之间的潜在映射关系,实现了对低光照场景下图像质量的显著改善。首先,提出新的特征提取网络,该网络整合了多个跳跃连接与卷积层,实现低层与高层特征的有效融合,从而学习到弱光图像中的关键特征,增强网络对弱光图像的学习能力。其次,设计一组联合的无参考损失函数,强调优化过程中与亮度相关的特性,从而更有利于图像增强模型的参数更新,提高图像增强的质量和效果。为了验证所提出算法的有效性,在5个公开数据集上进行了对比实验,与次优算法Zero-DCE相比,有参考数据集SICE上的峰值信噪比(PSNR)和结构相似性(SSIM)分别提升了9.4%、21%。无参考数据集LIME、DICM、MEF、NPE上NIQE分别达到了4.04、3.04、3.35、3.83。实验结果表明,所提出算法表现出色,增强后的图像色彩自然,亮度均衡且细节清晰。无论是主观视觉评价还是客观定量指标,均显著优于对比算法,充分体现了在图像增强效果上的卓越性和先进性。 展开更多
关键词 弱光图像增强 深度曲线估计 无参考损失函数 多层卷积神经网络 无监督学习
原文传递
基于GCN-LSTM的多交叉口信号灯控制 被引量:1
18
作者 徐东伟 朱宏俊 +2 位作者 郭海锋 周晓刚 汤立新 《高技术通讯》 北大核心 2025年第5期472-479,共8页
强化学习(reinforcement learning,RL)由于其解决高度动态环境中复杂决策问题的能力,成为信号灯控制中一种具有前景的解决方案。大多数基于强化学习的方法独立生成智能体的动作,它们可能导致交叉口的动作冲突、道路资源浪费。因此,本文... 强化学习(reinforcement learning,RL)由于其解决高度动态环境中复杂决策问题的能力,成为信号灯控制中一种具有前景的解决方案。大多数基于强化学习的方法独立生成智能体的动作,它们可能导致交叉口的动作冲突、道路资源浪费。因此,本文提出了基于图卷积网络和长短期记忆(graph convolution network-long short-term memory,GCNLSTM)的多交叉口信号灯控制方法。首先,基于二进制权重网络对多交叉口进行构图。其次,通过图卷积网络聚合周围交叉口的空间状态信息,利用长短期记忆(long short-term memory,LSTM)获得交叉口的历史状态信息。最后,通过基于竞争网络框架的Q值网络进行动作的选择,实现对交叉口相位的控制。实验结果表明,与其他强化学习方法相比,本文方法在多交叉口的信号灯控制中能够减少交叉口的队列长度,并使道路网络中的车辆获得更少的等待时间。 展开更多
关键词 智能交通系统 交通信号灯控制 多智能体强化学习 长短期记忆 图卷积网络
在线阅读 下载PDF
不良光照场景下的交通标志识别算法 被引量:1
19
作者 党宏社 肖利霞 张选德 《半导体光电》 北大核心 2025年第1期142-148,共7页
交通标志识别技术作为自动驾驶系统的核心组件,在保障行车安全方面具有重要作用。为改善不良光照场景下交通标志的自动识别效果,提出了一种基于改进NanoDet的交通标志识别算法。该算法以NanoDet模型为基础,首先,在主干网络集成SSM模块与... 交通标志识别技术作为自动驾驶系统的核心组件,在保障行车安全方面具有重要作用。为改善不良光照场景下交通标志的自动识别效果,提出了一种基于改进NanoDet的交通标志识别算法。该算法以NanoDet模型为基础,首先,在主干网络集成SSM模块与CBAM注意力机制,有效提高模型在不良光照场景下的识别精度;其次,构建加权双向特征金字塔网络强化多尺度特征融合;最后,将AGM模块中的标准卷积替换为深度可分离卷积,在保证感受野的同时显著降低模型参数量。基于扩充版CCTSDB数据集的实验表明,该算法在保持138.2帧/s实时处理速度的前提下,识别精度为90.2%,相较基准模型提升4.7个百分点。 展开更多
关键词 交通标志识别 不良光照 注意力模块 特征金字塔 深度可分离卷积
原文传递
基于CNN与堆叠LightGBM的多模态OSA检测方法
20
作者 林美娜 郑和裕 《自动化与信息工程》 2022年第3期25-30,共6页
提出一种基于血氧饱和度和心电图信号的多模态阻塞性睡眠呼吸暂停(OSA)检测方法。首先,提取血氧饱和度和心电图信号的经验特征,并利用皮尔逊相关系数获得最优特征集;然后,利用卷积网络(CNN)生成深层特征以挖掘不同模态间的潜在相关性;最... 提出一种基于血氧饱和度和心电图信号的多模态阻塞性睡眠呼吸暂停(OSA)检测方法。首先,提取血氧饱和度和心电图信号的经验特征,并利用皮尔逊相关系数获得最优特征集;然后,利用卷积网络(CNN)生成深层特征以挖掘不同模态间的潜在相关性;最后,构建堆叠的轻量级梯度提升机(LightGBM),以提高分类器检测精度。在公开数据集Apnea-ECG上进行四折交叉验证,平均准确性、敏感性和特异性分别为96.04%、96.44%和96.22%,相较于决策层融合有较高的分类性能。 展开更多
关键词 阻塞性睡眠呼吸暂停 卷积网络 轻量级梯度提升机 血氧 心电图
在线阅读 下载PDF
上一页 1 2 15 下一页 到第
使用帮助 返回顶部