期刊文献+
共找到7篇文章
< 1 >
每页显示 20 50 100
Specific Emitter Identification Based on RepVGG and Gramian Angular Field
1
作者 Deguo Zeng Fuyuan Xu +2 位作者 Jin Qin Zhenyi Yao Zuyue Shang 《Journal of Beijing Institute of Technology》 2025年第6期545-551,共7页
This paper presents a new method for specific emitter identification(SEI)using the reparameterization visual geometry group(RepVGG)neural network model and Gramian angular summation field(GASF).It converts in-phase an... This paper presents a new method for specific emitter identification(SEI)using the reparameterization visual geometry group(RepVGG)neural network model and Gramian angular summation field(GASF).It converts in-phase and quadrature(IQ)signals into 2D feature maps,retaining both time and frequency domain features.Compared to residual network 18-layer(ResNet18)and Hilbert transform methods,this approach offers higher accuracy,faster training,and a smaller model size,making it ideal for hardware deployment. 展开更多
关键词 specific emitter identification re-parameterization visual geometry group(RepVGG) Gramian angular field
在线阅读 下载PDF
基于改进YOLOX-m的安全帽佩戴检测 被引量:8
2
作者 王晓龙 江波 《计算机工程》 CAS CSCD 北大核心 2023年第12期252-261,共10页
安全帽佩戴检测是安全监控系统中的重要组成部分,其检测精度取决于目标分类、小目标检测、域迁移差异等因素。针对现有基于YOLOX-m模型的安全帽佩戴检测算法通常存在分类精度较低、检测目标不完整、轻量化模型性能下降等问题,构建一种... 安全帽佩戴检测是安全监控系统中的重要组成部分,其检测精度取决于目标分类、小目标检测、域迁移差异等因素。针对现有基于YOLOX-m模型的安全帽佩戴检测算法通常存在分类精度较低、检测目标不完整、轻量化模型性能下降等问题,构建一种基于多阶段网络训练策略的改进YOLOX-m模型。首先对YOLOX-m主干特征网络卷积块的堆叠次数进行重新设计,在减小网络规模的同时最大化模型性能,然后将残差化重参视觉几何组与快速空间金字塔池化相结合,提高检测精度和推理速度。设计一种多阶段网络训练策略,将训练集和测试集拆分成多个组,并结合推理阶段生成的伪标签进行多次网络训练,以减少域迁移差异,获得更高的检测精度。实验结果表明,与YOLOX-m模型相比,改进YOLOX-m模型的推理延迟降低了5 ms,模型大小减少了4.7 MB,检测精度提高了1.26个百分点。 展开更多
关键词 安全帽佩戴检测 深度学习 残差化重参视觉几何组 快速空间金字塔池化 多阶段网络训练策略
在线阅读 下载PDF
基于VGGNet-plus的路面裂痕自动分类识别方法 被引量:1
3
作者 肖彭昊 杨修伟 范媛媛 《电子器件》 CAS 北大核心 2022年第2期490-493,共4页
为了有效自动化地识别路面裂缝,在深度学习神经网络VGGNet基础上,提出了基于VGGNet-plus网络的路面裂痕自动分类识别方法。该网络增加了Dropout层和残差层,并在每个卷积层后连接Batch_normalize(BN)层和LeakyReLu层,解决了训练参数过多... 为了有效自动化地识别路面裂缝,在深度学习神经网络VGGNet基础上,提出了基于VGGNet-plus网络的路面裂痕自动分类识别方法。该网络增加了Dropout层和残差层,并在每个卷积层后连接Batch_normalize(BN)层和LeakyReLu层,解决了训练参数过多、深度神经网络的过拟合等问题,简化计算同时减少训练时间。为了增加训练样本的数量,同时使该方法对采集光照条件、角度、噪声等造成的影响具有更强的适应性和鲁棒性,通过灰度处理,上下翻转,左右翻转,灰度二值处理,均值滤波,灰度gamma处理,高斯滤波,中值滤波等方法来进行数据增容。通过Bagging模型集成方法,对预测的数据进行综合评估后选取最佳的预测结果。实验结果表明,VGGNet-plus网络在路面裂缝分类中的准确率可达92%,有效提升了路面裂缝自动检测精度。 展开更多
关键词 深度学习 裂缝分类 残差 数据增容 VGGNet 模型集成
在线阅读 下载PDF
基于深度学习的农田害虫识别研究 被引量:2
4
作者 马鑫鑫 张巧雨 +2 位作者 马越 孙绪程 陈浩 《信息与电脑》 2022年第24期180-182,共3页
农田害虫降低了农作物的产量和质量,如何有效区分和治理农田害虫成为首要解决的问题。文章紧抓农田环境需求和农民对农作物的产量需求不匹配的痛点,基于卷积神经网络技术识别农田害虫,为农业提供有效的识别方式。采用MobileNetV1、残差... 农田害虫降低了农作物的产量和质量,如何有效区分和治理农田害虫成为首要解决的问题。文章紧抓农田环境需求和农民对农作物的产量需求不匹配的痛点,基于卷积神经网络技术识别农田害虫,为农业提供有效的识别方式。采用MobileNetV1、残差神经网络(Residual Network,ResNet)50、视觉几何群网络(Visual Geometry Group Network,VGG)16以及微调预训练模型VGG16共4种网络模型二分类农田害虫图片集。由于样本数据量较少,为防止出现过拟合,使用了数据增强技术,即通过现有训练图片生成更多的训练图片,从而提高泛化能力。实验表明,4种网络模型的准确率分别为88.63%、91.73%、86.49%和90.13%,在农田害虫识别中均具有较好的实际应用效果。 展开更多
关键词 MobileNetV1 视觉几何群网络(VGG)16 残差神经网络(ResNet)50 过拟合
在线阅读 下载PDF
Multi-CNN Fusion Framework for Predictive Violence Detection in Animated Media
5
作者 Tahira Khalil Sadeeq Jan +1 位作者 Rania M.Ghoniem Muhammad Imran Khan Khalil 《Computers, Materials & Continua》 2026年第2期2167-2186,共20页
The contemporary era is characterized by rapid technological advancements,particularly in the fields of communication and multimedia.Digital media has significantly influenced the daily lives of individuals of all age... The contemporary era is characterized by rapid technological advancements,particularly in the fields of communication and multimedia.Digital media has significantly influenced the daily lives of individuals of all ages.One of the emerging domains in digital media is the creation of cartoons and animated videos.The accessibility of the internet has led to a surge in the consumption of cartoons among young children,presenting challenges in monitoring and controlling the content they view.The prevalence of cartoon videos containing potentially violent scenes has raised concerns regarding their impact,especially on young and impressionableminds.This article contributes to the growing concerns about the impact of animated media on children’s mental health and offers solutions to help mitigate these effects.To address this issue,an intelligent,multi-CNN fusion framework is proposed for detecting and predicting violent content in upcoming frames of animated videos.The framework integrates probabilistic and deep learning methodologies by leveraging a combination of visual and temporal features for violence prediction in future scenes.Two specific convolutional neural network classifiers i.e.,VGG16 and ResNet18 are employed to classify scenes from animated content as violent or non-violent.To enhance decision robustness,this study introduces a fusion strategy based on weighted averaging,combining the outputs of both Convolutional Neural Networks(CNNs)into a single decision stream.The resulting classifications are subsequently fed into Naive Bayes classifier,which analyzes sequential patterns to forecast violence in future scenes.The experimental findings demonstrate that the proposed framework achieved predictive accuracy of 92.84%,highlighting its effectiveness for intelligent content moderation.These results underscore the potential of intelligent data fusion techniques in enhancing the reliability and robustness of automated violence detection systems in animated content.This framework offers a promising solution for safeguarding young audiences by enabling proactive and accurate moderation of animated videos. 展开更多
关键词 Violence prediction multi-model fusion cartoon videos residual network(ResNet) visual geometry group(VGG) CNN
在线阅读 下载PDF
基于双支路特征融合的MRI颅脑肿瘤图像分割研究 被引量:2
6
作者 熊炜 周蕾 +2 位作者 乐玲 张开 李利荣 《光电子.激光》 CAS CSCD 北大核心 2022年第4期383-392,共10页
针对磁共振成像(magnetic resonance imaging, MRI)颅脑肿瘤区域误识别与分割网络空间信息丢失问题,提出一种基于双支路特征融合的MRI脑肿瘤图像分割方法。首先通过主支路的重构VGG与注意力模型(re-parameterization visual geometry gr... 针对磁共振成像(magnetic resonance imaging, MRI)颅脑肿瘤区域误识别与分割网络空间信息丢失问题,提出一种基于双支路特征融合的MRI脑肿瘤图像分割方法。首先通过主支路的重构VGG与注意力模型(re-parameterization visual geometry group and attention model, RVAM)提取网络的上下文信息,然后使用可变形卷积与金字塔池化模型(deformable convolution and pyramid pooling model, DCPM)在副支路获取丰富的空间信息,之后使用特征融合模块对两支路的特征信息进行融合。最后引入注意力模型,在上采样过程中加强分割目标在解码时的权重。提出的方法在Kaggle_3m数据集和BraTS2019数据集上进行了实验验证,实验结果表明该方法具有良好的脑肿瘤分割性能,其中在Kaggle_3m上,Dice相似系数、杰卡德系数分别达到了91.45%和85.19%。 展开更多
关键词 磁共振成像(magnetic resonance imaging MRI)颅脑肿瘤图像分割 双支路特征融合 重构VGG与注意力模型(re-parameterization visual geometry group and attention model RVAM) 可变形卷积与金字塔池化模型(deformable convolution and pyramid pooling model DCPM)
原文传递
Glaucoma Detection with Retinal Fundus Images Using Segmentation and Classification 被引量:2
7
作者 Thisara Shyamalee Dulani Meedeniya 《Machine Intelligence Research》 EI CSCD 2022年第6期563-580,共18页
Glaucoma is a prevalent cause of blindness worldwide.If not treated promptly,it can cause vision and quality of life to deteriorate.According to statistics,glaucoma affects approximately 65 million individuals globall... Glaucoma is a prevalent cause of blindness worldwide.If not treated promptly,it can cause vision and quality of life to deteriorate.According to statistics,glaucoma affects approximately 65 million individuals globally.Fundus image segmentation depends on the optic disc(OD)and optic cup(OC).This paper proposes a computational model to segment and classify retinal fundus images for glaucoma detection.Different data augmentation techniques were applied to prevent overfitting while employing several data pre-processing approaches to improve the image quality and achieve high accuracy.The segmentation models are based on an attention U-Net with three separate convolutional neural networks(CNNs)backbones:Inception-v3,visual geometry group 19(VGG19),and residual neural network 50(ResNet50).The classification models also employ a modified version of the above three CNN architectures.Using the RIM-ONE dataset,the attention U-Net with the ResNet50 model as the encoder backbone,achieved the best accuracy of 99.58%in segmenting OD.The Inception-v3 model had the highest accuracy of 98.79%for glaucoma classification among the evaluated segmentation,followed by the modified classification architectures. 展开更多
关键词 Attention U-Net SEGMENTATION classification Inception-v3 visual geometry group 19(VGG19) residual neural network 50(ResNet50) GLAUCOMA fundus images
原文传递
上一页 1 下一页 到第
使用帮助 返回顶部