This paper presents a new method for specific emitter identification(SEI)using the reparameterization visual geometry group(RepVGG)neural network model and Gramian angular summation field(GASF).It converts in-phase an...This paper presents a new method for specific emitter identification(SEI)using the reparameterization visual geometry group(RepVGG)neural network model and Gramian angular summation field(GASF).It converts in-phase and quadrature(IQ)signals into 2D feature maps,retaining both time and frequency domain features.Compared to residual network 18-layer(ResNet18)and Hilbert transform methods,this approach offers higher accuracy,faster training,and a smaller model size,making it ideal for hardware deployment.展开更多
农田害虫降低了农作物的产量和质量,如何有效区分和治理农田害虫成为首要解决的问题。文章紧抓农田环境需求和农民对农作物的产量需求不匹配的痛点,基于卷积神经网络技术识别农田害虫,为农业提供有效的识别方式。采用MobileNetV1、残差...农田害虫降低了农作物的产量和质量,如何有效区分和治理农田害虫成为首要解决的问题。文章紧抓农田环境需求和农民对农作物的产量需求不匹配的痛点,基于卷积神经网络技术识别农田害虫,为农业提供有效的识别方式。采用MobileNetV1、残差神经网络(Residual Network,ResNet)50、视觉几何群网络(Visual Geometry Group Network,VGG)16以及微调预训练模型VGG16共4种网络模型二分类农田害虫图片集。由于样本数据量较少,为防止出现过拟合,使用了数据增强技术,即通过现有训练图片生成更多的训练图片,从而提高泛化能力。实验表明,4种网络模型的准确率分别为88.63%、91.73%、86.49%和90.13%,在农田害虫识别中均具有较好的实际应用效果。展开更多
The contemporary era is characterized by rapid technological advancements,particularly in the fields of communication and multimedia.Digital media has significantly influenced the daily lives of individuals of all age...The contemporary era is characterized by rapid technological advancements,particularly in the fields of communication and multimedia.Digital media has significantly influenced the daily lives of individuals of all ages.One of the emerging domains in digital media is the creation of cartoons and animated videos.The accessibility of the internet has led to a surge in the consumption of cartoons among young children,presenting challenges in monitoring and controlling the content they view.The prevalence of cartoon videos containing potentially violent scenes has raised concerns regarding their impact,especially on young and impressionableminds.This article contributes to the growing concerns about the impact of animated media on children’s mental health and offers solutions to help mitigate these effects.To address this issue,an intelligent,multi-CNN fusion framework is proposed for detecting and predicting violent content in upcoming frames of animated videos.The framework integrates probabilistic and deep learning methodologies by leveraging a combination of visual and temporal features for violence prediction in future scenes.Two specific convolutional neural network classifiers i.e.,VGG16 and ResNet18 are employed to classify scenes from animated content as violent or non-violent.To enhance decision robustness,this study introduces a fusion strategy based on weighted averaging,combining the outputs of both Convolutional Neural Networks(CNNs)into a single decision stream.The resulting classifications are subsequently fed into Naive Bayes classifier,which analyzes sequential patterns to forecast violence in future scenes.The experimental findings demonstrate that the proposed framework achieved predictive accuracy of 92.84%,highlighting its effectiveness for intelligent content moderation.These results underscore the potential of intelligent data fusion techniques in enhancing the reliability and robustness of automated violence detection systems in animated content.This framework offers a promising solution for safeguarding young audiences by enabling proactive and accurate moderation of animated videos.展开更多
Glaucoma is a prevalent cause of blindness worldwide.If not treated promptly,it can cause vision and quality of life to deteriorate.According to statistics,glaucoma affects approximately 65 million individuals globall...Glaucoma is a prevalent cause of blindness worldwide.If not treated promptly,it can cause vision and quality of life to deteriorate.According to statistics,glaucoma affects approximately 65 million individuals globally.Fundus image segmentation depends on the optic disc(OD)and optic cup(OC).This paper proposes a computational model to segment and classify retinal fundus images for glaucoma detection.Different data augmentation techniques were applied to prevent overfitting while employing several data pre-processing approaches to improve the image quality and achieve high accuracy.The segmentation models are based on an attention U-Net with three separate convolutional neural networks(CNNs)backbones:Inception-v3,visual geometry group 19(VGG19),and residual neural network 50(ResNet50).The classification models also employ a modified version of the above three CNN architectures.Using the RIM-ONE dataset,the attention U-Net with the ResNet50 model as the encoder backbone,achieved the best accuracy of 99.58%in segmenting OD.The Inception-v3 model had the highest accuracy of 98.79%for glaucoma classification among the evaluated segmentation,followed by the modified classification architectures.展开更多
基金supported by the National Natural Science Foundation of China(No.62027801).
文摘This paper presents a new method for specific emitter identification(SEI)using the reparameterization visual geometry group(RepVGG)neural network model and Gramian angular summation field(GASF).It converts in-phase and quadrature(IQ)signals into 2D feature maps,retaining both time and frequency domain features.Compared to residual network 18-layer(ResNet18)and Hilbert transform methods,this approach offers higher accuracy,faster training,and a smaller model size,making it ideal for hardware deployment.
文摘农田害虫降低了农作物的产量和质量,如何有效区分和治理农田害虫成为首要解决的问题。文章紧抓农田环境需求和农民对农作物的产量需求不匹配的痛点,基于卷积神经网络技术识别农田害虫,为农业提供有效的识别方式。采用MobileNetV1、残差神经网络(Residual Network,ResNet)50、视觉几何群网络(Visual Geometry Group Network,VGG)16以及微调预训练模型VGG16共4种网络模型二分类农田害虫图片集。由于样本数据量较少,为防止出现过拟合,使用了数据增强技术,即通过现有训练图片生成更多的训练图片,从而提高泛化能力。实验表明,4种网络模型的准确率分别为88.63%、91.73%、86.49%和90.13%,在农田害虫识别中均具有较好的实际应用效果。
基金funded by Princess Nourah bint Abdulrahman University Researchers Supporting Project number(PNURSP2025R138),Princess Nourah bint Abdulrahman University,Riyadh,Saudi Arabia.
文摘The contemporary era is characterized by rapid technological advancements,particularly in the fields of communication and multimedia.Digital media has significantly influenced the daily lives of individuals of all ages.One of the emerging domains in digital media is the creation of cartoons and animated videos.The accessibility of the internet has led to a surge in the consumption of cartoons among young children,presenting challenges in monitoring and controlling the content they view.The prevalence of cartoon videos containing potentially violent scenes has raised concerns regarding their impact,especially on young and impressionableminds.This article contributes to the growing concerns about the impact of animated media on children’s mental health and offers solutions to help mitigate these effects.To address this issue,an intelligent,multi-CNN fusion framework is proposed for detecting and predicting violent content in upcoming frames of animated videos.The framework integrates probabilistic and deep learning methodologies by leveraging a combination of visual and temporal features for violence prediction in future scenes.Two specific convolutional neural network classifiers i.e.,VGG16 and ResNet18 are employed to classify scenes from animated content as violent or non-violent.To enhance decision robustness,this study introduces a fusion strategy based on weighted averaging,combining the outputs of both Convolutional Neural Networks(CNNs)into a single decision stream.The resulting classifications are subsequently fed into Naive Bayes classifier,which analyzes sequential patterns to forecast violence in future scenes.The experimental findings demonstrate that the proposed framework achieved predictive accuracy of 92.84%,highlighting its effectiveness for intelligent content moderation.These results underscore the potential of intelligent data fusion techniques in enhancing the reliability and robustness of automated violence detection systems in animated content.This framework offers a promising solution for safeguarding young audiences by enabling proactive and accurate moderation of animated videos.
文摘Glaucoma is a prevalent cause of blindness worldwide.If not treated promptly,it can cause vision and quality of life to deteriorate.According to statistics,glaucoma affects approximately 65 million individuals globally.Fundus image segmentation depends on the optic disc(OD)and optic cup(OC).This paper proposes a computational model to segment and classify retinal fundus images for glaucoma detection.Different data augmentation techniques were applied to prevent overfitting while employing several data pre-processing approaches to improve the image quality and achieve high accuracy.The segmentation models are based on an attention U-Net with three separate convolutional neural networks(CNNs)backbones:Inception-v3,visual geometry group 19(VGG19),and residual neural network 50(ResNet50).The classification models also employ a modified version of the above three CNN architectures.Using the RIM-ONE dataset,the attention U-Net with the ResNet50 model as the encoder backbone,achieved the best accuracy of 99.58%in segmenting OD.The Inception-v3 model had the highest accuracy of 98.79%for glaucoma classification among the evaluated segmentation,followed by the modified classification architectures.