期刊文献+
共找到39篇文章
< 1 2 >
每页显示 20 50 100
CLF-YOLOv8:Lightweight Multi-Scale Fusion with Focal Geometric Loss for Real-Time Night Maritime Detection
1
作者 Zhonghao Wang Xin Liu +1 位作者 Changhua Yue Haiwen Yuan 《Computers, Materials & Continua》 2026年第2期1667-1689,共23页
To address critical challenges in nighttime ship detection—high small-target missed detection(over 20%),insufficient lightweighting,and limited generalization due to scarce,low-quality datasets—this study proposes a... To address critical challenges in nighttime ship detection—high small-target missed detection(over 20%),insufficient lightweighting,and limited generalization due to scarce,low-quality datasets—this study proposes a systematic solution.First,a high-quality Night-Ships dataset is constructed via CycleGAN-based day-night transfer,combined with a dual-threshold cleaning strategy(Laplacian variance sharpness filtering and brightness-color deviation screening).Second,a Cross-stage Lightweight Fusion-You Only Look Once version 8(CLF-YOLOv8)is proposed with key improvements:the Neck network is reconstructed by replacing Cross Stage Partial(CSP)structure with the Cross Stage Partial Multi-Scale Convolutional Block(CSP-MSCB)and integrating Bidirectional Feature Pyramid Network(BiFPN)for weighted multi-scale fusion to enhance small-target detection;a Lightweight Shared Convolutional and Separated Batch Normalization Detection-Head(LSCSBD-Head)with shared convolutions and layer-wise Batch Normalization(BN)reduces parameters to 1.8M(42% fewer than YOLOv8n);and the FocalMinimum Point Distance Intersection over Union(Focal-MPDIoU)loss combines Minimum Point Distance Intersection over Union(MPDIoU)geometric constraints and Focal weighting to optimize low-overlap targets.Experiments show CLFYOLOv8 achieves 97.6%mAP@0.5(0.7% higher than YOLOv8n)with 1.8 M parameters,outperforming mainstream models in small-target detection,overlapping target discrimination,and adaptability to complex lighting. 展开更多
关键词 Nighttime ship detection lightweight model small object detection BiFPN LSCSBD-Head Focal-MPDIoU YOLOv8
在线阅读 下载PDF
A Knowledge-Distilled CharacterBERT-BiLSTM-ATT Framework for Lightweight DGA Detection in IoT Devices
2
作者 Chengqi Liu YongtaoLi +1 位作者 Weiping Zou Deyu Lin 《Computers, Materials & Continua》 2026年第4期2049-2068,共20页
With the large-scale deployment of the Internet ofThings(IoT)devices,their weak securitymechanisms make them prime targets for malware attacks.Attackers often use Domain Generation Algorithm(DGA)to generate random dom... With the large-scale deployment of the Internet ofThings(IoT)devices,their weak securitymechanisms make them prime targets for malware attacks.Attackers often use Domain Generation Algorithm(DGA)to generate random domain names,hiding the real IP of Command and Control(C&C)servers to build botnets.Due to the randomness and dynamics of DGA,traditional methods struggle to detect them accurately,increasing the difficulty of network defense.This paper proposes a lightweight DGA detection model based on knowledge distillation for resource-constrained IoT environments.Specifically,a teacher model combining CharacterBERT,a bidirectional long short-term memory(BiLSTM)network,and attention mechanism(ATT)is constructed:it extracts character-level semantic features viaCharacterBERT,captures sequence dependencieswith the BiLSTM,and integrates theATT for key feature weighting,formingmulti-granularity feature fusion.An improved knowledge distillation approach transfers the teacher model’s learned knowledge to the simplified DistilBERT student model.Experimental results show the teacher model achieves 98.68%detection accuracy.The student modelmaintains slightly improved accuracy while significantly compressing parameters to approximately 38.4%of the teacher model’s scale,greatly reducing computational overhead for IoT deployment. 展开更多
关键词 IoT security DGA detection knowledge distillation lightweight model edge computing
在线阅读 下载PDF
A lightweight pure visual BEV perception method based on dual distillation of spatial-temporal knowledge
3
作者 LIU Bingdong YU Ruihang +1 位作者 XIONG Zhiming WU Meiping 《Journal of Systems Engineering and Electronics》 2026年第1期36-44,共9页
Bird's-eye-view(BEV)perception is a core technology for autonomous driving systems.However,existing solutions face the dilemma of high costs associated with multimodal methods and limited performance of vision-onl... Bird's-eye-view(BEV)perception is a core technology for autonomous driving systems.However,existing solutions face the dilemma of high costs associated with multimodal methods and limited performance of vision-only approaches.To address this issue,this paper proposes a framework named“a lightweight pure visual BEV perception method based on dual distillation of spatial-temporal knowledge”.This framework innovatively designs a lightweight vision-only student model based on Res Net,which leverages a dual distillation mechanism to learn from a powerful teacher model that integrates temporal information from both image and light detection and ranging(LiDAR)modalities.Specifically,we distill efficient multi-modal feature extraction and spatial fusion capabilities from the BEVFusion model,and distill advanced temporal information fusion and spatiotemporal attention mechanisms from the BEVFormer model.This dual distillation strategy enables the student model to achieve perception performance close to that of multi-modal models without relying on Li DAR.Experimental results on the nu Scenes dataset demonstrate that the proposed model significantly outperforms classical vision-only algorithms,achieves comparable performance to current state-of-the-art vision-only methods on the nu Scenes detection leaderboard in terms of both mean average precision(mAP)and the nu Scenes detection score(NDS)metrics,and exhibits notable advantages in inference computational efficiency.Although the proposed dual-teacher paradigm incurs higher offline training costs compared to single-model approaches,it yields a streamlined and highly efficient student model suitable for resource-constrained real-time deployment.This provides an effective pathway toward low-cost,high-performance autonomous driving perception systems. 展开更多
关键词 3D object detection bird's-eye-view(BEV) knowledge distillation multimodal fusion lightweight model
在线阅读 下载PDF
Lightweight Complex-Valued Neural Network for Indoor Positioning
4
作者 Le Wang Bing Xu +1 位作者 Peng Liu En Yuan 《Computers, Materials & Continua》 2026年第2期1770-1783,共14页
Deep learning has been recognized as an effective method for indoor positioning.However,most existing real-valued neural networks(RVNNs)treat the two constituent components of complex-valued channel state information(... Deep learning has been recognized as an effective method for indoor positioning.However,most existing real-valued neural networks(RVNNs)treat the two constituent components of complex-valued channel state information(CSI)as real-valued inputs,potentially discarding useful information embedded in the original CSI.In addition,existing positioning models generally face the contradiction between computational complexity and positioning accuracy.To address these issues,we combine graph neural network(GNN)with complex-valued neural network(CVNN)to construct a lightweight indoor positioning model named CGNet.CGNet employs complexvalued convolution operation to directly process the original CSI data,fully exploiting the correlation between real and imaginary parts of CSI while extracting local features.Subsequently,the feature values are treated as nodes,and conditional position encoding(CPE)module is applied to add positional information.To reduce the number of connections in the graph structure and lower themodel complexity,feature information is mapped to an efficient graph structure through a dynamic axial graph construction(DAGC)method,with global features extracted usingmaximum relative graph convolution(MRConv).Experimental results show that,on the CTW dataset,CGNet achieves a 10%improvement in positioning accuracy compared to existing methods,while the number of model parameters is only 0.8 M.CGNet achieves excellent positioning accuracy with very few parameters. 展开更多
关键词 Indoor positioning complex-valued neural network channel state information lightweight model
在线阅读 下载PDF
Enhancing Lightweight Mango Disease Detection Model Performance through a Combined Attention Module
5
作者 Wen-Tsai Sung Indra Griha TofikIsa Sung-Jung Hsiao 《Computers, Materials & Continua》 2026年第2期986-1016,共31页
Mango is a plant with high economic value in the agricultural industry;thus,it is necessary to maximize the productivity performance of the mango plant,which can be done by implementing artificial intelligence.In this... Mango is a plant with high economic value in the agricultural industry;thus,it is necessary to maximize the productivity performance of the mango plant,which can be done by implementing artificial intelligence.In this study,a lightweight object detection model will be developed that can detect mango plant conditions based on disease potential,so that it becomes an early detection warning system that has an impact on increasing agricultural productivity.The proposed lightweight model integrates YOLOv7-Tiny and the proposed modules,namely the C2S module.The C2S module consists of three sub-modules such as the convolutional block attention module(CBAM),the coordinate attention(CA)module,and the squeeze-and-excitation(SE)module.The dataset is constructed by eight classes,including seven classes of disease conditions and one class of health conditions.The experimental result shows that the proposed lightweight model has the optimal results,which increase by 13.15% of mAP50 compared to the original model YOLOv7-Tiny.While the mAP50:95 also achieved the highest results compared to other models,including YOLOv3-Tiny,YOLOv4-Tiny,YOLOv5,and YOLOv7-Tiny.The advantage of the proposed lightweightmodel is the adaptability that supports it in constrained environments,such as edge computing systems.This proposedmodel can support a robust,precise,and convenient precision agriculture system for the user. 展开更多
关键词 Mango lightweight model combined attention module C2S module precision agriculture
在线阅读 下载PDF
Lightweight Small Defect Detection with YOLOv8 Using Cascaded Multi-Receptive Fields and Enhanced Detection Heads
6
作者 Shengran Zhao Zhensong Li +2 位作者 Xiaotan Wei Yutong Wang Kai Zhao 《Computers, Materials & Continua》 2026年第1期1278-1291,共14页
In printed circuit board(PCB)manufacturing,surface defects can significantly affect product quality.To address the performance degradation,high false detection rates,and missed detections caused by complex backgrounds... In printed circuit board(PCB)manufacturing,surface defects can significantly affect product quality.To address the performance degradation,high false detection rates,and missed detections caused by complex backgrounds in current intelligent inspection algorithms,this paper proposes CG-YOLOv8,a lightweight and improved model based on YOLOv8n for PCB surface defect detection.The proposed method optimizes the network architecture and compresses parameters to reduce model complexity while maintaining high detection accuracy,thereby enhancing the capability of identifying diverse defects under complex conditions.Specifically,a cascaded multi-receptive field(CMRF)module is adopted to replace the SPPF module in the backbone to improve feature perception,and an inverted residual mobile block(IRMB)is integrated into the C2f module to further enhance performance.Additionally,conventional convolution layers are replaced with GSConv to reduce computational cost,and a lightweight Convolutional Block Attention Module based Convolution(CBAMConv)module is introduced after Grouped Spatial Convolution(GSConv)to preserve accuracy through attention mechanisms.The detection head is also optimized by removing medium and large-scale detection layers,thereby enhancing the model’s ability to detect small-scale defects and further reducing complexity.Experimental results show that,compared to the original YOLOv8n,the proposed CG-YOLOv8 reduces parameter count by 53.9%,improves mAP@0.5 by 2.2%,and increases precision and recall by 2.0%and 1.8%,respectively.These improvements demonstrate that CG-YOLOv8 offers an efficient and lightweight solution for PCB surface defect detection. 展开更多
关键词 YOLOv8n PCB surface defect detection lightweight model small object detection
在线阅读 下载PDF
TELL-Me:A time-series-decomposition-based ensembled lightweight learning model for diverse battery prognosis and diagnosis 被引量:1
7
作者 Kun-Yu Liu Ting-Ting Wang +2 位作者 Bo-Bo Zou Hong-Jie Peng Xinyan Liu 《Journal of Energy Chemistry》 2025年第7期1-8,共8页
As batteries become increasingly essential for energy storage technologies,battery prognosis,and diagnosis remain central to ensure reliable operation and effective management,as well as to aid the in-depth investigat... As batteries become increasingly essential for energy storage technologies,battery prognosis,and diagnosis remain central to ensure reliable operation and effective management,as well as to aid the in-depth investigation of degradation mechanisms.However,dynamic operating conditions,cell-to-cell inconsistencies,and limited availability of labeled data have posed significant challenges to accurate and robust prognosis and diagnosis.Herein,we introduce a time-series-decomposition-based ensembled lightweight learning model(TELL-Me),which employs a synergistic dual-module framework to facilitate accurate and reliable forecasting.The feature module formulates features with physical implications and sheds light on battery aging mechanisms,while the gradient module monitors capacity degradation rates and captures aging trend.TELL-Me achieves high accuracy in end-of-life prediction using minimal historical data from a single battery without requiring offline training dataset,and demonstrates impressive generality and robustness across various operating conditions and battery types.Additionally,by correlating feature contributions with degradation mechanisms across different datasets,TELL-Me is endowed with the diagnostic ability that not only enhances prediction reliability but also provides critical insights into the design and optimization of next-generation batteries. 展开更多
关键词 Battery prognosis Interpretable machine learning Degradation diagnosis Ensemble learning Online prediction lightweight model
在线阅读 下载PDF
Improved lightweight road damage detection based on YOLOv5
8
作者 LIU Chang SUN Yu +2 位作者 CHEN Jin YANG Jing WANG Fengchao 《Optoelectronics Letters》 2025年第5期314-320,共7页
There is a problem of real-time detection difficulty in road surface damage detection. This paper proposes an improved lightweight model based on you only look once version 5(YOLOv5). Firstly, this paper fully utilize... There is a problem of real-time detection difficulty in road surface damage detection. This paper proposes an improved lightweight model based on you only look once version 5(YOLOv5). Firstly, this paper fully utilized the convolutional neural network(CNN) + ghosting bottleneck(G_bneck) architecture to reduce redundant feature maps. Afterwards, we upgraded the original upsampling algorithm to content-aware reassembly of features(CARAFE) and increased the receptive field. Finally, we replaced the spatial pyramid pooling fast(SPPF) module with the basic receptive field block(Basic RFB) pooling module and added dilated convolution. After comparative experiments, we can see that the number of parameters and model size of the improved algorithm in this paper have been reduced by nearly half compared to the YOLOv5s. The frame rate per second(FPS) has been increased by 3.25 times. The mean average precision(m AP@0.5: 0.95) has increased by 8%—17% compared to other lightweight algorithms. 展开更多
关键词 road surface damage detection convolutional neural network feature maps convolutional neural network cnn lightweight model yolov improved lightweight model spatial pyram
原文传递
Lightweight Residual Multi-Head Convolution with Channel Attention(ResMHCNN)for End-to-End Classification of Medical Images
9
作者 Sudhakar Tummala Sajjad Hussain Chauhdary +3 位作者 Vikash Singh Roshan Kumar Seifedine Kadry Jungeun Kim 《Computer Modeling in Engineering & Sciences》 2025年第9期3585-3605,共21页
Lightweight deep learning models are increasingly required in resource-constrained environments such as mobile devices and the Internet of Medical Things(IoMT).Multi-head convolution with channel attention can facilit... Lightweight deep learning models are increasingly required in resource-constrained environments such as mobile devices and the Internet of Medical Things(IoMT).Multi-head convolution with channel attention can facilitate learning activations relevant to different kernel sizes within a multi-head convolutional layer.Therefore,this study investigates the capability of novel lightweight models incorporating residual multi-head convolution with channel attention(ResMHCNN)blocks to classify medical images.We introduced three novel lightweight deep learning models(BT-Net,LCC-Net,and BC-Net)utilizing the ResMHCNN block as their backbone.These models were crossvalidated and tested on three publicly available medical image datasets:a brain tumor dataset from Figshare consisting of T1-weighted magnetic resonance imaging slices of meningioma,glioma,and pituitary tumors;the LC25000 dataset,which includes microscopic images of lung and colon cancers;and the BreaKHis dataset,containing benign and malignant breast microscopic images.The lightweight models achieved accuracies of 96.9%for 3-class brain tumor classification using BT-Net,and 99.7%for 5-class lung and colon cancer classification using LCC-Net.For 2-class breast cancer classification,BC-Net achieved an accuracy of 96.7%.The parameter counts for the proposed lightweight models—LCC-Net,BC-Net,and BT-Net—are 0.528,0.226,and 1.154 million,respectively.The presented lightweight models,featuring ResMHCNN blocks,may be effectively employed for accurate medical image classification.In the future,these models might be tested for viability in resource-constrained systems such as mobile devices and IoMT platforms. 展开更多
关键词 lightweight models brain tumor breast cancer lung cancer colon cancer multi-head CNN
在线阅读 下载PDF
MSAMamba-UNet:A Lightweight Multi-Scale Adaptive Mamba Network for Skin Lesion Segmentation
10
作者 Shouming Hou Jianchao Hou +2 位作者 Yuteng Pang Aoyu Xia Beibei Hou 《Journal of Bionic Engineering》 2025年第6期3209-3225,共17页
Segmenting skin lesions is critical for early skin cancer detection.Existing CNN and Transformer-based methods face challenges such as high computational complexity and limited adaptability to variations in lesion siz... Segmenting skin lesions is critical for early skin cancer detection.Existing CNN and Transformer-based methods face challenges such as high computational complexity and limited adaptability to variations in lesion sizes.To overcome these limitations,we introduce MSAMamba-UNet,a lightweight model that integrates two novel architectures:Multi-Scale Mamba(MSMamba)and Adaptive Dynamic Gating Block(ADGB).MSMamba utilizes multi-scale decomposition and a parallel hierarchical structure to enhance the delineation of irregular lesion boundaries and sensitivity to small targets.ADGB dynamically selects convolutional kernels with varying receptive fields based on input features,improving the model’s capacity to accommodate diverse lesion textures and scales.Additionally,we introduce a Mix Attention Fusion Block(MAF)to enhance shallow feature representation by integrating parallel channel and pixel attention mechanisms.Extensive evaluation of MSAMamba-UNet on the ISIC 2016,ISIC 2017,and ISIC 2018 datasets demonstrates competitive segmentation accuracy with only 0.056 M parameters and 0.069 GFLOPs.Our experiments revealed that MSAMamba-UNet achieved IoU scores of 85.53%,85.47%,and 82.22%,as well as DSC scores of 92.20%,92.17%,and 90.24%,respectively.These results underscore the lightweight design and effectiveness of MSAMamba-UNet. 展开更多
关键词 TRANSFORMER Segmenting skin lesions Mamba lightweight model MULTI-SCALE
在线阅读 下载PDF
KN-YOLOv8:A Lightweight Deep Learning Model for Real-Time Coffee Bean Defect Detection
11
作者 Tesfaye Adisu Tarekegn Taye Girma Debelee 《Journal on Artificial Intelligence》 2025年第1期585-613,共29页
The identification of defect types and their reduction values is the most crucial step in coffee grading.In Ethiopia,the current coffee defect investigation techniques rely on manual screening,which requires substanti... The identification of defect types and their reduction values is the most crucial step in coffee grading.In Ethiopia,the current coffee defect investigation techniques rely on manual screening,which requires substantial human resources,time-consuming,and prone to errors.Recently,the deep learning driven object detection has shown promising results in coffee defect identification and grading tasks.In this study,we propose KN-YOLOv8,a modified You Only Look Once version-8(YOLOv8)model optimized for real-time detection of coffee bean defects.This lightweight network incorporates effective feature fusion techniques to accurately detect and locate defects,even among overlapping beans.We have compiled a custom dataset of 562 images comprising thirteen distinct types of defects.The model achieved exceptional performance,with training dataset metrics of 97% recall,100% precision,and 98% mean average precision(mAP).On the test dataset,it maintained outstanding results with 99% recall,100% precision,and 98.9% mAP.The model outperforms existing approaches by achieving a 97.7%m AP for all classes at a 0.5 threshold,while maintaining an optimal precision-recall balance.The model outperforms new approaches by achieving a balance between precision and recall,achieving a mean average precision of 97.7% for all classes.This solution significantly reduces reliance on labor-intensivemanual inspection while improving accuracy.Its lightweight design and high speed make it suitable for real-time industrial applications,transforming coffee quality inspection. 展开更多
关键词 KN-YOLOv8 coffee-bean lightweight model defect detection optimization
在线阅读 下载PDF
Mineral identification in thin sections using a lightweight and attention mechanism
12
作者 Xin Zhang Wei Dang +4 位作者 Jun Liu Zijuan Yin Guichao Du Yawen He Yankai Xue 《Natural Gas Industry B》 2025年第2期135-146,共12页
Mineral identification is foundational to geological survey research,mineral resource exploration,and mining engineering.Considering the diversity of mineral types and the challenge of achieving high recognition accur... Mineral identification is foundational to geological survey research,mineral resource exploration,and mining engineering.Considering the diversity of mineral types and the challenge of achieving high recognition accuracy for similar features,this study introduces a mineral detection method based on YOLOv8-SBI.This work enhances feature extraction by integrating spatial pyramid pooling-fast(SPPF)with the simplified self-attention module(SimAM),significantly improving the precision of mineral feature detection.In the feature fusion network,a weighted bidirectional feature pyramid network is employed for advanced cross-channel feature integration,effectively reducing feature redundancy.Additionally,Inner-Intersection Over Union(InnerIOU)is used as the loss function to improve the average quality localization performance of anchor boxes.Experimental results show that the YOLOv8-SBI model achieves an accuracy of 67.9%,a recall of 74.3%,a mAP@0.5 of 75.8%,and a mAP@0.5:0.95 of 56.7%,with a real-time detection speed of 244.2 frames per second.Compared to YOLOv8,YOLOv8-SBI demonstrates a significant improvement with 15.4%increase in accuracy,28.5%increase in recall,and increases of 28.1%and 20.9%in mAP@0.5 and mAP@0.5:0.95,respectively.Furthermore,relative to other models,such as YOLOv3,YOLOv5,YOLOv6,YOLOv8,YOLOv9,and YOLOv10,YOLOv8-SBI has a smaller parameter size of only 3.01×10^(6).This highlights the optimal balance between detection accuracy and speed,thereby offering robust technical support for intelligent mineral classification. 展开更多
关键词 Deep learning Neural networks lightweight models Attention mechanisms Mineral identification
在线阅读 下载PDF
Efficient and lightweight 3D building reconstruction from drone imagery using sparse line and point clouds
13
作者 Xiongjie YIN Jinquan HE Zhanglin CHENG 《虚拟现实与智能硬件(中英文)》 2025年第2期111-126,共16页
Efficient three-dimensional(3D)building reconstruction from drone imagery often faces data acquisition,storage,and computational challenges because of its reliance on dense point clouds.In this study,we introduced a n... Efficient three-dimensional(3D)building reconstruction from drone imagery often faces data acquisition,storage,and computational challenges because of its reliance on dense point clouds.In this study,we introduced a novel method for efficient and lightweight 3D building reconstruction from drone imagery using line clouds and sparse point clouds.Our approach eliminates the need to generate dense point clouds,and thus significantly reduces the computational burden by reconstructing 3D models directly from sparse data.We addressed the limitations of line clouds for plane detection and reconstruction by using a new algorithm.This algorithm projects 3D line clouds onto a 2D plane,clusters the projections to identify potential planes,and refines them using sparse point clouds to ensure an accurate and efficient model reconstruction.Extensive qualitative and quantitative experiments demonstrated the effectiveness of our method,demonstrating its superiority over existing techniques in terms of simplicity and efficiency. 展开更多
关键词 3D reconstruction Line clouds Sparse clouds lightweight models
在线阅读 下载PDF
AW-HRNet:A Lightweight High-Resolution Crack Segmentation Network Integrating Spatial Robustness and Frequency-Domain Enhancement
14
作者 Dewang Ma Tong Lu 《Journal of Electronic Research and Application》 2025年第6期7-17,共11页
The study presents AW-HRNet,a lightweight high-resolution crack segmentation network that couples Adaptive residual enhancement(AREM)in the spatial domain with Wavelet-based decomposition-reconstruction(WDRM)in the fr... The study presents AW-HRNet,a lightweight high-resolution crack segmentation network that couples Adaptive residual enhancement(AREM)in the spatial domain with Wavelet-based decomposition-reconstruction(WDRM)in the frequency domain.AREM introduces a learnable channel-wise scaling after standard 3×3 convolution and merges it through a residual path to stabilize crack-sensitive responses while suppressing noise.WDRM performs DWT to decouple LL/LH/HL/HH sub-bands,conducts lightweight cross-band fusion,and applies IDWT to restore detail-enhanced features,unifying global topology and boundary sharpness without deformable offsets.Integrated into a high-resolution backbone with auxiliary deep supervision,AW-HRNet attains 79.07%mIoU on CrackSeg9k with only 1.24M parameters and 0.73 GFLOPs,offering an excellent accuracy-efficiency trade-off and strong robustness for real-world deployment. 展开更多
关键词 Crack segmentation lightweight model Wavelet decomposition and reconstruction Feature enhancement
在线阅读 下载PDF
Lightweight Classroom Student Action Recognition Method Based on Spatiotemporal Multimodal Feature Fusion
15
作者 Shaodong Zou Di Wu +2 位作者 Jianhou Gan Juxiang Zhou Jiatian Mei 《Computers, Materials & Continua》 2025年第4期1101-1116,共16页
The task of student action recognition in the classroom is to precisely capture and analyze the actions of students in classroom videos,providing a foundation for realizing intelligent and accurate teaching.However,th... The task of student action recognition in the classroom is to precisely capture and analyze the actions of students in classroom videos,providing a foundation for realizing intelligent and accurate teaching.However,the complex nature of the classroom environment has added challenges and difficulties in the process of student action recognition.In this research article,with regard to the circumstances where students are prone to be occluded and classroom computing resources are restricted in real classroom scenarios,a lightweight multi-modal fusion action recognition approach is put forward.This proposed method is capable of enhancing the accuracy of student action recognition while concurrently diminishing the number of parameters of the model and the Computation Amount,thereby achieving a more efficient and accurate recognition performance.In the feature extraction stage,this method fuses the keypoint heatmap with the RGB(Red-Green-Blue color model)image.In order to fully utilize the unique information of different modalities for feature complementarity,a Feature Fusion Module(FFE)is introduced.The FFE encodes and fuses the unique features of the two modalities during the feature extraction process.This fusion strategy not only achieves fusion and complementarity between modalities,but also improves the overall model performance.Furthermore,to reduce the computational load and parameter scale of the model,we use keypoint information to crop RGB images.At the same time,the first three networks of the lightweight feature extraction network X3D are used to extract dual-branch features.These methods significantly reduce the computational load and parameter scale.The number of parameters of the model is 1.40 million,and the computation amount is 5.04 billion floating-point operations per second(GFLOPs),achieving an efficient lightweight design.In the Student Classroom Action Dataset(SCAD),the accuracy of the model is 88.36%.In NTU 60(Nanyang Technological University Red-Green-Blue-Depth RGB+Ddataset with 60 categories),the accuracies on X-Sub(The people in the training set are different from those in the test set)and X-View(The perspectives of the training set and the test set are different)are 95.76%and 98.82%,respectively.On the NTU 120 dataset(Nanyang Technological University Red-Green-Blue-Depth dataset with 120 categories),RGB+Dthe accuracies on X-Sub and X-Set(the perspectives of the training set and the test set are different)are 91.97%and 93.45%,respectively.The model has achieved a balance in terms of accuracy,computation amount,and the number of parameters. 展开更多
关键词 Action recognition student classroom action multimodal fusion lightweight model design
在线阅读 下载PDF
YOLO-SPDNet:Multi-Scale Sequence and Attention-Based Tomato Leaf Disease Detection Model
16
作者 Meng Wang Jinghan Cai +6 位作者 Wenzheng Liu Xue Yang Jingjing Zhang Qiangmin Zhou Fanzhen Wang Hang Zhang Tonghai Liu 《Phyton-International Journal of Experimental Botany》 2026年第1期290-308,共19页
Tomato is a major economic crop worldwide,and diseases on tomato leaves can significantly reduce both yield and quality.Traditional manual inspection is inefficient and highly subjective,making it difficult to meet th... Tomato is a major economic crop worldwide,and diseases on tomato leaves can significantly reduce both yield and quality.Traditional manual inspection is inefficient and highly subjective,making it difficult to meet the requirements of early disease identification in complex natural environments.To address this issue,this study proposes an improved YOLO11-based model,YOLO-SPDNet(Scale Sequence Fusion,Position-Channel Attention,and Dual Enhancement Network).The model integrates the SEAM(Self-Ensembling Attention Mechanism)semantic enhancement module,the MLCA(Mixed Local Channel Attention)lightweight attention mechanism,and the SPA(Scale-Position-Detail Awareness)module composed of SSFF(Scale Sequence Feature Fusion),TFE(Triple Feature Encoding),and CPAM(Channel and Position Attention Mechanism).These enhancements strengthen fine-grained lesion detection while maintaining model lightweightness.Experimental results show that YOLO-SPDNet achieves an accuracy of 91.8%,a recall of 86.5%,and an mAP@0.5 of 90.6%on the test set,with a computational complexity of 12.5 GFLOPs.Furthermore,the model reaches a real-time inference speed of 987 FPS,making it suitable for deployment on mobile agricultural terminals and online monitoring systems.Comparative analysis and ablation studies further validate the reliability and practical applicability of the proposed model in complex natural scenes. 展开更多
关键词 Tomato disease detection YOLO multi-scale feature fusion attention mechanism lightweight model
在线阅读 下载PDF
Ultra-lightweight CNN design based on neural architecture search and knowledge distillation: A novel method to build the automatic recognition model of space target ISAR images 被引量:7
17
作者 Hong Yang Ya-sheng Zhang +1 位作者 Can-bin Yin Wen-zhe Ding 《Defence Technology(防务技术)》 SCIE EI CAS CSCD 2022年第6期1073-1095,共23页
In this paper,a novel method of ultra-lightweight convolution neural network(CNN)design based on neural architecture search(NAS)and knowledge distillation(KD)is proposed.It can realize the automatic construction of th... In this paper,a novel method of ultra-lightweight convolution neural network(CNN)design based on neural architecture search(NAS)and knowledge distillation(KD)is proposed.It can realize the automatic construction of the space target inverse synthetic aperture radar(ISAR)image recognition model with ultra-lightweight and high accuracy.This method introduces the NAS method into the radar image recognition for the first time,which solves the time-consuming and labor-consuming problems in the artificial design of the space target ISAR image automatic recognition model(STIIARM).On this basis,the NAS model’s knowledge is transferred to the student model with lower computational complexity by the flow of the solution procedure(FSP)distillation method.Thus,the decline of recognition accuracy caused by the direct compression of model structural parameters can be effectively avoided,and the ultralightweight STIIARM can be obtained.In the method,the Inverted Linear Bottleneck(ILB)and Inverted Residual Block(IRB)are firstly taken as each block’s basic structure in CNN.And the expansion ratio,output filter size,number of IRBs,and convolution kernel size are set as the search parameters to construct a hierarchical decomposition search space.Then,the recognition accuracy and computational complexity are taken as the objective function and constraint conditions,respectively,and the global optimization model of the CNN architecture search is established.Next,the simulated annealing(SA)algorithm is used as the search strategy to search out the lightweight and high accuracy STIIARM directly.After that,based on the three principles of similar block structure,the same corresponding channel number,and the minimum computational complexity,the more lightweight student model is designed,and the FSP matrix pairing between the NAS model and student model is completed.Finally,by minimizing the loss between the FSP matrix pairs of the NAS model and student model,the student model’s weight adjustment is completed.Thus the ultra-lightweight and high accuracy STIIARM is obtained.The proposed method’s effectiveness is verified by the simulation experiments on the ISAR image dataset of five types of space targets. 展开更多
关键词 Space target ISAR image Neural architecture search Knowledge distillation lightweight model
在线阅读 下载PDF
Identification of Weather Phenomena Based on Lightweight Convolutional Neural Networks 被引量:3
18
作者 Congcong Wang Pengyu Liu +2 位作者 Kebin Jia Xiaowei Jia Yaoyao Li 《Computers, Materials & Continua》 SCIE EI 2020年第9期2043-2055,共13页
Weather phenomenon recognition plays an important role in the field of meteorology.Nowadays,weather radars and weathers sensor have been widely used for weather recognition.However,given the high cost in deploying and... Weather phenomenon recognition plays an important role in the field of meteorology.Nowadays,weather radars and weathers sensor have been widely used for weather recognition.However,given the high cost in deploying and maintaining the devices,it is difficult to apply them to intensive weather phenomenon recognition.Moreover,advanced machine learning models such as Convolutional Neural Networks(CNNs)have shown a lot of promise in meteorology,but these models also require intensive computation and large memory,which make it difficult to use them in reality.In practice,lightweight models are often used to solve such problems.However,lightweight models often result in significant performance losses.To this end,after taking a deep dive into a large number of lightweight models and summarizing their shortcomings,we propose a novel lightweight CNNs model which is constructed based on new building blocks.The experimental results show that the model proposed in this paper has comparable performance with the mainstream non-lightweight model while also saving 25 times of memory consumption.Such memory reduction is even better than that of existing lightweight models. 展开更多
关键词 Deep learning convolution neural networks lightweight models weather identification
在线阅读 下载PDF
Unstructured Road Extraction in UAV Images based on Lightweight Model 被引量:1
19
作者 Di Zhang Qichao An +3 位作者 Xiaoxue Feng Ronghua Liu Jun Han Feng Pan 《Chinese Journal of Mechanical Engineering》 SCIE EI CAS CSCD 2024年第2期372-384,共13页
There is no unified planning standard for unstructured roads,and the morphological structures of these roads are complex and varied.It is important to maintain a balance between accuracy and speed for unstructured roa... There is no unified planning standard for unstructured roads,and the morphological structures of these roads are complex and varied.It is important to maintain a balance between accuracy and speed for unstructured road extraction models.Unstructured road extraction algorithms based on deep learning have problems such as high model complexity,high computational cost,and the inability to adapt to current edge computing devices.Therefore,it is best to use lightweight network models.Considering the need for lightweight models and the characteristics of unstructured roads with different pattern shapes,such as blocks and strips,a TMB(Triple Multi-Block)feature extraction module is proposed,and the overall structure of the TMBNet network is described.The TMB module was compared with SS-nbt,Non-bottleneck-1D,and other modules via experiments.The feasibility and effectiveness of the TMB module design were proven through experiments and visualizations.The comparison experiment,using multiple convolution kernel categories,proved that the TMB module can improve the segmentation accuracy of the network.The comparison with different semantic segmentation networks demonstrates that the TMBNet network has advantages in terms of unstructured road extraction. 展开更多
关键词 Unstructured road lightweight model Triple Multi-Block(TMB) Semantic segmentation net
在线阅读 下载PDF
Tree Detection Algorithm Based on Embedded YOLO Lightweight Network
20
作者 吕峰 王新彦 +2 位作者 李磊 江泉 易政洋 《Journal of Shanghai Jiaotong university(Science)》 EI 2024年第3期518-527,共10页
To avoid colliding with trees during its operation,a lawn mower robot must detect the trees.Existing tree detection methods suffer from low detection accuracy(missed detection)and the lack of a lightweight model.In th... To avoid colliding with trees during its operation,a lawn mower robot must detect the trees.Existing tree detection methods suffer from low detection accuracy(missed detection)and the lack of a lightweight model.In this study,a dataset of trees was constructed on the basis of a real lawn environment.According to the theory of channel incremental depthwise convolution and residual suppression,the Embedded-A module is proposed,which expands the depth of the feature map twice to form a residual structure to improve the lightweight degree of the model.According to residual fusion theory,the Embedded-B module is proposed,which improves the accuracy of feature-map downsampling by depthwise convolution and pooling fusion.The Embedded YOLO object detection network is formed by stacking the embedded modules and the fusion of feature maps of different resolutions.Experimental results on the testing set show that the Embedded YOLO tree detection algorithm has 84.17%and 69.91%average precision values respectively for trunk and spherical tree,and 77.04% mean average precision value.The number of convolution parameters is 1.78×10^(6),and the calculation amount is 3.85 billion float operations per second.The size of weight file is 7.11MB,and the detection speed can reach 179 frame/s.This study provides a theoretical basis for the lightweight application of the object detection algorithm based on deep learning for lawn mower robots. 展开更多
关键词 Embedded YOLO algorithm lightweight model machine vision tree detection mowing robot
原文传递
上一页 1 2 下一页 到第
使用帮助 返回顶部