期刊文献+
共找到13篇文章
< 1 >
每页显示 20 50 100
Multiscale Feature Fusion for Gesture Recognition Using Commodity Millimeter-Wave Radar 被引量:1
1
作者 Lingsheng Li Weiqing Bai Chong Han 《Computers, Materials & Continua》 SCIE EI 2024年第10期1613-1640,共28页
Gestures are one of the most natural and intuitive approach for human-computer interaction.Compared with traditional camera-based or wearable sensors-based solutions,gesture recognition using the millimeter wave radar... Gestures are one of the most natural and intuitive approach for human-computer interaction.Compared with traditional camera-based or wearable sensors-based solutions,gesture recognition using the millimeter wave radar has attracted growing attention for its characteristics of contact-free,privacy-preserving and less environmentdependence.Although there have been many recent studies on hand gesture recognition,the existing hand gesture recognition methods still have recognition accuracy and generalization ability shortcomings in shortrange applications.In this paper,we present a hand gesture recognition method named multiscale feature fusion(MSFF)to accurately identify micro hand gestures.In MSFF,not only the overall action recognition of the palm but also the subtle movements of the fingers are taken into account.Specifically,we adopt hand gesture multiangle Doppler-time and gesture trajectory range-angle map multi-feature fusion to comprehensively extract hand gesture features and fuse high-level deep neural networks to make it pay more attention to subtle finger movements.We evaluate the proposed method using data collected from 10 users and our proposed solution achieves an average recognition accuracy of 99.7%.Extensive experiments on a public mmWave gesture dataset demonstrate the superior effectiveness of the proposed system. 展开更多
关键词 Gesture recognition millimeter-wave(mmWave)radar radio frequency(RF)sensing human-computer interaction multiscale feature fusion
在线阅读 下载PDF
Multiscale parallel feature aggregation network with attention fusion(MPFAN-AF):A novel approach to cataract disease classification
2
作者 Mohd Aquib Ansari Shahnawaz Ahmad Arvind Mewada 《Medical Data Mining》 2025年第4期17-28,共12页
Background:Early and accurate diagnosis of cataracts,which ranks among the leading preventable causes of blindness,is critical to securing positive outcomes for patients.Recently,eye image analyses have used deep lear... Background:Early and accurate diagnosis of cataracts,which ranks among the leading preventable causes of blindness,is critical to securing positive outcomes for patients.Recently,eye image analyses have used deep learning(DL)approaches to automate cataract classification more precisely,leading to the development of the Multiscale Parallel Feature Aggregation Network with Attention Fusion(MPFAN-AF).Focused on improving a model’s performance,this approach applies multiscale feature extraction,parallel feature fusion,along with attention-based fusion to sharpen its focus on salient features,which are crucial in detecting cataracts.Methods:Coarse-level features are captured through the application of convolutional layers,and these features undergo refinement through layered kernels of varying sizes.Moreover,this method captures all the diverse representations of cataracts accurately by parallel feature aggregation.Utilizing the Cataract Eye Dataset available on Kaggle,containing 612 labelled images of eyes with and without cataracts proportionately(normal vs.pathological),this model was trained and tested.Results:Results using the proposed model reflect greater precision over traditional convolutional neural networks(CNNs)models,achieving a classification accuracy of 97.52%.Additionally,the model demonstrated exceptional performance in classification tasks.The ablation studies validated that all applications added value to the prediction process,particularly emphasizing the attention fusion module.Conclusion:The MPFAN-AF model demonstrates high efficiency together with interpretability because it shows promise as an integration solution for real-time mobile cataract detection screening systems.Standard performance indicators indicate that AI-based ophthalmology tools have a promising future for use in remote conditions that lack medical resources. 展开更多
关键词 cataract classification deep learning multiscale feature extraction attention mechanism medical image analysis
在线阅读 下载PDF
Nonlinear frequency prediction and uncertainty analysis for fully clamped laminates by using a self-developed multi-scale neural networks system
3
作者 Yuan LIU Xuan ZHANG +6 位作者 Xibin CAO Jinsheng GUO Zhongxi SHAO Qingyang DENG Pengbo FU Yaodong HOU Haipeng CHEN 《Chinese Journal of Aeronautics》 2025年第9期225-250,共26页
To improve design accuracy and reliability of structures,this study solves the uncertain natural frequencies with consideration for geometric nonlinearity and structural uncertainty.Frequencies of the laminated plate ... To improve design accuracy and reliability of structures,this study solves the uncertain natural frequencies with consideration for geometric nonlinearity and structural uncertainty.Frequencies of the laminated plate with all four edges clamped(CCCC)are derived based on Navier's method and Galerkin's method.The novelty of the current work is that the number of unknowns in the displacement field model of a CCCC plate with free midsurface(CCCC-2 plate)is only three compared with four or five in cases of other exposed methods.The present analytical method is proved to be accurate and reliable by comparing linear natural frequencies and nonlinear natural frequencies with other models available in the open literature.Furthermore,a novel method for analyzing effects of mean values and tolerance zones of uncertain structural parameters on random frequencies is proposed based on a self-developed Multiscale Feature Extraction and Fusion Network(MFEFN)system.Compared with a direct Monte Carlo Simulation(MCS),the MFEFNbased procedure significantly reduces the calculation burden with a guarantee of accuracy.Our research provides a method to calculate nonlinear natural frequencies under two boundary conditions and presentes a surrogate model to predict frequencies for accuracy analysis and optimization design. 展开更多
关键词 Geometric nonlinearity LAMINATES multiscale feature extraction and fusion networks(MFEFN) Natural frequency Uncertainty analysis
原文传递
Feature Fusion Multi_XMNet Convolution Neural Network for Clothing Image Classification 被引量:2
4
作者 ZHOU Honglei PENG Zhifei +1 位作者 TAO Ran ZHANG Lu 《Journal of Donghua University(English Edition)》 CAS 2021年第6期519-526,共8页
Faced with the massive amount of online shopping clothing images,how to classify them quickly and accurately is a challenging task in image classification.In this paper,we propose a novel method,named Multi_XMNet,to s... Faced with the massive amount of online shopping clothing images,how to classify them quickly and accurately is a challenging task in image classification.In this paper,we propose a novel method,named Multi_XMNet,to solve the clothing images classification problem.The proposed method mainly consists of two convolution neural network(CNN)branches.One branch extracts multiscale features from the whole expressional image by Multi_X which is designed by improving the Xception network,while the other extracts attention mechanism features from the whole expressional image by MobileNetV3-small network.Both multiscale and attention mechanism features are aggregated before making classification.Additionally,in the training stage,global average pooling(GAP),convolutional layers,and softmax classifiers are used instead of the fully connected layer to classify the final features,which speed up model training and alleviate the problem of overfitting caused by too many parameters.Experimental comparisons are made in the public DeepFashion dataset.The experimental results show that the classification accuracy of this method is 95.38%,which is better than InceptionV3,Xception and InceptionV3_Xception by 5.58%,3.32%,and 2.22%,respectively.The proposed Multi_XMNet image classification model can help enterprises and researchers in the field of clothing e-commerce to automaticly,efficiently and accurately classify massive clothing images. 展开更多
关键词 feature extraction feature fusion multiscale feature convolution neural network(CNN) clothing image classification
在线阅读 下载PDF
Multidimensional attention and multiscale upsampling for semantic segmentation
5
作者 LU Zhongda ZHANG Chunda +1 位作者 WANG Lijing XU Fengxia 《Journal of Measurement Science and Instrumentation》 CAS CSCD 2022年第1期68-78,共11页
Semantic segmentation is for pixel-level classification tasks,and contextual information has an important impact on the performance of segmentation.In order to capture richer contextual information,we adopt ResNet as ... Semantic segmentation is for pixel-level classification tasks,and contextual information has an important impact on the performance of segmentation.In order to capture richer contextual information,we adopt ResNet as the backbone network and designs an encoder-decoder architecture based on multidimensional attention(MDA)module and multiscale upsampling(MSU)module.The MDA module calculates the attention matrices of the three dimensions to capture the dependency of each position,and adaptively captures the image features.The MSU module adopts parallel branches to capture the multiscale features of the images,and multiscale feature aggregation can enhance contextual information.A series of experiments demonstrate the validity of the model on Cityscapes and Camvid datasets. 展开更多
关键词 semantic segmentation attention mechanism multiscale feature convolutional neural network(CNN) residual network(ResNet)
在线阅读 下载PDF
Grid Side Distributed Energy Storage Cloud Group End Region Hierarchical Time-Sharing Configuration Algorithm Based onMulti-Scale and Multi Feature Convolution Neural Network 被引量:1
6
作者 Wen Long Bin Zhu +3 位作者 Huaizheng Li Yan Zhu Zhiqiang Chen Gang Cheng 《Energy Engineering》 EI 2023年第5期1253-1269,共17页
There is instability in the distributed energy storage cloud group end region on the power grid side.In order to avoid large-scale fluctuating charging and discharging in the power grid environment and make the capaci... There is instability in the distributed energy storage cloud group end region on the power grid side.In order to avoid large-scale fluctuating charging and discharging in the power grid environment and make the capacitor components showa continuous and stable charging and discharging state,a hierarchical time-sharing configuration algorithm of distributed energy storage cloud group end region on the power grid side based on multi-scale and multi feature convolution neural network is proposed.Firstly,a voltage stability analysis model based onmulti-scale and multi feature convolution neural network is constructed,and the multi-scale and multi feature convolution neural network is optimized based on Self-OrganizingMaps(SOM)algorithm to analyze the voltage stability of the cloud group end region of distributed energy storage on the grid side under the framework of credibility.According to the optimal scheduling objectives and network size,the distributed robust optimal configuration control model is solved under the framework of coordinated optimal scheduling at multiple time scales;Finally,the time series characteristics of regional power grid load and distributed generation are analyzed.According to the regional hierarchical time-sharing configuration model of“cloud”,“group”and“end”layer,the grid side distributed energy storage cloud group end regional hierarchical time-sharing configuration algorithm is realized.The experimental results show that after applying this algorithm,the best grid side distributed energy storage configuration scheme can be determined,and the stability of grid side distributed energy storage cloud group end region layered timesharing configuration can be improved. 展开更多
关键词 multiscale and multi feature convolution neural network distributed energy storage at grid side cloud group end region layered time-sharing configuration algorithm
在线阅读 下载PDF
An Industrial Intrusion Detection Method Based on Hybrid Convolutional Neural Networks with Improved TCN
7
作者 Zhihua Liu Shengquan Liu Jian Zhang 《Computers, Materials & Continua》 SCIE EI 2024年第1期411-433,共23页
Network intrusion detection systems(NIDS)based on deep learning have continued to make significant advances.However,the following challenges remain:on the one hand,simply applying only Temporal Convolutional Networks(... Network intrusion detection systems(NIDS)based on deep learning have continued to make significant advances.However,the following challenges remain:on the one hand,simply applying only Temporal Convolutional Networks(TCNs)can lead to models that ignore the impact of network traffic features at different scales on the detection performance.On the other hand,some intrusion detection methods considermulti-scale information of traffic data,but considering only forward network traffic information can lead to deficiencies in capturing multi-scale temporal features.To address both of these issues,we propose a hybrid Convolutional Neural Network that supports a multi-output strategy(BONUS)for industrial internet intrusion detection.First,we create a multiscale Temporal Convolutional Network by stacking TCN of different scales to capture the multiscale information of network traffic.Meanwhile,we propose a bi-directional structure and dynamically set the weights to fuse the forward and backward contextual information of network traffic at each scale to enhance the model’s performance in capturing the multi-scale temporal features of network traffic.In addition,we introduce a gated network for each of the two branches in the proposed method to assist the model in learning the feature representation of each branch.Extensive experiments reveal the effectiveness of the proposed approach on two publicly available traffic intrusion detection datasets named UNSW-NB15 and NSL-KDD with F1 score of 85.03% and 99.31%,respectively,which also validates the effectiveness of enhancing the model’s ability to capture multi-scale temporal features of traffic data on detection performance. 展开更多
关键词 Intrusion detection industrial internet channel spatial attention multiscale features dynamic fusion multi-output learning strategy
在线阅读 下载PDF
Efficient 3D Biomedical Image Segmentation by Parallelly Multiscale Transformer−CNN Aggregation Network
8
作者 Wei Liu Yuxiao He +8 位作者 Tiantian Man Fulin Zhu Qiaoliang Chen Yaqi Huang Xuyu Feng Bin Li Ying Wan Jian He Shengyuan Deng 《Chemical & Biomedical Imaging》 2025年第8期522-533,共12页
Accurate and automated segmentation of 3D biomedical images is a sophisticated imperative in clinical diagnosis,imaging-guided surgery,and prognosis judgment.Although the burgeoning of deep learning technologies has f... Accurate and automated segmentation of 3D biomedical images is a sophisticated imperative in clinical diagnosis,imaging-guided surgery,and prognosis judgment.Although the burgeoning of deep learning technologies has fostered smart segmentators,the successive and simultaneous garnering global and local features still remains challenging,which is essential for an exact and efficient imageological assay.To this end,a segmentation solution dubbed the mixed parallel shunted transformer(MPSTrans)is developed here,highlighting 3DMPST blocks in a U-form framework.It enabled not only comprehensive characteristic capture and multiscale slice synchronization but also deep supervision in the decoder to facilitate the fetching of hierarchical representations.Performing on an unpublished colon cancer data set,this model achieved an impressive increase in dice similarity coefficient(DSC)and a 1.718 mm decease in Hausdorff distance at 95%(HD95),alongside a substantial shrink of computational load of 56.7%in giga floating-point operations per second(GFLOPs).Meanwhile,MPSTrans outperforms other mainstream methods(Swin UNETR,UNETR,nnU-Net,PHTrans,and 3D U-Net)on three public multiorgan(aorta,gallbladder,kidney,liver,pancreas,spleen,stomach,etc.)and multimodal(CT,PET-CT,and MRI)data sets of medical segmentation decathlon(MSD)brain tumor,multiatlas labeling beyond cranial vault(BCV),and automated cardiac diagnosis challenge(ACDC),accentuating its adaptability.These results reflect the potential of MPSTrans to advance the state-of-the-art in biomedical imaging analysis,which would offer a robust tool for enhanced diagnostic capacity. 展开更多
关键词 3D biomedical image segmentation shunted transformer convolutional neural networks parallel architecture multiscale feature extraction
在线阅读 下载PDF
RF-Net: Unsupervised Low-Light Image Enhancement Based on Retinex and Exposure Fusion 被引量:2
9
作者 Tian Ma Chenhui Fu +2 位作者 Jiayi Yang Jiehui Zhang Chuyang Shang 《Computers, Materials & Continua》 SCIE EI 2023年第10期1103-1122,共20页
Low-light image enhancement methods have limitations in addressing issues such as color distortion,lack of vibrancy,and uneven light distribution and often require paired training data.To address these issues,we propo... Low-light image enhancement methods have limitations in addressing issues such as color distortion,lack of vibrancy,and uneven light distribution and often require paired training data.To address these issues,we propose a two-stage unsupervised low-light image enhancement algorithm called Retinex and Exposure Fusion Network(RFNet),which can overcome the problems of over-enhancement of the high dynamic range and under-enhancement of the low dynamic range in existing enhancement algorithms.This algorithm can better manage the challenges brought about by complex environments in real-world scenarios by training with unpaired low-light images and regular-light images.In the first stage,we design a multi-scale feature extraction module based on Retinex theory,capable of extracting details and structural information at different scales to generate high-quality illumination and reflection images.In the second stage,an exposure image generator is designed through the camera response mechanism function to acquire exposure images containing more dark features,and the generated images are fused with the original input images to complete the low-light image enhancement.Experiments show the effectiveness and rationality of each module designed in this paper.And the method reconstructs the details of contrast and color distribution,outperforms the current state-of-the-art methods in both qualitative and quantitative metrics,and shows excellent performance in the real world. 展开更多
关键词 Low-light image enhancement multiscale feature extraction module exposure generator exposure fusion
在线阅读 下载PDF
Convolutional Neural Network Based on Spatial Pyramid for Image Classification 被引量:2
10
作者 Gaihua Wang Meng Lu +2 位作者 Tao Li Guoliang Yuan Wenzhou Liu 《Journal of Beijing Institute of Technology》 EI CAS 2018年第4期630-636,共7页
A novel convolutional neural network based on spatial pyramid for image classification is proposed.The network exploits image features with spatial pyramid representation.First,it extracts global features from an orig... A novel convolutional neural network based on spatial pyramid for image classification is proposed.The network exploits image features with spatial pyramid representation.First,it extracts global features from an original image,and then different layers of grids are utilized to extract feature maps from different convolutional layers.Inspired by the spatial pyramid,the new network contains two parts,one of which is just like a standard convolutional neural network,composing of alternating convolutions and subsampling layers.But those convolution layers would be averagely pooled by the grid way to obtain feature maps,and then concatenated into a feature vector individually.Finally,those vectors are sequentially concatenated into a total feature vector as the last feature to the fully connection layer.This generated feature vector derives benefits from the classic and previous convolution layer,while the size of the grid adjusting the weight of the feature maps improves the recognition efficiency of the network.Experimental results demonstrate that this model improves the accuracy and applicability compared with the traditional model. 展开更多
关键词 convolutional neural network multiscale feature extraction image classification
在线阅读 下载PDF
A point cloud segmentation method for power lines and towersbased on a combination of multiscale density features andpoint-based deep learning 被引量:1
11
作者 Wenbo Zhao Qing Dong Zhengli Zuo 《International Journal of Digital Earth》 SCIE EI 2023年第1期620-644,共25页
The point segmentation of power lines and towers aims to use unmanned aerial vehicles(UAVs)for the inspection of power facilities,risk detection and modelling.Because of the unclear spatial relationship between the po... The point segmentation of power lines and towers aims to use unmanned aerial vehicles(UAVs)for the inspection of power facilities,risk detection and modelling.Because of the unclear spatial relationship between the point clouds,the point segmentation of power lines and towers is challenging.In this paper,the power line and tower point datasets are constructed using Light Detection and Ranging(LiDAR)and a point segmentation method is proposed based on multiscale density features and a point-based deep learning network.First,the data are blocked and the neighbourhood is constructed.Second,the point clouds are downsampled to produce sparse point clouds.The point clouds before and after sampling are rotated,and their density is calculated.Next,a direct mapping method is selected to fuse the density information;a lightweight network is built to learn the features.Finally,the point clouds are segmented by concatenating the local features provided by PointCNN.The algorithm performs effectively on different types of power lines and towers.The mean interaction over union is 82.73%,and the overall accuracy can reach 91.76%.This approach can achieve the end-to-end integration of segmentation and provide theoretical support for the segmentation of large scenic point clouds. 展开更多
关键词 Power lines and power towers point cloud segmentation multiscale density features PointCNN
原文传递
MSCANet: multiscale context information aggregation network for Tibetan Plateau lake extraction from remote sensing images 被引量:1
12
作者 Zhihui Tian Xiaoyu Guo +3 位作者 Xiaohui He Panle Li Xijjie Cheng Guangsheng Zhou 《International Journal of Digital Earth》 SCIE EI 2023年第1期1-30,共30页
Qinghai-Tibet Plateau lakes are important carriers of water resources in the‘Asian’s Water Tower’,and it is of great significance to grasp the spatial distribution of plateau lakes for the climate,ecological enviro... Qinghai-Tibet Plateau lakes are important carriers of water resources in the‘Asian’s Water Tower’,and it is of great significance to grasp the spatial distribution of plateau lakes for the climate,ecological environment,and regional water cycle.However,the differences in spatial-spectral characteristics of various types of plateau lakes,and the complex background information of plateau both influence the extraction effect of lakes.Therefore,it is a great challenge to completely and effectively extract plateau lakes.In this study,we proposed a multiscale contextual information aggregation network,termed MSCANet,to automatically extract Plateau lake regions.It consists of three main components:a multiscale lake feature encoder,a feature decoder,and a Multicore Pyramid Pooling Module(MPPM).The multiscale lake feature encoder suppressed noise interference to capture multiscale spatial-spectral information from heterogeneous scenes.The MPPM module aggregated the contextual information of various lakes globally.We applied the MSCANet to the lake extraction of the Qinghai-Tibet Plateau based on Google data;additionally,comparative experiments showed that the MSCANet proposed had obvious improvement in lake detection accuracy and morphological integrity.Finally,we transferred the pre-trained optimal model to the Landsat-8 and Sentinel-2A dataset to verify the generalization of the MSCANet. 展开更多
关键词 Remote sensing imagery The Qinghai-Tibet Plateau lake extraction deep learning multiscale feature context information aggregation
原文传递
VMMAO-YOLO:an ultra-lightweight and scale-aware detector for real-time defect detection of avionics thermistor wire solder joints
13
作者 Xiaoqi YANG Xingyue LIU +4 位作者 Qian WU Guojun WEN Shuang MEI Guanglan LIAO Tielin SHI 《Frontiers of Mechanical Engineering》 SCIE CSCD 2024年第3期77-92,共16页
The quality of the exposed avionics solder joints has a significant impact on the stable operation of the inorbit spacecrafts.Nevertheless,the previously reported inspection methods for multi-scale solder joint defect... The quality of the exposed avionics solder joints has a significant impact on the stable operation of the inorbit spacecrafts.Nevertheless,the previously reported inspection methods for multi-scale solder joint defects generally suffer low accuracy and slow detection speed.Herein,a novel real-time detector VMMAO-YOLO is demonstrated based on variable multi-scale concurrency and multi-depth aggregation network(VMMANet)backbone and“one-stop”global information gather-distribute(OS-GD)module.Combined with infrared thermography technology,it can achieve fast and high-precision detection of both internal and external solder joint defects.Specifically,VMMANet is designed for efficient multi-scale feature extraction,which mainly comprises variable multi-scale feature concurrency(VMC)and multi-depth feature aggregation-alignment(MAA)modules.VMC can extract multi-scale features via multiple fix-sized and deformable convolutions,while MAA can aggregate and align multi-depth features on the same order for feature inference.This allows the low-level features with more spatial details to be transmitted in depth-wise,enabling the deeper network to selectively utilize the preceding inference information.The VMMANet replaces inefficient highdensity deep convolution by increasing the width of intermediate feature levels,leading to a salient decline in parameters.The OS-GD is developed for efficacious feature extraction,aggregation and distribution,further enhancing the global information gather and deployment capability of the network.On a self-made solder joint image data set,the VMMAOYOLO achieves a mean average precision mAP@0.5 of 91.6%,surpassing all the mainstream YOLO-series models.Moreover,the VMMAO-YOLO has a body size of merely 19.3 MB and a detection speed up to 119 frame per second,far superior to the prevalent YOLO-series detectors. 展开更多
关键词 defect detection of solder joints VMMAO-YOLO ultra-lightweight and high-performance multiscale feature extraction VMC and MAA modules OS-GD
原文传递
上一页 1 下一页 到第
使用帮助 返回顶部