期刊文献+
共找到170篇文章
< 1 2 9 >
每页显示 20 50 100
SEFormer:A Lightweight CNN-Transformer Based on Separable Multiscale Depthwise Convolution and Efficient Self-Attention for Rotating Machinery Fault Diagnosis 被引量:1
1
作者 Hongxing Wang Xilai Ju +1 位作者 Hua Zhu Huafeng Li 《Computers, Materials & Continua》 SCIE EI 2025年第1期1417-1437,共21页
Traditional data-driven fault diagnosis methods depend on expert experience to manually extract effective fault features of signals,which has certain limitations.Conversely,deep learning techniques have gained promine... Traditional data-driven fault diagnosis methods depend on expert experience to manually extract effective fault features of signals,which has certain limitations.Conversely,deep learning techniques have gained prominence as a central focus of research in the field of fault diagnosis by strong fault feature extraction ability and end-to-end fault diagnosis efficiency.Recently,utilizing the respective advantages of convolution neural network(CNN)and Transformer in local and global feature extraction,research on cooperating the two have demonstrated promise in the field of fault diagnosis.However,the cross-channel convolution mechanism in CNN and the self-attention calculations in Transformer contribute to excessive complexity in the cooperative model.This complexity results in high computational costs and limited industrial applicability.To tackle the above challenges,this paper proposes a lightweight CNN-Transformer named as SEFormer for rotating machinery fault diagnosis.First,a separable multiscale depthwise convolution block is designed to extract and integrate multiscale feature information from different channel dimensions of vibration signals.Then,an efficient self-attention block is developed to capture critical fine-grained features of the signal from a global perspective.Finally,experimental results on the planetary gearbox dataset and themotor roller bearing dataset prove that the proposed framework can balance the advantages of robustness,generalization and lightweight compared to recent state-of-the-art fault diagnosis models based on CNN and Transformer.This study presents a feasible strategy for developing a lightweight rotating machinery fault diagnosis framework aimed at economical deployment. 展开更多
关键词 CNN-Transformer separable multiscale depthwise convolution efficient self-attention fault diagnosis
在线阅读 下载PDF
Spatio-temporal prediction of groundwater vulnerability based on CNN-LSTM model with self-attention mechanism:A case study in Hetao Plain,northern China 被引量:2
2
作者 Yifu Zhao Liangping Yang +4 位作者 Hongjie Pan Yanlong Li Yongxu Shao Junxia Li Xianjun Xie 《Journal of Environmental Sciences》 2025年第7期128-142,共15页
Located in northern China,the Hetao Plain is an important agro-economic zone and population centre.The deterioration of local groundwater quality has had a serious impact on human health and economic development.Nowad... Located in northern China,the Hetao Plain is an important agro-economic zone and population centre.The deterioration of local groundwater quality has had a serious impact on human health and economic development.Nowadays,the groundwater vulnerability assessment(GVA)has become an essential task to identify the current status and development trend of groundwater quality.In this study,the Convolutional Neural Network(CNN)and Long Short-Term Memory(LSTM)models are integrated to realize the spatio-temporal prediction of regional groundwater vulnerability by introducing the Self-attention mechanism.The study firstly builds the CNN-LSTM modelwith self-attention(SA)mechanism and evaluates the prediction accuracy of the model for groundwater vulnerability compared to other common machine learning models such as Support Vector Machine(SVM),Random Forest(RF),and Extreme Gradient Boosting(XGBoost).The results indicate that the CNNLSTM model outperforms thesemodels,demonstrating its significance in groundwater vulnerability assessment.It can be posited that the predictions indicate an increased risk of groundwater vulnerability in the study area over the coming years.This increase can be attributed to the synergistic impact of global climate anomalies and intensified local human activities.Moreover,the overall groundwater vulnerability risk in the entire region has increased,evident fromboth the notably high value and standard deviation.This suggests that the spatial variability of groundwater vulnerability in the area is expected to expand in the future due to the sustained progression of climate change and human activities.The model can be optimized for diverse applications across regional environmental assessment,pollution prediction,and risk statistics.This study holds particular significance for ecological protection and groundwater resource management. 展开更多
关键词 Groundwater vulnerability assessment Convolutional Neural Network Long Short-Term Memory self-attention mechanism
原文传递
Multi-View Seizure Classification Based on Attention-Based Adaptive Graph ProbSparse Hybrid Network
3
作者 Changxu Dong Yanqing Liu Dengdi Sun 《CAAI Transactions on Intelligence Technology》 2025年第6期1783-1798,共16页
Epilepsy is a neurological disorder characterised by recurrent seizures due to abnormal neuronal discharges.Seizure detection via EEG signals has progressed,but two main challenges are still encountered.First,EEG data... Epilepsy is a neurological disorder characterised by recurrent seizures due to abnormal neuronal discharges.Seizure detection via EEG signals has progressed,but two main challenges are still encountered.First,EEG data can be distorted by physiological factors and external variables,resulting in noisy brain networks.Static adjacency matrices are typically used in current mainstream methods,which neglect the need for dynamic updates and feature refinement.The second challenge stems from the strong reliance on long-range dependencies through self-attention in current methods,which can introduce redundant noise and increase computational complexity,especially in long-duration data.To address these challenges,the Attention-based Adaptive Graph ProbSparse Hybrid Network(AA-GPHN)is proposed.Brain network structures are dynamically optimised using variational inference and the information bottleneck principle,refining the adjacency matrix for improved epilepsy classification.A Linear Graph Convolutional Network(LGCN)is incorporated to focus on first-order neighbours,minimising the aggregation of distant information.Furthermore,a ProbSparse attention-based Informer(PAT)is introduced to adaptively filter long-range dependencies,enhancing efficiency.A joint optimisation loss function is applied to improve robustness in noisy environments.Experimental results on both patient-specific and cross-subject datasets demonstrate that AA-GPHN outperforms existing methods in seizure detection,showing superior effectiveness and generalisation. 展开更多
关键词 BIOINFORMATICS deep learning dynamically EEG ELECTROENCEPHALOGRAPHY probsparse attention seizure classification
在线阅读 下载PDF
The brief self-attention module for lightweight convolution neural networks
4
作者 YAN Jie WEI Yingmei +3 位作者 XIE Yuxiang GONG Quanzhi ZOU Shiwei LUAN Xidao 《Journal of Systems Engineering and Electronics》 2025年第6期1389-1397,共9页
Lightweight convolutional neural networks(CNNs)have simple structures but struggle to comprehensively and accurately extract important semantic information from images.While attention mechanisms can enhance CNNs by le... Lightweight convolutional neural networks(CNNs)have simple structures but struggle to comprehensively and accurately extract important semantic information from images.While attention mechanisms can enhance CNNs by learning distinctive representations,most existing spatial and hybrid attention methods focus on local regions with extensive parameters,making them unsuitable for lightweight CNNs.In this paper,we propose a self-attention mechanism tailored for lightweight networks,namely the brief self-attention module(BSAM).BSAM consists of the brief spatial attention(BSA)and advanced channel attention blocks.Unlike conventional self-attention methods with many parameters,our BSA block improves the performance of lightweight networks by effectively learning global semantic representations.Moreover,BSAM can be seamlessly integrated into lightweight CNNs for end-to-end training,maintaining the network’s lightweight and mobile characteristics.We validate the effectiveness of the proposed method on image classification tasks using the Food-101,Caltech-256,and Mini-ImageNet datasets. 展开更多
关键词 self-attention lightweight neural network deep learning
在线阅读 下载PDF
A Novel Dynamic Residual Self-Attention Transfer Adaptive Learning Fusion Approach for Brain Tumor Diagnosis
5
作者 Tawfeeq Shawly Ahmed A.Alsheikhy 《Computers, Materials & Continua》 2025年第3期4161-4179,共19页
A healthy brain is vital to every person since the brain controls every movement and emotion.Sometimes,some brain cells grow unexpectedly to be uncontrollable and cancerous.These cancerous cells are called brain tumor... A healthy brain is vital to every person since the brain controls every movement and emotion.Sometimes,some brain cells grow unexpectedly to be uncontrollable and cancerous.These cancerous cells are called brain tumors.For diagnosed patients,their lives depend mainly on the early diagnosis of these tumors to provide suitable treatment plans.Nowadays,Physicians and radiologists rely on Magnetic Resonance Imaging(MRI)pictures for their clinical evaluations of brain tumors.These evaluations are time-consuming,expensive,and require expertise with high skills to provide an accurate diagnosis.Scholars and industrials have recently partnered to implement automatic solutions to diagnose the disease with high accuracy.Due to their accuracy,some of these solutions depend on deep-learning(DL)methodologies.These techniques have become important due to their roles in the diagnosis process,which includes identification and classification.Therefore,there is a need for a solid and robust approach based on a deep-learning method to diagnose brain tumors.The purpose of this study is to develop an intelligent automatic framework for brain tumor diagnosis.The proposed solution is based on a novel dense dynamic residual self-attention transfer adaptive learning fusion approach(NDDRSATALFA),carried over two implemented deep-learning networks:VGG19 and UNET to identify and classify brain tumors.In addition,this solution applies a transfer learning approach to exchange extracted features and data within the two neural networks.The presented framework is trained,validated,and tested on six public datasets of MRIs to detect brain tumors and categorize these tumors into three suitable classes,which are glioma,meningioma,and pituitary.The proposed framework yielded remarkable findings on variously evaluated performance indicators:99.32%accuracy,98.74%sensitivity,98.89%specificity,99.01%Dice,98.93%Area Under the Curve(AUC),and 99.81%F1-score.In addition,a comparative analysis with recent state-of-the-art methods was performed and according to the comparative analysis,NDDRSATALFA shows an admirable level of reliability in simplifying the timely identification of diverse brain tumors.Moreover,this framework can be applied by healthcare providers to assist radiologists,pathologists,and physicians in their evaluations.The attained outcomes open doors for advanced automatic solutions that improve clinical evaluations and provide reasonable treatment plans. 展开更多
关键词 Brain tumor deep learning transfer learning RESIDUAL self-attention VGG19 UNET
在线阅读 下载PDF
An Overlapped Multihead Self-Attention-Based Feature Enhancement Approach for Ocular Disease Image Recognition
6
作者 Peng Xiao Haiyu Xu +3 位作者 Peng Xu Zhiwei Guo Amr Tolba Osama Alfarraj 《Computers, Materials & Continua》 2025年第11期2999-3022,共24页
Medical image analysis based on deep learning has become an important technical requirement in the field of smart healthcare.In view of the difficulties in collaborative modeling of local details and global features i... Medical image analysis based on deep learning has become an important technical requirement in the field of smart healthcare.In view of the difficulties in collaborative modeling of local details and global features in multimodal image analysis of ophthalmology,as well as the existence of information redundancy in cross-modal data fusion,this paper proposes amultimodal fusion framework based on cross-modal collaboration and weighted attention mechanism.In terms of feature extraction,the framework collaboratively extracts local fine-grained features and global structural dependencies through a parallel dual-branch architecture,overcoming the limitations of traditional single-modality models in capturing either local or global information;in terms of fusion strategy,the framework innovatively designs a cross-modal dynamic fusion strategy,combining overlappingmulti-head self-attention modules with a bidirectional feature alignment mechanism,addressing the bottlenecks of low feature interaction efficiency and excessive attention fusion computations in traditional parallel fusion,and further introduces cross-domain local integration technology,which enhances the representation ability of the lesion area through pixel-level feature recalibration and optimizes the diagnostic robustness of complex cases.Experiments show that the framework exhibits excellent feature expression and generalization performance in cross-domain scenarios of ophthalmic medical images and natural images,providing a high-precision,low-redundancy fusion paradigm for multimodal medical image analysis,and promoting the upgrade of intelligent diagnosis and treatment fromsingle-modal static analysis to dynamic decision-making. 展开更多
关键词 Overlapping multi-head self-attention deep learning cross-modal dynamic fusion multi-level fusion
在线阅读 下载PDF
EFI-SATL:An Efficient Net and Self-Attention Based Biometric Recognition for Finger-Vein Using Deep Transfer Learning
7
作者 Manjit Singh Sunil Kumar Singla 《Computer Modeling in Engineering & Sciences》 2025年第3期3003-3029,共27页
Deep Learning-based systems for Finger vein recognition have gained rising attention in recent years due to improved efficiency and enhanced security.The performance of existing CNN-based methods is limited by the pun... Deep Learning-based systems for Finger vein recognition have gained rising attention in recent years due to improved efficiency and enhanced security.The performance of existing CNN-based methods is limited by the puny generalization of learned features and deficiency of the finger vein image training data.Considering the concerns of existing methods,in this work,a simplified deep transfer learning-based framework for finger-vein recognition is developed using an EfficientNet model of deep learning with a self-attention mechanism.Data augmentation using various geometrical methods is employed to address the problem of training data shortage required for a deep learning model.The proposed model is tested using K-fold cross-validation on three publicly available datasets:HKPU,FVUSM,and SDUMLA.Also,the developed network is compared with other modern deep nets to check its effectiveness.In addition,a comparison of the proposed method with other existing Finger vein recognition(FVR)methods is also done.The experimental results exhibited superior recognition accuracy of the proposed method compared to other existing methods.In addition,the developed method proves to be more effective and less sophisticated at extracting robust features.The proposed EffAttenNet achieves an accuracy of 98.14%on HKPU,99.03%on FVUSM,and 99.50%on SDUMLA databases. 展开更多
关键词 Biometrics finger-vein recognition(FVR) deep net self-attention Efficient Nets transfer learning
在线阅读 下载PDF
Image compressed sensing reconstruction network based on self-attention mechanism
8
作者 LIU Yuhong LIU Xiaoyan CHEN Manyin 《Journal of Measurement Science and Instrumentation》 2025年第4期537-546,共10页
For image compression sensing reconstruction,most algorithms use the method of reconstructing image blocks one by one and stacking many convolutional layers,which usually have defects of obvious block effects,high com... For image compression sensing reconstruction,most algorithms use the method of reconstructing image blocks one by one and stacking many convolutional layers,which usually have defects of obvious block effects,high computational complexity,and long reconstruction time.An image compressed sensing reconstruction network based on self-attention mechanism(SAMNet)was proposed.For the compressed sampling,self-attention convolution was designed,which was conducive to capturing richer features,so that the compressed sensing measurement value retained more image structure information.For the reconstruction,a self-attention mechanism was introduced in the convolutional neural network.A reconstruction network including residual blocks,bottleneck transformer(BoTNet),and dense blocks was proposed,which strengthened the transfer of image features and reduced the amount of parameters dramatically.Under the Set5 dataset,when the measurement rates are 0.01,0.04,0.10,and 0.25,the average peak signal-to-noise ratio(PSNR)of SAMNet is improved by 1.27,1.23,0.50,and 0.15 dB,respectively,compared to the CSNet+.The running time of reconstructing a 256×256 image is reduced by 0.1473,0.1789,0.2310,and 0.2524 s compared to ReconNet.Experimental results showed that SAMNet improved the quality of reconstructed images and reduced the reconstruction time. 展开更多
关键词 convolutional neural network compressed sensing self-attention mechanism dense block image reconstruction
在线阅读 下载PDF
A precise magnetic modeling method for scientific satellites based on a self-attention mechanism and Kolmogorov-Arnold Networks
9
作者 Ye Liu Xingjian Shi +2 位作者 Wenzhe Yang Zhiming Cai Huawang Li 《Astronomical Techniques and Instruments》 2025年第1期1-9,共9页
As the complexity of scientific satellite missions increases,the requirements for their magnetic fields,magnetic field fluctuations,and even magnetic field gradients and variations become increasingly stringent.Additi... As the complexity of scientific satellite missions increases,the requirements for their magnetic fields,magnetic field fluctuations,and even magnetic field gradients and variations become increasingly stringent.Additionally,there is a growing need to address the alternating magnetic fields produced by the spacecraft itself.This paper introduces a novel modeling method for spacecraft magnetic dipoles using an integrated self-attention mechanism and a transformer combined with Kolmogorov-Arnold Networks.The self-attention mechanism captures correlations among globally sparse data,establishing dependencies b.etween sparse magnetometer readings.Concurrently,the Kolmogorov-Arnold Network,proficient in modeling implicit numerical relationships between data features,enhances the ability to learn subtle patterns.Comparative experiments validate the capability of the proposed method to precisely model magnetic dipoles,achieving maximum Root Mean Square Errors of 24.06 mA·m^(2)and 0.32 cm for size and location modeling,respectively.The spacecraft magnetic model established using this method accurately computes magnetic fields and alternating magnetic fields at designated surfaces or points.This approach facilitates the rapid and precise construction of individual and complete spacecraft magnetic models,enabling the verification of magnetic specifications from the spacecraft design phase. 展开更多
关键词 Magnetic dipole model self-attention mechanism Kolmogorov-Arnold networks Alternating current magnetic fields
在线阅读 下载PDF
Dual Self-attention Fusion Message Neural Network for Virtual Screening in Drug Discovery by Molecular Property Prediction
10
作者 Jingjing Wang Kangming Hou +2 位作者 Hao Chen Jing Fang Hongzhen Li 《Journal of Bionic Engineering》 2025年第1期354-369,共16页
The development of deep learning has made non-biochemical methods for molecular property prediction screening a reality,which can increase the experimental speed and reduce the experimental cost of relevant experiment... The development of deep learning has made non-biochemical methods for molecular property prediction screening a reality,which can increase the experimental speed and reduce the experimental cost of relevant experiments.There are currently two main approaches to representing molecules:(a)representing molecules by fixing molecular descriptors,and(b)representing molecules by graph convolutional neural networks.Currently,both of these Representative methods have achieved some results in their respective experiments.Based on past efforts,we propose a Dual Self-attention Fusion Message Neural Network(DSFMNN).DSFMNN uses a combination of dual self-attention mechanism and graph convolutional neural network.Advantages of DSFMNN:(1)The dual self-attention mechanism focuses not only on the relationship between individual subunits in a molecule but also on the relationship between the atoms and chemical bonds contained in each subunit.(2)On the directed molecular graph,a message delivery approach centered on directed molecular bonds is used.We test the performance of the model on eight publicly available datasets and compare the performance with several models.Based on the current experimental results,DSFMNN has superior performance compared to previous models on the datasets applied in this paper. 展开更多
关键词 Directed message passing network Deep learning Molecular property prediction self-attention mechanism
暂未订购
Self-attention and convolutional feature fusion for real-time intelligent fault detection of high-speed railway pantographs
11
作者 Xufeng LI Jien MA +3 位作者 Ping TAN Lanfen LIN Lin QIU Youtong FANG 《Journal of Zhejiang University-Science A(Applied Physics & Engineering)》 2025年第10期997-1009,共13页
Currently,most trains are equipped with dedicated cameras for capturing pantograph videos.Pantographs are core to the high-speed-railway pantograph-catenary system,and their failure directly affects the normal operati... Currently,most trains are equipped with dedicated cameras for capturing pantograph videos.Pantographs are core to the high-speed-railway pantograph-catenary system,and their failure directly affects the normal operation of high-speed trains.However,given the complex and variable real-world operational conditions of high-speed railways,there is no real-time and robust pantograph fault-detection method capable of handling large volumes of surveillance video.Hence,it is of paramount importance to maintain real-time monitoring and analysis of pantographs.Our study presents a real-time intelligent detection technology for identifying faults in high-speed railway pantographs,utilizing a fusion of self-attention and convolution features.We delved into lightweight multi-scale feature-extraction and fault-detection models based on deep learning to detect pantograph anomalies.Compared with traditional methods,this approach achieves high recall and accuracy in pantograph recognition,accurately pinpointing issues like discharge sparks,pantograph horns,and carbon pantograph-slide malfunctions.After experimentation and validation with actual surveillance videos of electric multiple-unit train,our algorithmic model demonstrates real-time,high-accuracy performance even under complex operational conditions. 展开更多
关键词 High-speed railway pantograph self-attention Convolutional neural network(CNN) REAL-TIME Feature fusion Faultdetection
原文传递
结合LDA与Self-Attention的短文本情感分类方法 被引量:9
12
作者 陈欢 黄勃 +2 位作者 朱翌民 俞雷 余宇新 《计算机工程与应用》 CSCD 北大核心 2020年第18期165-170,共6页
在对短文本进行情感分类任务的过程中,由于文本长度过短导致数据稀疏,降低了分类任务的准确率。针对这个问题,提出了一种基于潜在狄利克雷分布(LDA)与Self-Attention的短文本情感分类方法。使用LDA获得每个评论的主题词分布作为该条评... 在对短文本进行情感分类任务的过程中,由于文本长度过短导致数据稀疏,降低了分类任务的准确率。针对这个问题,提出了一种基于潜在狄利克雷分布(LDA)与Self-Attention的短文本情感分类方法。使用LDA获得每个评论的主题词分布作为该条评论信息的扩展,将扩展信息和原评论文本一起输入到word2vec模型,进行词向量训练,使得该评论文本在高维向量空间实现同一主题的聚类,使用Self-Attention进行动态权重分配并进行分类。通过在谭松波酒店评论数据集上的实验表明,该算法与当前主流的短文本分类情感算法相比,有效地提高了分类性能。 展开更多
关键词 主题词 短文本 self-attention 潜在狄利克雷分布(LDA) word2vec
在线阅读 下载PDF
结合TFIDF的Self-Attention-Based Bi-LSTM的垃圾短信识别 被引量:11
13
作者 吴思慧 陈世平 《计算机系统应用》 2020年第9期171-177,共7页
随着手机短信成为人们日常生活交往的重要手段,垃圾短信的识别具有重要的现实意义.针对此提出一种结合TFIDF的self-attention-based Bi-LSTM的神经网络模型.该模型首先将短信文本以词向量的方式输入到Bi-LSTM层,经过特征提取并结合TFIDF... 随着手机短信成为人们日常生活交往的重要手段,垃圾短信的识别具有重要的现实意义.针对此提出一种结合TFIDF的self-attention-based Bi-LSTM的神经网络模型.该模型首先将短信文本以词向量的方式输入到Bi-LSTM层,经过特征提取并结合TFIDF和self-attention层的信息聚焦获得最后的特征向量,最后将特征向量通过Softmax分类器进行分类得到短信文本分类结果.实验结果表明,结合TFIDF的self-attention-based Bi-LSTM模型相比于传统分类模型的短信文本识别准确率提高了2.1%–4.6%,运行时间减少了0.6 s–10.2 s. 展开更多
关键词 垃圾短信 文本分类 self-attention Bi-LSTM TFIDF
在线阅读 下载PDF
基于Self-Attention模型的机器翻译系统 被引量:10
14
作者 师岩 王宇 吴水清 《计算机与现代化》 2019年第7期9-14,共6页
近几年来神经机器翻译(Neural Machine Translation,NMT)发展迅速,Seq2Seq框架的提出为机器翻译带来了很大的优势,可以在观测到整个输入句子后生成任意输出序列。但是该模型对于长距离信息的捕获能力仍有很大的局限,循环神经网络(RNN)、... 近几年来神经机器翻译(Neural Machine Translation,NMT)发展迅速,Seq2Seq框架的提出为机器翻译带来了很大的优势,可以在观测到整个输入句子后生成任意输出序列。但是该模型对于长距离信息的捕获能力仍有很大的局限,循环神经网络(RNN)、LSTM网络都是为了改善这一问题提出的,但是效果并不明显。注意力机制的提出与运用则有效地弥补了该缺陷。Self-Attention模型就是在注意力机制的基础上提出的,本文使用Self-Attention为基础构建编码器-解码器框架。本文通过探讨以往的神经网络翻译模型,分析Self-Attention模型的机制与原理,通过TensorFlow深度学习框架对基于Self-Attention模型的翻译系统进行实现,在英文到中文的翻译实验中与以往的神经网络翻译模型进行对比,表明该模型取得了较好的翻译效果。 展开更多
关键词 神经机器翻译 Seq2Seq框架 注意力机制 self-attention模型
在线阅读 下载PDF
引入Self-Attention的电力作业违规穿戴智能检测技术研究 被引量:5
15
作者 莫蓓蓓 吴克河 《计算机与现代化》 2020年第2期115-121,126,共8页
随着电网建设的高速发展,作业现场技术支撑人员规模不断扩大。电力现场属于高危作业场所,违规穿戴安全防护用品将会严重危及作业人员的人身安全,为了改善传统人工监管方式效率低下的问题,本文采用实时深度学习算法进行违规穿戴行为检测... 随着电网建设的高速发展,作业现场技术支撑人员规模不断扩大。电力现场属于高危作业场所,违规穿戴安全防护用品将会严重危及作业人员的人身安全,为了改善传统人工监管方式效率低下的问题,本文采用实时深度学习算法进行违规穿戴行为检测。检测模型结合实时目标检测网络YOLOv3和Self-Attention机制,借鉴DANet结构,在YOLOv3网络高层嵌入自注意力模块,更好地挖掘和学习特征位置和通道关系。实验结果表明,该模型在违规穿戴检测任务中mAP达到了94.58%,Recall达到了96.67%,与YOLOv3相比,mAP提高了12.66%,Recall提高了2.69%,显著提高模型的精度,可以满足任务的检测需求,提升了电网智能化水平。 展开更多
关键词 电力作业 违规穿戴 YOLOv3技术 self-attention机制 目标检测
在线阅读 下载PDF
融合Self-Attention机制和n-gram卷积核的印尼语复合名词自动识别方法研究 被引量:2
16
作者 丘心颖 陈汉武 +3 位作者 陈源 谭立聪 张皓 肖莉娴 《湖南工业大学学报》 2020年第3期1-9,共9页
针对印尼语复合名词短语自动识别,提出一种融合Self-Attention机制、n-gram卷积核的神经网络和统计模型相结合的方法,改进现有的多词表达抽取模型。在现有SHOMA模型的基础上,使用多层CNN和Self-Attention机制进行改进。对Universal Depe... 针对印尼语复合名词短语自动识别,提出一种融合Self-Attention机制、n-gram卷积核的神经网络和统计模型相结合的方法,改进现有的多词表达抽取模型。在现有SHOMA模型的基础上,使用多层CNN和Self-Attention机制进行改进。对Universal Dependencies公开的印尼语数据进行复合名词短语自动识别的对比实验,结果表明:TextCNN+Self-Attention+CRF模型取得32.20的短语多词识别F1值和32.34的短语单字识别F1值,比SHOMA模型分别提升了4.93%和3.04%。 展开更多
关键词 印尼语复合名词短语 self-attention机制 卷积神经网络 自动识别 条件随机场
在线阅读 下载PDF
Hierarchical multihead self-attention for time-series-based fault diagnosis 被引量:3
17
作者 Chengtian Wang Hongbo Shi +1 位作者 Bing Song Yang Tao 《Chinese Journal of Chemical Engineering》 SCIE EI CAS CSCD 2024年第6期104-117,共14页
Fault diagnosis is important for maintaining the safety and effectiveness of chemical process.Considering the multivariate,nonlinear,and dynamic characteristic of chemical process,many time-series-based data-driven fa... Fault diagnosis is important for maintaining the safety and effectiveness of chemical process.Considering the multivariate,nonlinear,and dynamic characteristic of chemical process,many time-series-based data-driven fault diagnosis methods have been developed in recent years.However,the existing methods have the problem of long-term dependency and are difficult to train due to the sequential way of training.To overcome these problems,a novel fault diagnosis method based on time-series and the hierarchical multihead self-attention(HMSAN)is proposed for chemical process.First,a sliding window strategy is adopted to construct the normalized time-series dataset.Second,the HMSAN is developed to extract the time-relevant features from the time-series process data.It improves the basic self-attention model in both width and depth.With the multihead structure,the HMSAN can pay attention to different aspects of the complicated chemical process and obtain the global dynamic features.However,the multiple heads in parallel lead to redundant information,which cannot improve the diagnosis performance.With the hierarchical structure,the redundant information is reduced and the deep local time-related features are further extracted.Besides,a novel many-to-one training strategy is introduced for HMSAN to simplify the training procedure and capture the long-term dependency.Finally,the effectiveness of the proposed method is demonstrated by two chemical cases.The experimental results show that the proposed method achieves a great performance on time-series industrial data and outperforms the state-of-the-art approaches. 展开更多
关键词 self-attention mechanism Deep learning Chemical process Time-series Fault diagnosis
在线阅读 下载PDF
Aerial target threat assessment based on gated recurrent unit and self-attention mechanism 被引量:4
18
作者 CHEN Chen QUAN Wei SHAO Zhuang 《Journal of Systems Engineering and Electronics》 SCIE CSCD 2024年第2期361-373,共13页
Aerial threat assessment is a crucial link in modern air combat, whose result counts a great deal for commanders to make decisions. With the consideration that the existing threat assessment methods have difficulties ... Aerial threat assessment is a crucial link in modern air combat, whose result counts a great deal for commanders to make decisions. With the consideration that the existing threat assessment methods have difficulties in dealing with high dimensional time series target data, a threat assessment method based on self-attention mechanism and gated recurrent unit(SAGRU) is proposed. Firstly, a threat feature system including air combat situations and capability features is established. Moreover, a data augmentation process based on fractional Fourier transform(FRFT) is applied to extract more valuable information from time series situation features. Furthermore, aiming to capture key characteristics of battlefield evolution, a bidirectional GRU and SA mechanisms are designed for enhanced features.Subsequently, after the concatenation of the processed air combat situation and capability features, the target threat level will be predicted by fully connected neural layers and the softmax classifier. Finally, in order to validate this model, an air combat dataset generated by a combat simulation system is introduced for model training and testing. The comparison experiments show the proposed model has structural rationality and can perform threat assessment faster and more accurately than the other existing models based on deep learning. 展开更多
关键词 target threat assessment gated recurrent unit(GRU) self-attention(SA) fractional Fourier transform(FRFT)
在线阅读 下载PDF
Hashtag Recommendation Using LSTM Networks with Self-Attention 被引量:2
19
作者 Yatian Shen Yan Li +5 位作者 Jun Sun Wenke Ding Xianjin Shi Lei Zhang Xiajiong Shen Jing He 《Computers, Materials & Continua》 SCIE EI 2019年第9期1261-1269,共9页
On Twitter,people often use hashtags to mark the subject of a tweet.Tweets have specific themes or content that are easy for people to manage.With the increase in the number of tweets,how to automatically recommend ha... On Twitter,people often use hashtags to mark the subject of a tweet.Tweets have specific themes or content that are easy for people to manage.With the increase in the number of tweets,how to automatically recommend hashtags for tweets has received wide attention.The previous hashtag recommendation methods were to convert the task into a multi-class classification problem.However,these methods can only recommend hashtags that appeared in historical information,and cannot recommend the new ones.In this work,we extend the self-attention mechanism to turn the hashtag recommendation task into a sequence labeling task.To train and evaluate the proposed method,we used the real tweet data which is collected from Twitter.Experimental results show that the proposed method can be significantly better than the most advanced method.Compared with the state-of-the-art methods,the accuracy of our method has been increased 4%. 展开更多
关键词 Hashtags recommendation self-attention neural networks sequence labeling
在线阅读 下载PDF
Sentiment classification model for bullet screen based on self-attention mechanism 被引量:3
20
作者 ZHAO Shuxu LIU Lijiao MA Qinjing 《Journal of Measurement Science and Instrumentation》 CAS CSCD 2021年第4期479-488,共10页
With the development of short video industry,video and bullet screen have become important ways to spread public opinions.Public attitudes can be timely obtained through emotional analysis on bullet screen,which can a... With the development of short video industry,video and bullet screen have become important ways to spread public opinions.Public attitudes can be timely obtained through emotional analysis on bullet screen,which can also reduce difficulties in management of online public opinions.A convolutional neural network model based on multi-head attention is proposed to solve the problem of how to effectively model relations among words and identify key words in emotion classification tasks with short text contents and lack of complete context information.Firstly,encode word positions so that order information of input sequences can be used by the model.Secondly,use a multi-head attention mechanism to obtain semantic expressions in different subspaces,effectively capture internal relevance and enhance dependent relationships among words,as well as highlight emotional weights of key emotional words.Then a dilated convolution is used to increase the receptive field and extract more features.On this basis,the above multi-attention mechanism is combined with a convolutional neural network to model and analyze the seven emotional categories of bullet screens.Testing from perspectives of model and dataset,experimental results can validate effectiveness of our approach.Finally,emotions of bullet screens are visualized to provide data supports for hot event controls and other fields. 展开更多
关键词 bullet screen text sentiment classification self-attention mechanism visual analysis hot events control
在线阅读 下载PDF
上一页 1 2 9 下一页 到第
使用帮助 返回顶部