期刊文献+
共找到213篇文章
< 1 2 11 >
每页显示 20 50 100
SMSTracker:A Self-Calibration Multi-Head Self-Attention Transformer for Visual Object Tracking 被引量:1
1
作者 Zhongyang Wang Hu Zhu Feng Liu 《Computers, Materials & Continua》 SCIE EI 2024年第7期605-623,共19页
Visual object tracking plays a crucial role in computer vision.In recent years,researchers have proposed various methods to achieve high-performance object tracking.Among these,methods based on Transformers have becom... Visual object tracking plays a crucial role in computer vision.In recent years,researchers have proposed various methods to achieve high-performance object tracking.Among these,methods based on Transformers have become a research hotspot due to their ability to globally model and contextualize information.However,current Transformer-based object tracking methods still face challenges such as low tracking accuracy and the presence of redundant feature information.In this paper,we introduce self-calibration multi-head self-attention Transformer(SMSTracker)as a solution to these challenges.It employs a hybrid tensor decomposition self-organizing multihead self-attention transformermechanism,which not only compresses and accelerates Transformer operations but also significantly reduces redundant data,thereby enhancing the accuracy and efficiency of tracking.Additionally,we introduce a self-calibration attention fusion block to resolve common issues of attention ambiguities and inconsistencies found in traditional trackingmethods,ensuring the stability and reliability of tracking performance across various scenarios.By integrating a hybrid tensor decomposition approach with a self-organizingmulti-head self-attentive transformer mechanism,SMSTracker enhances the efficiency and accuracy of the tracking process.Experimental results show that SMSTracker achieves competitive performance in visual object tracking,promising more robust and efficient tracking systems,demonstrating its potential to providemore robust and efficient tracking solutions in real-world applications. 展开更多
关键词 Visual object tracking tensor decomposition TRANSFORMER self-attention
在线阅读 下载PDF
Enhanced Classification of Brain Tumor Types Using Multi-Head Self-Attention and ResNeXt CNN
2
作者 Muhammad Naeem Abdul Majid 《Journal on Artificial Intelligence》 2025年第1期115-141,共27页
Brain tumor identification is a challenging task in neuro-oncology.The brain’s complex anatomy makes it a crucial part of the central nervous system.Accurate tumor classification is crucial for clinical diagnosis and... Brain tumor identification is a challenging task in neuro-oncology.The brain’s complex anatomy makes it a crucial part of the central nervous system.Accurate tumor classification is crucial for clinical diagnosis and treatment planning.This research presents a significant advancement in the multi-classification of brain tumors.This paper proposed a novel architecture that integrates Enhanced ResNeXt 101_32×8d,a Convolutional Neural Network(CNN)with a multi-head self-attention(MHSA)mechanism.This combination harnesses the strengths of the feature extraction,feature representation by CNN,and long-range dependencies by MHSA.Magnetic Resonance Imaging(MRI)datasets were employed to check the effectiveness of the proposed architecture.The first dataset(DS-1,Msoud)included four brain tumor classes,and the second dataset(DS-2)contained seven brain tumor classes.This methodology effectively distinguished various tumor classes,achieving high accuracies of 99.75% on DS-1 and 98.80% on DS-2.These impressive results indicate the superior performance and adaptability of our model for multiclass brain tumor classification.Evaluationmetrics such as accuracy,precision,recall,F1 score,and ROC(receiver operating characteristic)curve were utilized to comprehensively evaluate model validity.The performance results showed that the model is well-suited for clinical applications,with reduced errors and high accuracy. 展开更多
关键词 Brain tumor classification multi-head self-attention module(MHSA) ResNeXt 101_32×8d deep learning medical imaging
在线阅读 下载PDF
Geological information prediction for shield machine using an enhanced multi-head self-attention convolution neural network with two-stage feature extraction 被引量:7
3
作者 Chengjin Qin Guoqiang Huang +3 位作者 Honggan Yu Ruihong Wu Jianfeng Tao Chengliang Liu 《Geoscience Frontiers》 SCIE CAS CSCD 2023年第2期86-104,共19页
Due to the closed working environment of shield machines,the construction personnel cannot observe the construction geological environment,which seriously restricts the safety and efficiency of the tunneling process.I... Due to the closed working environment of shield machines,the construction personnel cannot observe the construction geological environment,which seriously restricts the safety and efficiency of the tunneling process.In this study,we present an enhanced multi-head self-attention convolution neural network(EMSACNN)with two-stage feature extraction for geological condition prediction of shield machine.Firstly,we select 30 important parameters according to statistical analysis method and the working principle of the shield machine.Then,we delete the non-working sample data,and combine 10 consecutive data as the input of the model.Thereafter,to deeply mine and extract essential and relevant features,we build a novel model combined with the particularity of the geological type recognition task,in which an enhanced multi-head self-attention block is utilized as the first feature extractor to fully extract the correlation of geological information of adjacent working face of tunnel,and two-dimensional CNN(2dCNN)is utilized as the second feature extractor.The performance and superiority of proposed EMSACNN are verified by the actual data collected by the shield machine used in the construction of a double-track tunnel in Guangzhou,China.The results show that EMSACNN achieves at least 96%accuracy on the test sets of the two tunnels,and all the evaluation indicators of EMSACNN are much better than those of classical AI model and the model that use only the second-stage feature extractor.Therefore,the proposed EMSACNN achieves high accuracy and strong generalization for geological information prediction of shield machine,which is of great guiding significance to engineering practice. 展开更多
关键词 Geological information prediction Shield machine Enhanced multi-head self-attention CNN
在线阅读 下载PDF
An Overlapped Multihead Self-Attention-Based Feature Enhancement Approach for Ocular Disease Image Recognition
4
作者 Peng Xiao Haiyu Xu +3 位作者 Peng Xu Zhiwei Guo Amr Tolba Osama Alfarraj 《Computers, Materials & Continua》 2025年第11期2999-3022,共24页
Medical image analysis based on deep learning has become an important technical requirement in the field of smart healthcare.In view of the difficulties in collaborative modeling of local details and global features i... Medical image analysis based on deep learning has become an important technical requirement in the field of smart healthcare.In view of the difficulties in collaborative modeling of local details and global features in multimodal image analysis of ophthalmology,as well as the existence of information redundancy in cross-modal data fusion,this paper proposes amultimodal fusion framework based on cross-modal collaboration and weighted attention mechanism.In terms of feature extraction,the framework collaboratively extracts local fine-grained features and global structural dependencies through a parallel dual-branch architecture,overcoming the limitations of traditional single-modality models in capturing either local or global information;in terms of fusion strategy,the framework innovatively designs a cross-modal dynamic fusion strategy,combining overlappingmulti-head self-attention modules with a bidirectional feature alignment mechanism,addressing the bottlenecks of low feature interaction efficiency and excessive attention fusion computations in traditional parallel fusion,and further introduces cross-domain local integration technology,which enhances the representation ability of the lesion area through pixel-level feature recalibration and optimizes the diagnostic robustness of complex cases.Experiments show that the framework exhibits excellent feature expression and generalization performance in cross-domain scenarios of ophthalmic medical images and natural images,providing a high-precision,low-redundancy fusion paradigm for multimodal medical image analysis,and promoting the upgrade of intelligent diagnosis and treatment fromsingle-modal static analysis to dynamic decision-making. 展开更多
关键词 Overlapping multi-head self-attention deep learning cross-modal dynamic fusion multi-level fusion
在线阅读 下载PDF
SEFormer:A Lightweight CNN-Transformer Based on Separable Multiscale Depthwise Convolution and Efficient Self-Attention for Rotating Machinery Fault Diagnosis 被引量:1
5
作者 Hongxing Wang Xilai Ju +1 位作者 Hua Zhu Huafeng Li 《Computers, Materials & Continua》 SCIE EI 2025年第1期1417-1437,共21页
Traditional data-driven fault diagnosis methods depend on expert experience to manually extract effective fault features of signals,which has certain limitations.Conversely,deep learning techniques have gained promine... Traditional data-driven fault diagnosis methods depend on expert experience to manually extract effective fault features of signals,which has certain limitations.Conversely,deep learning techniques have gained prominence as a central focus of research in the field of fault diagnosis by strong fault feature extraction ability and end-to-end fault diagnosis efficiency.Recently,utilizing the respective advantages of convolution neural network(CNN)and Transformer in local and global feature extraction,research on cooperating the two have demonstrated promise in the field of fault diagnosis.However,the cross-channel convolution mechanism in CNN and the self-attention calculations in Transformer contribute to excessive complexity in the cooperative model.This complexity results in high computational costs and limited industrial applicability.To tackle the above challenges,this paper proposes a lightweight CNN-Transformer named as SEFormer for rotating machinery fault diagnosis.First,a separable multiscale depthwise convolution block is designed to extract and integrate multiscale feature information from different channel dimensions of vibration signals.Then,an efficient self-attention block is developed to capture critical fine-grained features of the signal from a global perspective.Finally,experimental results on the planetary gearbox dataset and themotor roller bearing dataset prove that the proposed framework can balance the advantages of robustness,generalization and lightweight compared to recent state-of-the-art fault diagnosis models based on CNN and Transformer.This study presents a feasible strategy for developing a lightweight rotating machinery fault diagnosis framework aimed at economical deployment. 展开更多
关键词 CNN-Transformer separable multiscale depthwise convolution efficient self-attention fault diagnosis
在线阅读 下载PDF
Spatio-temporal prediction of groundwater vulnerability based on CNN-LSTM model with self-attention mechanism:A case study in Hetao Plain,northern China 被引量:2
6
作者 Yifu Zhao Liangping Yang +4 位作者 Hongjie Pan Yanlong Li Yongxu Shao Junxia Li Xianjun Xie 《Journal of Environmental Sciences》 2025年第7期128-142,共15页
Located in northern China,the Hetao Plain is an important agro-economic zone and population centre.The deterioration of local groundwater quality has had a serious impact on human health and economic development.Nowad... Located in northern China,the Hetao Plain is an important agro-economic zone and population centre.The deterioration of local groundwater quality has had a serious impact on human health and economic development.Nowadays,the groundwater vulnerability assessment(GVA)has become an essential task to identify the current status and development trend of groundwater quality.In this study,the Convolutional Neural Network(CNN)and Long Short-Term Memory(LSTM)models are integrated to realize the spatio-temporal prediction of regional groundwater vulnerability by introducing the Self-attention mechanism.The study firstly builds the CNN-LSTM modelwith self-attention(SA)mechanism and evaluates the prediction accuracy of the model for groundwater vulnerability compared to other common machine learning models such as Support Vector Machine(SVM),Random Forest(RF),and Extreme Gradient Boosting(XGBoost).The results indicate that the CNNLSTM model outperforms thesemodels,demonstrating its significance in groundwater vulnerability assessment.It can be posited that the predictions indicate an increased risk of groundwater vulnerability in the study area over the coming years.This increase can be attributed to the synergistic impact of global climate anomalies and intensified local human activities.Moreover,the overall groundwater vulnerability risk in the entire region has increased,evident fromboth the notably high value and standard deviation.This suggests that the spatial variability of groundwater vulnerability in the area is expected to expand in the future due to the sustained progression of climate change and human activities.The model can be optimized for diverse applications across regional environmental assessment,pollution prediction,and risk statistics.This study holds particular significance for ecological protection and groundwater resource management. 展开更多
关键词 Groundwater vulnerability assessment Convolutional Neural Network Long Short-Term Memory self-attention mechanism
原文传递
The brief self-attention module for lightweight convolution neural networks
7
作者 YAN Jie WEI Yingmei +3 位作者 XIE Yuxiang GONG Quanzhi ZOU Shiwei LUAN Xidao 《Journal of Systems Engineering and Electronics》 2025年第6期1389-1397,共9页
Lightweight convolutional neural networks(CNNs)have simple structures but struggle to comprehensively and accurately extract important semantic information from images.While attention mechanisms can enhance CNNs by le... Lightweight convolutional neural networks(CNNs)have simple structures but struggle to comprehensively and accurately extract important semantic information from images.While attention mechanisms can enhance CNNs by learning distinctive representations,most existing spatial and hybrid attention methods focus on local regions with extensive parameters,making them unsuitable for lightweight CNNs.In this paper,we propose a self-attention mechanism tailored for lightweight networks,namely the brief self-attention module(BSAM).BSAM consists of the brief spatial attention(BSA)and advanced channel attention blocks.Unlike conventional self-attention methods with many parameters,our BSA block improves the performance of lightweight networks by effectively learning global semantic representations.Moreover,BSAM can be seamlessly integrated into lightweight CNNs for end-to-end training,maintaining the network’s lightweight and mobile characteristics.We validate the effectiveness of the proposed method on image classification tasks using the Food-101,Caltech-256,and Mini-ImageNet datasets. 展开更多
关键词 self-attention lightweight neural network deep learning
在线阅读 下载PDF
A Novel Dynamic Residual Self-Attention Transfer Adaptive Learning Fusion Approach for Brain Tumor Diagnosis
8
作者 Tawfeeq Shawly Ahmed A.Alsheikhy 《Computers, Materials & Continua》 2025年第3期4161-4179,共19页
A healthy brain is vital to every person since the brain controls every movement and emotion.Sometimes,some brain cells grow unexpectedly to be uncontrollable and cancerous.These cancerous cells are called brain tumor... A healthy brain is vital to every person since the brain controls every movement and emotion.Sometimes,some brain cells grow unexpectedly to be uncontrollable and cancerous.These cancerous cells are called brain tumors.For diagnosed patients,their lives depend mainly on the early diagnosis of these tumors to provide suitable treatment plans.Nowadays,Physicians and radiologists rely on Magnetic Resonance Imaging(MRI)pictures for their clinical evaluations of brain tumors.These evaluations are time-consuming,expensive,and require expertise with high skills to provide an accurate diagnosis.Scholars and industrials have recently partnered to implement automatic solutions to diagnose the disease with high accuracy.Due to their accuracy,some of these solutions depend on deep-learning(DL)methodologies.These techniques have become important due to their roles in the diagnosis process,which includes identification and classification.Therefore,there is a need for a solid and robust approach based on a deep-learning method to diagnose brain tumors.The purpose of this study is to develop an intelligent automatic framework for brain tumor diagnosis.The proposed solution is based on a novel dense dynamic residual self-attention transfer adaptive learning fusion approach(NDDRSATALFA),carried over two implemented deep-learning networks:VGG19 and UNET to identify and classify brain tumors.In addition,this solution applies a transfer learning approach to exchange extracted features and data within the two neural networks.The presented framework is trained,validated,and tested on six public datasets of MRIs to detect brain tumors and categorize these tumors into three suitable classes,which are glioma,meningioma,and pituitary.The proposed framework yielded remarkable findings on variously evaluated performance indicators:99.32%accuracy,98.74%sensitivity,98.89%specificity,99.01%Dice,98.93%Area Under the Curve(AUC),and 99.81%F1-score.In addition,a comparative analysis with recent state-of-the-art methods was performed and according to the comparative analysis,NDDRSATALFA shows an admirable level of reliability in simplifying the timely identification of diverse brain tumors.Moreover,this framework can be applied by healthcare providers to assist radiologists,pathologists,and physicians in their evaluations.The attained outcomes open doors for advanced automatic solutions that improve clinical evaluations and provide reasonable treatment plans. 展开更多
关键词 Brain tumor deep learning transfer learning RESIDUAL self-attention VGG19 UNET
在线阅读 下载PDF
EFI-SATL:An Efficient Net and Self-Attention Based Biometric Recognition for Finger-Vein Using Deep Transfer Learning
9
作者 Manjit Singh Sunil Kumar Singla 《Computer Modeling in Engineering & Sciences》 2025年第3期3003-3029,共27页
Deep Learning-based systems for Finger vein recognition have gained rising attention in recent years due to improved efficiency and enhanced security.The performance of existing CNN-based methods is limited by the pun... Deep Learning-based systems for Finger vein recognition have gained rising attention in recent years due to improved efficiency and enhanced security.The performance of existing CNN-based methods is limited by the puny generalization of learned features and deficiency of the finger vein image training data.Considering the concerns of existing methods,in this work,a simplified deep transfer learning-based framework for finger-vein recognition is developed using an EfficientNet model of deep learning with a self-attention mechanism.Data augmentation using various geometrical methods is employed to address the problem of training data shortage required for a deep learning model.The proposed model is tested using K-fold cross-validation on three publicly available datasets:HKPU,FVUSM,and SDUMLA.Also,the developed network is compared with other modern deep nets to check its effectiveness.In addition,a comparison of the proposed method with other existing Finger vein recognition(FVR)methods is also done.The experimental results exhibited superior recognition accuracy of the proposed method compared to other existing methods.In addition,the developed method proves to be more effective and less sophisticated at extracting robust features.The proposed EffAttenNet achieves an accuracy of 98.14%on HKPU,99.03%on FVUSM,and 99.50%on SDUMLA databases. 展开更多
关键词 Biometrics finger-vein recognition(FVR) deep net self-attention Efficient Nets transfer learning
在线阅读 下载PDF
Image compressed sensing reconstruction network based on self-attention mechanism
10
作者 LIU Yuhong LIU Xiaoyan CHEN Manyin 《Journal of Measurement Science and Instrumentation》 2025年第4期537-546,共10页
For image compression sensing reconstruction,most algorithms use the method of reconstructing image blocks one by one and stacking many convolutional layers,which usually have defects of obvious block effects,high com... For image compression sensing reconstruction,most algorithms use the method of reconstructing image blocks one by one and stacking many convolutional layers,which usually have defects of obvious block effects,high computational complexity,and long reconstruction time.An image compressed sensing reconstruction network based on self-attention mechanism(SAMNet)was proposed.For the compressed sampling,self-attention convolution was designed,which was conducive to capturing richer features,so that the compressed sensing measurement value retained more image structure information.For the reconstruction,a self-attention mechanism was introduced in the convolutional neural network.A reconstruction network including residual blocks,bottleneck transformer(BoTNet),and dense blocks was proposed,which strengthened the transfer of image features and reduced the amount of parameters dramatically.Under the Set5 dataset,when the measurement rates are 0.01,0.04,0.10,and 0.25,the average peak signal-to-noise ratio(PSNR)of SAMNet is improved by 1.27,1.23,0.50,and 0.15 dB,respectively,compared to the CSNet+.The running time of reconstructing a 256×256 image is reduced by 0.1473,0.1789,0.2310,and 0.2524 s compared to ReconNet.Experimental results showed that SAMNet improved the quality of reconstructed images and reduced the reconstruction time. 展开更多
关键词 convolutional neural network compressed sensing self-attention mechanism dense block image reconstruction
在线阅读 下载PDF
A precise magnetic modeling method for scientific satellites based on a self-attention mechanism and Kolmogorov-Arnold Networks
11
作者 Ye Liu Xingjian Shi +2 位作者 Wenzhe Yang Zhiming Cai Huawang Li 《Astronomical Techniques and Instruments》 2025年第1期1-9,共9页
As the complexity of scientific satellite missions increases,the requirements for their magnetic fields,magnetic field fluctuations,and even magnetic field gradients and variations become increasingly stringent.Additi... As the complexity of scientific satellite missions increases,the requirements for their magnetic fields,magnetic field fluctuations,and even magnetic field gradients and variations become increasingly stringent.Additionally,there is a growing need to address the alternating magnetic fields produced by the spacecraft itself.This paper introduces a novel modeling method for spacecraft magnetic dipoles using an integrated self-attention mechanism and a transformer combined with Kolmogorov-Arnold Networks.The self-attention mechanism captures correlations among globally sparse data,establishing dependencies b.etween sparse magnetometer readings.Concurrently,the Kolmogorov-Arnold Network,proficient in modeling implicit numerical relationships between data features,enhances the ability to learn subtle patterns.Comparative experiments validate the capability of the proposed method to precisely model magnetic dipoles,achieving maximum Root Mean Square Errors of 24.06 mA·m^(2)and 0.32 cm for size and location modeling,respectively.The spacecraft magnetic model established using this method accurately computes magnetic fields and alternating magnetic fields at designated surfaces or points.This approach facilitates the rapid and precise construction of individual and complete spacecraft magnetic models,enabling the verification of magnetic specifications from the spacecraft design phase. 展开更多
关键词 Magnetic dipole model self-attention mechanism Kolmogorov-Arnold networks Alternating current magnetic fields
在线阅读 下载PDF
Dual Self-attention Fusion Message Neural Network for Virtual Screening in Drug Discovery by Molecular Property Prediction
12
作者 Jingjing Wang Kangming Hou +2 位作者 Hao Chen Jing Fang Hongzhen Li 《Journal of Bionic Engineering》 2025年第1期354-369,共16页
The development of deep learning has made non-biochemical methods for molecular property prediction screening a reality,which can increase the experimental speed and reduce the experimental cost of relevant experiment... The development of deep learning has made non-biochemical methods for molecular property prediction screening a reality,which can increase the experimental speed and reduce the experimental cost of relevant experiments.There are currently two main approaches to representing molecules:(a)representing molecules by fixing molecular descriptors,and(b)representing molecules by graph convolutional neural networks.Currently,both of these Representative methods have achieved some results in their respective experiments.Based on past efforts,we propose a Dual Self-attention Fusion Message Neural Network(DSFMNN).DSFMNN uses a combination of dual self-attention mechanism and graph convolutional neural network.Advantages of DSFMNN:(1)The dual self-attention mechanism focuses not only on the relationship between individual subunits in a molecule but also on the relationship between the atoms and chemical bonds contained in each subunit.(2)On the directed molecular graph,a message delivery approach centered on directed molecular bonds is used.We test the performance of the model on eight publicly available datasets and compare the performance with several models.Based on the current experimental results,DSFMNN has superior performance compared to previous models on the datasets applied in this paper. 展开更多
关键词 Directed message passing network Deep learning Molecular property prediction self-attention mechanism
暂未订购
Self-attention and convolutional feature fusion for real-time intelligent fault detection of high-speed railway pantographs
13
作者 Xufeng LI Jien MA +3 位作者 Ping TAN Lanfen LIN Lin QIU Youtong FANG 《Journal of Zhejiang University-Science A(Applied Physics & Engineering)》 2025年第10期997-1009,共13页
Currently,most trains are equipped with dedicated cameras for capturing pantograph videos.Pantographs are core to the high-speed-railway pantograph-catenary system,and their failure directly affects the normal operati... Currently,most trains are equipped with dedicated cameras for capturing pantograph videos.Pantographs are core to the high-speed-railway pantograph-catenary system,and their failure directly affects the normal operation of high-speed trains.However,given the complex and variable real-world operational conditions of high-speed railways,there is no real-time and robust pantograph fault-detection method capable of handling large volumes of surveillance video.Hence,it is of paramount importance to maintain real-time monitoring and analysis of pantographs.Our study presents a real-time intelligent detection technology for identifying faults in high-speed railway pantographs,utilizing a fusion of self-attention and convolution features.We delved into lightweight multi-scale feature-extraction and fault-detection models based on deep learning to detect pantograph anomalies.Compared with traditional methods,this approach achieves high recall and accuracy in pantograph recognition,accurately pinpointing issues like discharge sparks,pantograph horns,and carbon pantograph-slide malfunctions.After experimentation and validation with actual surveillance videos of electric multiple-unit train,our algorithmic model demonstrates real-time,high-accuracy performance even under complex operational conditions. 展开更多
关键词 High-speed railway pantograph self-attention Convolutional neural network(CNN) REAL-TIME Feature fusion Faultdetection
原文传递
结合LDA与Self-Attention的短文本情感分类方法 被引量:9
14
作者 陈欢 黄勃 +2 位作者 朱翌民 俞雷 余宇新 《计算机工程与应用》 CSCD 北大核心 2020年第18期165-170,共6页
在对短文本进行情感分类任务的过程中,由于文本长度过短导致数据稀疏,降低了分类任务的准确率。针对这个问题,提出了一种基于潜在狄利克雷分布(LDA)与Self-Attention的短文本情感分类方法。使用LDA获得每个评论的主题词分布作为该条评... 在对短文本进行情感分类任务的过程中,由于文本长度过短导致数据稀疏,降低了分类任务的准确率。针对这个问题,提出了一种基于潜在狄利克雷分布(LDA)与Self-Attention的短文本情感分类方法。使用LDA获得每个评论的主题词分布作为该条评论信息的扩展,将扩展信息和原评论文本一起输入到word2vec模型,进行词向量训练,使得该评论文本在高维向量空间实现同一主题的聚类,使用Self-Attention进行动态权重分配并进行分类。通过在谭松波酒店评论数据集上的实验表明,该算法与当前主流的短文本分类情感算法相比,有效地提高了分类性能。 展开更多
关键词 主题词 短文本 self-attention 潜在狄利克雷分布(LDA) word2vec
在线阅读 下载PDF
结合TFIDF的Self-Attention-Based Bi-LSTM的垃圾短信识别 被引量:11
15
作者 吴思慧 陈世平 《计算机系统应用》 2020年第9期171-177,共7页
随着手机短信成为人们日常生活交往的重要手段,垃圾短信的识别具有重要的现实意义.针对此提出一种结合TFIDF的self-attention-based Bi-LSTM的神经网络模型.该模型首先将短信文本以词向量的方式输入到Bi-LSTM层,经过特征提取并结合TFIDF... 随着手机短信成为人们日常生活交往的重要手段,垃圾短信的识别具有重要的现实意义.针对此提出一种结合TFIDF的self-attention-based Bi-LSTM的神经网络模型.该模型首先将短信文本以词向量的方式输入到Bi-LSTM层,经过特征提取并结合TFIDF和self-attention层的信息聚焦获得最后的特征向量,最后将特征向量通过Softmax分类器进行分类得到短信文本分类结果.实验结果表明,结合TFIDF的self-attention-based Bi-LSTM模型相比于传统分类模型的短信文本识别准确率提高了2.1%–4.6%,运行时间减少了0.6 s–10.2 s. 展开更多
关键词 垃圾短信 文本分类 self-attention Bi-LSTM TFIDF
在线阅读 下载PDF
基于Self-Attention模型的机器翻译系统 被引量:10
16
作者 师岩 王宇 吴水清 《计算机与现代化》 2019年第7期9-14,共6页
近几年来神经机器翻译(Neural Machine Translation,NMT)发展迅速,Seq2Seq框架的提出为机器翻译带来了很大的优势,可以在观测到整个输入句子后生成任意输出序列。但是该模型对于长距离信息的捕获能力仍有很大的局限,循环神经网络(RNN)、... 近几年来神经机器翻译(Neural Machine Translation,NMT)发展迅速,Seq2Seq框架的提出为机器翻译带来了很大的优势,可以在观测到整个输入句子后生成任意输出序列。但是该模型对于长距离信息的捕获能力仍有很大的局限,循环神经网络(RNN)、LSTM网络都是为了改善这一问题提出的,但是效果并不明显。注意力机制的提出与运用则有效地弥补了该缺陷。Self-Attention模型就是在注意力机制的基础上提出的,本文使用Self-Attention为基础构建编码器-解码器框架。本文通过探讨以往的神经网络翻译模型,分析Self-Attention模型的机制与原理,通过TensorFlow深度学习框架对基于Self-Attention模型的翻译系统进行实现,在英文到中文的翻译实验中与以往的神经网络翻译模型进行对比,表明该模型取得了较好的翻译效果。 展开更多
关键词 神经机器翻译 Seq2Seq框架 注意力机制 self-attention模型
在线阅读 下载PDF
引入Self-Attention的电力作业违规穿戴智能检测技术研究 被引量:5
17
作者 莫蓓蓓 吴克河 《计算机与现代化》 2020年第2期115-121,126,共8页
随着电网建设的高速发展,作业现场技术支撑人员规模不断扩大。电力现场属于高危作业场所,违规穿戴安全防护用品将会严重危及作业人员的人身安全,为了改善传统人工监管方式效率低下的问题,本文采用实时深度学习算法进行违规穿戴行为检测... 随着电网建设的高速发展,作业现场技术支撑人员规模不断扩大。电力现场属于高危作业场所,违规穿戴安全防护用品将会严重危及作业人员的人身安全,为了改善传统人工监管方式效率低下的问题,本文采用实时深度学习算法进行违规穿戴行为检测。检测模型结合实时目标检测网络YOLOv3和Self-Attention机制,借鉴DANet结构,在YOLOv3网络高层嵌入自注意力模块,更好地挖掘和学习特征位置和通道关系。实验结果表明,该模型在违规穿戴检测任务中mAP达到了94.58%,Recall达到了96.67%,与YOLOv3相比,mAP提高了12.66%,Recall提高了2.69%,显著提高模型的精度,可以满足任务的检测需求,提升了电网智能化水平。 展开更多
关键词 电力作业 违规穿戴 YOLOv3技术 self-attention机制 目标检测
在线阅读 下载PDF
融合Self-Attention机制和n-gram卷积核的印尼语复合名词自动识别方法研究 被引量:2
18
作者 丘心颖 陈汉武 +3 位作者 陈源 谭立聪 张皓 肖莉娴 《湖南工业大学学报》 2020年第3期1-9,共9页
针对印尼语复合名词短语自动识别,提出一种融合Self-Attention机制、n-gram卷积核的神经网络和统计模型相结合的方法,改进现有的多词表达抽取模型。在现有SHOMA模型的基础上,使用多层CNN和Self-Attention机制进行改进。对Universal Depe... 针对印尼语复合名词短语自动识别,提出一种融合Self-Attention机制、n-gram卷积核的神经网络和统计模型相结合的方法,改进现有的多词表达抽取模型。在现有SHOMA模型的基础上,使用多层CNN和Self-Attention机制进行改进。对Universal Dependencies公开的印尼语数据进行复合名词短语自动识别的对比实验,结果表明:TextCNN+Self-Attention+CRF模型取得32.20的短语多词识别F1值和32.34的短语单字识别F1值,比SHOMA模型分别提升了4.93%和3.04%。 展开更多
关键词 印尼语复合名词短语 self-attention机制 卷积神经网络 自动识别 条件随机场
在线阅读 下载PDF
Hierarchical multihead self-attention for time-series-based fault diagnosis 被引量:3
19
作者 Chengtian Wang Hongbo Shi +1 位作者 Bing Song Yang Tao 《Chinese Journal of Chemical Engineering》 SCIE EI CAS CSCD 2024年第6期104-117,共14页
Fault diagnosis is important for maintaining the safety and effectiveness of chemical process.Considering the multivariate,nonlinear,and dynamic characteristic of chemical process,many time-series-based data-driven fa... Fault diagnosis is important for maintaining the safety and effectiveness of chemical process.Considering the multivariate,nonlinear,and dynamic characteristic of chemical process,many time-series-based data-driven fault diagnosis methods have been developed in recent years.However,the existing methods have the problem of long-term dependency and are difficult to train due to the sequential way of training.To overcome these problems,a novel fault diagnosis method based on time-series and the hierarchical multihead self-attention(HMSAN)is proposed for chemical process.First,a sliding window strategy is adopted to construct the normalized time-series dataset.Second,the HMSAN is developed to extract the time-relevant features from the time-series process data.It improves the basic self-attention model in both width and depth.With the multihead structure,the HMSAN can pay attention to different aspects of the complicated chemical process and obtain the global dynamic features.However,the multiple heads in parallel lead to redundant information,which cannot improve the diagnosis performance.With the hierarchical structure,the redundant information is reduced and the deep local time-related features are further extracted.Besides,a novel many-to-one training strategy is introduced for HMSAN to simplify the training procedure and capture the long-term dependency.Finally,the effectiveness of the proposed method is demonstrated by two chemical cases.The experimental results show that the proposed method achieves a great performance on time-series industrial data and outperforms the state-of-the-art approaches. 展开更多
关键词 self-attention mechanism Deep learning Chemical process Time-series Fault diagnosis
在线阅读 下载PDF
Aerial target threat assessment based on gated recurrent unit and self-attention mechanism 被引量:4
20
作者 CHEN Chen QUAN Wei SHAO Zhuang 《Journal of Systems Engineering and Electronics》 SCIE CSCD 2024年第2期361-373,共13页
Aerial threat assessment is a crucial link in modern air combat, whose result counts a great deal for commanders to make decisions. With the consideration that the existing threat assessment methods have difficulties ... Aerial threat assessment is a crucial link in modern air combat, whose result counts a great deal for commanders to make decisions. With the consideration that the existing threat assessment methods have difficulties in dealing with high dimensional time series target data, a threat assessment method based on self-attention mechanism and gated recurrent unit(SAGRU) is proposed. Firstly, a threat feature system including air combat situations and capability features is established. Moreover, a data augmentation process based on fractional Fourier transform(FRFT) is applied to extract more valuable information from time series situation features. Furthermore, aiming to capture key characteristics of battlefield evolution, a bidirectional GRU and SA mechanisms are designed for enhanced features.Subsequently, after the concatenation of the processed air combat situation and capability features, the target threat level will be predicted by fully connected neural layers and the softmax classifier. Finally, in order to validate this model, an air combat dataset generated by a combat simulation system is introduced for model training and testing. The comparison experiments show the proposed model has structural rationality and can perform threat assessment faster and more accurately than the other existing models based on deep learning. 展开更多
关键词 target threat assessment gated recurrent unit(GRU) self-attention(SA) fractional Fourier transform(FRFT)
在线阅读 下载PDF
上一页 1 2 11 下一页 到第
使用帮助 返回顶部