期刊文献+
共找到399篇文章
< 1 2 20 >
每页显示 20 50 100
DA-ViT:Deformable Attention Vision Transformer for Alzheimer’s Disease Classification from MRI Scans
1
作者 Abdullah G.M.Almansour Faisal Alshomrani +4 位作者 Abdulaziz T.M.Almutairi Easa Alalwany Mohammed S.Alshuhri Hussein Alshaari Abdullah Alfahaid 《Computer Modeling in Engineering & Sciences》 2025年第8期2395-2418,共24页
The early and precise identification of Alzheimer’s Disease(AD)continues to pose considerable clinical difficulty due to subtle structural alterations and overlapping symptoms across the disease phases.This study pre... The early and precise identification of Alzheimer’s Disease(AD)continues to pose considerable clinical difficulty due to subtle structural alterations and overlapping symptoms across the disease phases.This study presents a novel Deformable Attention Vision Transformer(DA-ViT)architecture that integrates deformable Multi-Head Self-Attention(MHSA)with a Multi-Layer Perceptron(MLP)block for efficient classification of Alzheimer’s disease(AD)using Magnetic resonance imaging(MRI)scans.In contrast to traditional vision transformers,our deformable MHSA module preferentially concentrates on spatially pertinent patches through learned offset predictions,markedly diminishing processing demands while improving localized feature representation.DA-ViT contains only 0.93 million parameters,making it exceptionally suitable for implementation in resource-limited settings.We evaluate the model using a class-imbalanced Alzheimer’s MRI dataset comprising 6400 images across four categories,achieving a test accuracy of 80.31%,a macro F1-score of 0.80,and an area under the receiver operating characteristic curve(AUC)of 1.00 for the Mild Demented category.Thorough ablation studies validate the ideal configuration of transformer depth,headcount,and embedding dimensions.Moreover,comparison research indicates that DA-ViT surpasses state-of-theart pre-trained Convolutional Neural Network(CNN)models in terms of accuracy and parameter efficiency. 展开更多
关键词 Alzheimer disease classification vision transformer deformable attention MRI analysis bayesian optimization
在线阅读 下载PDF
Interactive Dynamic Graph Convolution with Temporal Attention for Traffic Flow Forecasting
2
作者 Zitong Zhao Zixuan Zhang Zhenxing Niu 《Computers, Materials & Continua》 2026年第1期1049-1064,共16页
Reliable traffic flow prediction is crucial for mitigating urban congestion.This paper proposes Attentionbased spatiotemporal Interactive Dynamic Graph Convolutional Network(AIDGCN),a novel architecture integrating In... Reliable traffic flow prediction is crucial for mitigating urban congestion.This paper proposes Attentionbased spatiotemporal Interactive Dynamic Graph Convolutional Network(AIDGCN),a novel architecture integrating Interactive Dynamic Graph Convolution Network(IDGCN)with Temporal Multi-Head Trend-Aware Attention.Its core innovation lies in IDGCN,which uniquely splits sequences into symmetric intervals for interactive feature sharing via dynamic graphs,and a novel attention mechanism incorporating convolutional operations to capture essential local traffic trends—addressing a critical gap in standard attention for continuous data.For 15-and 60-min forecasting on METR-LA,AIDGCN achieves MAEs of 0.75%and 0.39%,and RMSEs of 1.32%and 0.14%,respectively.In the 60-min long-term forecasting of the PEMS-BAY dataset,the AIDGCN out-performs the MRA-BGCN method by 6.28%,4.93%,and 7.17%in terms of MAE,RMSE,and MAPE,respectively.Experimental results demonstrate the superiority of our pro-posed model over state-of-the-art methods. 展开更多
关键词 Traffic flow prediction interactive dynamic graph convolution graph convolution temporal multi-head trend-aware attention self-attention mechanism
在线阅读 下载PDF
Event-Aware Sarcasm Detection in Chinese Social Media Using Multi-Head Attention and Contrastive Learning
3
作者 Kexuan Niu Xiameng Si +1 位作者 Xiaojie Qi Haiyan Kang 《Computers, Materials & Continua》 2025年第10期2051-2070,共20页
Sarcasm detection is a complex and challenging task,particularly in the context of Chinese social media,where it exhibits strong contextual dependencies and cultural specificity.To address the limitations of existing ... Sarcasm detection is a complex and challenging task,particularly in the context of Chinese social media,where it exhibits strong contextual dependencies and cultural specificity.To address the limitations of existing methods in capturing the implicit semantics and contextual associations in sarcastic expressions,this paper proposes an event-aware model for Chinese sarcasm detection,leveraging a multi-head attention(MHA)mechanism and contrastive learning(CL)strategies.The proposed model employs a dual-path Bidirectional Encoder Representations from Transformers(BERT)encoder to process comment text and event context separately and integrates an MHA mechanism to facilitate deep interactions between the two,thereby capturing multidimensional semantic associations.Additionally,a CL strategy is introduced to enhance feature representation capabilities,further improving the model’s performance in handling class imbalance and complex contextual scenarios.The model achieves state-of-the-art performance on the Chinese sarcasm dataset,with significant improvements in accuracy(79.55%),F1-score(84.22%),and an area under the curve(AUC,84.35%). 展开更多
关键词 Sarcasm detection event-aware multi-head attention contrastive learning NLP
在线阅读 下载PDF
Self-reduction multi-head attention module for defect recognition of power equipment in substation
4
作者 Yifeng Han Donglian Qi Yunfeng Yan 《Global Energy Interconnection》 2025年第1期82-91,共10页
Safety maintenance of power equipment is of great importance in power grids,in which image-processing-based defect recognition is supposed to classify abnormal conditions during daily inspection.However,owing to the b... Safety maintenance of power equipment is of great importance in power grids,in which image-processing-based defect recognition is supposed to classify abnormal conditions during daily inspection.However,owing to the blurred features of defect images,the current defect recognition algorithm has poor fine-grained recognition ability.Visual attention can achieve fine-grained recognition with its abil-ity to model long-range dependencies while introducing extra computational complexity,especially for multi-head attention in vision transformer structures.Under these circumstances,this paper proposes a self-reduction multi-head attention module that can reduce computational complexity and be easily combined with a Convolutional Neural Network(CNN).In this manner,local and global fea-tures can be calculated simultaneously in our proposed structure,aiming to improve the defect recognition performance.Specifically,the proposed self-reduction multi-head attention can reduce redundant parameters,thereby solving the problem of limited computational resources.Experimental results were obtained based on the defect dataset collected from the substation.The results demonstrated the efficiency and superiority of the proposed method over other advanced algorithms. 展开更多
关键词 multi-head attention Defect recognition Power equipment Computational complexity
在线阅读 下载PDF
Multi-Head Attention Enhanced Parallel Dilated Convolution and Residual Learning for Network Traffic Anomaly Detection
5
作者 Guorong Qi Jian Mao +2 位作者 Kai Huang Zhengxian You Jinliang Lin 《Computers, Materials & Continua》 2025年第2期2159-2176,共18页
Abnormal network traffic, as a frequent security risk, requires a series of techniques to categorize and detect it. Existing network traffic anomaly detection still faces challenges: the inability to fully extract loc... Abnormal network traffic, as a frequent security risk, requires a series of techniques to categorize and detect it. Existing network traffic anomaly detection still faces challenges: the inability to fully extract local and global features, as well as the lack of effective mechanisms to capture complex interactions between features;Additionally, when increasing the receptive field to obtain deeper feature representations, the reliance on increasing network depth leads to a significant increase in computational resource consumption, affecting the efficiency and performance of detection. Based on these issues, firstly, this paper proposes a network traffic anomaly detection model based on parallel dilated convolution and residual learning (Res-PDC). To better explore the interactive relationships between features, the traffic samples are converted into two-dimensional matrix. A module combining parallel dilated convolutions and residual learning (res-pdc) was designed to extract local and global features of traffic at different scales. By utilizing res-pdc modules with different dilation rates, we can effectively capture spatial features at different scales and explore feature dependencies spanning wider regions without increasing computational resources. Secondly, to focus and integrate the information in different feature subspaces, further enhance and extract the interactions among the features, multi-head attention is added to Res-PDC, resulting in the final model: multi-head attention enhanced parallel dilated convolution and residual learning (MHA-Res-PDC) for network traffic anomaly detection. Finally, comparisons with other machine learning and deep learning algorithms are conducted on the NSL-KDD and CIC-IDS-2018 datasets. The experimental results demonstrate that the proposed method in this paper can effectively improve the detection performance. 展开更多
关键词 Network traffic anomaly detection multi-head attention parallel dilated convolution residual learning
在线阅读 下载PDF
MAMGBR: Group-Buying Recommendation Model Based on Multi-Head Attention Mechanism and Multi-Task Learning
6
作者 Zongzhe Xu Ming Yu 《Computers, Materials & Continua》 2025年第8期2805-2826,共22页
As the group-buying model shows significant progress in attracting new users,enhancing user engagement,and increasing platform profitability,providing personalized recommendations for group-buying users has emerged as... As the group-buying model shows significant progress in attracting new users,enhancing user engagement,and increasing platform profitability,providing personalized recommendations for group-buying users has emerged as a new challenge in the field of recommendation systems.This paper introduces a group-buying recommendation model based on multi-head attention mechanisms and multi-task learning,termed the Multi-head Attention Mechanisms and Multi-task Learning Group-Buying Recommendation(MAMGBR)model,specifically designed to optimize group-buying recommendations on e-commerce platforms.The core dataset of this study comes from the Chinese maternal and infant e-commerce platform“Beibei,”encompassing approximately 430,000 successful groupbuying actions and over 120,000 users.Themodel focuses on twomain tasks:recommending items for group organizers(Task Ⅰ)and recommending participants for a given group-buying event(Task Ⅱ).In model evaluation,MAMGBR achieves an MRR@10 of 0.7696 for Task I,marking a 20.23%improvement over baseline models.Furthermore,in Task II,where complex interaction patterns prevail,MAMGBR utilizes auxiliary loss functions to effectively model the multifaceted roles of users,items,and participants,leading to a 24.08%increase in MRR@100 under a 1:99 sample ratio.Experimental results show that compared to benchmark models,such as NGCF and EATNN,MAMGBR’s integration ofmulti-head attentionmechanisms,expert networks,and gating mechanisms enables more accurate modeling of user preferences and social associations within group-buying scenarios,significantly enhancing recommendation accuracy and platform group-buying success rates. 展开更多
关键词 Group-buying recommendation multi-head attention mechanism multi-task learning
在线阅读 下载PDF
SSA-LSTM-Multi-Head Attention Modelling Approach for Prediction of Coal Dust Maximum Explosion Pressure Based on the Synergistic Effect of Particle Size and Concentration
7
作者 Yongli Liu Weihao Li +1 位作者 Haitao Wang Taoren Du 《Computer Modeling in Engineering & Sciences》 2025年第5期2261-2286,共26页
Coal dust explosions are severe safety accidents in coal mine production,posing significant threats to life and property.Predicting the maximum explosion pressure(Pm)of coal dust using deep learning models can effecti... Coal dust explosions are severe safety accidents in coal mine production,posing significant threats to life and property.Predicting the maximum explosion pressure(Pm)of coal dust using deep learning models can effectively assess potential risks and provide a scientific basis for preventing coal dust explosions.In this study,a 20-L explosion sphere apparatus was used to test the maximum explosion pressure of coal dust under seven different particle sizes and ten mass concentrations(Cdust),resulting in a dataset of 70 experimental groups.Through Spearman correlation analysis and random forest feature selection methods,particle size(D_(10),D_(20),D_(50))and mass concentration(Cdust)were identified as critical feature parameters from the ten initial parameters of the coal dust samples.Based on this,a hybrid Long Short-Term Memory(LSTM)network model incorporating a Multi-Head Attention Mechanism and the Sparrow Search Algorithm(SSA)was proposed to predict the maximum explosion pressure of coal dust.The results demonstrate that the SSA-LSTM-Multi-Head Attention model excels in predicting the maximum explosion pressure of coal dust.The four evaluation metrics indicate that the model achieved a coefficient of determination(R^(2)),root mean square error(RMSE),mean absolute percentage error(MAPE),and mean absolute error(MAE)of 0.9841,0.0030,0.0074,and 0.0049,respectively,in the training set.In the testing set,these values were 0.9743,0.0087,0.0108,and 0.0069,respectively.Compared to artificial neural networks(ANN),random forest(RF),support vector machines(SVM),particle swarm optimized-SVM(PSO-SVM)neural networks,and the traditional single-model LSTM,the SSA-LSTM-Multi-Head Attention model demonstrated superior generalization capability and prediction accuracy.The findings of this study not only advance the application of deep learning in coal dust explosion prediction but also provide robust technical support for the prevention and risk assessment of coal dust explosions. 展开更多
关键词 Coal dust explosion deep learning maximum explosion pressure predictive model SSA-LSTM multi-head attention mechanism
在线阅读 下载PDF
A local-global dynamic hypergraph convolution with multi-head flow attention for traffic flow forecasting
8
作者 ZHANG Hong LI Yang +3 位作者 LUO Shengjun ZHANG Pengcheng ZHANG Xijun YI Min 《High Technology Letters》 2025年第3期246-256,共11页
Traffic flow prediction is a crucial element of intelligent transportation systems.However,accu-rate traffic flow prediction is quite challenging because of its highly nonlinear,complex,and dynam-ic characteristics.To... Traffic flow prediction is a crucial element of intelligent transportation systems.However,accu-rate traffic flow prediction is quite challenging because of its highly nonlinear,complex,and dynam-ic characteristics.To address the difficulties in simultaneously capturing local and global dynamic spatiotemporal correlations in traffic flow,as well as the high time complexity of existing models,a multi-head flow attention-based local-global dynamic hypergraph convolution(MFA-LGDHC)pre-diction model is proposed.which consists of multi-head flow attention(MHFA)mechanism,graph convolution network(GCN),and local-global dynamic hypergraph convolution(LGHC).MHFA is utilized to extract the time dependency of traffic flow and reduce the time complexity of the model.GCN is employed to catch the spatial dependency of traffic flow.LGHC utilizes down-sampling con-volution and isometric convolution to capture the local and global spatial dependencies of traffic flow.And dynamic hypergraph convolution is used to model the dynamic higher-order relationships of the traffic road network.Experimental results indicate that the MFA-LGDHC model outperforms current popular baseline models and exhibits good prediction performance. 展开更多
关键词 traffic flow prediction multi-head flow attention graph convolution hypergraph learning dynamic spatio-temporal properties
在线阅读 下载PDF
A Deformable Network with Attention Mechanism for Retinal Vessel Segmentation
9
作者 Xiaolong Zhu Wenjian Li +2 位作者 Weihang Zhang Dongwei Li Huiqi Li 《Journal of Beijing Institute of Technology》 EI CAS 2024年第3期186-193,共8页
The intensive application of deep learning in medical image processing has facilitated the advancement of automatic retinal vessel segmentation research.To overcome the limitation that traditional U-shaped vessel segm... The intensive application of deep learning in medical image processing has facilitated the advancement of automatic retinal vessel segmentation research.To overcome the limitation that traditional U-shaped vessel segmentation networks fail to extract features in fundus image sufficiently,we propose a novel network(DSeU-net)based on deformable convolution and squeeze excitation residual module.The deformable convolution is utilized to dynamically adjust the receptive field for the feature extraction of retinal vessel.And the squeeze excitation residual module is used to scale the weights of the low-level features so that the network learns the complex relationships of the different feature layers efficiently.We validate the DSeU-net on three public retinal vessel segmentation datasets including DRIVE,CHASEDB1,and STARE,and the experimental results demonstrate the satisfactory segmentation performance of the network. 展开更多
关键词 retinal vessel segmentation deformable convolution attention mechanism deep learning
暂未订购
基于Multi-head Attention和Bi-LSTM的实体关系分类 被引量:12
10
作者 刘峰 高赛 +1 位作者 于碧辉 郭放达 《计算机系统应用》 2019年第6期118-124,共7页
关系分类是自然语言处理领域的一项重要任务,能够为知识图谱的构建、问答系统和信息检索等提供技术支持.与传统关系分类方法相比较,基于神经网络和注意力机制的关系分类模型在各种关系分类任务中都获得了更出色的表现.以往的模型大多采... 关系分类是自然语言处理领域的一项重要任务,能够为知识图谱的构建、问答系统和信息检索等提供技术支持.与传统关系分类方法相比较,基于神经网络和注意力机制的关系分类模型在各种关系分类任务中都获得了更出色的表现.以往的模型大多采用单层注意力机制,特征表达相对单一.因此本文在已有研究基础上,引入多头注意力机制(Multi-head attention),旨在让模型从不同表示空间上获取关于句子更多层面的信息,提高模型的特征表达能力.同时在现有的词向量和位置向量作为网络输入的基础上,进一步引入依存句法特征和相对核心谓词依赖特征,其中依存句法特征包括当前词的依存关系值和所依赖的父节点位置,从而使模型进一步获取更多的文本句法信息.在SemEval-2010 任务8 数据集上的实验结果证明,该方法相较之前的深度学习模型,性能有进一步提高. 展开更多
关键词 关系分类 Bi-LSTM 句法特征 self-attention multi-head attention
在线阅读 下载PDF
基于VMD-Self-attention-LSTM的水闸深基坑变形智能预测方法 被引量:2
11
作者 张伟 仇建春 +5 位作者 夏国春 姚兆仁 吴昊 刘占午 王昱锦 朱新宇 《水电能源科学》 北大核心 2025年第1期99-102,196,共5页
针对水闸深基坑变形监测数据具有非稳定性的特点,提出了基于VMD-Self-attention-LSTM的水闸深基坑变形预测方法。该方法分为三个主要模块,第一模块采用VMD算法自适应调整变形分解模数,将原始变形数据分解为若干具有明显周期规律的分量,... 针对水闸深基坑变形监测数据具有非稳定性的特点,提出了基于VMD-Self-attention-LSTM的水闸深基坑变形预测方法。该方法分为三个主要模块,第一模块采用VMD算法自适应调整变形分解模数,将原始变形数据分解为若干具有明显周期规律的分量,有效解决变形数据的非稳定性,为提升变形预测精度奠定初步基础;第二模块,在传统LSTM算法基础上进一步发展Self-attention-LSTM方法,提升模型对基坑变形样本的时序关系挖掘能力,进而提升基坑变形预测精度;第三模块,将各分量对应的变形预测结果重构得到最终预测值。实例分析可知,所提方法有效解决了非稳定特性给变形预测精度带来的不利影响,与VMD-LSTM、Self-attention-LSTM、LSTM等深度学习方法相比,VMD-Self-attention-LSTM的预测精度最大提升41.49%,与BP、ELM等传统机器学习算法相比,预测精度最大提升50.43%,为水闸深基坑安全监控模型的构建提供了新思路。 展开更多
关键词 水闸深基坑 变形预测 VMD Self-attention-LSTM
原文传递
Application of the improved dung beetle optimizer,muti-head attention and hybrid deep learning algorithms to groundwater depth prediction in the Ningxia area,China 被引量:1
12
作者 Jiarui Cai Bo Sun +5 位作者 Huijun Wang Yi Zheng Siyu Zhou Huixin Li Yanyan Huang Peishu Zong 《Atmospheric and Oceanic Science Letters》 2025年第1期18-23,共6页
Due to the lack of accurate data and complex parameterization,the prediction of groundwater depth is a chal-lenge for numerical models.Machine learning can effectively solve this issue and has been proven useful in th... Due to the lack of accurate data and complex parameterization,the prediction of groundwater depth is a chal-lenge for numerical models.Machine learning can effectively solve this issue and has been proven useful in the prediction of groundwater depth in many areas.In this study,two new models are applied to the prediction of groundwater depth in the Ningxia area,China.The two models combine the improved dung beetle optimizer(DBO)algorithm with two deep learning models:The Multi-head Attention-Convolution Neural Network-Long Short Term Memory networks(MH-CNN-LSTM)and the Multi-head Attention-Convolution Neural Network-Gated Recurrent Unit(MH-CNN-GRU).The models with DBO show better prediction performance,with larger R(correlation coefficient),RPD(residual prediction deviation),and lower RMSE(root-mean-square error).Com-pared with the models with the original DBO,the R and RPD of models with the improved DBO increase by over 1.5%,and the RMSE decreases by over 1.8%,indicating better prediction results.In addition,compared with the multiple linear regression model,a traditional statistical model,deep learning models have better prediction performance. 展开更多
关键词 Groundwater depth multi-head attention Improved dung beetle optimizer CNN-LSTM CNN-GRU Ningxia
在线阅读 下载PDF
基于CNN-BiLSTM-Attention的深基坑变形预测方法 被引量:2
13
作者 孟飞 郑卓然 +2 位作者 黄文聪 岳学军 张伟锋 《地下空间与工程学报》 北大核心 2025年第S1期87-94,共8页
深基坑变形的准确预测一直是基坑工程面临的关键问题之一,地下空间的复杂性和基坑施工环境的多样性使得传统的预测方法在应对这一问题时显得力不从心。为了提高深基坑变形预测的精准度,提出一种结合卷积神经网络(CNN)、双向长短期记忆... 深基坑变形的准确预测一直是基坑工程面临的关键问题之一,地下空间的复杂性和基坑施工环境的多样性使得传统的预测方法在应对这一问题时显得力不从心。为了提高深基坑变形预测的精准度,提出一种结合卷积神经网络(CNN)、双向长短期记忆神经网络(BiLSTM)和注意力机制(Attention)的深基坑变形预测模型。通过构建时空网格,利用卷积神经网络提取基坑变形的空间特征,结合双向长短时记忆网络对时序特征进行建模,引入注意力机制提高模型对关键时空位置的关注度,最后通过全连接层整合特征,输出预测的监测值。基于广州市某人才公寓深基坑的监测数据进行工程案例验证,通过消融试验和对比试验结果表明,所提方法在深基坑变形预测中精度较高。 展开更多
关键词 深基坑 变形预测 卷积神经网络(CNN) 双向长短期记忆神经网络(BiLSTM) 注意力机制
原文传递
DCA-YOLO:Detection Algorithm for YOLOv8 Pulmonary Nodules Based on Attention Mechanism Optimization 被引量:1
14
作者 SONG Yongsheng LIU Guohua 《Journal of Donghua University(English Edition)》 2025年第1期78-87,共10页
Pulmonary nodules represent an early manifestation of lung cancer.However,pulmonary nodules only constitute a small portion of the overall image,posing challenges for physicians in image interpretation and potentially... Pulmonary nodules represent an early manifestation of lung cancer.However,pulmonary nodules only constitute a small portion of the overall image,posing challenges for physicians in image interpretation and potentially leading to false positives or missed detections.To solve these problems,the YOLOv8 network is enhanced by adding deformable convolution and atrous spatial pyramid pooling(ASPP),along with the integration of a coordinate attention(CA)mechanism.This allows the network to focus on small targets while expanding the receptive field without losing resolution.At the same time,context information on the target is gathered and feature expression is enhanced by attention modules in different directions.It effectively improves the positioning accuracy and achieves good results on the LUNA16 dataset.Compared with other detection algorithms,it improves the accuracy of pulmonary nodule detection to a certain extent. 展开更多
关键词 pulmonary nodule YOLOv8 network object detection deformable convolution atrous spatial pyramid pooling(ASPP) coordinate attention(CA)mechanism
在线阅读 下载PDF
An Intelligent Framework for Resilience Recovery of FANETs with Spatio-Temporal Aggregation and Multi-Head Attention Mechanism 被引量:1
15
作者 Zhijun Guo Yun Sun +2 位作者 YingWang Chaoqi Fu Jilong Zhong 《Computers, Materials & Continua》 SCIE EI 2024年第5期2375-2398,共24页
Due to the time-varying topology and possible disturbances in a conflict environment,it is still challenging to maintain the mission performance of flying Ad hoc networks(FANET),which limits the application of Unmanne... Due to the time-varying topology and possible disturbances in a conflict environment,it is still challenging to maintain the mission performance of flying Ad hoc networks(FANET),which limits the application of Unmanned Aerial Vehicle(UAV)swarms in harsh environments.This paper proposes an intelligent framework to quickly recover the cooperative coveragemission by aggregating the historical spatio-temporal network with the attention mechanism.The mission resilience metric is introduced in conjunction with connectivity and coverage status information to simplify the optimization model.A spatio-temporal node pooling method is proposed to ensure all node location features can be updated after destruction by capturing the temporal network structure.Combined with the corresponding Laplacian matrix as the hyperparameter,a recovery algorithm based on the multi-head attention graph network is designed to achieve rapid recovery.Simulation results showed that the proposed framework can facilitate rapid recovery of the connectivity and coverage more effectively compared to the existing studies.The results demonstrate that the average connectivity and coverage results is improved by 17.92%and 16.96%,respectively compared with the state-of-the-art model.Furthermore,by the ablation study,the contributions of each different improvement are compared.The proposed model can be used to support resilient network design for real-time mission execution. 展开更多
关键词 RESILIENCE cooperative mission FANET spatio-temporal node pooling multi-head attention graph network
在线阅读 下载PDF
Using Recurrent Neural Network Structure and Multi-Head Attention with Convolution for Fraudulent Phone Text Recognition 被引量:1
16
作者 Junjie Zhou Hongkui Xu +3 位作者 Zifeng Zhang Jiangkun Lu Wentao Guo Zhenye Li 《Computer Systems Science & Engineering》 SCIE EI 2023年第8期2277-2297,共21页
Fraud cases have been a risk in society and people’s property security has been greatly threatened.In recent studies,many promising algorithms have been developed for social media offensive text recognition as well a... Fraud cases have been a risk in society and people’s property security has been greatly threatened.In recent studies,many promising algorithms have been developed for social media offensive text recognition as well as sentiment analysis.These algorithms are also suitable for fraudulent phone text recognition.Compared to these tasks,the semantics of fraudulent words are more complex and more difficult to distinguish.Recurrent Neural Networks(RNN),the variants ofRNN,ConvolutionalNeuralNetworks(CNN),and hybrid neural networks to extract text features are used by most text classification research.However,a single network or a simple network combination cannot obtain rich characteristic knowledge of fraudulent phone texts relatively.Therefore,a new model is proposed in this paper.In the fraudulent phone text,the knowledge that can be learned by the model includes the sequence structure of sentences,the correlation between words,the correlation of contextual semantics,the feature of keywords in sentences,etc.The new model combines a bidirectional Long-Short Term Memory Neural Network(BiLSTM)or a bidirectional Gate Recurrent United(BiGRU)and a Multi-Head attention mechanism module with convolution.A normalization layer is added after the output of the final hidden layer.BiLSTM or BiGRU is used to build the encoding and decoding layer.Multi-head attention mechanism module with convolution(MHAC)enhances the ability of the model to learn global interaction information and multi-granularity local interaction information in fraudulent sentences.A fraudulent phone text dataset is produced by us in this paper.The THUCNews data sets and fraudulent phone text data sets are used in experiments.Experiment results show that compared with the baseline model,the proposed model(LMHACL)has the best experiment results in terms of Accuracy,Precision,Recall,and F1 score on the two data sets.And the performance indexes on fraudulent phone text data sets are all above 0.94. 展开更多
关键词 BiLSTM BiGRU multi-head attention mechanism CNN
在线阅读 下载PDF
Multi-head attention-based long short-term memory model for speech emotion recognition 被引量:1
17
作者 Zhao Yan Zhao Li +3 位作者 Lu Cheng Li Sunan Tang Chuangao Lian Hailun 《Journal of Southeast University(English Edition)》 EI CAS 2022年第2期103-109,共7页
To fully make use of information from different representation subspaces,a multi-head attention-based long short-term memory(LSTM)model is proposed in this study for speech emotion recognition(SER).The proposed model ... To fully make use of information from different representation subspaces,a multi-head attention-based long short-term memory(LSTM)model is proposed in this study for speech emotion recognition(SER).The proposed model uses frame-level features and takes the temporal information of emotion speech as the input of the LSTM layer.Here,a multi-head time-dimension attention(MHTA)layer was employed to linearly project the output of the LSTM layer into different subspaces for the reduced-dimension context vectors.To provide relative vital information from other dimensions,the output of MHTA,the output of feature-dimension attention,and the last time-step output of LSTM were utilized to form multiple context vectors as the input of the fully connected layer.To improve the performance of multiple vectors,feature-dimension attention was employed for the all-time output of the first LSTM layer.The proposed model was evaluated on the eNTERFACE and GEMEP corpora,respectively.The results indicate that the proposed model outperforms LSTM by 14.6%and 10.5%for eNTERFACE and GEMEP,respectively,proving the effectiveness of the proposed model in SER tasks. 展开更多
关键词 speech emotion recognition long short-term memory(LSTM) multi-head attention mechanism frame-level features self-attention
在线阅读 下载PDF
Pyramid–MixNet: Integrate Attention into Encoder-Decoder Transformer Framework for Automatic Railway Surface Damage Segmentation
18
作者 Hui Luo Wenqing Li Wei Zeng 《Computers, Materials & Continua》 2025年第7期1567-1580,共14页
Rail surface damage is a critical component of high-speed railway infrastructure,directly affecting train operational stability and safety.Existing methods face limitations in accuracy and speed for small-sample,multi... Rail surface damage is a critical component of high-speed railway infrastructure,directly affecting train operational stability and safety.Existing methods face limitations in accuracy and speed for small-sample,multi-category,and multi-scale target segmentation tasks.To address these challenges,this paper proposes Pyramid-MixNet,an intelligent segmentation model for high-speed rail surface damage,leveraging dataset construction and expansion alongside a feature pyramid-based encoder-decoder network with multi-attention mechanisms.The encoding net-work integrates Spatial Reduction Masked Multi-Head Attention(SRMMHA)to enhance global feature extraction while reducing trainable parameters.The decoding network incorporates Mix-Attention(MA),enabling multi-scale structural understanding and cross-scale token group correlation learning.Experimental results demonstrate that the proposed method achieves 62.17%average segmentation accuracy,80.28%Damage Dice Coefficient,and 56.83 FPS,meeting real-time detection requirements.The model’s high accuracy and scene adaptability significantly improve the detection of small-scale and complex multi-scale rail damage,offering practical value for real-time monitoring in high-speed railway maintenance systems. 展开更多
关键词 Pyramid vision transformer encoder–decoder architecture railway damage segmentation masked multi-head attention mix-attention
在线阅读 下载PDF
MA-VoxelMorph:Multi-scale attention-based VoxelMorph for nonrigid registration of thoracoabdominal CT images
19
作者 Qing Huang Lei Ren +3 位作者 Tingwei Quan Minglei Yang Hongmei Yuan Kai Cao 《Journal of Innovative Optical Health Sciences》 2025年第1期135-151,共17页
This paper aims to develop a nonrigid registration method of preoperative and intraoperative thoracoabdominal CT images in computer-assisted interventional surgeries for accurate tumor localization and tissue visualiz... This paper aims to develop a nonrigid registration method of preoperative and intraoperative thoracoabdominal CT images in computer-assisted interventional surgeries for accurate tumor localization and tissue visualization enhancement.However,fine structure registration of complex thoracoabdominal organs and large deformation registration caused by respiratory motion is challenging.To deal with this problem,we propose a 3D multi-scale attention VoxelMorph(MAVoxelMorph)registration network.To alleviate the large deformation problem,a multi-scale axial attention mechanism is utilized by using a residual dilated pyramid pooling for multi-scale feature extraction,and position-aware axial attention for long-distance dependencies between pixels capture.To further improve the large deformation and fine structure registration results,a multi-scale context channel attention mechanism is employed utilizing content information via adjacent encoding layers.Our method was evaluated on four public lung datasets(DIR-Lab dataset,Creatis dataset,Learn2Reg dataset,OASIS dataset)and a local dataset.Results proved that the proposed method achieved better registration performance than current state-of-the-art methods,especially in handling the registration of large deformations and fine structures.It also proved to be fast in 3D image registration,using about 1.5 s,and faster than most methods.Qualitative and quantitative assessments proved that the proposed MA-VoxelMorph has the potential to realize precise and fast tumor localization in clinical interventional surgeries. 展开更多
关键词 Thoracoabdominal CT image registration large deformation fine structure MULTI-SCALE attention mechanism
原文传递
基于可变形CNN LSTM Attention网络的瓦斯抽采有效半径预测模型及应用
20
作者 杨帅 《中国煤炭》 北大核心 2025年第7期124-133,共10页
瓦斯抽采有效半径是确定井下煤层抽采钻孔间距的基础参数,因受煤层赋存条件和技术条件影响,其确定方法尚未统一、普适性差。提出了基于可变形CNN LSTM MulAttention网络的瓦斯抽采有效半径预测模型,将煤层埋深、煤层初始瓦斯压力和渗透... 瓦斯抽采有效半径是确定井下煤层抽采钻孔间距的基础参数,因受煤层赋存条件和技术条件影响,其确定方法尚未统一、普适性差。提出了基于可变形CNN LSTM MulAttention网络的瓦斯抽采有效半径预测模型,将煤层埋深、煤层初始瓦斯压力和渗透率等作为特征量,有效半径作为目标量,构建有效半径预测数据集,用于模型的训练和测试。训练和测试结果表明,可变形CNN LSTM Attention网络预测模型训练和测试中的平均绝对误差(MAE)、均方误差(RMSE)、决定系数(R^(2))分别为0.3182、0.6226、0.8727和0.4242、0.6426、0.8496,预测精度良好。与LSTM、CNN、CNN LSTM和CNN LSTM Attention预测模型相比,可变形CNN LSTM Attention网络预测模型的总体MAE和RMSE分别降低29.62%、17.30%、15.01%、4.61%和75.26%、63.37%、59.88%、54.23%,决定系数R^(2)分别提高66.12%、16.36%、11.70%、6.73%,采用可变形CNN LSTM Attention网络预测模型具有更高的精度和更好的预测效果。利用训练的机器学习预测模型对顺层钻孔瓦斯抽采有效半径进行了预测,结果与瓦斯含量现场测定法和数值模拟法确定的有效半径相差较小,验证了机器学习预测瓦斯抽采有效半径的可行性。 展开更多
关键词 瓦斯抽采 有效半径 预测模型 可变形CNN LSTM attention网络
在线阅读 下载PDF
上一页 1 2 20 下一页 到第
使用帮助 返回顶部