期刊文献+
共找到251,241篇文章
< 1 2 250 >
每页显示 20 50 100
基于改进时间融合Transformers的中国大豆需求预测方法
1
作者 刘佳佳 秦晓婧 +5 位作者 李乾川 许世卫 赵继春 王一罡 熊露 梁晓贺 《智慧农业(中英文)》 2025年第4期187-199,共13页
[目的/意义]精准预测大豆需求对保障国家粮食安全、优化产业决策与应对国际贸易变局有着重要的现实意义,而利用时间融合Transformers(Temporal Fusion Transformers,TFT)模型开展中国大豆需求预测时,在特征交互层与注意力权重分配等方... [目的/意义]精准预测大豆需求对保障国家粮食安全、优化产业决策与应对国际贸易变局有着重要的现实意义,而利用时间融合Transformers(Temporal Fusion Transformers,TFT)模型开展中国大豆需求预测时,在特征交互层与注意力权重分配等方面仍存在一定局限。为此,亟需探索一种基于改进TFT模型的预测方法,以提升需求预测的准确性与可解释性。[方法]本研究将深度学习的TFT模型应用到中国大豆需求预测中,提出了一种基于多层动态特征交互(Multi-layer Dynamic Feature Interaction,MDFI)与自适应注意力权重优化(Adaptive Attention Weight Optimization,AAWO)改进的MA-TFT(Improved TFT Model Based on MDFI and AAWO)模型。对包含1980—2024年4652个相关指标的中国大豆需求分析数据集进行数据预处理和特征工程,设计实验将MA-TFT模型分别与自回归差分移动平均模型(Autoregressive Integrated Moving Average Model,ARIMA)、长短期记忆网络(Long Short-Term Memory,LSTM)模型及TFT模型进行预测性能对比,进行了消融实验,同时利用SHAP(SHapley Additive exPlanations)工具可解释性分析影响中国大豆需求的关键特征变量,开展了未来10年的中国大豆需求量预测。[结果和讨论]MA-TFT模型的均方误差(Mean Squared Error,MSE)、平均绝对百分比误差(Mean Absolute Percentage Error,MAPE)分别为0.036和5.89%,决定系数R^(2)为0.91,均高于对比模型,均方根误差(Root Mean Square Error,RMSE)和MAPE分别较基准模型TFT累计降低21.84%和3.44%,表明改进TFT的MA-TFT模型能够捕捉特征间复杂关系,提升预测性能;研究利用SHAP工具可解释性分析发现,MA-TFT模型对影响中国大豆需求关键特征变量的解释稳定性较高;预计2025、2030和2034年中国大豆需求量分别达到11799万吨、11033万吨和11378万吨。[结论]基于改进TFT的MA-TFT模型方法为解决现有大豆需求预测方法精度不足、可解释性不强的实际问题提供了解决思路,也为其他农产品时间序列预测的方法优化与应用提供了参考和借鉴。 展开更多
关键词 时间融合transformers(TFT) 大豆需求预测 多层动态特征交互 自适应注意力权重优化 可解释性分析
在线阅读 下载PDF
新解码器的CNNs-Transformers融合网络及其病理图像肿瘤分割应用 被引量:1
2
作者 马丽晶 王朝立 +2 位作者 孙占全 程树群 王康 《小型微型计算机系统》 北大核心 2025年第6期1442-1449,共8页
病理图像是肿瘤诊断的"金标准",但超高分辨率的病理图像使得医生需要消耗大量的精力和时间,而且诊断结果主观性比较强.随着人工智能技术的发展,深度学习模型提供了计算机代替人对病理图像进行快速、准确和可靠诊断的可能性.然... 病理图像是肿瘤诊断的"金标准",但超高分辨率的病理图像使得医生需要消耗大量的精力和时间,而且诊断结果主观性比较强.随着人工智能技术的发展,深度学习模型提供了计算机代替人对病理图像进行快速、准确和可靠诊断的可能性.然而,目前大多数的网络更注重如何在编码器部分提取更准确的特征,而对于同等重要的解码器部分的结构设计研究则稍显不足.针对该问题,本文提出了由三类上采样模块组成的新网络,而编码器部分采用Swin Transformer和ConvNeXt作为网络的双分支并行独立结构.三类上采样模块分别是多重转置卷积采样、双线性上采样和Swin Transformer上采样,其特点是可以充分利用病理图像特征之间局部和全局的依赖关系.该网络分别在肝癌数据集和GLAS数据集上进行了验证,并与不同类型的主流网络进行了对比,性能指标皆达到比较好的结果. 展开更多
关键词 医学图像分割 深度学习 卷积神经网络 Swin transformer
在线阅读 下载PDF
Generating Abstractive Summaries from Social Media Discussions Using Transformers
3
作者 Afrodite Papagiannopoulou Chrissanthi Angeli Mazida Ahmad 《Open Journal of Applied Sciences》 2025年第1期239-258,共20页
The rise of social media platforms has revolutionized communication, enabling the exchange of vast amounts of data through text, audio, images, and videos. These platforms have become critical for sharing opinions and... The rise of social media platforms has revolutionized communication, enabling the exchange of vast amounts of data through text, audio, images, and videos. These platforms have become critical for sharing opinions and insights, influencing daily habits, and driving business, political, and economic decisions. Text posts are particularly significant, and natural language processing (NLP) has emerged as a powerful tool for analyzing such data. While traditional NLP methods have been effective for structured media, social media content poses unique challenges due to its informal and diverse nature. This has spurred the development of new techniques tailored for processing and extracting insights from unstructured user-generated text. One key application of NLP is the summarization of user comments to manage overwhelming content volumes. Abstractive summarization has proven highly effective in generating concise, human-like summaries, offering clear overviews of key themes and sentiments. This enhances understanding and engagement while reducing cognitive effort for users. For businesses, summarization provides actionable insights into customer preferences and feedback, enabling faster trend analysis, improved responsiveness, and strategic adaptability. By distilling complex data into manageable insights, summarization plays a vital role in improving user experiences and empowering informed decision-making in a data-driven landscape. This paper proposes a new implementation framework by fine-tuning and parameterizing Transformer Large Language Models to manage and maintain linguistic and semantic components in abstractive summary generation. The system excels in transforming large volumes of data into meaningful summaries, as evidenced by its strong performance across metrics like fluency, consistency, readability, and semantic coherence. 展开更多
关键词 Abstractive Summarization transformers Social Media Summarization transformer Language Models
在线阅读 下载PDF
Token Masked Pose Transformers Are Efficient Learners
4
作者 Xinyi Song Haixiang Zhang Shaohua Li 《Computers, Materials & Continua》 2025年第5期2735-2750,共16页
In recent years,Transformer has achieved remarkable results in the field of computer vision,with its built-in attention layers effectively modeling global dependencies in images by transforming image features into tok... In recent years,Transformer has achieved remarkable results in the field of computer vision,with its built-in attention layers effectively modeling global dependencies in images by transforming image features into token forms.However,Transformers often face high computational costs when processing large-scale image data,which limits their feasibility in real-time applications.To address this issue,we propose Token Masked Pose Transformers(TMPose),constructing an efficient Transformer network for pose estimation.This network applies semantic-level masking to tokens and employs three different masking strategies to optimize model performance,aiming to reduce computational complexity.Experimental results show that TMPose reduces computational complexity by 61.1%on the COCO validation dataset,with negligible loss in accuracy.Additionally,our performance on the MPII dataset is also competitive.This research not only enhances the accuracy of pose estimation but also significantly reduces the demand for computational resources,providing new directions for further studies in this field. 展开更多
关键词 Pattern recognition image processing neural network pose transformer
在线阅读 下载PDF
Leveraging Transformers for Detection of Arabic Cyberbullying on Social Media: Hybrid Arabic Transformers
5
作者 Amjad A.Alsuwaylimi Zaid S.Alenezi 《Computers, Materials & Continua》 2025年第5期3165-3185,共21页
Cyberbullying is a remarkable issue in the Arabic-speaking world,affecting children,organizations,and businesses.Various efforts have been made to combat this problem through proposed models using machine learning(ML)... Cyberbullying is a remarkable issue in the Arabic-speaking world,affecting children,organizations,and businesses.Various efforts have been made to combat this problem through proposed models using machine learning(ML)and deep learning(DL)approaches utilizing natural language processing(NLP)methods and by proposing relevant datasets.However,most of these endeavors focused predominantly on the English language,leaving a substantial gap in addressing Arabic cyberbullying.Given the complexities of the Arabic language,transfer learning techniques and transformers present a promising approach to enhance the detection and classification of abusive content by leveraging large and pretrained models that use a large dataset.Therefore,this study proposes a hybrid model using transformers trained on extensive Arabic datasets.It then fine-tunes the hybrid model on a newly curated Arabic cyberbullying dataset collected from social media platforms,in particular Twitter.Additionally,the following two hybrid transformer models are introduced:the first combines CAmelid Morphologically-aware pretrained Bidirectional Encoder Representations from Transformers(CAMeLBERT)with Arabic Generative Pre-trained Transformer 2(AraGPT2)and the second combines Arabic BERT(AraBERT)with Cross-lingual Language Model-RoBERTa(XLM-R).Two strategies,namely,feature fusion and ensemble voting,are employed to improve the model performance accuracy.Experimental results,measured through precision,recall,F1-score,accuracy,and AreaUnder the Curve-Receiver Operating Characteristic(AUC-ROC),demonstrate that the combined CAMeLBERT and AraGPT2 models using feature fusion outperformed traditional DL models,such as Long Short-Term Memory(LSTM)and Bidirectional Long Short-Term Memory(BiLSTM),as well as other independent Arabic-based transformer models. 展开更多
关键词 CYBERBULLYING transformers pre-trained models arabic cyberbullying detection deep learning
在线阅读 下载PDF
Research on the Selection and Layout Scheme of Main Transformers in the Primary Electrical Design of New Energy Step-Up Stations
6
作者 Yuekai Liao 《Journal of Electronic Research and Application》 2025年第4期254-260,共7页
This paper focuses on the research of the main transformer selection and layout scheme for new energy step-up substations.From the perspective of engineering design,it analyzes the principles of main transformer selec... This paper focuses on the research of the main transformer selection and layout scheme for new energy step-up substations.From the perspective of engineering design,it analyzes the principles of main transformer selection,key parameters,and their matching with the characteristics of new energy.It also explores the layout methods and optimization strategies.Combined with typical case studies,optimization suggestions are proposed for the design of main transformers in new energy step-up substations.The research shows that rational main transformer selection and scientific layout schemes can better adapt to the characteristics of new energy projects while effectively improving land use efficiency and economic viability.This study can provide technical experience support for the design of new energy projects. 展开更多
关键词 New energy step-up substation Engineering design Main transformer selection
在线阅读 下载PDF
Analysis of the effects of strong stray magnetic fields generated by tokamak device on transformers assembled in electronic power converters
7
作者 Xingjian ZHAO Ge GAO +2 位作者 Li JIANG Yong YANG Hong LEI 《Plasma Science and Technology》 2025年第5期81-93,共13页
As the plasma current power in tokamak devices increases,a significant number of stray magnetic fields are generated around the equipment.These stray magnetic fields can disrupt the operation of electronic power devic... As the plasma current power in tokamak devices increases,a significant number of stray magnetic fields are generated around the equipment.These stray magnetic fields can disrupt the operation of electronic power devices,particularly transformers in switched-mode power supplies.Testing flyback converters with transformers under strong background magnetic fields highlights electromagnetic compatibility(EMC)issues for such switched-mode power supplies.This study utilizes finite element analysis software to simulate the electromagnetic environment of switched-mode power supply transformers and investigates the impact of variations in different magnetic field parameters on the performance of switched-mode power supplies under strong stray magnetic fields.The findings indicate that EMC issues are associated with transformer core saturation and can be alleviated through appropriate configurations of the core size,air gap,fillet radius,and installation direction.This study offers novel solutions for addressing EMC issues in high magnetic field environments. 展开更多
关键词 transformers magnetic field interference magnetic components power electronics magnetic field simulation
在线阅读 下载PDF
Improving Fashion Sentiment Detection on X through Hybrid Transformers and RNNs
8
作者 Bandar Alotaibi Aljawhara Almutarie +1 位作者 Shuaa Alotaibi Munif Alotaibi 《Computers, Materials & Continua》 2025年第9期4451-4467,共17页
X(formerly known as Twitter)is one of the most prominent social media platforms,enabling users to share short messages(tweets)with the public or their followers.It serves various purposes,from real-time news dissemina... X(formerly known as Twitter)is one of the most prominent social media platforms,enabling users to share short messages(tweets)with the public or their followers.It serves various purposes,from real-time news dissemination and political discourse to trend spotting and consumer engagement.X has emerged as a key space for understanding shifting brand perceptions,consumer preferences,and product-related sentiment in the fashion industry.However,the platform’s informal,dynamic,and context-dependent language poses substantial challenges for sentiment analysis,mainly when attempting to detect sarcasm,slang,and nuanced emotional tones.This study introduces a hybrid deep learning framework that integrates Transformer encoders,recurrent neural networks(i.e.,Long Short-Term Memory(LSTM)and Gated Recurrent Unit(GRU)),and attention mechanisms to improve the accuracy of fashion-related sentiment classification.These methods were selected due to their proven strength in capturing both contextual dependencies and sequential structures,which are essential for interpreting short-form text.Our model was evaluated on a dataset of 20,000 fashion tweets.The experimental results demonstrate a classification accuracy of 92.25%,outperforming conventional models such as Logistic Regression,Linear Support Vector Machine(SVM),and even standalone LSTM by a margin of up to 8%.This improvement highlights the importance of hybrid architectures in handling noisy,informal social media data.This study’s findings offer strong implications for digital marketing and brand management,where timely sentiment detection is critical.Despite the promising results,challenges remain regarding the precise identification of negative sentiments,indicating that further work is needed to detect subtle and contextually embedded expressions. 展开更多
关键词 Sentiment analysis deep learning natural language processing transformers recurrent neural networks
在线阅读 下载PDF
Data-driven measurement performance evaluation of voltage transformers in electric railway traction power supply systems
9
作者 Zhaoyang Li Muqi Sun +5 位作者 Jun Zhu Haoyu Luo Qi Wang Haitao Hu Zhengyou He Ke Wang 《Railway Engineering Science》 2025年第2期311-323,共13页
Critical for metering and protection in electric railway traction power supply systems(TPSSs),the measurement performance of voltage transformers(VTs)must be timely and reliably monitored.This paper outlines a three-s... Critical for metering and protection in electric railway traction power supply systems(TPSSs),the measurement performance of voltage transformers(VTs)must be timely and reliably monitored.This paper outlines a three-step,RMS data only method for evaluating VTs in TPSSs.First,a kernel principal component analysis approach is used to diagnose the VT exhibiting significant measurement deviations over time,mitigating the influence of stochastic fluctuations in traction loads.Second,a back propagation neural network is employed to continuously estimate the measurement deviations of the targeted VT.Third,a trend analysis method is developed to assess the evolution of the measurement performance of VTs.Case studies conducted on field data from an operational TPSS demonstrate the effectiveness of the proposed method in detecting VTs with measurement deviations exceeding 1%relative to their original accuracy levels.Additionally,the method accurately tracks deviation trends,enabling the identification of potential early-stage faults in VTs and helping prevent significant economic losses in TPSS operations. 展开更多
关键词 Voltage transformer Traction power supply system Measurement performance Data-driven evaluation Abrupt change detection Bootstrap confidence interval
在线阅读 下载PDF
Transformers for Multi-Modal Image Analysis in Healthcare
10
作者 Sameera V Mohd Sagheer Meghana K H +2 位作者 P M Ameer Muneer Parayangat Mohamed Abbas 《Computers, Materials & Continua》 2025年第9期4259-4297,共39页
Integrating multiple medical imaging techniques,including Magnetic Resonance Imaging(MRI),Computed Tomography,Positron Emission Tomography(PET),and ultrasound,provides a comprehensive view of the patient health status... Integrating multiple medical imaging techniques,including Magnetic Resonance Imaging(MRI),Computed Tomography,Positron Emission Tomography(PET),and ultrasound,provides a comprehensive view of the patient health status.Each of these methods contributes unique diagnostic insights,enhancing the overall assessment of patient condition.Nevertheless,the amalgamation of data from multiple modalities presents difficulties due to disparities in resolution,data collection methods,and noise levels.While traditional models like Convolutional Neural Networks(CNNs)excel in single-modality tasks,they struggle to handle multi-modal complexities,lacking the capacity to model global relationships.This research presents a novel approach for examining multi-modal medical imagery using a transformer-based system.The framework employs self-attention and cross-attention mechanisms to synchronize and integrate features across various modalities.Additionally,it shows resilience to variations in noise and image quality,making it adaptable for real-time clinical use.To address the computational hurdles linked to transformer models,particularly in real-time clinical applications in resource-constrained environments,several optimization techniques have been integrated to boost scalability and efficiency.Initially,a streamlined transformer architecture was adopted to minimize the computational load while maintaining model effectiveness.Methods such as model pruning,quantization,and knowledge distillation have been applied to reduce the parameter count and enhance the inference speed.Furthermore,efficient attention mechanisms such as linear or sparse attention were employed to alleviate the substantial memory and processing requirements of traditional self-attention operations.For further deployment optimization,researchers have implemented hardware-aware acceleration strategies,including the use of TensorRT and ONNX-based model compression,to ensure efficient execution on edge devices.These optimizations allow the approach to function effectively in real-time clinical settings,ensuring viability even in environments with limited resources.Future research directions include integrating non-imaging data to facilitate personalized treatment and enhancing computational efficiency for implementation in resource-limited environments.This study highlights the transformative potential of transformer models in multi-modal medical imaging,offering improvements in diagnostic accuracy and patient care outcomes. 展开更多
关键词 Multi-modal image analysis medical imaging deep learning image segmentation disease detection multi-modal fusion Vision transformers(ViTs) precision medicine clinical decision support
在线阅读 下载PDF
基于Transformer的时间序列预测方法综述 被引量:4
11
作者 陈嘉俊 刘波 +2 位作者 林伟伟 郑剑文 谢家晨 《计算机科学》 北大核心 2025年第6期96-105,共10页
时间序列预测作为分析历史数据以预测未来趋势的关键技术,已广泛应用于金融、气象等领域。然而,传统方法如自回归移动平均模型和指数平滑法等在处理非线性模式、捕捉长期依赖性时存在局限。最近,基于Transformer的方法因其自注意力机制... 时间序列预测作为分析历史数据以预测未来趋势的关键技术,已广泛应用于金融、气象等领域。然而,传统方法如自回归移动平均模型和指数平滑法等在处理非线性模式、捕捉长期依赖性时存在局限。最近,基于Transformer的方法因其自注意力机制,在自然语言处理与计算机视觉领域取得突破,也开始拓展至时间序列预测领域并取得显著成果。因此,探究如何将Transformer高效运用于时间序列预测,成为推动该领域发展的关键。首先,介绍了时间序列的特性,阐述了时间序列预测的常见任务类别及评估指标。接着,深入解析Transformer的基本架构,并挑选了近年来在时间序列预测中广受关注的Transfo-rmer衍生模型,从模块及架构层面进行分类,并分别从问题解决、创新点及局限性3个维度进行比较和分析。最后,进一步探讨了时间序列预测Transformer在未来可能的研究方向。 展开更多
关键词 时间序列 transformer模型 深度学习 注意力机制 预测
在线阅读 下载PDF
基于Transformer模型的时序数据预测方法综述 被引量:13
12
作者 孟祥福 石皓源 《计算机科学与探索》 北大核心 2025年第1期45-64,共20页
时序数据预测(TSF)是指通过分析历史数据的趋势性、季节性等潜在信息,预测未来时间点或时间段的数值和趋势。时序数据由传感器生成,在金融、医疗、能源、交通、气象等众多领域都发挥着重要作用。随着物联网传感器的发展,海量的时序数据... 时序数据预测(TSF)是指通过分析历史数据的趋势性、季节性等潜在信息,预测未来时间点或时间段的数值和趋势。时序数据由传感器生成,在金融、医疗、能源、交通、气象等众多领域都发挥着重要作用。随着物联网传感器的发展,海量的时序数据难以使用传统的机器学习解决,而Transformer在自然语言处理和计算机视觉等领域的诸多任务表现优秀,学者们利用Transformer模型有效捕获长期依赖关系,使得时序数据预测任务取得了飞速发展。综述了基于Transformer模型的时序数据预测方法,按时间梳理了时序数据预测的发展进程,系统介绍了时序数据预处理过程和方法,介绍了常用的时序预测评价指标和数据集。以算法框架为研究内容系统阐述了基于Transformer的各类模型在TSF任务中的应用方法和工作原理。通过实验对比了各个模型的性能、优点和局限性,并对实验结果展开了分析与讨论。结合Transformer模型在时序数据预测任务中现有工作存在的挑战提出了该方向未来发展趋势。 展开更多
关键词 深度学习 时序数据预测 数据预处理 transformer模型
在线阅读 下载PDF
多变量时序标记Transformer及其在电潜泵故障诊断中的应用 被引量:2
13
作者 李康 李爽 +2 位作者 高小永 李强 张来斌 《控制与决策》 北大核心 2025年第4期1145-1153,共9页
电潜泵故障诊断对于确保安全可靠采油至关重要,但是,电潜泵数据呈现出的多变量、非线性和动态变化等复杂特性为该任务带来了严峻挑战.近年来,深度学习在复杂数据特征提取方面表现出的强大能力催生了一系列基于神经网络的电潜泵故障诊断... 电潜泵故障诊断对于确保安全可靠采油至关重要,但是,电潜泵数据呈现出的多变量、非线性和动态变化等复杂特性为该任务带来了严峻挑战.近年来,深度学习在复杂数据特征提取方面表现出的强大能力催生了一系列基于神经网络的电潜泵故障诊断方法.然而,多数方法忽略了电潜泵数据的动态特性以及长时依赖特征提取困难的问题.针对上述问题,提出一种多变量时序标记Transformer神经网络来实现电潜泵故障诊断.该模型设计新的多变量时间序列标记策略,继承引入多头注意力机制和残差连接的传统Transformer神经网络编码器在长时依赖特征提取方面的优势,用前向神经网络替代传统Transformer神经网络解码器来简化模型复杂度.通过对油田现场故障数据分析,验证所提出方法的有效性.实验结果表明,所提出方法实现了10类电潜泵故障的精确诊断,相比于流行的深度学习方法诊断性能更优. 展开更多
关键词 电潜泵 transformer神经网络 深度学习 特征提取 故障诊断 多变量时序标记
原文传递
基于改进Transformer结构的电力绝缘子运动模糊图像复原网络 被引量:1
14
作者 李鹏 常乐 +2 位作者 覃发富 孟庆伟 陈继明 《电网技术》 北大核心 2025年第6期2623-2631,I0143-I0146,共13页
针对高压输电线路巡检航拍过程中产生的电力绝缘子图像运动模糊的失真情形,影响后续绝缘子定位及缺陷检测的问题,提出了一种基于改进Transformer结构的电力绝缘子图像运动模糊复原方法。为了适应电力绝缘子航拍图像中全局与局部模糊的... 针对高压输电线路巡检航拍过程中产生的电力绝缘子图像运动模糊的失真情形,影响后续绝缘子定位及缺陷检测的问题,提出了一种基于改进Transformer结构的电力绝缘子图像运动模糊复原方法。为了适应电力绝缘子航拍图像中全局与局部模糊的复原需求,在Transformer网络结构上引入条带注意力模块,结合卷积神经网络,在减小内存空间需求和不依赖大量训练数据的同时实现高效的模糊绝缘子图像复原;同时,在网络目标函数中引入对比学习损失,充分地挖掘和利用清晰与模糊电力绝缘子图像的关联信息。构建运动模糊绝缘子图像数据集进行图像复原与缺陷检测实验,结果表明,该文的运动模糊绝缘子图像复原方法在峰值信噪比(peak signal-to-noise ratio,PSNR)和结构相似度(structure similarity index measure,SSIM)这两个指标上均高于Deblur GAN-v2、MIMO-UNet等主流算法,使用目标检测算法YOLOv5和YOLOv7对去模糊前后的绝缘子进行定位与自爆缺陷检测后显示该文方法在提升高压输电线路巡检任务中绝缘子定位与缺陷检测的准确率上具有实际应用意义。 展开更多
关键词 运动模糊图像复原 transformER 对比学习 绝缘子及缺陷检测
原文传递
融合梯度预测和无参注意力的高效地震去噪Transformer 被引量:1
15
作者 高磊 乔昊炜 +2 位作者 梁东升 闵帆 杨梅 《计算机科学与探索》 北大核心 2025年第5期1342-1352,共11页
压制随机噪声能够有效提升地震数据的信噪比(SNR)。近年来,基于卷积神经网络(CNN)的深度学习方法在地震数据去噪领域展现出显著性能。然而,CNN中的卷积操作由于感受野的限制通常只能捕获局部信息而不能建立全局信息的长距离连接,可能会... 压制随机噪声能够有效提升地震数据的信噪比(SNR)。近年来,基于卷积神经网络(CNN)的深度学习方法在地震数据去噪领域展现出显著性能。然而,CNN中的卷积操作由于感受野的限制通常只能捕获局部信息而不能建立全局信息的长距离连接,可能会导致细节信息的丢失。针对地震数据去噪问题,提出了一种融合梯度预测和无参注意力的高效Transformer模型(ETGP)。引入多头“转置”注意力来代替传统的多头注意力,它能在通道间计算注意力来表示全局信息,缓解了传统多头注意力复杂度过高的问题。提出了无参注意力前馈神经网络,它能同时考虑空间和通道维度计算注意力权重,而不向网络增加参数。设计了梯度预测网络以提取边缘信息,并将信息自适应地添加到并行Transformer的输入中,从而获得高质量的地震数据。在合成数据和野外数据上进行了实验,并与经典和先进的去噪方法进行了比较。结果表明,ETGP去噪方法不仅能更有效地压制随机噪声,并且在弱信号保留和同相轴连续性方面具有显著优势。 展开更多
关键词 地震数据去噪 卷积神经网络 transformER 注意力模块 梯度融合
在线阅读 下载PDF
基于改进Swin Transformer的人脸活体检测 被引量:2
16
作者 王旭光 卜辰宇 时泽宇 《中国测试》 北大核心 2025年第6期31-39,共9页
随着人脸识别技术的发展,人脸活体检测作为人脸识别系统的安全保障变得更加重要。但当前主流的人脸活体检测模型仅针对特定的检测场景及欺诈攻击方式,面对未知攻击的鲁棒性和泛化能力较差。为此,该文提出一种改进的Swin Transformer模型... 随着人脸识别技术的发展,人脸活体检测作为人脸识别系统的安全保障变得更加重要。但当前主流的人脸活体检测模型仅针对特定的检测场景及欺诈攻击方式,面对未知攻击的鲁棒性和泛化能力较差。为此,该文提出一种改进的Swin Transformer模型,即CDCSwin-T(central difference convolution Swin Transformer)模型。该模型以Swin Transformer为主干,利用其滑动窗口注意力机制提取人脸全局信息,同时引入中心差分卷积(central difference convolution,CDC)模块提取人脸局部信息,加强主干模型捕获真假人脸差异的能力,从而增强其面对未知攻击的鲁棒性;另外在主干模型中引入瓶颈注意力模块,引导模型关注人脸关键信息,加速模型训练;最终将主干模型不同阶段的多尺度信息进行自适应融合,进一步提升该文模型的泛化能力。CDCSwin-T模型在OULU-NPU数据集4个协议上的平均分类错误率(ACER)分别为0.2%,1.1%,(1.1±0.6)%,(2.8±1.4)%,在CASIA-MFSD和REPLAYATTACK数据集跨库测试上的半错误率(HTER)分别为14.1%,22.9%,均优于当前的主流模型,表明其面对未知攻击的鲁棒性和泛化能力均有所提升。 展开更多
关键词 人脸活体检测 Swin transformer 瓶颈注意力模块 特征融合
在线阅读 下载PDF
多尺度特征提取的Transformer短期风电功率预测 被引量:5
17
作者 徐武 范鑫豪 +1 位作者 沈智方 刘洋 《太阳能学报》 北大核心 2025年第2期640-648,共9页
针对短期风电功率预测特征提取尺度单一问题,设计一种基于多尺度特征提取的Transformer短期风电功率预测模型(MTPNet)。首先,在Transformer构架的基础上,利用维数不变嵌入,设计多尺度特征提取网络挖掘风电功率序列本身时序特征,保证了... 针对短期风电功率预测特征提取尺度单一问题,设计一种基于多尺度特征提取的Transformer短期风电功率预测模型(MTPNet)。首先,在Transformer构架的基础上,利用维数不变嵌入,设计多尺度特征提取网络挖掘风电功率序列本身时序特征,保证了特征提取时维数不被破坏;其次,利用融合自注意力机制的长短期记忆网络挖掘气象条件与功率之间的全局依赖关系;最后,融合风电功率序列本身时序特征和气象条件依赖关系,实现短期风电功率预测。实例仿真结果表明,MTPNet模型预测精度得到提升;消融实验证明了模型各模块的可靠性和有效性,具有一定的实用价值。 展开更多
关键词 风电功率预测 transformER 注意力机制 特征提取 长短期记忆网络 维数不变嵌入层
原文传递
融合Gabor滤波与Transformer的图像水印方法 被引量:1
18
作者 张天骐 谭霜 +1 位作者 沈夕文 唐娟 《信号处理》 北大核心 2025年第4期694-705,共12页
图像水印在数字版权保护和身份验证领域中具有关键意义,是保护图像信息安全和确保数据可信性的重要技术手段。目前,大多数已发表的基于深度学习的图像水印方法都是基于卷积神经网络设计的,此类方法存在无法充分捕捉图像的全局信息和细... 图像水印在数字版权保护和身份验证领域中具有关键意义,是保护图像信息安全和确保数据可信性的重要技术手段。目前,大多数已发表的基于深度学习的图像水印方法都是基于卷积神经网络设计的,此类方法存在无法充分捕捉图像的全局信息和细节信息,以及忽略图像高频信息具备稳定和不可感知特点等问题,为了克服上述问题,该论文提出一种融合Gabor滤波与Transformer的图像水印模型。该模型由嵌入网络、提取网络和判别网络组成:在嵌入网络设计了水印信息处理模块对水印信息引入冗余和扩展操作,以增加水印信息在传输过程中的鲁棒性;在嵌入网络引入Gabor滤波的思想在特征提取模块通过卷积分支来捕捉局部特征,通过Transformer分支捕捉全局信息,来充分挖掘图像的稳定特征;在提取网络中融合标准卷积和差分卷积,来准确感知图像的细微信息,进而提高水印的提取精度;引入判别网络与嵌入网络形成对抗训练关系,评估生成水印图像的真实性和质量,从而提升嵌入网络生成水印图像的视觉质量。分别在COCO、ImageNet和VOC2012数据集下进行综合对比实验,结果表明,该文方法针对不可感知性和鲁棒性,相比于相关水印模型取得了更优的指标,具有较为突出的增强性能与泛化能力。此外,还进行了相关的消融实验,结果进一步验证了该模型的可靠性和有效性。 展开更多
关键词 图像水印 不可感知 鲁棒性 卷积神经网络 transformER
在线阅读 下载PDF
基于时序二维变换和多尺度Transformer的电能质量扰动分类方法 被引量:1
19
作者 王守相 李慧强 +3 位作者 赵倩宇 郭陆阳 王同勋 王洋 《电力系统自动化》 北大核心 2025年第7期198-207,共10页
随着新能源渗透率的不断提高,电网面临的电能质量扰动(PQD)问题变得更加复杂,基于一维PQD信号的传统分类方法难以同时提取并辨识周期性与趋势性扰动。针对此问题,提出了一种基于时序二维变换和多尺度Transformer的PQD分类方法。首先,利... 随着新能源渗透率的不断提高,电网面临的电能质量扰动(PQD)问题变得更加复杂,基于一维PQD信号的传统分类方法难以同时提取并辨识周期性与趋势性扰动。针对此问题,提出了一种基于时序二维变换和多尺度Transformer的PQD分类方法。首先,利用时序二维变换将一维PQD时间序列转换为一组基于多个周期的二维张量,以实现在二维空间中深入挖掘PQD信号中所包含的特征信息。然后,通过多尺度Transformer编码器模块提取PQD信号的多尺度特征图,利用多尺度Transformer解码器模块对多尺度特征图进行拼接和特征融合,有效合并在不同尺度上提取的特征图。最后,通过全连接层和Softmax分类器完成PQD分类任务。为验证所提方法的有效性,建立了含24种PQD的数据集对模型进行测试,结果表明所提方法对PQD信号具有较高的分类准确率和噪声鲁棒性。 展开更多
关键词 电能质量 扰动 分类 时序二维变换 多尺度transformer 特征提取 特征融合
在线阅读 下载PDF
基于融合卷积Transformer的航空发动机故障诊断 被引量:2
20
作者 赵洪利 杨佳强 《北京航空航天大学学报》 北大核心 2025年第4期1117-1126,共10页
航空发动机长期处于恶劣的气路环境下工作会面临腐蚀、侵蚀等问题,且故障参数特征不明显,因此,精准的航空发动机故障诊断方法对保证飞机安全运行具有重要意义。为提高预测准确性,提出了一种基于融合卷积Transformer的航空发动机故障诊... 航空发动机长期处于恶劣的气路环境下工作会面临腐蚀、侵蚀等问题,且故障参数特征不明显,因此,精准的航空发动机故障诊断方法对保证飞机安全运行具有重要意义。为提高预测准确性,提出了一种基于融合卷积Transformer的航空发动机故障诊断方法。利用自注意力机制提取有用特征,抑制冗余信息,并将最大池化层引入Transformer模型中,进一步降低模型内存消耗及参数量,缓解过拟合现象。采用基于GasTurb建模的涡扇发动机仿真数据集进行验证,结果与Transformer模型和反向传播(BP)神经网络、卷积神经网络(CNN)、循环神经网络(RNN)等传统深度学习模型相比,准确率分别提高了6.552%和28.117%、13.189%、10.29%,证明了所提方法的有效性,可为航空发动机故障诊断提供一定的参考。 展开更多
关键词 航空发动机 故障诊断 自注意力机制 融合卷积transformer 深度神经网络
原文传递
上一页 1 2 250 下一页 到第
使用帮助 返回顶部