期刊文献+
共找到94篇文章
< 1 2 5 >
每页显示 20 50 100
Enhancing Multi-Class Cyberbullying Classification with Hybrid Feature Extraction and Transformer-Based Models
1
作者 Suliman Mohamed Fati Mohammed A.Mahdi +4 位作者 Mohamed A.G.Hazber Shahanawaj Ahamad Sawsan A.Saad Mohammed Gamal Ragab Mohammed Al-Shalabi 《Computer Modeling in Engineering & Sciences》 2025年第5期2109-2131,共23页
Cyberbullying on social media poses significant psychological risks,yet most detection systems over-simplify the task by focusing on binary classification,ignoring nuanced categories like passive-aggressive remarks or... Cyberbullying on social media poses significant psychological risks,yet most detection systems over-simplify the task by focusing on binary classification,ignoring nuanced categories like passive-aggressive remarks or indirect slurs.To address this gap,we propose a hybrid framework combining Term Frequency-Inverse Document Frequency(TF-IDF),word-to-vector(Word2Vec),and Bidirectional Encoder Representations from Transformers(BERT)based models for multi-class cyberbullying detection.Our approach integrates TF-IDF for lexical specificity and Word2Vec for semantic relationships,fused with BERT’s contextual embeddings to capture syntactic and semantic complexities.We evaluate the framework on a publicly available dataset of 47,000 annotated social media posts across five cyberbullying categories:age,ethnicity,gender,religion,and indirect aggression.Among BERT variants tested,BERT Base Un-Cased achieved the highest performance with 93%accuracy(standard deviation across±1%5-fold cross-validation)and an average AUC of 0.96,outperforming standalone TF-IDF(78%)and Word2Vec(82%)models.Notably,it achieved near-perfect AUC scores(0.99)for age and ethnicity-based bullying.A comparative analysis with state-of-the-art benchmarks,including Generative Pre-trained Transformer 2(GPT-2)and Text-to-Text Transfer Transformer(T5)models highlights BERT’s superiority in handling ambiguous language.This work advances cyberbullying detection by demonstrating how hybrid feature extraction and transformer models improve multi-class classification,offering a scalable solution for moderating nuanced harmful content. 展开更多
关键词 Cyberbullying classification multi-class classification BERT models machine learning TF-IDF Word2Vec social media analysis transformer models
在线阅读 下载PDF
Combining transformer and 3DCNN models to achieve co-design of structures and sequences of antibodies in a diffusional manner
2
作者 Yue Hu Feng Tao +3 位作者 Jiajie Xu Wen-Jun Lan Jing Zhang Wei Lan 《Journal of Pharmaceutical Analysis》 2025年第6期1406-1408,共3页
AlphaPanda(AlphaFold2[1]inspired protein-specific antibody design in a diffusional manner)is an advanced algorithm for designing complementary determining regions(CDRs)of the antibody targeted the specific epitope,com... AlphaPanda(AlphaFold2[1]inspired protein-specific antibody design in a diffusional manner)is an advanced algorithm for designing complementary determining regions(CDRs)of the antibody targeted the specific epitope,combining transformer[2]models,3DCNN[3],and diffusion[4]generative models. 展开更多
关键词 advanced algorithm diffusion generative models dcnn epitope targeting antibody design complementary determining regions complementary determining regions cdrs transformer models
在线阅读 下载PDF
IG-3D:Integrated-Gradients 3D Optimization for Private Transformer Inference
3
作者 Lei Sun Jingwen Wang +3 位作者 Peng Hu Xiuqing Mao Cuiyun Hu Zhihong Wang 《Computers, Materials & Continua》 2026年第5期1158-1176,共19页
Transformer models face significant computational challenges in private inference(PI).Existing optimization methods often rely on isolated techniques,neglecting joint structural and operational improvements.We propose... Transformer models face significant computational challenges in private inference(PI).Existing optimization methods often rely on isolated techniques,neglecting joint structural and operational improvements.We propose IG-3D,a unified framework that integrates structured compression and operator approximation through accurate importance assessment.Our approach first evaluates attention head importance using Integrated Gradients(IG),offering greater stability and theoretical soundness than gradient-based methods.We then apply a threedimensional optimization:(1)structurally pruning redundant attention heads;(2)replacing Softmax with adaptive polynomial approximation to avoid exponential computations;(3)implementing layer-wise GELU substitution to accommodate different layer characteristics.A joint thresholdmechanism coordinates compression across dimensions under accuracy constraints.Experimental results on the GLUE benchmark show that our method achieves an average 2.9×speedup in inference latency and a 50%reduction in communication cost,while controlling the accuracy loss within 2.3%,demonstrating significant synergistic effects and a superior accuracy-efficiency trade-off compared to single-technique optimization strategies. 展开更多
关键词 Private inference transformer attention-head pruning integrated gradients transformer model optimization
在线阅读 下载PDF
Why Transformers Outperform LSTMs:A Comparative Study on Sarcasm Detection
4
作者 Palak Bari Gurnur Bedi +1 位作者 Khushi Joshi Anupama Jawale 《Journal on Artificial Intelligence》 2025年第1期499-508,共10页
This study investigates sarcasm detection in text using a dataset of 8095 sentences compiled from MUStARD and HuggingFace repositories,balanced across sarcastic and non-sarcastic classes.A sequential baseline model(LS... This study investigates sarcasm detection in text using a dataset of 8095 sentences compiled from MUStARD and HuggingFace repositories,balanced across sarcastic and non-sarcastic classes.A sequential baseline model(LSTM)is compared with transformer-based models(RoBERTa and XLNet),integrated with attention mechanisms.Transformers were chosen for their proven ability to capture long-range contextual dependencies,whereas LSTM serves as a traditional benchmark for sequential modeling.Experimental results show that RoBERTa achieves 0.87 accuracy,XLNet 0.83,and LSTM 0.52.These findings confirm that transformer architectures significantly outperform recurrent models in sarcasm detection.Future work will incorporate multimodal features and error analysis to further improve robustness. 展开更多
关键词 Attention mechanism LSTM natural language processing sarcasm detection sentiment analysis transformer models RoBERTa XLNet
在线阅读 下载PDF
The 3D-Geoformer for ENSO studies:a Transformer-based model with integrated gradient methods for enhanced explainability 被引量:2
5
作者 Lu ZHOU Rong-Hua ZHANG 《Journal of Oceanology and Limnology》 2025年第6期1688-1708,共21页
Deep learning(DL)has become a crucial technique for predicting the El Niño-Southern Oscillation(ENSO)and evaluating its predictability.While various DL-based models have been developed for ENSO predictions,many f... Deep learning(DL)has become a crucial technique for predicting the El Niño-Southern Oscillation(ENSO)and evaluating its predictability.While various DL-based models have been developed for ENSO predictions,many fail to capture the coherent multivariate evolution within the coupled ocean-atmosphere system of the tropical Pacific.To address this three-dimensional(3D)limitation and represent ENSO-related ocean-atmosphere interactions more accurately,a novel this 3D multivariate prediction model was proposed based on a Transformer architecture,which incorporates a spatiotemporal self-attention mechanism.This model,named 3D-Geoformer,offers several advantages,enabling accurate ENSO predictions up to one and a half years in advance.Furthermore,an integrated gradient method was introduced into the model to identify the sources of predictability for sea surface temperature(SST)variability in the eastern equatorial Pacific.Results reveal that the 3D-Geoformer effectively captures ENSO-related precursors during the evolution of ENSO events,particularly the thermocline feedback processes and ocean temperature anomaly pathways on and off the equator.By extending DL-based ENSO predictions from one-dimensional Niño time series to 3D multivariate fields,the 3D-Geoformer represents a significant advancement in ENSO prediction.This study provides details in the model formulation,analysis procedures,sensitivity experiments,and illustrative examples,offering practical guidance for the application of the model in ENSO research. 展开更多
关键词 transformer model 3 D-Geoformer El Niño-Southern Oscillation(ENSO)prediction explainable artificial intelligence(XAI) integrated gradient method
在线阅读 下载PDF
Millimeter-wave modeling based on transformer model for InP high electron mobility transistor
6
作者 ZHANG Ya-Xue ZHANG Ao GAO Jian-Jun 《红外与毫米波学报》 北大核心 2025年第4期534-539,共6页
In this paper,the small-signal modeling of the Indium Phosphide High Electron Mobility Transistor(InP HEMT)based on the Transformer neural network model is investigated.The AC S-parameters of the HEMT device are train... In this paper,the small-signal modeling of the Indium Phosphide High Electron Mobility Transistor(InP HEMT)based on the Transformer neural network model is investigated.The AC S-parameters of the HEMT device are trained and validated using the Transformer model.In the proposed model,the eight-layer transformer encoders are connected in series and the encoder layer of each Transformer consists of the multi-head attention layer and the feed-forward neural network layer.The experimental results show that the measured and modeled S-parameters of the HEMT device match well in the frequency range of 0.5-40 GHz,with the errors versus frequency less than 1%.Compared with other models,good accuracy can be achieved to verify the effectiveness of the proposed model. 展开更多
关键词 transformer model neural network high electron mobility transistor(HEMT) small signal model
在线阅读 下载PDF
Mobility-Aware Edge Caching with Transformer-DQN in D2D-Enabled Heterogeneous Networks
7
作者 Yiming Guo Hongyu Ma 《Computers, Materials & Continua》 2025年第11期3485-3505,共21页
In dynamic 5G network environments,user mobility and heterogeneous network topologies pose dual challenges to the effort of improving performance of mobile edge caching.Existing studies often overlook the dynamic natu... In dynamic 5G network environments,user mobility and heterogeneous network topologies pose dual challenges to the effort of improving performance of mobile edge caching.Existing studies often overlook the dynamic nature of user locations and the potential of device-to-device(D2D)cooperative caching,limiting the reduction of transmission latency.To address this issue,this paper proposes a joint optimization scheme for edge caching that integrates user mobility prediction with deep reinforcement learning.First,a Transformer-based geolocation prediction model is designed,leveraging multi-head attention mechanisms to capture correlations in historical user trajectories for accurate future location prediction.Then,within a three-tier heterogeneous network,we formulate a latency minimization problem under a D2D cooperative caching architecture and develop a mobility-aware Deep Q-Network(DQN)caching strategy.This strategy takes predicted location information as state input and dynamically adjusts the content distribution across small base stations(SBSs)andmobile users(MUs)to reduce end-to-end delay inmulti-hop content retrieval.Simulation results show that the proposed DQN-based method outperforms other baseline strategies across variousmetrics,achieving a 17.2%reduction in transmission delay compared to DQNmethods withoutmobility integration,thus validating the effectiveness of the joint optimization of location prediction and caching decisions. 展开更多
关键词 Mobile edge caching D2D heterogeneous networks deep reinforcement learning transformer model transmission delay optimization
在线阅读 下载PDF
High-precision copper-grade identification via a vision transformer with PGNAA
8
作者 Jie Cao Chong-Gui Zhong +6 位作者 Han-Ting You Yan Zhang Ren-Bo Wang Shu-Min Zhou Jin-Hui Qu Rui Chen Shi-Liang Liu 《Nuclear Science and Techniques》 2025年第7期89-99,共11页
The identification of ore grades is a critical step in mineral resource exploration and mining.Prompt gamma neutron activation analysis(PGNAA)technology employs gamma rays generated by the nuclear reactions between ne... The identification of ore grades is a critical step in mineral resource exploration and mining.Prompt gamma neutron activation analysis(PGNAA)technology employs gamma rays generated by the nuclear reactions between neutrons and samples to achieve the qualitative and quantitative detection of sample components.In this study,we present a novel method for identifying copper grade by combining the vision transformer(ViT)model with the PGNAA technique.First,a Monte Carlo simulation is employed to determine the optimal sizes of the neutron moderator,thermal neutron absorption material,and dimensions of the device.Subsequently,based on the parameters obtained through optimization,a PGNAA copper ore measurement model is established.The gamma spectrum of the copper ore is analyzed using the ViT model.The ViT model is optimized for hyperparameters using a grid search.To ensure the reliability of the identification results,the test results are obtained through five repeated tenfold cross-validations.Long short-term memory and convolutional neural network models are compared with the ViT method.These results indicate that the ViT method is efficient in identifying copper ore grades with average accuracy,precision,recall,F_(1)score,and F_(1)(-)score values of 0.9795,0.9637,0.9614,0.9625,and 0.9942,respectively.When identifying associated minerals,the ViT model can identify Pb,Zn,Fe,and Co minerals with identification accuracies of 0.9215,0.9396,0.9966,and 0.8311,respectively. 展开更多
关键词 Copper-grade identification Vision transformer model Prompt gamma neutron activation analysis Monte Carlo N-particle
在线阅读 下载PDF
Multi⁃Step Short⁃Term Traffic Flow Prediction of Urban Road Network Based on ISTA⁃Transformer Model
9
作者 Leyao Xiao Qian Chen 《Journal of Harbin Institute of Technology(New Series)》 2025年第6期1-14,共14页
Short⁃term traffic flow prediction plays a crucial role in the planning of intelligent transportation systems.Nowadays,there is a large amount of traffic flow data generated from the monitoring devices of urban road n... Short⁃term traffic flow prediction plays a crucial role in the planning of intelligent transportation systems.Nowadays,there is a large amount of traffic flow data generated from the monitoring devices of urban road networks,which contains road network traffic information with high application value.In this study,an improved spatio⁃temporal attention transformer model(ISTA⁃transformer model)is proposed to provide a more accurate method for predicting multi⁃step short⁃term traffic flow based on monitoring data.By embedding a temporal attention layer and a spatial attention layer in the model,the model learns the relationship between traffic flows at different time intervals and different geographic locations,and realizes more accurate multi⁃step short⁃time flow prediction.Finally,we validate the superiority of the model with monitoring data spanning 15 days from 620 monitoring points in Qingdao,China.In the four time steps of prediction,the MAPE(Mean Absolute Percentage Error)values of ISTA⁃transformers prediction results are 0.22,0.29,0.37,and 0.38,respectively,and its prediction accuracy is usually better than that of six baseline models(Transformer,GRU,CNN,LSTM,Seq2Seq and LightGBM),which indicates that the proposed model in this paper always has a better ability to explain the prediction results with the time steps in the multi⁃step prediction. 展开更多
关键词 urban road network traffic flow prediction spatio⁃temporal feature ISTA⁃transformer model
在线阅读 下载PDF
Retinexformer+:Retinex-Based Dual-Channel Transformer for Low-Light Image Enhancement
10
作者 Song Liu Hongying Zhang +1 位作者 Xue Li Xi Yang 《Computers, Materials & Continua》 2025年第2期1969-1984,共16页
Enhancing low-light images with color distortion and uneven multi-light source distribution presents challenges. Most advanced methods for low-light image enhancement are based on the Retinex model using deep learning... Enhancing low-light images with color distortion and uneven multi-light source distribution presents challenges. Most advanced methods for low-light image enhancement are based on the Retinex model using deep learning. Retinexformer introduces channel self-attention mechanisms in the IG-MSA. However, it fails to effectively capture long-range spatial dependencies, leaving room for improvement. Based on the Retinexformer deep learning framework, we designed the Retinexformer+ network. The “+” signifies our advancements in extracting long-range spatial dependencies. We introduced multi-scale dilated convolutions in illumination estimation to expand the receptive field. These convolutions effectively capture the weakening semantic dependency between pixels as distance increases. In illumination restoration, we used Unet++ with multi-level skip connections to better integrate semantic information at different scales. The designed Illumination Fusion Dual Self-Attention (IF-DSA) module embeds multi-scale dilated convolutions to achieve spatial self-attention. This module captures long-range spatial semantic relationships within acceptable computational complexity. Experimental results on the Low-Light (LOL) dataset show that Retexformer+ outperforms other State-Of-The-Art (SOTA) methods in both quantitative and qualitative evaluations, with the computational complexity increased to an acceptable 51.63 G FLOPS. On the LOL_v1 dataset, RetinexFormer+ shows an increase of 1.15 in Peak Signal-to-Noise Ratio (PSNR) and a decrease of 0.39 in Root Mean Square Error (RMSE). On the LOL_v2_real dataset, the PSNR increases by 0.42 and the RMSE decreases by 0.18. Experimental results on the Exdark dataset show that Retexformer+ can effectively enhance real-scene images and maintain their semantic information. 展开更多
关键词 Low-light image enhancement RETINEX transformer model
在线阅读 下载PDF
Generating Abstractive Summaries from Social Media Discussions Using Transformers
11
作者 Afrodite Papagiannopoulou Chrissanthi Angeli Mazida Ahmad 《Open Journal of Applied Sciences》 2025年第1期239-258,共20页
The rise of social media platforms has revolutionized communication, enabling the exchange of vast amounts of data through text, audio, images, and videos. These platforms have become critical for sharing opinions and... The rise of social media platforms has revolutionized communication, enabling the exchange of vast amounts of data through text, audio, images, and videos. These platforms have become critical for sharing opinions and insights, influencing daily habits, and driving business, political, and economic decisions. Text posts are particularly significant, and natural language processing (NLP) has emerged as a powerful tool for analyzing such data. While traditional NLP methods have been effective for structured media, social media content poses unique challenges due to its informal and diverse nature. This has spurred the development of new techniques tailored for processing and extracting insights from unstructured user-generated text. One key application of NLP is the summarization of user comments to manage overwhelming content volumes. Abstractive summarization has proven highly effective in generating concise, human-like summaries, offering clear overviews of key themes and sentiments. This enhances understanding and engagement while reducing cognitive effort for users. For businesses, summarization provides actionable insights into customer preferences and feedback, enabling faster trend analysis, improved responsiveness, and strategic adaptability. By distilling complex data into manageable insights, summarization plays a vital role in improving user experiences and empowering informed decision-making in a data-driven landscape. This paper proposes a new implementation framework by fine-tuning and parameterizing Transformer Large Language Models to manage and maintain linguistic and semantic components in abstractive summary generation. The system excels in transforming large volumes of data into meaningful summaries, as evidenced by its strong performance across metrics like fluency, consistency, readability, and semantic coherence. 展开更多
关键词 Abstractive Summarization transformerS Social Media Summarization transformer Language models
在线阅读 下载PDF
CT-NET: A Novel Convolutional Transformer-Based Network for Short-Term Solar Energy Forecasting Using Climatic Information 被引量:1
12
作者 Muhammad Munsif Fath U Min Ullah +2 位作者 Samee Ullah Khan Noman Khan Sung Wook Baik 《Computer Systems Science & Engineering》 SCIE EI 2023年第11期1751-1773,共23页
Photovoltaic(PV)systems are environmentally friendly,generate green energy,and receive support from policies and organizations.However,weather fluctuations make large-scale PV power integration and management challeng... Photovoltaic(PV)systems are environmentally friendly,generate green energy,and receive support from policies and organizations.However,weather fluctuations make large-scale PV power integration and management challenging despite the economic benefits.Existing PV forecasting techniques(sequential and convolutional neural networks(CNN))are sensitive to environmental conditions,reducing energy distribution system performance.To handle these issues,this article proposes an efficient,weather-resilient convolutional-transformer-based network(CT-NET)for accurate and efficient PV power forecasting.The network consists of three main modules.First,the acquired PV generation data are forwarded to the pre-processing module for data refinement.Next,to carry out data encoding,a CNNbased multi-head attention(MHA)module is developed in which a single MHA is used to decode the encoded data.The encoder module is mainly composed of 1D convolutional and MHA layers,which extract local as well as contextual features,while the decoder part includes MHA and feedforward layers to generate the final prediction.Finally,the performance of the proposed network is evaluated using standard error metrics,including the mean squared error(MSE),root mean squared error(RMSE),and mean absolute percentage error(MAPE).An ablation study and comparative analysis with several competitive state-of-the-art approaches revealed a lower error rate in terms of MSE(0.0471),RMSE(0.2167),and MAPE(0.6135)over publicly available benchmark data.In addition,it is demonstrated that our proposed model is less complex,with the lowest number of parameters(0.0135 M),size(0.106 MB),and inference time(2 ms/step),suggesting that it is easy to integrate into the smart grid. 展开更多
关键词 Solar energy forecasting renewable energy systems photovoltaic generation forecasting time series data transformer models deep learning machine learning
在线阅读 下载PDF
E2ETCA:End-to-end training of CNN and attention ensembles for rice disease diagnosis
13
作者 Md.Zasim Uddin Md.Nadim Mahamood +3 位作者 Ausrukona Ray Md.Ileas Pramanik Fady Alnajjar Md Atiqur Rahman Ahad 《Journal of Integrative Agriculture》 2026年第2期756-768,共13页
Rice is one of the most important staple crops globally.Rice plant diseases can severely reduce crop yields and,in extreme cases,lead to total production loss.Early diagnosis enables timely intervention,mitigates dise... Rice is one of the most important staple crops globally.Rice plant diseases can severely reduce crop yields and,in extreme cases,lead to total production loss.Early diagnosis enables timely intervention,mitigates disease severity,supports effective treatment strategies,and reduces reliance on excessive pesticide use.Traditional machine learning approaches have been applied for automated rice disease diagnosis;however,these methods depend heavily on manual image preprocessing and handcrafted feature extraction,which are labor-intensive and time-consuming and often require domain expertise.Recently,end-to-end deep learning(DL) models have been introduced for this task,but they often lack robustness and generalizability across diverse datasets.To address these limitations,we propose a novel end-toend training framework for convolutional neural network(CNN) and attention-based model ensembles(E2ETCA).This framework integrates features from two state-of-the-art(SOTA) CNN models,Inception V3 and DenseNet-201,and an attention-based vision transformer(ViT) model.The fused features are passed through an additional fully connected layer with softmax activation for final classification.The entire process is trained end-to-end,enhancing its suitability for realworld deployment.Furthermore,we extract and analyze the learned features using a support vector machine(SVM),a traditional machine learning classifier,to provide comparative insights.We evaluate the proposed E2ETCA framework on three publicly available datasets,the Mendeley Rice Leaf Disease Image Samples dataset,the Kaggle Rice Diseases Image dataset,the Bangladesh Rice Research Institute dataset,and a combined version of all three.Using standard evaluation metrics(accuracy,precision,recall,and F1-score),our framework demonstrates superior performance compared to existing SOTA methods in rice disease diagnosis,with potential applicability to other agricultural disease detection tasks. 展开更多
关键词 rice disease diagnosis ensemble method CNN-based model end-to-end model Inception model DenseNet model vision transformer model attention-based model support vector machine
在线阅读 下载PDF
Knowledge graph-enhanced framework for electric power engineering report generation
14
作者 Chen Qian Yu-Yan Chen +3 位作者 Jia-Ying Yang Xiao-Wen Le Xiao-Yang Shen Yi-Heng Zeng 《Journal of Electronic Science and Technology》 2026年第1期46-64,共19页
Due to the complex structural hierarchy,with deeply nested associative relations between entities such as equipment,specifications,and business processes,intelligent power grid engineering is challenging.Meanwhile,lim... Due to the complex structural hierarchy,with deeply nested associative relations between entities such as equipment,specifications,and business processes,intelligent power grid engineering is challenging.Meanwhile,limited by the fragmented data and loss of contextual information,the generated reports are prone to the problems such as content redundancy and omission of critical information,failing to meet the demands of efficient decision-making and accurate management in modern power systems.To address these issues,this paper proposes a knowledge graph(KG)-enhanced framework to automatically generate electric power engineering reports.In the KG construction phase,a feature-fused entity recognition model named BERT-BiLSTM-CRF is adopted to improve the accuracy of entity recognition in scenarios involving power engineering professional terminology,thereby solving the problem of ambiguous entity boundaries in traditional models;then a BERT-attention relation extraction model is proposed to enhance the completeness of extracting complex hierarchical and implicit relations in power grid data.In the report generation phase,an improved Transformer architecture is adopted to accurately transform structured knowledge into natural language reports that comply with engineering specifications,addressing the issue of semantic inconsistency caused by the loss of structural information in existing models.By validating with real-world projects,the results show that the proposed framework significantly outperforms existing baseline models in entity recognition,confirming its superiority and applicability in practical engineering. 展开更多
关键词 Entity recognition Improved transformer model Knowledge graph enhancement Power grid engineering report generation Relation extraction
在线阅读 下载PDF
Micro-expression recognition algorithm based on graph convolutional network and Transformer model 被引量:1
15
作者 吴进 PANG Wenting +1 位作者 WANG Lei ZHAO Bo 《High Technology Letters》 EI CAS 2023年第2期213-222,共10页
Micro-expressions are spontaneous, unconscious movements that reveal true emotions.Accurate facial movement information and network training learning methods are crucial for micro-expression recognition.However, most ... Micro-expressions are spontaneous, unconscious movements that reveal true emotions.Accurate facial movement information and network training learning methods are crucial for micro-expression recognition.However, most existing micro-expression recognition technologies so far focus on modeling the single category of micro-expression images and neural network structure.Aiming at the problems of low recognition rate and weak model generalization ability in micro-expression recognition, a micro-expression recognition algorithm is proposed based on graph convolution network(GCN) and Transformer model.Firstly, action unit(AU) feature detection is extracted and facial muscle nodes in the neighborhood are divided into three subsets for recognition.Then, graph convolution layer is used to find the layout of dependencies between AU nodes of micro-expression classification.Finally, multiple attentional features of each facial action are enriched with Transformer model to include more sequence information before calculating the overall correlation of each region.The proposed method is validated in CASME II and CAS(ME)^2 datasets, and the recognition rate reached 69.85%. 展开更多
关键词 micro-expression recognition graph convolutional network(GCN) action unit(AU)detection transformer model
在线阅读 下载PDF
Formal Verification of TASM Models by Translating into UPPAAL 被引量:1
16
作者 胡凯 张腾 +3 位作者 杨志斌 顾斌 蒋树 姜泮昌 《Journal of Donghua University(English Edition)》 EI CAS 2012年第1期51-54,共4页
Timed abstract state machine(TASM) is a formal specification language used to specify and simulate the behavior of real-time systems. Formal verification of TASM model can be fulfilled through model checking activitie... Timed abstract state machine(TASM) is a formal specification language used to specify and simulate the behavior of real-time systems. Formal verification of TASM model can be fulfilled through model checking activities by translating into UPPAAL. Firstly, the translational semantics from TASM to UPPAAL is presented through atlas transformation language(ATL). Secondly, the implementation of the proposed model transformation tool TASM2UPPAAL is provided. Finally, a case study is given to illustrate the automatic transformation from TASM model to UPPAAL model. 展开更多
关键词 timed abstract state machine(TASM) formal verification model transformation atlas transformation language(ATL) UPPAAL
在线阅读 下载PDF
Model Transformer Evaluation of High-Permeability Grain-Oriented Electrical Steels 被引量:1
17
作者 Masayoshi Ishida, Seiji Okabe, Takeshi Imamura and Michiro Komatsubara (Kawasaki Steel Corporation, Kurashiki 712-8511, Japan) 《Journal of Materials Science & Technology》 SCIE EI CAS CSCD 2000年第2期223-227,共5页
The dependence of transformer performance on the material properties was investigated using two laboratory-processed 0.23 mm thick grain-oriented electrical steels domain-refined with elec-trolytically etched grooves ... The dependence of transformer performance on the material properties was investigated using two laboratory-processed 0.23 mm thick grain-oriented electrical steels domain-refined with elec-trolytically etched grooves having different magnetic properties. The iron loss at 1.7 T, 50 Hz and the flux density at 800 A/m of material A were 0.73 W/kg and 1.89 T, respectively; and those of material B, 0.83 W/kg and 1.88 T. Model stacked and wound transformer core experiments using the tested materials exhibited performance well reflecting the material characteristics. In a three-phase stacked core with step-lap joints excited to 1.7 T, 50 Hz, the core loss, the exciting current and the noise level were 0.86 W/kg, 0.74 A and 52 dB, respectively, with material A; and 0.97 W/kg, 1.0 A and 54 dB with material B. The building factors for the core losses of the two materials were almost the same in both core configurations. The effect of higher harmonics on transformer performance was also investigated. 展开更多
关键词 Model transformer Evaluation of High-Permeability Grain-Oriented Electrical Steels
在线阅读 下载PDF
A Deep Learning Ensemble Method for Forecasting Daily Crude Oil Price Based on Snapshot Ensemble of Transformer Model
18
作者 Ahmed Fathalla Zakaria Alameer +1 位作者 Mohamed Abbas Ahmed Ali 《Computer Systems Science & Engineering》 SCIE EI 2023年第7期929-950,共22页
The oil industries are an important part of a country’s economy.The crude oil’s price is influenced by a wide range of variables.Therefore,how accurately can countries predict its behavior and what predictors to emp... The oil industries are an important part of a country’s economy.The crude oil’s price is influenced by a wide range of variables.Therefore,how accurately can countries predict its behavior and what predictors to employ are two main questions.In this view,we propose utilizing deep learning and ensemble learning techniques to boost crude oil’s price forecasting performance.The suggested method is based on a deep learning snapshot ensemble method of the Transformer model.To examine the superiority of the proposed model,this paper compares the proposed deep learning ensemble model against different machine learning and statistical models for daily Organization of the Petroleum Exporting Countries(OPEC)oil price forecasting.Experimental results demonstrated the outperformance of the proposed method over statistical and machine learning methods.More precisely,the proposed snapshot ensemble of Transformer method achieved relative improvement in the forecasting performance compared to autoregressive integrated moving average ARIMA(1,1,1),ARIMA(0,1,1),autoregressive moving average(ARMA)(0,1),vector autoregression(VAR),random walk(RW),support vector machine(SVM),and random forests(RF)models by 99.94%,99.62%,99.87%,99.65%,7.55%,98.38%,and 99.35%,respectively,according to mean square error metric. 展开更多
关键词 Deep learning ensemble learning transformer model crude oil price
在线阅读 下载PDF
Vehicle Density Prediction in Low Quality Videos with Transformer Timeseries Prediction Model(TTPM)
19
作者 D.Suvitha M.Vijayalakshmi 《Computer Systems Science & Engineering》 SCIE EI 2023年第1期873-894,共22页
Recent advancement in low-cost cameras has facilitated surveillance in various developing towns in India.The video obtained from such surveillance are of low quality.Still counting vehicles from such videos are necess... Recent advancement in low-cost cameras has facilitated surveillance in various developing towns in India.The video obtained from such surveillance are of low quality.Still counting vehicles from such videos are necessity to avoid traf-fic congestion and allows drivers to plan their routes more precisely.On the other hand,detecting vehicles from such low quality videos are highly challenging with vision based methodologies.In this research a meticulous attempt is made to access low-quality videos to describe traffic in Salem town in India,which is mostly an un-attempted entity by most available sources.In this work profound Detection Transformer(DETR)model is used for object(vehicle)detection.Here vehicles are anticipated in a rush-hour traffic video using a set of loss functions that carry out bipartite coordinating among estimated and information acquired on real attributes.Every frame in the traffic footage has its date and time which is detected and retrieved using Tesseract Optical Character Recognition.The date and time extricated and perceived from the input image are incorporated with the length of the recognized objects acquired from the DETR model.This furnishes the vehicles report with timestamp.Transformer Timeseries Prediction Model(TTPM)is proposed to predict the density of the vehicle for future prediction,here the regular NLP layers have been removed and the encoding temporal layer has been modified.The proposed TTPM error rate outperforms the existing models with RMSE of 4.313 and MAE of 3.812. 展开更多
关键词 Detection transformer self-attention tesseract optical character recognition transformer timeseries prediction model time encoding vector
在线阅读 下载PDF
AADL2TASM: a Verification and Analysis Tool for AADL Models
20
作者 蒋树 胡凯 +3 位作者 杨志斌 顾斌 张腾 姜泮昌 《Journal of Donghua University(English Edition)》 EI CAS 2012年第1期94-98,共5页
Architecture analysis and design language (AADL) is an architecture description language standard for embedded real-time systems and it is widely used in safety-critical applications. For facilitating verifcafion an... Architecture analysis and design language (AADL) is an architecture description language standard for embedded real-time systems and it is widely used in safety-critical applications. For facilitating verifcafion and analysis, model transformation is one of the methods. A synchronous subset of AADL and a general methodology for translating the AADL subset into timed abstract state machine (TASM) were studied. Based on the arias transformation language ( ATL ) framework, the associated translating tool AADL2TASM was implemented by defining the meta-model of both AADL and TASM, and the ATL transformation rules. A case study with property verification of the AADL model was also presented for validating the tool. 展开更多
关键词 architecture analysis and design language AADL timed abstract state machine TASM model transformation atlas transformation languaee( ATL
在线阅读 下载PDF
上一页 1 2 5 下一页 到第
使用帮助 返回顶部