期刊文献+
共找到621篇文章
< 1 2 32 >
每页显示 20 50 100
KitWaSor:Pioneering pre-trained model for kitchen waste sorting with an innovative million-level benchmark dataset
1
作者 Leyuan Fang Shuaiyu Ding +3 位作者 Hao Feng Junwu Yu Lin Tang Pedram Ghamisi 《CAAI Transactions on Intelligence Technology》 2025年第1期94-114,共21页
Intelligent sorting is an important prerequisite for the full quantitative consumption and harmless disposal of kitchen waste.The existing object detection method based on an ImageNet pre-trained model is an effective... Intelligent sorting is an important prerequisite for the full quantitative consumption and harmless disposal of kitchen waste.The existing object detection method based on an ImageNet pre-trained model is an effective way of sorting.Owing to significant domain gaps between natural images and kitchen waste images,it is difficult to reflect the characteristics of diverse scales and dense distribution in kitchen waste based on an ImageNet pre-trained model,leading to poor generalisation.In this article,the authors propose the first pre-trained model for kitchen waste sorting called KitWaSor,which combines both contrastive learning(CL)and masked image modelling(MIM)through self-supervised learning(SSL).First,to address the issue of diverse scales,the authors propose a mixed masking strategy by introducing an incomplete masking branch based on the original random masking branch.It prevents the complete loss of small-scale objects while avoiding excessive leakage of large-scale object pixels.Second,to address the issue of dense distribution,the authors introduce semantic consistency constraints on the basis of the mixed masking strategy.That is,object semantic reasoning is performed through semantic consistency constraints to compensate for the lack of contextual information.To train KitWaSor,the authors construct the first million-level kitchen waste dataset across seasonal and regional distributions,named KWD-Million.Extensive experiments show that KitWaSor achieves state-of-the-art(SOTA)performance on the two most relevant downstream tasks for kitchen waste sorting(i.e.image classification and object detection),demonstrating the effectiveness of the proposed KitWaSor. 展开更多
关键词 contrastive learning kitchen waste masked image modeling pre-trained model self-supervised learning
在线阅读 下载PDF
DPCIPI: A pre-trained deep learning model for predicting cross-immunity between drifted strains of Influenza A/H3N2
2
作者 Yiming Du Zhuotian Li +8 位作者 Qian He Thomas Wetere Tulu Kei Hang Katie Chan Lin Wang Sen Pei Zhanwei Du Zhen Wang Xiao-Ke Xu Xiao Fan Liu 《Journal of Automation and Intelligence》 2025年第2期115-124,共10页
Predicting cross-immunity between viral strains is vital for public health surveillance and vaccine development.Traditional neural network methods,such as BiLSTM,could be ineffective due to the lack of lab data for mo... Predicting cross-immunity between viral strains is vital for public health surveillance and vaccine development.Traditional neural network methods,such as BiLSTM,could be ineffective due to the lack of lab data for model training and the overshadowing of crucial features within sequence concatenation.The current work proposes a less data-consuming model incorporating a pre-trained gene sequence model and a mutual information inference operator.Our methodology utilizes gene alignment and deduplication algorithms to preprocess gene sequences,enhancing the model’s capacity to discern and focus on distinctions among input gene pairs.The model,i.e.,DNA Pretrained Cross-Immunity Protection Inference model(DPCIPI),outperforms state-of-theart(SOTA)models in predicting hemagglutination inhibition titer from influenza viral gene sequences only.Improvement in binary cross-immunity prediction is 1.58%in F1,2.34%in precision,1.57%in recall,and 1.57%in Accuracy.For multilevel cross-immunity improvements,the improvement is 2.12%in F1,3.50%in precision,2.19%in recall,and 2.19%in Accuracy.Our study showcases the potential of pre-trained gene models to improve predictions of antigenic variation and cross-immunity.With expanding gene data and advancements in pre-trained models,this approach promises significant impacts on vaccine development and public health. 展开更多
关键词 Cross-immunity prediction pre-trained model Deep learning Influenza strains Hemagglutination inhibition
在线阅读 下载PDF
Multilingual Text Summarization in Healthcare Using Pre-Trained Transformer-Based Language Models
3
作者 Josua Käser Thomas Nagy +1 位作者 Patrick Stirnemann Thomas Hanne 《Computers, Materials & Continua》 2025年第4期201-217,共17页
We analyze the suitability of existing pre-trained transformer-based language models(PLMs)for abstractive text summarization on German technical healthcare texts.The study focuses on the multilingual capabilities of t... We analyze the suitability of existing pre-trained transformer-based language models(PLMs)for abstractive text summarization on German technical healthcare texts.The study focuses on the multilingual capabilities of these models and their ability to perform the task of abstractive text summarization in the healthcare field.The research hypothesis was that large language models could perform high-quality abstractive text summarization on German technical healthcare texts,even if the model is not specifically trained in that language.Through experiments,the research questions explore the performance of transformer language models in dealing with complex syntax constructs,the difference in performance between models trained in English and German,and the impact of translating the source text to English before conducting the summarization.We conducted an evaluation of four PLMs(GPT-3,a translation-based approach also utilizing GPT-3,a German language Model,and a domain-specific bio-medical model approach).The evaluation considered the informativeness using 3 types of metrics based on Recall-Oriented Understudy for Gisting Evaluation(ROUGE)and the quality of results which is manually evaluated considering 5 aspects.The results show that text summarization models could be used in the German healthcare domain and that domain-independent language models achieved the best results.The study proves that text summarization models can simplify the search for pre-existing German knowledge in various domains. 展开更多
关键词 Text summarization pre-trained transformer-based language models large language models technical healthcare texts natural language processing
在线阅读 下载PDF
A Classification–Detection Approach of COVID-19 Based on Chest X-ray and CT by Using Keras Pre-Trained Deep Learning Models 被引量:10
4
作者 Xing Deng Haijian Shao +2 位作者 Liang Shi Xia Wang Tongling Xie 《Computer Modeling in Engineering & Sciences》 SCIE EI 2020年第11期579-596,共18页
The Coronavirus Disease 2019(COVID-19)is wreaking havoc around the world,bring out that the enormous pressure on national health and medical staff systems.One of the most effective and critical steps in the fight agai... The Coronavirus Disease 2019(COVID-19)is wreaking havoc around the world,bring out that the enormous pressure on national health and medical staff systems.One of the most effective and critical steps in the fight against COVID-19,is to examine the patient’s lungs based on the Chest X-ray and CT generated by radiation imaging.In this paper,five keras-related deep learning models:ResNet50,InceptionResNetV2,Xception,transfer learning and pre-trained VGGNet16 is applied to formulate an classification-detection approaches of COVID-19.Two benchmark methods SVM(Support Vector Machine),CNN(Conventional Neural Networks)are provided to compare with the classification-detection approaches based on the performance indicators,i.e.,precision,recall,F1 scores,confusion matrix,classification accuracy and three types of AUC(Area Under Curve).The highest classification accuracy derived by classification-detection based on 5857 Chest X-rays and 767 Chest CTs are respectively 84%and 75%,which shows that the keras-related deep learning approaches facilitate accurate and effective COVID-19-assisted detection. 展开更多
关键词 COVID-19 detection deep learning transfer learning pre-trained models
在线阅读 下载PDF
Fine-Tuning Pre-Trained CodeBERT for Code Search in Smart Contract 被引量:1
5
作者 JIN Huan LI Qinying 《Wuhan University Journal of Natural Sciences》 CAS CSCD 2023年第3期237-245,共9页
Smart contracts,which automatically execute on decentralized platforms like Ethereum,require high security and low gas consumption.As a result,developers have a strong demand for semantic code search tools that utiliz... Smart contracts,which automatically execute on decentralized platforms like Ethereum,require high security and low gas consumption.As a result,developers have a strong demand for semantic code search tools that utilize natural language queries to efficiently search for existing code snippets.However,existing code search models face a semantic gap between code and queries,which requires a large amount of training data.In this paper,we propose a fine-tuning approach to bridge the semantic gap in code search and improve the search accuracy.We collect 80723 different pairs of<comment,code snippet>from Etherscan.io and use these pairs to fine-tune,validate,and test the pre-trained CodeBERT model.Using the fine-tuned model,we develop a code search engine specifically for smart contracts.We evaluate the Recall@k and Mean Reciprocal Rank(MRR)of the fine-tuned CodeBERT model using different proportions of the finetuned data.It is encouraging that even a small amount of fine-tuned data can produce satisfactory results.In addition,we perform a comparative analysis between the fine-tuned CodeBERT model and the two state-of-the-art models.The experimental results show that the finetuned CodeBERT model has superior performance in terms of Recall@k and MRR.These findings highlight the effectiveness of our finetuning approach and its potential to significantly improve the code search accuracy. 展开更多
关键词 code search smart contract pre-trained code models program analysis machine learning
原文传递
Construction and application of knowledge graph for grid dispatch fault handling based on pre-trained model 被引量:1
6
作者 Zhixiang Ji Xiaohui Wang +1 位作者 Jie Zhang Di Wu 《Global Energy Interconnection》 EI CSCD 2023年第4期493-504,共12页
With the construction of new power systems,the power grid has become extremely large,with an increasing proportion of new energy and AC/DC hybrid connections.The dynamic characteristics and fault patterns of the power... With the construction of new power systems,the power grid has become extremely large,with an increasing proportion of new energy and AC/DC hybrid connections.The dynamic characteristics and fault patterns of the power grid are complex;additionally,power grid control is difficult,operation risks are high,and the task of fault handling is arduous.Traditional power-grid fault handling relies primarily on human experience.The difference in and lack of knowledge reserve of control personnel restrict the accuracy and timeliness of fault handling.Therefore,this mode of operation is no longer suitable for the requirements of new systems.Based on the multi-source heterogeneous data of power grid dispatch,this paper proposes a joint entity–relationship extraction method for power-grid dispatch fault processing based on a pre-trained model,constructs a knowledge graph of power-grid dispatch fault processing and designs,and develops a fault-processing auxiliary decision-making system based on the knowledge graph.It was applied to study a provincial dispatch control center,and it effectively improved the accident processing ability and intelligent level of accident management and control of the power grid. 展开更多
关键词 Power-grid dispatch fault handling Knowledge graph pre-trained model Auxiliary decision-making
在线阅读 下载PDF
Geometry-based BERT:An experimentally validated deep learning model for molecular property prediction in drug discovery
7
作者 Xiang Zhang Chenliang Qian +5 位作者 Bochao Yang Hongwei Jin Song Wu Jie Xia Fan Yang Liangren Zhang 《Journal of Pharmaceutical Analysis》 2025年第12期2960-2974,共15页
Various deep learning based methods have significantlyimpacted the realm of drug discovery.The development of deep learning methods for identifying novel structural types of active compounds has become an urgent chall... Various deep learning based methods have significantlyimpacted the realm of drug discovery.The development of deep learning methods for identifying novel structural types of active compounds has become an urgent challenge.In this paper,we introduce a self-supervised representation learning framework,i.e.,Geometry-based Bidirectional Encoder Representations from Transformers(GEO-BERT).GEO-BERT considers the information of atoms and chemical bonds in chemical structures as the input,and integrates the positional information of the three-dimensional conformation of the molecule for training.Specifically,GEO-BERT enhances its ability to characterize molecular structures by introducing three different positional relationships:atom-atom,bond-bond,and atom-bond.By benchmarking study,GEO-BERT has demonstrated optimal performance on multiple benchmarks.We also performed prospective study to validate the GEO-BERT model,with screening for DYRK1A inhibitors as a case.Two potent and novel DYRK1A inhibitors(IC_(50):<1μM)were ultimately discovered.Taken together,we have developed an open-source GEO-BERT model for molecular property prediction(https://github.com/drug-designer/GEO-BERT)and proved its practical utility in early-stage drug discovery. 展开更多
关键词 Drug discovery Chemical pre-trained model Self-supervised learning bert DYRK1A inhibitor
在线阅读 下载PDF
基于BERT微调与特征融合的在线日志异常检测方法
8
作者 彭思源 齐金鹏 +2 位作者 戴理 夏奕麒 成晨 《重庆邮电大学学报(自然科学版)》 北大核心 2026年第1期128-136,共9页
针对日志异常检测中快速性和检测精度难以协同优化的问题,提出了一种基于双向编码器表征模型微调与多特征融合的在线日志异常检测方法(log triple feature representation using BERT,LogTri-BERT)。设计了一种双阶段处理架构,通过TinyB... 针对日志异常检测中快速性和检测精度难以协同优化的问题,提出了一种基于双向编码器表征模型微调与多特征融合的在线日志异常检测方法(log triple feature representation using BERT,LogTri-BERT)。设计了一种双阶段处理架构,通过TinyBERT微调实现离线阶段日志语义的深度表征学习,利用多路搜索树(ternary search tree,TST)和滑动窗口机制,提取在线阶段的事件突变点密度特征,融合时序信息与语义嵌入特征形成联合表征向量,最终通过XGBoost(extreme gradient boosting)分类器实现高效判别。实验结果表明,LogTri-BERT在公开基准数据集上的分类速度和准确性均优于基线方法。 展开更多
关键词 日志分析 bert模型 TST算法 日志异常检测
在线阅读 下载PDF
融合提示工程与BERT-LSTM模型的“卡脖子”替代技术识别
9
作者 仵轩 李广建 +1 位作者 王楚涵 潘佳立 《图书馆论坛》 北大核心 2026年第3期135-147,共13页
针对“卡脖子”技术研究存在替代技术识别机制缺失与技术要素解析精度不足等局限,文章提出融合提示工程与BERT-LSTM模型的“卡脖子”替代技术识别方法。首先,基于商业管制清单(Commercial Control List,CCL)对ECCN物项进行解析,并开展... 针对“卡脖子”技术研究存在替代技术识别机制缺失与技术要素解析精度不足等局限,文章提出融合提示工程与BERT-LSTM模型的“卡脖子”替代技术识别方法。首先,基于商业管制清单(Commercial Control List,CCL)对ECCN物项进行解析,并开展专利检索工作,通过SPC算法提取技术主路径的关键核心专利;其次,运用大语言模型提示工程抽取“问题-方案对”,借此解析技术功效,并结合功能导向搜索(Function-Oriented Search,FOS)初步查找可能具备技术替代功效的相关专利;再次,采用BERT-LSTM模型对专利文本实施二元分类,精准识别出具备技术替代功效的专利样本;通过提示工程抽取“方案-类别对”,系统识别替代技术方案;最后,建立科学-产业双维度评估体系完成替代技术潜力分级。文章以光刻技术为例,阐述该识别方法的应用流程,系统识别出极紫外(Extreme Ultra-violet,EUV)光刻技术的五种替代技术及其替代潜力。 展开更多
关键词 “卡脖子”技术 替代技术 提示工程 bert-LSTM模型 光刻技术
在线阅读 下载PDF
基于Stroke BERT的中风知识图谱与智能诊疗研究
10
作者 于惠 祖红月 +1 位作者 王卫东 刘洪运 《计算机技术与发展》 2026年第2期126-131,共6页
为促进中医精准医疗发展,本研究聚焦中风病,构建了“领域自适应预训练—跨模态知识对齐—混合增强推理”框架。首先,基于中医文献与病历构建Stroke BERT预训练模型,采用动态全词掩码增强中医隐喻语义表征。实验表明,Stroke BERT在中医... 为促进中医精准医疗发展,本研究聚焦中风病,构建了“领域自适应预训练—跨模态知识对齐—混合增强推理”框架。首先,基于中医文献与病历构建Stroke BERT预训练模型,采用动态全词掩码增强中医隐喻语义表征。实验表明,Stroke BERT在中医特异性任务上显著优于通用BERT(F 1值达83.24%和98.15%),有效提升信息提取精度。其次,构建中西医融合中风知识图谱,通过BERT-INT模型实现中西医实体对齐,突破语义壁垒。在此基础上,提出KG-LLM协同推理架构,通过动态提示工程将知识图谱路径与GLM-4模型深度耦合。在Stroke QA和Stroke CQ数据集上,该架构的诊断符合率分别提升至59.05%和78.10%,较独立大语言模型显著提高(分别提升10.48和7.62百分点),实现了诊疗可解释性与准确性的双重提升。为中医知识图谱构建提供了领域适配方法,为多模态医学数据融合与智能诊疗提供了创新技术支撑。 展开更多
关键词 中风 Stroke bert 知识图谱 大语言模型 智能诊疗
在线阅读 下载PDF
Optimizing Airline Review Sentiment Analysis:A Comparative Analysis of LLaMA and BERT Models through Fine-Tuning and Few-Shot Learning
11
作者 Konstantinos I.Roumeliotis Nikolaos D.Tselikas Dimitrios K.Nasiopoulos 《Computers, Materials & Continua》 2025年第2期2769-2792,共24页
In the rapidly evolving landscape of natural language processing(NLP)and sentiment analysis,improving the accuracy and efficiency of sentiment classification models is crucial.This paper investigates the performance o... In the rapidly evolving landscape of natural language processing(NLP)and sentiment analysis,improving the accuracy and efficiency of sentiment classification models is crucial.This paper investigates the performance of two advanced models,the Large Language Model(LLM)LLaMA model and NLP BERT model,in the context of airline review sentiment analysis.Through fine-tuning,domain adaptation,and the application of few-shot learning,the study addresses the subtleties of sentiment expressions in airline-related text data.Employing predictive modeling and comparative analysis,the research evaluates the effectiveness of Large Language Model Meta AI(LLaMA)and Bidirectional Encoder Representations from Transformers(BERT)in capturing sentiment intricacies.Fine-tuning,including domain adaptation,enhances the models'performance in sentiment classification tasks.Additionally,the study explores the potential of few-shot learning to improve model generalization using minimal annotated data for targeted sentiment analysis.By conducting experiments on a diverse airline review dataset,the research quantifies the impact of fine-tuning,domain adaptation,and few-shot learning on model performance,providing valuable insights for industries aiming to predict recommendations and enhance customer satisfaction through a deeper understanding of sentiment in user-generated content(UGC).This research contributes to refining sentiment analysis models,ultimately fostering improved customer satisfaction in the airline industry. 展开更多
关键词 Sentiment classification review sentiment analysis user-generated content domain adaptation customer satisfaction LLaMA model bert model airline reviews LLM classification fine-tuning
在线阅读 下载PDF
基于BERT和GRU的Web攻击检测模型研究
12
作者 谭攀 王海珍 成雨玫 《齐齐哈尔大学学报(自然科学版)》 2026年第1期30-37,共8页
目前Web攻击日益频繁和复杂,现有入侵检测系统的检测准确率、实时性需要进一步改进。为此,提出一种基于BERT和GRU的Web攻击检测模型BERT_GRU。通过Web攻击检测脚本Scalp进行攻击类型标注,利用预训练的BERT模型对HTTP请求中的文本数据进... 目前Web攻击日益频繁和复杂,现有入侵检测系统的检测准确率、实时性需要进一步改进。为此,提出一种基于BERT和GRU的Web攻击检测模型BERT_GRU。通过Web攻击检测脚本Scalp进行攻击类型标注,利用预训练的BERT模型对HTTP请求中的文本数据进行上下文语义编码、全连接层处理,将获得的数据与原始文本数据拼接,输入GRU网络提取序列特征,输出通过Softmax计算,最终实现多分类。在CSIC 2010数据集和FWAF数据集上的实验结果表明,BERT_GRU模型在准确率、召回率、F1值等评价指标以及训练时间方面均优于对比模型。 展开更多
关键词 Web攻击检测 bert模型 GRU
在线阅读 下载PDF
Classification of Conversational Sentences Using an Ensemble Pre-Trained Language Model with the Fine-Tuned Parameter
13
作者 R.Sujatha K.Nimala 《Computers, Materials & Continua》 SCIE EI 2024年第2期1669-1686,共18页
Sentence classification is the process of categorizing a sentence based on the context of the sentence.Sentence categorization requires more semantic highlights than other tasks,such as dependence parsing,which requir... Sentence classification is the process of categorizing a sentence based on the context of the sentence.Sentence categorization requires more semantic highlights than other tasks,such as dependence parsing,which requires more syntactic elements.Most existing strategies focus on the general semantics of a conversation without involving the context of the sentence,recognizing the progress and comparing impacts.An ensemble pre-trained language model was taken up here to classify the conversation sentences from the conversation corpus.The conversational sentences are classified into four categories:information,question,directive,and commission.These classification label sequences are for analyzing the conversation progress and predicting the pecking order of the conversation.Ensemble of Bidirectional Encoder for Representation of Transformer(BERT),Robustly Optimized BERT pretraining Approach(RoBERTa),Generative Pre-Trained Transformer(GPT),DistilBERT and Generalized Autoregressive Pretraining for Language Understanding(XLNet)models are trained on conversation corpus with hyperparameters.Hyperparameter tuning approach is carried out for better performance on sentence classification.This Ensemble of Pre-trained Language Models with a Hyperparameter Tuning(EPLM-HT)system is trained on an annotated conversation dataset.The proposed approach outperformed compared to the base BERT,GPT,DistilBERT and XLNet transformer models.The proposed ensemble model with the fine-tuned parameters achieved an F1_score of 0.88. 展开更多
关键词 Bidirectional encoder for representation of transformer conversation ensemble model fine-tuning generalized autoregressive pretraining for language understanding generative pre-trained transformer hyperparameter tuning natural language processing robustly optimized bert pretraining approach sentence classification transformer models
在线阅读 下载PDF
Leveraging Vision-Language Pre-Trained Model and Contrastive Learning for Enhanced Multimodal Sentiment Analysis
14
作者 Jieyu An Wan Mohd Nazmee Wan Zainon Binfen Ding 《Intelligent Automation & Soft Computing》 SCIE 2023年第8期1673-1689,共17页
Multimodal sentiment analysis is an essential area of research in artificial intelligence that combines multiple modes,such as text and image,to accurately assess sentiment.However,conventional approaches that rely on... Multimodal sentiment analysis is an essential area of research in artificial intelligence that combines multiple modes,such as text and image,to accurately assess sentiment.However,conventional approaches that rely on unimodal pre-trained models for feature extraction from each modality often overlook the intrinsic connections of semantic information between modalities.This limitation is attributed to their training on unimodal data,and necessitates the use of complex fusion mechanisms for sentiment analysis.In this study,we present a novel approach that combines a vision-language pre-trained model with a proposed multimodal contrastive learning method.Our approach harnesses the power of transfer learning by utilizing a vision-language pre-trained model to extract both visual and textual representations in a unified framework.We employ a Transformer architecture to integrate these representations,thereby enabling the capture of rich semantic infor-mation in image-text pairs.To further enhance the representation learning of these pairs,we introduce our proposed multimodal contrastive learning method,which leads to improved performance in sentiment analysis tasks.Our approach is evaluated through extensive experiments on two publicly accessible datasets,where we demonstrate its effectiveness.We achieve a significant improvement in sentiment analysis accuracy,indicating the supe-riority of our approach over existing techniques.These results highlight the potential of multimodal sentiment analysis and underscore the importance of considering the intrinsic semantic connections between modalities for accurate sentiment assessment. 展开更多
关键词 Multimodal sentiment analysis vision–language pre-trained model contrastive learning sentiment classification
在线阅读 下载PDF
人工智能技术创新如何赋能企业成长——来自BERT模型的证据
15
作者 何小钢 毛莘娅 《财经问题研究》 北大核心 2026年第1期44-58,共15页
人工智能技术创新是促进新质生产力发展的关键,对于推动高质量发展具有重要意义。本文利用双向编码器表示模型识别人工智能专利,以刻画企业人工智能技术创新水平,并基于2012-2022年中国A股上市公司数据,利用双向固定效应模型实证检验了... 人工智能技术创新是促进新质生产力发展的关键,对于推动高质量发展具有重要意义。本文利用双向编码器表示模型识别人工智能专利,以刻画企业人工智能技术创新水平,并基于2012-2022年中国A股上市公司数据,利用双向固定效应模型实证检验了人工智能技术创新对企业成长的影响及作用机制。研究结果显示:人工智能技术创新能够促进企业成长,且该结论在经过一系列内生性检验和稳健性检验后仍成立;人工智能技术创新通过创新溢出效应、降本增效效应、产品竞争效应和人力资本质量效应促进企业成长;人工智能技术创新对企业成长的促进作用发生在高管受教育程度较高企业、较年轻的高管团队企业、吸收能力较强的企业、劳动密集型行业企业、技术密集型行业企业中。本文研究拓宽了人工智能技术创新对微观企业经济效应的研究范围,为中国人工智能技术创新的政策制定和企业数智化战略决策提供了启示。 展开更多
关键词 人工智能技术创新 企业成长 bert模型
在线阅读 下载PDF
基于改进BERT的多头自注意力非侵入式负荷分解方法
16
作者 孙晓晴 李元诚 王庆乐 《电力信息与通信技术》 2026年第1期45-54,共10页
针对非侵入式负荷分解方法负荷特征捕捉不足、负荷分解精度不够等问题,文章提出一种基于改进BERT(bidirectional encoder representations from transformers)模型的多头自注意力非侵入式负荷分解方法(frequency and temporal attention... 针对非侵入式负荷分解方法负荷特征捕捉不足、负荷分解精度不够等问题,文章提出一种基于改进BERT(bidirectional encoder representations from transformers)模型的多头自注意力非侵入式负荷分解方法(frequency and temporal attention-BERT, FAT-BERT)。首先通过傅里叶变换将时域数据转换为频域数据,采用多尺度卷积全面捕捉负荷信号的时域和频域特征,从而增强模型对多样化负荷信号的表达能力;其次,在多头自注意力机制中引入频率注意力机制,从而增强模型对时序数据中频率成分的感知能力,进一步改善复杂负荷模式的表示,改进BERT模型中增加局部自注意力从而减少不必要的全局计算,提升模型的运行速度;接着将残差连接和正则化技术结合使模型在训练过程中更加稳定,并且能够更好地避免过拟合,最后在REDD和UK-DALE数据集上对提出的方法进行实验,实验结果验证了所提方法的有效性。 展开更多
关键词 非侵入式负荷监测 负荷分解 改进bert模型 多头自注意力机制 频率注意力
在线阅读 下载PDF
基于BERT大语言模型的广东省人工智能发明专利测度与时空演化特征
17
作者 陈奕嘉 谭俊涛 杨瑞霖 《热带地理》 北大核心 2026年第1期154-166,共13页
在人工智能(AI)日益成为推动区域高质量发展的关键技术背景下,如何科学测度AI技术活动的空间分布和演化趋势,成为经济地理学研究的重要议题。然而,现有的识别方法在语义有效性、识别精度以及覆盖广度等方面仍存在不足,有待进一步提升与... 在人工智能(AI)日益成为推动区域高质量发展的关键技术背景下,如何科学测度AI技术活动的空间分布和演化趋势,成为经济地理学研究的重要议题。然而,现有的识别方法在语义有效性、识别精度以及覆盖广度等方面仍存在不足,有待进一步提升与完善。为此,文章以广东省2001-2021年约120万条授权发明专利摘要为基础,引入BERT大语言模型,构建基于语义理解的AI授权发明专利识别方法,识别出约20万条AI发明专利文本。在此基础上,结合BERTopic主题建模方法,进一步揭示AI发明专利在广东省的关注领域与动态演化特征。研究发现:1)2001-2021年广东省AI发明专利数量整体经历了缓慢增长和快速扩张2个阶段,AI发明专利占比则呈持续上升趋势,反映AI技术在区域创新体系中的重要性不断提升。2)空间分布方面,AI发明专利在大湾区高度集聚,深圳和广州2市占全省AI发明专利的75.1%,珠海和东莞两地在AI发明专利占比上亦显示出较强活力。在合作网络方面,广东省AI发明专利合作网络呈现显著的“核心―边缘”结构特征,城市间在合作网络中的地位差异明显。3)在技术内容上,广东省AI发明专利主要聚焦于数据与图像处理、机器人与自动化装置、智能交通与故障检测、智能家居与环境控制和生物模拟与图像分析五大主题,不同主题的演化节奏存在差异,其中图像处理技术最为活跃,机器人和智能交通等领域近年来快速增长。此外,广东省内不同城市在AI发明专利布局上也各具特色,形成了差异化的技术优势。 展开更多
关键词 人工智能 AI发明专利 bert大语言模型 主题建模 技术演化 合作网络 广东省
在线阅读 下载PDF
Adapter Based on Pre-Trained Language Models for Classification of Medical Text
18
作者 Quan Li 《Journal of Electronic Research and Application》 2024年第3期129-134,共6页
We present an approach to classify medical text at a sentence level automatically.Given the inherent complexity of medical text classification,we employ adapters based on pre-trained language models to extract informa... We present an approach to classify medical text at a sentence level automatically.Given the inherent complexity of medical text classification,we employ adapters based on pre-trained language models to extract information from medical text,facilitating more accurate classification while minimizing the number of trainable parameters.Extensive experiments conducted on various datasets demonstrate the effectiveness of our approach. 展开更多
关键词 Classification of medical text ADAPTER pre-trained language model
在线阅读 下载PDF
基于BERT模型的微博文本细粒度情感分析
19
作者 张逸民 李野 《现代信息科技》 2026年第4期112-115,121,共5页
随着社交媒体的快速发展,微博作为用户信息交流与情感表达的重要平台,积累了海量丰富的文本数据。文章在调研微博情感分析相关技术的基础上,提出一种基于BERT模型的细粒度情感分析方法,结合微博数据抓取与预处理技术,构建了高效的分析... 随着社交媒体的快速发展,微博作为用户信息交流与情感表达的重要平台,积累了海量丰富的文本数据。文章在调研微博情感分析相关技术的基础上,提出一种基于BERT模型的细粒度情感分析方法,结合微博数据抓取与预处理技术,构建了高效的分析框架。该方法通过微博开放API完成数据采集,利用BERT预训练模型实现文本向量化,并依托Transformer架构完成愤怒、高兴、中性、惊讶、悲伤、恐惧六类情绪的精准分类;同时引入情感词典与数据增强技术提升模型性能,借助可视化工具展示分析结果。研究表明,该方法在SMP2020数据集上取得了较高的分类精度,为微博文本细粒度情感分析提供了新思路。 展开更多
关键词 微博情感分析 细粒度情感 bert模型 社交媒体
在线阅读 下载PDF
基金经理语调、基金收益与投资者行为——基于BERT人工智能模型的检验 被引量:2
20
作者 沈红波 洪康隆 王锴 《现代金融研究》 北大核心 2025年第4期20-38,共19页
本文以2014-2023年我国基金年度报告“展望”章节文本为样本,运用BERT人工智能模型和传统的词袋法,分别构建“词藻堆砌”与“真情实感”变量,对比基金经理语调对基金未来收益的预测效果及其对个人投资者行为的影响。研究发现:(1)相比传... 本文以2014-2023年我国基金年度报告“展望”章节文本为样本,运用BERT人工智能模型和传统的词袋法,分别构建“词藻堆砌”与“真情实感”变量,对比基金经理语调对基金未来收益的预测效果及其对个人投资者行为的影响。研究发现:(1)相比传统的词袋法,BERT人工智能模型度量的基金经理语调更能识别基金经理的真情实感,对基金未来收益、顺境中的基金业绩持续性及逆境中的基金崩盘风险的预测效果更好,且上述识别优势在复杂度较高、可读性较差的“展望”文本中更显著;(2)个人投资者的“有限理性”特征和公募基金存在的委托代理问题导致个人投资者更容易被基金经理“词藻堆砌”的文本所吸引,进而导致投资收益率降低。本文将BERT人工智能模型引入财经文本情感分析,为文本语调的度量方法提供了新思路。 展开更多
关键词 基金经理语调 基金收益 投资者行为 bert模型
在线阅读 下载PDF
上一页 1 2 32 下一页 到第
使用帮助 返回顶部