期刊文献+
共找到620篇文章
< 1 2 31 >
每页显示 20 50 100
Geometry-based BERT:An experimentally validated deep learning model for molecular property prediction in drug discovery
1
作者 Xiang Zhang Chenliang Qian +5 位作者 Bochao Yang Hongwei Jin Song Wu Jie Xia Fan Yang Liangren Zhang 《Journal of Pharmaceutical Analysis》 2025年第12期2960-2974,共15页
Various deep learning based methods have significantlyimpacted the realm of drug discovery.The development of deep learning methods for identifying novel structural types of active compounds has become an urgent chall... Various deep learning based methods have significantlyimpacted the realm of drug discovery.The development of deep learning methods for identifying novel structural types of active compounds has become an urgent challenge.In this paper,we introduce a self-supervised representation learning framework,i.e.,Geometry-based Bidirectional Encoder Representations from Transformers(GEO-BERT).GEO-BERT considers the information of atoms and chemical bonds in chemical structures as the input,and integrates the positional information of the three-dimensional conformation of the molecule for training.Specifically,GEO-BERT enhances its ability to characterize molecular structures by introducing three different positional relationships:atom-atom,bond-bond,and atom-bond.By benchmarking study,GEO-BERT has demonstrated optimal performance on multiple benchmarks.We also performed prospective study to validate the GEO-BERT model,with screening for DYRK1A inhibitors as a case.Two potent and novel DYRK1A inhibitors(IC_(50):<1μM)were ultimately discovered.Taken together,we have developed an open-source GEO-BERT model for molecular property prediction(https://github.com/drug-designer/GEO-BERT)and proved its practical utility in early-stage drug discovery. 展开更多
关键词 Drug discovery Chemical pre-trained model Self-supervised learning bert DYRK1A inhibitor
在线阅读 下载PDF
基于BERT微调与特征融合的在线日志异常检测方法
2
作者 彭思源 齐金鹏 +2 位作者 戴理 夏奕麒 成晨 《重庆邮电大学学报(自然科学版)》 北大核心 2026年第1期128-136,共9页
针对日志异常检测中快速性和检测精度难以协同优化的问题,提出了一种基于双向编码器表征模型微调与多特征融合的在线日志异常检测方法(log triple feature representation using BERT,LogTri-BERT)。设计了一种双阶段处理架构,通过TinyB... 针对日志异常检测中快速性和检测精度难以协同优化的问题,提出了一种基于双向编码器表征模型微调与多特征融合的在线日志异常检测方法(log triple feature representation using BERT,LogTri-BERT)。设计了一种双阶段处理架构,通过TinyBERT微调实现离线阶段日志语义的深度表征学习,利用多路搜索树(ternary search tree,TST)和滑动窗口机制,提取在线阶段的事件突变点密度特征,融合时序信息与语义嵌入特征形成联合表征向量,最终通过XGBoost(extreme gradient boosting)分类器实现高效判别。实验结果表明,LogTri-BERT在公开基准数据集上的分类速度和准确性均优于基线方法。 展开更多
关键词 日志分析 bert模型 TST算法 日志异常检测
在线阅读 下载PDF
融合提示工程与BERT-LSTM模型的“卡脖子”替代技术识别
3
作者 仵轩 李广建 +1 位作者 王楚涵 潘佳立 《图书馆论坛》 北大核心 2026年第3期135-147,共13页
针对“卡脖子”技术研究存在替代技术识别机制缺失与技术要素解析精度不足等局限,文章提出融合提示工程与BERT-LSTM模型的“卡脖子”替代技术识别方法。首先,基于商业管制清单(Commercial Control List,CCL)对ECCN物项进行解析,并开展... 针对“卡脖子”技术研究存在替代技术识别机制缺失与技术要素解析精度不足等局限,文章提出融合提示工程与BERT-LSTM模型的“卡脖子”替代技术识别方法。首先,基于商业管制清单(Commercial Control List,CCL)对ECCN物项进行解析,并开展专利检索工作,通过SPC算法提取技术主路径的关键核心专利;其次,运用大语言模型提示工程抽取“问题-方案对”,借此解析技术功效,并结合功能导向搜索(Function-Oriented Search,FOS)初步查找可能具备技术替代功效的相关专利;再次,采用BERT-LSTM模型对专利文本实施二元分类,精准识别出具备技术替代功效的专利样本;通过提示工程抽取“方案-类别对”,系统识别替代技术方案;最后,建立科学-产业双维度评估体系完成替代技术潜力分级。文章以光刻技术为例,阐述该识别方法的应用流程,系统识别出极紫外(Extreme Ultra-violet,EUV)光刻技术的五种替代技术及其替代潜力。 展开更多
关键词 “卡脖子”技术 替代技术 提示工程 bert-LSTM模型 光刻技术
在线阅读 下载PDF
KitWaSor:Pioneering pre-trained model for kitchen waste sorting with an innovative million-level benchmark dataset
4
作者 Leyuan Fang Shuaiyu Ding +3 位作者 Hao Feng Junwu Yu Lin Tang Pedram Ghamisi 《CAAI Transactions on Intelligence Technology》 2025年第1期94-114,共21页
Intelligent sorting is an important prerequisite for the full quantitative consumption and harmless disposal of kitchen waste.The existing object detection method based on an ImageNet pre-trained model is an effective... Intelligent sorting is an important prerequisite for the full quantitative consumption and harmless disposal of kitchen waste.The existing object detection method based on an ImageNet pre-trained model is an effective way of sorting.Owing to significant domain gaps between natural images and kitchen waste images,it is difficult to reflect the characteristics of diverse scales and dense distribution in kitchen waste based on an ImageNet pre-trained model,leading to poor generalisation.In this article,the authors propose the first pre-trained model for kitchen waste sorting called KitWaSor,which combines both contrastive learning(CL)and masked image modelling(MIM)through self-supervised learning(SSL).First,to address the issue of diverse scales,the authors propose a mixed masking strategy by introducing an incomplete masking branch based on the original random masking branch.It prevents the complete loss of small-scale objects while avoiding excessive leakage of large-scale object pixels.Second,to address the issue of dense distribution,the authors introduce semantic consistency constraints on the basis of the mixed masking strategy.That is,object semantic reasoning is performed through semantic consistency constraints to compensate for the lack of contextual information.To train KitWaSor,the authors construct the first million-level kitchen waste dataset across seasonal and regional distributions,named KWD-Million.Extensive experiments show that KitWaSor achieves state-of-the-art(SOTA)performance on the two most relevant downstream tasks for kitchen waste sorting(i.e.image classification and object detection),demonstrating the effectiveness of the proposed KitWaSor. 展开更多
关键词 contrastive learning kitchen waste masked image modeling pre-trained model self-supervised learning
在线阅读 下载PDF
基于Stroke BERT的中风知识图谱与智能诊疗研究
5
作者 于惠 祖红月 +1 位作者 王卫东 刘洪运 《计算机技术与发展》 2026年第2期126-131,共6页
为促进中医精准医疗发展,本研究聚焦中风病,构建了“领域自适应预训练—跨模态知识对齐—混合增强推理”框架。首先,基于中医文献与病历构建Stroke BERT预训练模型,采用动态全词掩码增强中医隐喻语义表征。实验表明,Stroke BERT在中医... 为促进中医精准医疗发展,本研究聚焦中风病,构建了“领域自适应预训练—跨模态知识对齐—混合增强推理”框架。首先,基于中医文献与病历构建Stroke BERT预训练模型,采用动态全词掩码增强中医隐喻语义表征。实验表明,Stroke BERT在中医特异性任务上显著优于通用BERT(F 1值达83.24%和98.15%),有效提升信息提取精度。其次,构建中西医融合中风知识图谱,通过BERT-INT模型实现中西医实体对齐,突破语义壁垒。在此基础上,提出KG-LLM协同推理架构,通过动态提示工程将知识图谱路径与GLM-4模型深度耦合。在Stroke QA和Stroke CQ数据集上,该架构的诊断符合率分别提升至59.05%和78.10%,较独立大语言模型显著提高(分别提升10.48和7.62百分点),实现了诊疗可解释性与准确性的双重提升。为中医知识图谱构建提供了领域适配方法,为多模态医学数据融合与智能诊疗提供了创新技术支撑。 展开更多
关键词 中风 Stroke bert 知识图谱 大语言模型 智能诊疗
在线阅读 下载PDF
DPCIPI: A pre-trained deep learning model for predicting cross-immunity between drifted strains of Influenza A/H3N2
6
作者 Yiming Du Zhuotian Li +8 位作者 Qian He Thomas Wetere Tulu Kei Hang Katie Chan Lin Wang Sen Pei Zhanwei Du Zhen Wang Xiao-Ke Xu Xiao Fan Liu 《Journal of Automation and Intelligence》 2025年第2期115-124,共10页
Predicting cross-immunity between viral strains is vital for public health surveillance and vaccine development.Traditional neural network methods,such as BiLSTM,could be ineffective due to the lack of lab data for mo... Predicting cross-immunity between viral strains is vital for public health surveillance and vaccine development.Traditional neural network methods,such as BiLSTM,could be ineffective due to the lack of lab data for model training and the overshadowing of crucial features within sequence concatenation.The current work proposes a less data-consuming model incorporating a pre-trained gene sequence model and a mutual information inference operator.Our methodology utilizes gene alignment and deduplication algorithms to preprocess gene sequences,enhancing the model’s capacity to discern and focus on distinctions among input gene pairs.The model,i.e.,DNA Pretrained Cross-Immunity Protection Inference model(DPCIPI),outperforms state-of-theart(SOTA)models in predicting hemagglutination inhibition titer from influenza viral gene sequences only.Improvement in binary cross-immunity prediction is 1.58%in F1,2.34%in precision,1.57%in recall,and 1.57%in Accuracy.For multilevel cross-immunity improvements,the improvement is 2.12%in F1,3.50%in precision,2.19%in recall,and 2.19%in Accuracy.Our study showcases the potential of pre-trained gene models to improve predictions of antigenic variation and cross-immunity.With expanding gene data and advancements in pre-trained models,this approach promises significant impacts on vaccine development and public health. 展开更多
关键词 Cross-immunity prediction pre-trained model Deep learning Influenza strains Hemagglutination inhibition
在线阅读 下载PDF
Multilingual Text Summarization in Healthcare Using Pre-Trained Transformer-Based Language Models
7
作者 Josua Käser Thomas Nagy +1 位作者 Patrick Stirnemann Thomas Hanne 《Computers, Materials & Continua》 2025年第4期201-217,共17页
We analyze the suitability of existing pre-trained transformer-based language models(PLMs)for abstractive text summarization on German technical healthcare texts.The study focuses on the multilingual capabilities of t... We analyze the suitability of existing pre-trained transformer-based language models(PLMs)for abstractive text summarization on German technical healthcare texts.The study focuses on the multilingual capabilities of these models and their ability to perform the task of abstractive text summarization in the healthcare field.The research hypothesis was that large language models could perform high-quality abstractive text summarization on German technical healthcare texts,even if the model is not specifically trained in that language.Through experiments,the research questions explore the performance of transformer language models in dealing with complex syntax constructs,the difference in performance between models trained in English and German,and the impact of translating the source text to English before conducting the summarization.We conducted an evaluation of four PLMs(GPT-3,a translation-based approach also utilizing GPT-3,a German language Model,and a domain-specific bio-medical model approach).The evaluation considered the informativeness using 3 types of metrics based on Recall-Oriented Understudy for Gisting Evaluation(ROUGE)and the quality of results which is manually evaluated considering 5 aspects.The results show that text summarization models could be used in the German healthcare domain and that domain-independent language models achieved the best results.The study proves that text summarization models can simplify the search for pre-existing German knowledge in various domains. 展开更多
关键词 Text summarization pre-trained transformer-based language models large language models technical healthcare texts natural language processing
在线阅读 下载PDF
Optimizing Airline Review Sentiment Analysis:A Comparative Analysis of LLaMA and BERT Models through Fine-Tuning and Few-Shot Learning
8
作者 Konstantinos I.Roumeliotis Nikolaos D.Tselikas Dimitrios K.Nasiopoulos 《Computers, Materials & Continua》 2025年第2期2769-2792,共24页
In the rapidly evolving landscape of natural language processing(NLP)and sentiment analysis,improving the accuracy and efficiency of sentiment classification models is crucial.This paper investigates the performance o... In the rapidly evolving landscape of natural language processing(NLP)and sentiment analysis,improving the accuracy and efficiency of sentiment classification models is crucial.This paper investigates the performance of two advanced models,the Large Language Model(LLM)LLaMA model and NLP BERT model,in the context of airline review sentiment analysis.Through fine-tuning,domain adaptation,and the application of few-shot learning,the study addresses the subtleties of sentiment expressions in airline-related text data.Employing predictive modeling and comparative analysis,the research evaluates the effectiveness of Large Language Model Meta AI(LLaMA)and Bidirectional Encoder Representations from Transformers(BERT)in capturing sentiment intricacies.Fine-tuning,including domain adaptation,enhances the models'performance in sentiment classification tasks.Additionally,the study explores the potential of few-shot learning to improve model generalization using minimal annotated data for targeted sentiment analysis.By conducting experiments on a diverse airline review dataset,the research quantifies the impact of fine-tuning,domain adaptation,and few-shot learning on model performance,providing valuable insights for industries aiming to predict recommendations and enhance customer satisfaction through a deeper understanding of sentiment in user-generated content(UGC).This research contributes to refining sentiment analysis models,ultimately fostering improved customer satisfaction in the airline industry. 展开更多
关键词 Sentiment classification review sentiment analysis user-generated content domain adaptation customer satisfaction LLaMA model bert model airline reviews LLM classification fine-tuning
在线阅读 下载PDF
基于BERT和GRU的Web攻击检测模型研究
9
作者 谭攀 王海珍 成雨玫 《齐齐哈尔大学学报(自然科学版)》 2026年第1期30-37,共8页
目前Web攻击日益频繁和复杂,现有入侵检测系统的检测准确率、实时性需要进一步改进。为此,提出一种基于BERT和GRU的Web攻击检测模型BERT_GRU。通过Web攻击检测脚本Scalp进行攻击类型标注,利用预训练的BERT模型对HTTP请求中的文本数据进... 目前Web攻击日益频繁和复杂,现有入侵检测系统的检测准确率、实时性需要进一步改进。为此,提出一种基于BERT和GRU的Web攻击检测模型BERT_GRU。通过Web攻击检测脚本Scalp进行攻击类型标注,利用预训练的BERT模型对HTTP请求中的文本数据进行上下文语义编码、全连接层处理,将获得的数据与原始文本数据拼接,输入GRU网络提取序列特征,输出通过Softmax计算,最终实现多分类。在CSIC 2010数据集和FWAF数据集上的实验结果表明,BERT_GRU模型在准确率、召回率、F1值等评价指标以及训练时间方面均优于对比模型。 展开更多
关键词 Web攻击检测 bert模型 GRU
在线阅读 下载PDF
GeoNER:Geological Named Entity Recognition with Enriched Domain Pre-Training Model and Adversarial Training
10
作者 MA Kai HU Xinxin +4 位作者 TIAN Miao TAN Yongjian ZHENG Shuai TAO Liufeng QIU Qinjun 《Acta Geologica Sinica(English Edition)》 SCIE CAS CSCD 2024年第5期1404-1417,共14页
As important geological data,a geological report contains rich expert and geological knowledge,but the challenge facing current research into geological knowledge extraction and mining is how to render accurate unders... As important geological data,a geological report contains rich expert and geological knowledge,but the challenge facing current research into geological knowledge extraction and mining is how to render accurate understanding of geological reports guided by domain knowledge.While generic named entity recognition models/tools can be utilized for the processing of geoscience reports/documents,their effectiveness is hampered by a dearth of domain-specific knowledge,which in turn leads to a pronounced decline in recognition accuracy.This study summarizes six types of typical geological entities,with reference to the ontological system of geological domains and builds a high quality corpus for the task of geological named entity recognition(GNER).In addition,Geo Wo BERT-adv BGP(Geological Word-base BERTadversarial training Bi-directional Long Short-Term Memory Global Pointer)is proposed to address the issues of ambiguity,diversity and nested entities for the geological entities.The model first uses the fine-tuned word granularitybased pre-training model Geo Wo BERT(Geological Word-base BERT)and combines the text features that are extracted using the Bi LSTM(Bi-directional Long Short-Term Memory),followed by an adversarial training algorithm to improve the robustness of the model and enhance its resistance to interference,the decoding finally being performed using a global association pointer algorithm.The experimental results show that the proposed model for the constructed dataset achieves high performance and is capable of mining the rich geological information. 展开更多
关键词 geological named entity recognition geological report adversarial training confrontation training global pointer pre-training model
在线阅读 下载PDF
基于改进BERT模型的城市轨道交通应急社交媒体文本分类
11
作者 张诺 史世盛 +1 位作者 吴波 朱广宇 《综合运输》 2026年第2期114-119,共6页
针对城市轨道交通应急社交媒体文本分类任务中存在的长序列处理效率低、噪声干扰强、局部特征捕捉不足等问题,本文提出一种基于Mamba-Transformer混合架构的多尺度门控BERT模型(MTH-BERT)。该模型通过交替堆叠选择性状态空间模型与多头... 针对城市轨道交通应急社交媒体文本分类任务中存在的长序列处理效率低、噪声干扰强、局部特征捕捉不足等问题,本文提出一种基于Mamba-Transformer混合架构的多尺度门控BERT模型(MTH-BERT)。该模型通过交替堆叠选择性状态空间模型与多头自注意力机制,构建层次化特征提取结构,在保持全局语义建模能力的同时显著提升长文本处理效率;引入多尺度空洞卷积组增强局部关键特征与短语级语义的捕获能力;设计门控多尺度交叉注意力模块,实现全局上下文与局部特征的动态融合与噪声抑制。在真实应急社交媒体文本数据集上的实验表明,MTH-BERT在分类准确率、推理时延及抗噪性(F1值)等指标上均优于BERT-base、RoBERTa-large、DistilBERT等基线模型,验证了该模型对大规模、高噪声、实时性应急文本分类任务的有效性与鲁棒性。 展开更多
关键词 城市轨道交通应急 社交媒体 文本分类 多尺度门控模型(MTH-bert)
原文传递
人工智能技术创新如何赋能企业成长——来自BERT模型的证据
12
作者 何小钢 毛莘娅 《财经问题研究》 北大核心 2026年第1期44-58,共15页
人工智能技术创新是促进新质生产力发展的关键,对于推动高质量发展具有重要意义。本文利用双向编码器表示模型识别人工智能专利,以刻画企业人工智能技术创新水平,并基于2012-2022年中国A股上市公司数据,利用双向固定效应模型实证检验了... 人工智能技术创新是促进新质生产力发展的关键,对于推动高质量发展具有重要意义。本文利用双向编码器表示模型识别人工智能专利,以刻画企业人工智能技术创新水平,并基于2012-2022年中国A股上市公司数据,利用双向固定效应模型实证检验了人工智能技术创新对企业成长的影响及作用机制。研究结果显示:人工智能技术创新能够促进企业成长,且该结论在经过一系列内生性检验和稳健性检验后仍成立;人工智能技术创新通过创新溢出效应、降本增效效应、产品竞争效应和人力资本质量效应促进企业成长;人工智能技术创新对企业成长的促进作用发生在高管受教育程度较高企业、较年轻的高管团队企业、吸收能力较强的企业、劳动密集型行业企业、技术密集型行业企业中。本文研究拓宽了人工智能技术创新对微观企业经济效应的研究范围,为中国人工智能技术创新的政策制定和企业数智化战略决策提供了启示。 展开更多
关键词 人工智能技术创新 企业成长 bert模型
在线阅读 下载PDF
基于改进BERT的多头自注意力非侵入式负荷分解方法
13
作者 孙晓晴 李元诚 王庆乐 《电力信息与通信技术》 2026年第1期45-54,共10页
针对非侵入式负荷分解方法负荷特征捕捉不足、负荷分解精度不够等问题,文章提出一种基于改进BERT(bidirectional encoder representations from transformers)模型的多头自注意力非侵入式负荷分解方法(frequency and temporal attention... 针对非侵入式负荷分解方法负荷特征捕捉不足、负荷分解精度不够等问题,文章提出一种基于改进BERT(bidirectional encoder representations from transformers)模型的多头自注意力非侵入式负荷分解方法(frequency and temporal attention-BERT, FAT-BERT)。首先通过傅里叶变换将时域数据转换为频域数据,采用多尺度卷积全面捕捉负荷信号的时域和频域特征,从而增强模型对多样化负荷信号的表达能力;其次,在多头自注意力机制中引入频率注意力机制,从而增强模型对时序数据中频率成分的感知能力,进一步改善复杂负荷模式的表示,改进BERT模型中增加局部自注意力从而减少不必要的全局计算,提升模型的运行速度;接着将残差连接和正则化技术结合使模型在训练过程中更加稳定,并且能够更好地避免过拟合,最后在REDD和UK-DALE数据集上对提出的方法进行实验,实验结果验证了所提方法的有效性。 展开更多
关键词 非侵入式负荷监测 负荷分解 改进bert模型 多头自注意力机制 频率注意力
在线阅读 下载PDF
基于BERT大语言模型的广东省人工智能发明专利测度与时空演化特征
14
作者 陈奕嘉 谭俊涛 杨瑞霖 《热带地理》 北大核心 2026年第1期154-166,共13页
在人工智能(AI)日益成为推动区域高质量发展的关键技术背景下,如何科学测度AI技术活动的空间分布和演化趋势,成为经济地理学研究的重要议题。然而,现有的识别方法在语义有效性、识别精度以及覆盖广度等方面仍存在不足,有待进一步提升与... 在人工智能(AI)日益成为推动区域高质量发展的关键技术背景下,如何科学测度AI技术活动的空间分布和演化趋势,成为经济地理学研究的重要议题。然而,现有的识别方法在语义有效性、识别精度以及覆盖广度等方面仍存在不足,有待进一步提升与完善。为此,文章以广东省2001-2021年约120万条授权发明专利摘要为基础,引入BERT大语言模型,构建基于语义理解的AI授权发明专利识别方法,识别出约20万条AI发明专利文本。在此基础上,结合BERTopic主题建模方法,进一步揭示AI发明专利在广东省的关注领域与动态演化特征。研究发现:1)2001-2021年广东省AI发明专利数量整体经历了缓慢增长和快速扩张2个阶段,AI发明专利占比则呈持续上升趋势,反映AI技术在区域创新体系中的重要性不断提升。2)空间分布方面,AI发明专利在大湾区高度集聚,深圳和广州2市占全省AI发明专利的75.1%,珠海和东莞两地在AI发明专利占比上亦显示出较强活力。在合作网络方面,广东省AI发明专利合作网络呈现显著的“核心―边缘”结构特征,城市间在合作网络中的地位差异明显。3)在技术内容上,广东省AI发明专利主要聚焦于数据与图像处理、机器人与自动化装置、智能交通与故障检测、智能家居与环境控制和生物模拟与图像分析五大主题,不同主题的演化节奏存在差异,其中图像处理技术最为活跃,机器人和智能交通等领域近年来快速增长。此外,广东省内不同城市在AI发明专利布局上也各具特色,形成了差异化的技术优势。 展开更多
关键词 人工智能 AI发明专利 bert大语言模型 主题建模 技术演化 合作网络 广东省
在线阅读 下载PDF
基于BERT模型的微博文本细粒度情感分析
15
作者 张逸民 李野 《现代信息科技》 2026年第4期112-115,121,共5页
随着社交媒体的快速发展,微博作为用户信息交流与情感表达的重要平台,积累了海量丰富的文本数据。文章在调研微博情感分析相关技术的基础上,提出一种基于BERT模型的细粒度情感分析方法,结合微博数据抓取与预处理技术,构建了高效的分析... 随着社交媒体的快速发展,微博作为用户信息交流与情感表达的重要平台,积累了海量丰富的文本数据。文章在调研微博情感分析相关技术的基础上,提出一种基于BERT模型的细粒度情感分析方法,结合微博数据抓取与预处理技术,构建了高效的分析框架。该方法通过微博开放API完成数据采集,利用BERT预训练模型实现文本向量化,并依托Transformer架构完成愤怒、高兴、中性、惊讶、悲伤、恐惧六类情绪的精准分类;同时引入情感词典与数据增强技术提升模型性能,借助可视化工具展示分析结果。研究表明,该方法在SMP2020数据集上取得了较高的分类精度,为微博文本细粒度情感分析提供了新思路。 展开更多
关键词 微博情感分析 细粒度情感 bert模型 社交媒体
在线阅读 下载PDF
基金经理语调、基金收益与投资者行为——基于BERT人工智能模型的检验 被引量:2
16
作者 沈红波 洪康隆 王锴 《现代金融研究》 北大核心 2025年第4期20-38,共19页
本文以2014-2023年我国基金年度报告“展望”章节文本为样本,运用BERT人工智能模型和传统的词袋法,分别构建“词藻堆砌”与“真情实感”变量,对比基金经理语调对基金未来收益的预测效果及其对个人投资者行为的影响。研究发现:(1)相比传... 本文以2014-2023年我国基金年度报告“展望”章节文本为样本,运用BERT人工智能模型和传统的词袋法,分别构建“词藻堆砌”与“真情实感”变量,对比基金经理语调对基金未来收益的预测效果及其对个人投资者行为的影响。研究发现:(1)相比传统的词袋法,BERT人工智能模型度量的基金经理语调更能识别基金经理的真情实感,对基金未来收益、顺境中的基金业绩持续性及逆境中的基金崩盘风险的预测效果更好,且上述识别优势在复杂度较高、可读性较差的“展望”文本中更显著;(2)个人投资者的“有限理性”特征和公募基金存在的委托代理问题导致个人投资者更容易被基金经理“词藻堆砌”的文本所吸引,进而导致投资收益率降低。本文将BERT人工智能模型引入财经文本情感分析,为文本语调的度量方法提供了新思路。 展开更多
关键词 基金经理语调 基金收益 投资者行为 bert模型
在线阅读 下载PDF
Masked Sentence Model Based on BERT for Move Recognition in Medical Scientific Abstracts 被引量:24
17
作者 Gaihong Yu Zhixiong Zhang +1 位作者 Huan Liu Liangping Ding 《Journal of Data and Information Science》 CSCD 2019年第4期42-55,共14页
Purpose:Mo ve recognition in scientific abstracts is an NLP task of classifying sentences of the abstracts into different types of language units.To improve the performance of move recognition in scientific abstracts,... Purpose:Mo ve recognition in scientific abstracts is an NLP task of classifying sentences of the abstracts into different types of language units.To improve the performance of move recognition in scientific abstracts,a novel model of move recognition is proposed that outperforms the BERT-based method.Design/methodology/approach:Prevalent models based on BERT for sentence classification often classify sentences without considering the context of the sentences.In this paper,inspired by the BERT masked language model(MLM),we propose a novel model called the masked sentence model that integrates the content and contextual information of the sentences in move recognition.Experiments are conducted on the benchmark dataset PubMed 20K RCT in three steps.Then,we compare our model with HSLN-RNN,BERT-based and SciBERT using the same dataset.Findings:Compared with the BERT-based and SciBERT models,the F1 score of our model outperforms them by 4.96%and 4.34%,respectively,which shows the feasibility and effectiveness of the novel model and the result of our model comes closest to the state-of-theart results of HSLN-RNN at present.Research limitations:The sequential features of move labels are not considered,which might be one of the reasons why HSLN-RNN has better performance.Our model is restricted to dealing with biomedical English literature because we use a dataset from PubMed,which is a typical biomedical database,to fine-tune our model.Practical implications:The proposed model is better and simpler in identifying move structures in scientific abstracts and is worthy of text classification experiments for capturing contextual features of sentences.Originality/value:T he study proposes a masked sentence model based on BERT that considers the contextual features of the sentences in abstracts in a new way.The performance of this classification model is significantly improved by rebuilding the input layer without changing the structure of neural networks. 展开更多
关键词 Move recognition bert Masked sentence model Scientific abstracts
在线阅读 下载PDF
基于BERT和图注意力网络的医疗文本因果关系抽取算法
18
作者 刘位龙 王玎 +6 位作者 赵超 王宁 张旭 苏萍 宋书典 张娜 迟蔚蔚 《山东大学学报(医学版)》 北大核心 2025年第8期61-68,共8页
目的提出一种能够有效抽取因果关系的算法,以提高医疗领域文本处理的准确性。方法提出基于Transformer的双向编码器(bidirectional encoder representations from Transformers,BERT)和因果图注意力网络(causal graph attention network... 目的提出一种能够有效抽取因果关系的算法,以提高医疗领域文本处理的准确性。方法提出基于Transformer的双向编码器(bidirectional encoder representations from Transformers,BERT)和因果图注意力网络(causal graph attention networks,CGAT)的BERT-CGAT算法。首先构建因果关系图,利用医疗文本对BERT模型进行微调,以获得优化的实体嵌入表示;随后通过知识融合通道整合文本编码信息与因果结构,输入至图注意力网络;采用多头注意力机制并行处理不同子空间信息,增强复杂语义关系捕捉能力;最后通过双通道解码层实现实体及因果关系的同步抽取。结果在自建的糖尿病因果实体数据集上的实验表明,模型在准确率(99.74%)与召回率(81.04%)上较传统BiLSTM-CRF基线提升0.65%和16.73%,F1分数达80.83%。结论BERT-CGAT算法通过结合BERT的语义特征提取能力和图神经网络的关系建模优势,有效提升了医疗文本因果关系抽取的准确性,验证了该方法的有效性。 展开更多
关键词 医疗文本 bert模型 图注意力网络 因果关系抽取
原文传递
基于BERT并融合法律事件信息的罪名预测方法
19
作者 邱一卉 喻瑶瑶 《厦门大学学报(自然科学版)》 北大核心 2025年第4期642-652,共11页
[目的]罪名预测是AI&Law领域的一个关键研究内容,对于提升司法领域的判决效率具有重要意义.由于法律文本的专业性和复杂性,传统罪名预测模型在提取法律文本特征时面临挑战,因此本文提出了一个基于预训练语言模型(BERT)并融合法律事... [目的]罪名预测是AI&Law领域的一个关键研究内容,对于提升司法领域的判决效率具有重要意义.由于法律文本的专业性和复杂性,传统罪名预测模型在提取法律文本特征时面临挑战,因此本文提出了一个基于预训练语言模型(BERT)并融合法律事件信息的罪名预测模型,通过利用法律事件信息为模型提供更多的法律案件特征,提升模型对案件的理解,从而提升罪名预测的表现.[方法]首先训练了一个全局上层事件类型信息增强的法律事件检测模型,利用其对案情描述中的法律事件类型进行检测,并在此基础上构建法律事件类型序列.其次,利用双向长短期记忆模型(BiLSTM)对法律事件类型序列进行编码获取法律事件信息,并将法律事件信息与经过BERT编码后的案情描述的语义表示拼接融合,最后利用一层全连接层对罪名进行预测.[结果]在公开的刑事案件数据集CAIL2018-small上的实验结果表明,相比于领域内的其他基线模型,本文提出的模型在各个关键指标上具备更好的性能,即在Mac.F_(1)上平均提升3.12个百分点,在Mac.P上平均提升1.94个百分点,在Mac.R上平均提升3.53个百分点.[结论]验证了法律事件信息在增强模型对案件理解方面的有效性,从而提高罪名预测的准确性. 展开更多
关键词 AI&Law bert模型 罪名预测 法律事件信息
在线阅读 下载PDF
A Classification–Detection Approach of COVID-19 Based on Chest X-ray and CT by Using Keras Pre-Trained Deep Learning Models 被引量:10
20
作者 Xing Deng Haijian Shao +2 位作者 Liang Shi Xia Wang Tongling Xie 《Computer Modeling in Engineering & Sciences》 SCIE EI 2020年第11期579-596,共18页
The Coronavirus Disease 2019(COVID-19)is wreaking havoc around the world,bring out that the enormous pressure on national health and medical staff systems.One of the most effective and critical steps in the fight agai... The Coronavirus Disease 2019(COVID-19)is wreaking havoc around the world,bring out that the enormous pressure on national health and medical staff systems.One of the most effective and critical steps in the fight against COVID-19,is to examine the patient’s lungs based on the Chest X-ray and CT generated by radiation imaging.In this paper,five keras-related deep learning models:ResNet50,InceptionResNetV2,Xception,transfer learning and pre-trained VGGNet16 is applied to formulate an classification-detection approaches of COVID-19.Two benchmark methods SVM(Support Vector Machine),CNN(Conventional Neural Networks)are provided to compare with the classification-detection approaches based on the performance indicators,i.e.,precision,recall,F1 scores,confusion matrix,classification accuracy and three types of AUC(Area Under Curve).The highest classification accuracy derived by classification-detection based on 5857 Chest X-rays and 767 Chest CTs are respectively 84%and 75%,which shows that the keras-related deep learning approaches facilitate accurate and effective COVID-19-assisted detection. 展开更多
关键词 COVID-19 detection deep learning transfer learning pre-trained models
在线阅读 下载PDF
上一页 1 2 31 下一页 到第
使用帮助 返回顶部