期刊文献+
共找到373篇文章
< 1 2 19 >
每页显示 20 50 100
PAL-BERT:An Improved Question Answering Model
1
作者 Wenfeng Zheng Siyu Lu +3 位作者 Zhuohang Cai Ruiyang Wang Lei Wang Lirong Yin 《Computer Modeling in Engineering & Sciences》 SCIE EI 2024年第6期2729-2745,共17页
In the field of natural language processing(NLP),there have been various pre-training language models in recent years,with question answering systems gaining significant attention.However,as algorithms,data,and comput... In the field of natural language processing(NLP),there have been various pre-training language models in recent years,with question answering systems gaining significant attention.However,as algorithms,data,and computing power advance,the issue of increasingly larger models and a growing number of parameters has surfaced.Consequently,model training has become more costly and less efficient.To enhance the efficiency and accuracy of the training process while reducing themodel volume,this paper proposes a first-order pruningmodel PAL-BERT based on the ALBERT model according to the characteristics of question-answering(QA)system and language model.Firstly,a first-order network pruning method based on the ALBERT model is designed,and the PAL-BERT model is formed.Then,the parameter optimization strategy of the PAL-BERT model is formulated,and the Mish function was used as an activation function instead of ReLU to improve the performance.Finally,after comparison experiments with traditional deep learning models TextCNN and BiLSTM,it is confirmed that PALBERT is a pruning model compression method that can significantly reduce training time and optimize training efficiency.Compared with traditional models,PAL-BERT significantly improves the NLP task’s performance. 展开更多
关键词 PAL-BERT question answering model pretraining language models ALBERT pruning model network pruning TextCNN BiLSTM
在线阅读 下载PDF
DPAL-BERT:A Faster and Lighter Question Answering Model
2
作者 Lirong Yin Lei Wang +8 位作者 Zhuohang Cai Siyu Lu Ruiyang Wang Ahmed AlSanad Salman A.AlQahtani Xiaobing Chen Zhengtong Yin Xiaolu Li Wenfeng Zheng 《Computer Modeling in Engineering & Sciences》 SCIE EI 2024年第10期771-786,共16页
Recent advancements in natural language processing have given rise to numerous pre-training language models in question-answering systems.However,with the constant evolution of algorithms,data,and computing power,the ... Recent advancements in natural language processing have given rise to numerous pre-training language models in question-answering systems.However,with the constant evolution of algorithms,data,and computing power,the increasing size and complexity of these models have led to increased training costs and reduced efficiency.This study aims to minimize the inference time of such models while maintaining computational performance.It also proposes a novel Distillation model for PAL-BERT(DPAL-BERT),specifically,employs knowledge distillation,using the PAL-BERT model as the teacher model to train two student models:DPAL-BERT-Bi and DPAL-BERTC.This research enhances the dataset through techniques such as masking,replacement,and n-gram sampling to optimize knowledge transfer.The experimental results showed that the distilled models greatly outperform models trained from scratch.In addition,although the distilled models exhibit a slight decrease in performance compared to PAL-BERT,they significantly reduce inference time to just 0.25%of the original.This demonstrates the effectiveness of the proposed approach in balancing model performance and efficiency. 展开更多
关键词 DPAL-BERT question answering systems knowledge distillation model compression BERT Bi-directional long short-term memory(BiLSTM) knowledge information transfer PAL-BERT training efficiency natural language processing
在线阅读 下载PDF
Question classification in question answering based on real-world web data sets 被引量:1
3
作者 袁晓洁 于士涛 +1 位作者 师建兴 陈秋双 《Journal of Southeast University(English Edition)》 EI CAS 2008年第3期272-275,共4页
To improve question answering (QA) performance based on real-world web data sets,a new set of question classes and a general answer re-ranking model are defined.With pre-defined dictionary and grammatical analysis,t... To improve question answering (QA) performance based on real-world web data sets,a new set of question classes and a general answer re-ranking model are defined.With pre-defined dictionary and grammatical analysis,the question classifier draws both semantic and grammatical information into information retrieval and machine learning methods in the form of various training features,including the question word,the main verb of the question,the dependency structure,the position of the main auxiliary verb,the main noun of the question,the top hypernym of the main noun,etc.Then the QA query results are re-ranked by question class information.Experiments show that the questions in real-world web data sets can be accurately classified by the classifier,and the QA results after re-ranking can be obviously improved.It is proved that with both semantic and grammatical information,applications such as QA, built upon real-world web data sets, can be improved,thus showing better performance. 展开更多
关键词 question classification question answering real-world web data sets question and answer web forums re-ranking model
在线阅读 下载PDF
ALBERT with Knowledge Graph Encoder Utilizing Semantic Similarity for Commonsense Question Answering 被引量:1
4
作者 Byeongmin Choi YongHyun Lee +1 位作者 Yeunwoong Kyung Eunchan Kim 《Intelligent Automation & Soft Computing》 SCIE 2023年第4期71-82,共12页
Recently,pre-trained language representation models such as bidirec-tional encoder representations from transformers(BERT)have been performing well in commonsense question answering(CSQA).However,there is a problem th... Recently,pre-trained language representation models such as bidirec-tional encoder representations from transformers(BERT)have been performing well in commonsense question answering(CSQA).However,there is a problem that the models do not directly use explicit information of knowledge sources existing outside.To augment this,additional methods such as knowledge-aware graph network(KagNet)and multi-hop graph relation network(MHGRN)have been proposed.In this study,we propose to use the latest pre-trained language model a lite bidirectional encoder representations from transformers(ALBERT)with knowledge graph information extraction technique.We also propose to applying the novel method,schema graph expansion to recent language models.Then,we analyze the effect of applying knowledge graph-based knowledge extraction techniques to recent pre-trained language models and confirm that schema graph expansion is effective in some extent.Furthermore,we show that our proposed model can achieve better performance than existing KagNet and MHGRN models in CommonsenseQA dataset. 展开更多
关键词 Commonsense reasoning question answering knowledge graph language representation model
在线阅读 下载PDF
A Dynamic Knowledge Base Updating Mechanism-Based Retrieval-Augmented Generation Framework for Intelligent Question-and-Answer Systems 被引量:1
5
作者 Yu Li 《Journal of Computer and Communications》 2025年第1期41-58,共18页
In the context of power generation companies, vast amounts of specialized data and expert knowledge have been accumulated. However, challenges such as data silos and fragmented knowledge hinder the effective utilizati... In the context of power generation companies, vast amounts of specialized data and expert knowledge have been accumulated. However, challenges such as data silos and fragmented knowledge hinder the effective utilization of this information. This study proposes a novel framework for intelligent Question-and-Answer (Q&A) systems based on Retrieval-Augmented Generation (RAG) to address these issues. The system efficiently acquires domain-specific knowledge by leveraging external databases, including Relational Databases (RDBs) and graph databases, without additional fine-tuning for Large Language Models (LLMs). Crucially, the framework integrates a Dynamic Knowledge Base Updating Mechanism (DKBUM) and a Weighted Context-Aware Similarity (WCAS) method to enhance retrieval accuracy and mitigate inherent limitations of LLMs, such as hallucinations and lack of specialization. Additionally, the proposed DKBUM dynamically adjusts knowledge weights within the database, ensuring that the most recent and relevant information is utilized, while WCAS refines the alignment between queries and knowledge items by enhanced context understanding. Experimental validation demonstrates that the system can generate timely, accurate, and context-sensitive responses, making it a robust solution for managing complex business logic in specialized industries. 展开更多
关键词 Retrieval-Augmented Generation question-and-answer Large Language models Dynamic Knowledge Base Updating Mechanism Weighted Context-Aware Similarity
在线阅读 下载PDF
Bootstrapping Large Language Models with Outsideknowledge for Knowledge-based Visual Question Answering
6
作者 Yanze Min Yawei Sun +2 位作者 Yin Zhu Jun Zhu Bo Zhang 《Machine Intelligence Research》 2026年第1期115-132,共18页
Knowledge-based visual question answering(KB-VQA),requiring external world knowledge beyond the image for reasoning,is more challenging than traditional visual question answering.Recent works have demonstrated the eff... Knowledge-based visual question answering(KB-VQA),requiring external world knowledge beyond the image for reasoning,is more challenging than traditional visual question answering.Recent works have demonstrated the effectiveness of using a large(vision)language model as an implicit knowledge source to acquire the necessary information.However,the knowledge stored in large models(LMs)is often coarse-grained and inaccurate,causing questions requiring finer-grained information to be answered incorrectly.In this work,we propose a variational expectation-maximization(EM)framework that bootstraps the VQA performance of LMs with its own answer.In contrast to former VQA pipelines,we treat the outside knowledge as a latent variable.In the E-step,we approximate the posterior with two components:First,a rough answer,e.g.,a general description of the image,which is usually the strength of LMs,and second,a multi-modal neural retriever to retrieve question-specific knowledge from an external knowledge base.In the M-step,the training objective optimizes the ability of the original LMs to generate rough answers as well as refined answers based on the retrieved information.Extensive experiments show that our proposed framework,BootLM,has a strong retrieval ability and achieves state-of-the-art performance on knowledge-based VQA tasks. 展开更多
关键词 Multi-modal large language models visual question answering(VQA) knowledge retrieval graphical models machine learning
原文传递
Benchmarking large multimodal models for ophthalmic visual question answering with OphthalWeChat
7
作者 Pusheng Xu Xia Gong +7 位作者 Xiaolan Chen Weiyi Zhang Jiancheng Yang Bingjie Yan Meng Yuan Yalin Zheng Mingguang He Danli Shi 《Advances in Ophthalmology Practice and Research》 2026年第1期33-41,共9页
Purposes:To develop a bilingual multimodal visual question answering(VQA)benchmark for evaluating Vision-language models(VLMs)in ophthalmology.Methods:In this cross-sectional study,ophthalmic image posts and associate... Purposes:To develop a bilingual multimodal visual question answering(VQA)benchmark for evaluating Vision-language models(VLMs)in ophthalmology.Methods:In this cross-sectional study,ophthalmic image posts and associated captions published between Jan 1,2016,and Dec 31,2024,were collected from WeChat Official Accounts.Based on these captions,bilingual question-answer(QA)pairs in Chinese and English were generated using GPT-4o-mini.QA pairs were categorized into six subsets by question type and language:binary(Binary_CN,Binary_EN),single-choice(Singlechoice_CN,Single-choice_EN),and open-ended(Open-ended_CN,Open-ended_EN).The benchmark was used to evaluate six VLMs:GPT-4o,Gemini 2.0 Flash,Qwen2.5-VL-72B-Instruct,Janus-Pro-7B,InternVL3-8B,and HealthGPT-L14.Primary outcome was overall accuracy;secondary outcomes included subset-,subspeciality-,and modality-specific accuracy.Performance on open-ended questions were also quantified using languagebased metrics,including AlignScore,BARTScore,BERTScore,BLEU,CIDEr,METEOR,and ROUGE_L.Error types in open-ended responses were manually analyzed through stratified sampling.Results:OphthalWeChat included 3469 images and 30120 QA pairs cover 9 ophthalmic subspecialties,548 conditions,29 imaging modalities,and 68 modality combinations.Gemini 2.0 Flash achieved the highest overall accuracy(0.555),significantly outperforming GPT-4o(0.527),Qwen2.5-VL-72B-Instruct(0.520),HealthGPTL14(0.502),InternVL3-L14(0.453),and Janus-Pro-7B(0.333)(all P<0.001).It also led in both Chinese(0.551)and English subsets(0.559).By subset,Gemini 2.0 Flash excelled in Binary_CN(0.687)and Singlechoice_CN(0.666);HealthGPT-L14 performed best in Single-choice_EN(0.739);while GPT-4o ranked highest in Binary_EN(0.717),Open-ended_CN(0.254),and Open-ended_EN(0.271).Language-based metrics showed inconsistent rankings relative to accuracy in open-ended subsets.Performance varied across subspecialties and modalities,with Gemini 2.0 Flash leading in 6 of 9 subspecialties and 11 of top-15 imaging modalities.Error types analysis revealed lesion/diagnosis errors as the most frequent(35.6%-50.6%),followed by anatomical location errors(28.3%-37.5%).Conclusions:This study presents the first bilingual VQA benchmark for ophthalmology,distinguished by its realworld context and inclusion of multiple examinations per patient.The dataset enables quantitative evaluation of VLMs,supporting the development of accurate and specialized AI systems for eye care. 展开更多
关键词 Visual question answering Ophthalmology Multimodal benchmark Multimodal dataset Vision-language models
原文传递
Zero-Shot Knowledge-Based Visual Question Answering with Frozen Language Models
8
作者 Jing Liu Lizong Zhang +3 位作者 Chenpeng Cao Yinong Shi Chong Mu Jiaxin Li 《Big Data Mining and Analytics》 2025年第6期1418-1431,共14页
Knowledge-based Visual Question Answering(VQA)is a challenging task that requires models to access external knowledge for reasoning.Large Language Models(LLMs)have recently been employed for zero-shot knowledge-based ... Knowledge-based Visual Question Answering(VQA)is a challenging task that requires models to access external knowledge for reasoning.Large Language Models(LLMs)have recently been employed for zero-shot knowledge-based VQA due to their inherent knowledge storage and in-context learning capabilities.However,LLMs are commonly perceived as implicit knowledge bases,and their generative and in-context learning potential remains underutilized.Existing works demonstrate that the performance of in-context learning strongly depends on the quality and order of demonstrations in prompts.In light of this,we propose Knowledge Generation with Frozen Language Models(KGFLM),a novel method for generating explicit knowledge statements to improve zero-shot knowledge-based VQA.Our knowledge generation strategy aims to identify effective demonstrations and determine their optimal order,thereby activating the frozen LLM to produce more useful knowledge statements for better predictions.The generated knowledge statements can also serve as interpretable rationales.In our method,the selection and arrangement of demonstrations are based on semantic similarity and quality of demonstrations for each question,without requiring additional annotations.Furthermore,a series of experiments are conducted on A-OKVQA and OKVQA datasets.The results show that our method outperforms some superior zero-shot knowledge-based VQA methods. 展开更多
关键词 knowledge-based Visual question answering(VQA) zero-shot learning Large Language models(LLMs)
原文传递
The question answer system based on natural language understanding
9
作者 郭庆琳 樊孝忠 《Journal of Harbin Institute of Technology(New Series)》 EI CAS 2007年第3期419-422,共4页
Automatic Question Answer System(QAS)is a kind of high-powered software system based on Internet.Its key technology is the interrelated technology based on natural language understanding,including the construction of ... Automatic Question Answer System(QAS)is a kind of high-powered software system based on Internet.Its key technology is the interrelated technology based on natural language understanding,including the construction of knowledge base and corpus,the Word Segmentation and POS Tagging of text,the Grammatical Analysis and Semantic Analysis of sentences etc.This thesis dissertated mainly the denotation of knowledge-information based on semantic network in QAS,the stochastic syntax-parse model named LSF of knowledge-information in QAS,the structure and constitution of QAS.And the LSF model's parameters were exercised,which proved that they were feasible.At the same time,through "the limited-domain QAS" which was exploited for banks by us,these technologies were proved effective and propagable. 展开更多
关键词 question answer system semantic network LSF model predicate logic
在线阅读 下载PDF
融合DeepSeek-R1和RAG技术的先秦文化元典智能问答研究 被引量:1
10
作者 张强 高颖 +2 位作者 任豆豆 韩牧哲 包平 《现代情报》 北大核心 2026年第1期173-186,共14页
[目的/意义]先秦文化元典是中华文明的源头文献,对其进行知识组织与智能应用,可以为建设中华民族现代文明提供历史依据和价值判断,增强国家文化软实力。本研究旨在基于检索增强生成(RAG)技术的先秦文化元典智能问答系统,推动相关知识的... [目的/意义]先秦文化元典是中华文明的源头文献,对其进行知识组织与智能应用,可以为建设中华民族现代文明提供历史依据和价值判断,增强国家文化软实力。本研究旨在基于检索增强生成(RAG)技术的先秦文化元典智能问答系统,推动相关知识的智能化应用与传承。[方法/过程]以中华书局出版的《春秋》三传为研究对象,构建先秦文化元典本体模型,采用DeepSeek-R1进行知识抽取并构建知识图谱。基于LangChain框架,运用GraphRAG、NaiveRAG、LightRAG、HybridRAG这4种RAG方法对大语言模型进行检索增强,并从定量和混合两方面评估问答能力。[结果/结论]研究结果显示,DeepSeek-R1抽取效果良好,生成的三元组能有效覆盖关键知识且质量较高。在智能问答评估中,不同RAG方法各有优劣。GraphRAG在各类问题和评估维度上表现较佳,尤其在考证溯源型、应用实践型等问题上表现突出;NaiveRAG在事实知识型问题上表现较好。综合定量与混合评估来看,根据实际应用场景选择合适的RAG技术至关重要。 展开更多
关键词 先秦文化元典 大语言模型 DeepSeek 检索增强生成 智能问答
在线阅读 下载PDF
基于协同专家系统的建筑施工大语言模型问答系统
11
作者 杨彬 肖鸿儒 +4 位作者 高尚 雷克 陈文硕 张其林 汪丛军 《同济大学学报(自然科学版)》 北大核心 2026年第1期13-21,30,共10页
为解决大型语言模型问答系统在建筑施工场景中存在的生成幻觉与部署成本高的问题,提出了一种基于协同专家机制的大型语言模型施工问答系统。该系统通过共享专家与路由专家的协同工作方式,在保证模型表达能力的同时,显著提升了问答生成... 为解决大型语言模型问答系统在建筑施工场景中存在的生成幻觉与部署成本高的问题,提出了一种基于协同专家机制的大型语言模型施工问答系统。该系统通过共享专家与路由专家的协同工作方式,在保证模型表达能力的同时,显著提升了问答生成的准确性与推理效率,并有效降低了计算开销。此外,设计了一种领域知识库注入的微调策略,在训练阶段引导模型深度学习施工领域专业语义,从而增强其对工程文本的理解能力,确保生成结果更加符合实际工程需求。实验结果表明,在仅激活约1/3模型参数的情况下,所提出系统仍可达到81.1%的生成语义相似度,兼顾了效率与性能,为建筑施工管理提供了一种高效、可靠且具备工程针对性的智能决策支持工具。 展开更多
关键词 建筑施工 智能建造 问答系统 大语言模型 本地知识库
在线阅读 下载PDF
面向医疗问答的KG与LLMs协同推理机制
12
作者 袁嵩 程芬 顾进广 《计算机工程与设计》 北大核心 2026年第1期252-259,共8页
针对现有大型语言模型(LLMs)在医学推理任务中存在的隐式知识利用不足、推理路径冗余及透明度缺失等问题,提出一种基于协同推理的医学问答方法。构建推理子图学习医学知识的显式关联,并利用LLMs的隐式知识进行初步诊断,扩展关键实体。... 针对现有大型语言模型(LLMs)在医学推理任务中存在的隐式知识利用不足、推理路径冗余及透明度缺失等问题,提出一种基于协同推理的医学问答方法。构建推理子图学习医学知识的显式关联,并利用LLMs的隐式知识进行初步诊断,扩展关键实体。引入剪枝技术去除冗余推理路径,并设计推理融合机制对LLMs诊断结果与子图推理结果进行对比,以优化推理过程。在GenMedGPT-5k和CMCQA两个数据集上进行了广泛实验,实验结果表明,所提方法在推理准确性上均优于现有基准模型。 展开更多
关键词 医疗问答 提示工程 知识图谱 大型语言模型 医疗诊断 知识图谱与LLMs结合 知识图谱增强推理
在线阅读 下载PDF
素养导向的中小学人工智能课程知识图谱构建与应用研究
13
作者 黄景修 郑孜譞 +3 位作者 赖飞宇 张舒冉 陈星宇 郑云翔 《中国电化教育》 北大核心 2026年第2期46-52,59,共8页
人工智能重构教育系统背景下,构建中小学人工智能课程知识图谱是智能化人才培养的重要举措。然而,现有研究多集中于高等教育领域,缺乏与核心素养目标的深度融合,难以满足中小学人工智能教育需求。为此,该文以人工智能素养框架为指导,依... 人工智能重构教育系统背景下,构建中小学人工智能课程知识图谱是智能化人才培养的重要举措。然而,现有研究多集中于高等教育领域,缺乏与核心素养目标的深度融合,难以满足中小学人工智能教育需求。为此,该文以人工智能素养框架为指导,依托广州市中小学人工智能课程教材,采用自顶向下方法构建面向中小学的课程知识图谱。为验证其有效性,研发课程知识图谱增强的大模型问答系统,并通过人工评估测试系统性能。研究结果表明,课程知识图谱通过结构化知识注入机制,显著提升了大语言模型在人工智能素养的情感、思维、知识三个维度上的问答表现。该文通过课程知识图谱与大语言模型的融合应用,探索其在教育场景中的增益效应,实现从知识体系重构到工程实践的范式跃迁,为人工智能素养教育的规模化推广提供了理论与实践耦合的技术框架。 展开更多
关键词 课程知识图谱 人工智能素养 人工智能教育 大语言模型 问答系统
在线阅读 下载PDF
基于检索增强生成和智能体的建筑材料碳排放单位换算问答模型
14
作者 阎俏 焦飞 +2 位作者 严毅 杜向华 刘鹏程 《山东大学学报(工学版)》 北大核心 2026年第1期97-104,共8页
为解决建筑材料生产及运输阶段碳排放计算时建筑材料计量单位与碳排放因子单位不匹配的问题,提出一种基于检索增强生成(retrieval-augmented generation,RAG)和智能体(Agent)的建筑材料碳排放单位换算问答模型。通过解析典型材料换算步... 为解决建筑材料生产及运输阶段碳排放计算时建筑材料计量单位与碳排放因子单位不匹配的问题,提出一种基于检索增强生成(retrieval-augmented generation,RAG)和智能体(Agent)的建筑材料碳排放单位换算问答模型。通过解析典型材料换算步骤构建本地知识库,设计RAG模块,为换算提供步骤参考;开发可调用计算工具的Agent,执行换算过程中的数学运算;设计提示词模板并接入大语言模型,实现基于本地知识库的文本问答。试验结果表明,所提模型能够准确回答建材的单位换算问题,支持Web端与本地控制台交互,实现单位换算结果及推理步骤的可视化。 展开更多
关键词 建筑材料碳排放 单位换算 检索增强生成 智能体 问答模型
原文传递
基于检索增强生成的化工领域大模型智能问答
15
作者 宋凯 陈泽华 +3 位作者 娄娟 陈建 董宇轩 魏啸然 《天津大学学报(自然科学与工程技术版)》 北大核心 2026年第2期212-220,共9页
化工设备设计需要严格依照标准规范.然而标准规范数量多、内容上相互引用,设计人员面对非常规的设计要求或设计问题时很难准确、全面地查找到所有涉及的标准规范条目.利用检索增强生成(RAG)技术结合大语言模型(LLM)可以对设计要求或设... 化工设备设计需要严格依照标准规范.然而标准规范数量多、内容上相互引用,设计人员面对非常规的设计要求或设计问题时很难准确、全面地查找到所有涉及的标准规范条目.利用检索增强生成(RAG)技术结合大语言模型(LLM)可以对设计要求或设计问题进行准确回答的同时分析并提供相应标准规范内容,从而避免遗漏相关的标准规范.然而,由于化工设备设计领域知识库中具有大量公式、图表等复杂数据,如何构建相应的结构化RAG数据库实现LLM在化工设备设计领域的智能问答尚不明确.针对上述问题,本文提出了一种垂直领域的复杂数据智能问答系统构建一体化框架,该框架结合提示工程方法与多个视觉语言模型以实现RAG数据库的构建,采用语义检索与重排序技术,并选取嵌入模型与大语言模型分别作为检索器与生成器,以实现基于RAG的智能问答.基于该框架,本文构建了化工设备设计领域的智能问答系统,并使用Qwen2.5-72b和Qwen2.5-7b模型在以GB/T 150—2011规范为主的压力容器设计问答数据集上进行实验.结果表明,本文所提出的框架在复杂数据提取的准确性上优于现有技术,并通过RAG技术显著提升了问答系统的性能.相比于未结合RAG的技术,Qwen2.5-72b和Qwen2.5-7b模型的准确率分别提高了19.3%和17.7%.此外还对生成器接受的文档块数量对问答系统准确性的影响与设备设计领域数据的泛化性能进行了研究. 展开更多
关键词 大语言模型 检索增强生成 化工设备设计 智能问答 复杂数据信息提取
在线阅读 下载PDF
基于多粒度知识的无监督常识问答
16
作者 杨陟卓 王年楷 《中北大学学报(自然科学版)》 2026年第1期62-70,共9页
常识性问答(Commonsense Question Answering,CQA)是一项比传统问答任务更具挑战性的自然语言理解任务,它要求模型具备更强的常识推理能力。目前,基于无监督方法的常识问答在若干数据集上取得了较好的性能,但这些方法难以充分挖掘和利... 常识性问答(Commonsense Question Answering,CQA)是一项比传统问答任务更具挑战性的自然语言理解任务,它要求模型具备更强的常识推理能力。目前,基于无监督方法的常识问答在若干数据集上取得了较好的性能,但这些方法难以充分挖掘和利用常识知识,限制了模型在复杂场景下的推理能力。针对这一问题,本文提出了一种新颖的无监督常识问答方法,其核心优势在于通过无监督学习有效整合外部常识知识,从而提升模型的泛化能力和推理深度。首先,该方法对问题进行分类,区分科学常识问题与日常事件问题;随后,根据问题类型生成相应的知识前缀;接着,将知识前缀输入预训练语言模型,通过大模型提示生成多粒度的常识知识;最后,利用多粒度知识辅助问答推理模块进行答案生成。采用无监督方法不仅可以减少对标注数据的依赖,还能更好地适应多样化的常识场景,体现了其在实际应用中的灵活性和普适性。实验结果表明,所提方法在相关数据集上显著优于基线模型,验证了其在无监督常识问答任务中的正确性和合理性。 展开更多
关键词 常识问答 大模型提示 知识生成 答案推理
在线阅读 下载PDF
表格问答研究综述
17
作者 吴贤杰 李彤亮 李舟军 《计算机科学》 北大核心 2026年第3期295-306,共12页
表格作为一种重要的数据载体,能以紧凑的形式承载大量高价值信息,被广泛应用于经济、金融及科研等领域。表格问答(Table Question Answering,TableQA)旨在针对用自然语言描述的问题,从表格数据中自动进行推理并生成相应的答案,是自然语... 表格作为一种重要的数据载体,能以紧凑的形式承载大量高价值信息,被广泛应用于经济、金融及科研等领域。表格问答(Table Question Answering,TableQA)旨在针对用自然语言描述的问题,从表格数据中自动进行推理并生成相应的答案,是自然语言处理与数据分析交叉领域的重要研究方向。与传统的文本问答和知识库问答相比,表格问答不仅需要理解自然语言,还须解析表格的二维结构,并处理数值计算与复杂逻辑推理,因此面临更大的挑战。近年来,随着多样化数据集的持续构建,表格问答技术不断取得进展。其研究范式经历了从基于规则与模板的方法,到统计学习与神经网络模型的应用,再到预训练语言模型的引入,整体性能不断提升。尤其是近年来大语言模型(Large Language Models,LLMs)的兴起,进一步推动了表格问答进入新的发展阶段。凭借卓越的跨任务泛化能力与推理能力,大语言模型加速了新型研究范式的形成与发展,为方法创新提供了有力支撑。文中系统梳理了表格问答技术的演进脉络与代表性方法,重点总结了大语言模型驱动下的最新研究进展,概述了当前研究面临的关键挑战,并对未来发展趋势进行了展望。 展开更多
关键词 表格问答 表格推理 大语言模型 自然语言处理
在线阅读 下载PDF
大模型时代自动问答系统及评价体系综述
18
作者 崔龙飞 王宗水 +1 位作者 鲍盈旭 赵红 《计算机工程与应用》 北大核心 2026年第5期1-17,共17页
大模型时代,自动问答系统呈现出诸多新的特征。通过文献阅读和梳理,对自动问答系统特征和评测体系进行总结与归纳,从问答模型推理训练的训练数据、预训练框架、模型后处理、模型高效微调等阶段,对比大模型发展初期“追求数据和参数规模... 大模型时代,自动问答系统呈现出诸多新的特征。通过文献阅读和梳理,对自动问答系统特征和评测体系进行总结与归纳,从问答模型推理训练的训练数据、预训练框架、模型后处理、模型高效微调等阶段,对比大模型发展初期“追求数据和参数规模”的训练方法和如今“注重数据和模型效率”之间的差异,系统分析基于大模型的自动问答系统新的特征。总结当前各种类型的自动问答大模型评测体系,并详细梳理自动化评价体系HELM(holistic evaluation of language model)在自动问答任务上的数据集、评价指标和量化计算方法。未来基于大模型的自动问答系统研究将会围绕多模态融合、高安全性、高可解释性、低资源消耗,以及结合大模型和自动化的综合评价体系这几个方面进一步拓展与深化。 展开更多
关键词 大模型(LMs) 自动问答(QA)系统 系统特征 HELM评价体系
在线阅读 下载PDF
智能问答系统逻辑推理测试
19
作者 沈庆超 李行健 +3 位作者 姜佳君 陈俊洁 齐一先 王赞 《软件学报》 北大核心 2026年第2期543-562,共20页
智能问答系统利用信息检索和自然语言处理技术,实现对问题的自动化回复.然而,与其他人工智能软件相似,智能问答系统同样存在缺陷.存在缺陷的智能问答系统会降低用户体验,造成企业的经济损失,甚至引发社会层面的恐慌.因此,及时检测并修... 智能问答系统利用信息检索和自然语言处理技术,实现对问题的自动化回复.然而,与其他人工智能软件相似,智能问答系统同样存在缺陷.存在缺陷的智能问答系统会降低用户体验,造成企业的经济损失,甚至引发社会层面的恐慌.因此,及时检测并修复智能问答系统中的缺陷至关重要.目前,智能问答系统自动测试方法主要分为两类.其一,基于问题与预测答案合成假定事实,并基于假定事实生成新问题和预期答案,以此揭示问答系统中的缺陷.其二,从现有数据集中提取不影响原问题答案的知识片段并融入原始测试输入中生成答案一致的新测试输入,实现对问答系统的缺陷检测任务.然而,这两类方法均着重于测试模型的语义理解能力,未能充分测试模型的逻辑推理能力.此外,这两类方法分别依赖于问答系统的回答范式和模型自带的数据集来生成新的测试用例,限制了其在基于大规模语言模型的问答系统中的测试效能.针对上述挑战,提出一种逻辑引导的蜕变测试技术QALT.QALT设计了3种逻辑相关的蜕变关系,并使用了语义相似度度量和依存句法分析等技术指导生成高质量的测试用例,实现对智能问答系统的精准测试.实验结果表明,QALT在两类智能问答系统上一共检测9247个缺陷,分别比当前两种最先进的技术(即QAQA和QAAskeR)多检测3150和3897个缺陷.基于人工采样标注结果的统计分析,QALT在两个智能问答系统上检测到真阳性缺陷的期望数量总和为8073,预期比QAQA和QAAskeR分别多检测2142和4867个真阳性缺陷.此外,使用QALT生成的测试输入通过模型微调对被测软件中的缺陷进行修复.微调后模型的错误率成功地从22.33%降至14.37%. 展开更多
关键词 智能问答系统 测试用例生成 蜕变测试 大型语言模型
在线阅读 下载PDF
面向遥感视觉问答的跨模态知识引入与提示推理框架
20
作者 董欣 俞鹏飞 顾晶晶 《计算机科学与探索》 北大核心 2026年第3期760-772,共13页
随着遥感技术的快速发展,遥感视觉问答(RSVQA)作为一种结合语言与视觉交互的新兴技术,显著提升了地球观测、环境监测等领域中遥感图像信息的解读效率和交互能力。然而,RSVQA仍面临遥感图像信息复杂度高、遥感图像-文本对齐数据稀缺,以... 随着遥感技术的快速发展,遥感视觉问答(RSVQA)作为一种结合语言与视觉交互的新兴技术,显著提升了地球观测、环境监测等领域中遥感图像信息的解读效率和交互能力。然而,RSVQA仍面临遥感图像信息复杂度高、遥感图像-文本对齐数据稀缺,以及文本问题表达形式多样等挑战。为了应对这些挑战,提出一种面向RSVQA的跨模态知识引入与提示推理框架(CMKIP)。针对遥感图像的高复杂度,CMKIP为大语言模型LLaMA构建可学习的图像特征适配器,以具备对复杂图像的表征能力;针对遥感图像-文本对齐数据稀缺问题,构建自动化数据生成管道,从公开遥感数据集中生成高质量的图像-文本对,实现高效的遥感领域知识注入;针对问题表达的多样性,创新性地提出一种大小模型协同推理机制,利用小模型进行知识库检索与中间推理校正,显著提升大语言模型对多样化问题的理解能力与推理准确性。此外,CMKIP支持根据任务需求灵活更换小模型,可广泛应用于遥感领域的多项下游任务。实验结果表明,CMKIP在RSVQA基准数据集上的性能显著优于现有方法,特别是在低样本场景下表现尤为突出,展示了其在RSVQA任务中的有效性和泛化性。 展开更多
关键词 遥感视觉问答 大语言模型 跨模态扩展 遥感微调指令集 轻量级模型 提示推理
在线阅读 下载PDF
上一页 1 2 19 下一页 到第
使用帮助 返回顶部