The rapid development of information technology and accelerated digitalization have led to an explosive growth of data across various fields.As a key technology for knowledge representation and sharing,knowledge graph...The rapid development of information technology and accelerated digitalization have led to an explosive growth of data across various fields.As a key technology for knowledge representation and sharing,knowledge graphs play a crucial role by constructing structured networks of relationships among entities.However,data sparsity and numerous unexplored implicit relations result in the widespread incompleteness of knowledge graphs.In static knowledge graph completion,most existing methods rely on linear operations or simple interaction mechanisms for triple encoding,making it difficult to fully capture the deep semantic associations between entities and relations.Moreover,many methods focus only on the local information of individual triples,ignoring the rich semantic dependencies embedded in the neighboring nodes of entities within the graph structure,which leads to incomplete embedding representations.To address these challenges,we propose Two-Stage Mixer Embedding(TSMixerE),a static knowledge graph completion method based on entity context.In the unit semantic extraction stage,TSMixerE leveragesmulti-scale circular convolution to capture local features atmultiple granularities,enhancing the flexibility and robustness of feature interactions.A channel attention mechanism amplifies key channel responses to suppress noise and irrelevant information,thereby improving the discriminative power and semantic depth of feature representations.For contextual information fusion,a multi-layer self-attentionmechanism enables deep interactions among contextual cues,effectively integrating local details with global context.Simultaneously,type embeddings clarify the semantic identities and roles of each component,enhancing the model’s sensitivity and fusion capabilities for diverse information sources.Furthermore,TSMixerE constructs contextual unit sequences for entities,fully exploring neighborhood information within the graph structure to model complex semantic dependencies,thus improving the completeness and generalization of embedding representations.展开更多
With the widespread use of social media,the propagation of health-related rumors has become a significant public health threat.Existing methods for detecting health rumors predominantly rely on external knowledge or p...With the widespread use of social media,the propagation of health-related rumors has become a significant public health threat.Existing methods for detecting health rumors predominantly rely on external knowledge or propagation structures,with only a few recent approaches attempting causal inference;however,these have not yet effectively integrated causal discovery with domain-specific knowledge graphs for detecting health rumors.In this study,we found that the combined use of causal discovery and domain-specific knowledge graphs can effectively identify implicit pseudo-causal logic embedded within texts,holding significant potential for health rumor detection.To this end,we propose CKDG—a dual-graph fusion framework based on causal logic and medical knowledge graphs.CKDG constructs a weighted causal graph to capture the implicit causal relationships in the text and introduces a medical knowledge graph to verify semantic consistency,thereby enhancing the ability to identify the misuse of professional terminology and pseudoscientific claims.In experiments conducted on a dataset comprising 8430 health rumors,CKDG achieved an accuracy of 91.28%and an F1 score of 90.38%,representing improvements of 5.11%and 3.29%over the best baseline,respectively.Our results indicate that the integrated use of causal discovery and domainspecific knowledge graphs offers significant advantages for health rumor detection systems.This method not only improves detection performance but also enhances the transparency and credibility of model decisions by tracing causal chains and sources of knowledge conflicts.We anticipate that this work will provide key technological support for the development of trustworthy health-information filtering systems,thereby improving the reliability of public health information on social media.展开更多
Objective To address the dual challenges of long-tail distribution and feature sparsity in traditional Chinese medicine(TCM)syndrome differentiation within real clinical settings,we propose a data-efficient learning f...Objective To address the dual challenges of long-tail distribution and feature sparsity in traditional Chinese medicine(TCM)syndrome differentiation within real clinical settings,we propose a data-efficient learning framework enhanced by knowledge graphs.Methods We developed Agent-GNN,a three-stage decoupled learning framework,and validated it on the Traditional Chinese Medicine Syndrome Diagnosis(TCM-SD)dataset containing 54152 clinical records across 148 syndrome categories.First,we constructed a comprehensive medical knowledge graph encoding the complete TCM reasoning system.Second,we proposed a Functional Patient Profiling(FPP)method that utilizes large language models(LLMs)combined with Graph Retrieval-Augmented Generation(RAG)to extract structured symptom-etiology-pathogenesis subgraphs from medical records.Third,we employed heterogeneous graph neural networks to learn structured combination patterns explicitly.We compared our method against multiple baselines including BERT,ZY-BERT,ZY-BERT+Know,GAT,and GPT-4 Few-shot,using macro-F1 score as the primary evaluation metric.Additionally,ablation experiments were conducted to validate the contribution of each key component to model performance.Results Agent-GNN achieved an overall macro-F1 score of 72.4%,representing an 8.7 percentage points improvement over ZY-BERT+Know(63.7%),the strongest baseline among traditional methods.For long-tail syndromes with fewer than 10 samples,Agent-GNN reached a macro-F1 score of 58.6%,compared with 39.3%for ZY-BERT+Know and 41.2%for GPT-4 Few-shot,representing relative improvements of 49.2%and 42.2%,respectively.Ablation experiments confirmed that the explicit modeling of etiology-pathogenesis nodes contributed 12.4 percentage points to this enhanced long-tail syndrome performance.Conclusion This study proposes Agent-GNN,a knowledge graph-enhanced framework that effectively addresses the long-tail distribution challenge in TCM syndrome differentiation.By explicitly modeling manifestation-mechanism-essence patterns through structured knowledge graphs,our approach achieves superior performance in data-scarce scenarios while providing interpretable reasoning paths for TCM intelligent diagnosis.展开更多
Knowledge graphs often suffer from sparsity and incompleteness.Knowledge graph reasoning is an effective way to address these issues.Unlike static knowledge graph reasoning,which is invariant over time,dynamic knowled...Knowledge graphs often suffer from sparsity and incompleteness.Knowledge graph reasoning is an effective way to address these issues.Unlike static knowledge graph reasoning,which is invariant over time,dynamic knowledge graph reasoning is more challenging due to its temporal nature.In essence,within each time step in a dynamic knowledge graph,there exists structural dependencies among entities and relations,whereas between adjacent time steps,there exists temporal continuity.Based on these structural and temporal characteristics,we propose a model named“DKGR-DR”to learn distributed representations of entities and relations by combining recurrent neural networks and graph neural networks to capture structural dependencies and temporal continuity in DKGs.In addition,we construct a static attribute graph to represent entities’inherent properties.DKGR-DR is capable of modeling both dynamic and static aspects of entities,enabling effective entity prediction and relation prediction.We conduct experiments on ICEWS05-15,ICEWS18,and ICEWS14 to demonstrate that DKGR-DR achieves competitive performance.展开更多
To integrate traditional culture and modern technology,Shandong University’s School of Software has promoted an interdisciplinary teaching project called IYAN&ITAN,the I Ching Knowledge Graph.The project,driven b...To integrate traditional culture and modern technology,Shandong University’s School of Software has promoted an interdisciplinary teaching project called IYAN&ITAN,the I Ching Knowledge Graph.The project,driven by I Ching texts,guides students to practice natural language processing(NLP)and knowledge graph technology in a task-oriented curriculum,based on constructivism,situated learning,and inquiry-based pedagogy,with a progressive and task-oriented teaching model.The platform established enables the retrieval of knowledge,parsing of text,symbolic-numeric analysis,and historical commentary integration,making possible multidimensional,structured representation of I Ching knowledge,and offering an extensible reference for interdisciplinary learning in the context of New Engineering Education.展开更多
This paper delves into effective pathways for transforming course ecosystems from resource provision to knowledge service and competency development through university-enterprise collaboration in co-building knowledge...This paper delves into effective pathways for transforming course ecosystems from resource provision to knowledge service and competency development through university-enterprise collaboration in co-building knowledge graphs and intelligent shared courses.This approach enables personalized,learning-driven teaching.Based on knowledge graphs and integrated teacher-machine-student smart teaching scenarios,it not only innovates autonomous learning environments and human-computer interaction models while optimizing teaching experiences for both instructors and students,but also effectively addresses the issues of students’“scattered,superficial,and fragmented learning”.This establishes the foundation for personalized teaching tailored to individual aptitudes.展开更多
The reliable operation of power grid secondary equipment is an important guarantee for the safety and stability of the power system.However,various defects could be produced in the secondary equipment during longtermo...The reliable operation of power grid secondary equipment is an important guarantee for the safety and stability of the power system.However,various defects could be produced in the secondary equipment during longtermoperation.The complex relationship between the defect phenomenon andmulti-layer causes and the probabilistic influence of secondary equipment cannot be described through knowledge extraction and fusion technology by existing methods,which limits the real-time and accuracy of defect identification.Therefore,a defect recognition method based on the Bayesian network and knowledge graph fusion is proposed.The defect data of secondary equipment is transformed into the structured knowledge graph through knowledge extraction and fusion technology.The knowledge graph of power grid secondary equipment is mapped to the Bayesian network framework,combined with historical defect data,and introduced Noisy-OR nodes.The prior and conditional probabilities of the Bayesian network are then reasonably assigned to build a model that reflects the probability dependence between defect phenomena and potential causes in power grid secondary equipment.Defect identification of power grid secondary equipment is achieved by defect subgraph search based on the knowledge graph,and defect inference based on the Bayesian network.Practical application cases prove this method’s effectiveness in identifying secondary equipment defect causes,improving identification accuracy and efficiency.展开更多
Large language models(LLMs)show great potential in educational scenarios but face challenges like hallucination,knowledge gaps,and reasoning discontinuities.This study proposes a dynamic knowledge enhancement framewor...Large language models(LLMs)show great potential in educational scenarios but face challenges like hallucination,knowledge gaps,and reasoning discontinuities.This study proposes a dynamic knowledge enhancement framework.By integrating local knowledge graphs and stepwise prompting mechanisms,it improves LLMs’accuracy and interpretability in solving professional domain problems.The framework has two core modules:an LLM-driven knowledge graph construction system for incremental updates and a unified reasoning engine for generating enhanced prompts.Experiments on 680 educational questions show that the method boosts accuracy by 4.5%and 4.3%for multi-step reasoning and knowledge-dependent questions respectively,and increases reasoning step completeness from 68.2%to 83.7%.It also reduces hallucination problems.Key contributions include the followings:①validation of an effective framework synergizing knowledge graphs with retrieval mechanisms to enhance LLM reliability;②a stepwise prompting strategy enforcing explicit reasoning chain generation,addressing pedagogical requirements for process interpretability;③a lightweight deployment solution for educational systems such as adaptive learning platforms.展开更多
Knowledge graphs,which combine structured representation with semantic modeling,have shown great potential in knowledge expression,causal inference,and automated reasoning,and are widely used in fields such as intelli...Knowledge graphs,which combine structured representation with semantic modeling,have shown great potential in knowledge expression,causal inference,and automated reasoning,and are widely used in fields such as intelligent question answering,decision support,and fault diagnosis.As high-speed train systems become increasingly intelligent and interconnected,fault patterns have grown more complex and dynamic.Knowledge graphs offer a promising solution to support the structured management and real-time reasoning of fault knowledge,addressing key requirements such as interpretability,accuracy,and continuous evolution in intelligent diagnostic systems.However,conventional knowledge graph construction relies heavily on domain expertise and specialized tools,resulting in high entry barriers for non-experts and limiting their practical application in frontline maintenance scenarios.To address this limitation,this paper proposes a fault knowledge modeling approach for high-speed trains that integrates structured logic diagrams with knowledge graphs.The method employs a seven-layer logic structure—comprising fault name,applicable vehicles,diagnostic logic,signal parameters,verification conditions,fault causes,and emergency measures—to transform unstructured knowledge into a visual and hierarchical representation.A semantic mapping mechanism is then used to automatically convert logic diagrams into machine-interpretable knowledge graphs,enabling dynamic reasoning and knowledge reuse.Furthermore,the proposed method establishes a three-layer architecture—logic structuring,knowledge graph transformation,and dynamic inference—to bridge human-expert logic with machinebased reasoning.Experimental validation and system implementation demonstrate that this approach not only improves knowledge interpretability and inference precision but also significantly enhances modeling efficiency and system maintainability.It provides a scalable and adaptable solution for intelligent operation and maintenance platforms in the high-speed rail domain.展开更多
Temporal knowledge graph completion(TKGC),which merges temporal information into traditional static knowledge graph completion(SKGC),has garnered increasing attention recently.Among numerous emerging approaches,transl...Temporal knowledge graph completion(TKGC),which merges temporal information into traditional static knowledge graph completion(SKGC),has garnered increasing attention recently.Among numerous emerging approaches,translation-based embedding models constitute a prominent approach in TKGC research.However,existing translation-based methods typically incorporate timestamps into entities or relations,rather than utilizing them independently.This practice fails to fully exploit the rich semantics inherent in temporal information,thereby weakening the expressive capability of models.To address this limitation,we propose embedding timestamps,like entities and relations,in one or more dedicated semantic spaces.After projecting all embeddings into a shared space,we use the relation-timestamp pair instead of the conventional relation embedding as the translation vector between head and tail entities.Our method elevates timestamps to the same representational significance as entities and relations.Based on this strategy,we introduce two novel translation-based embedding models:TE-TransR and TE-TransT.With the independent representation of timestamps,our method not only enhances capabilities in link prediction but also facilitates a relatively underexplored task,namely time prediction.To further bolster the precision and reliability of time prediction,we introduce a granular,time unit-based timestamp setting and a relation-specific evaluation protocol.Extensive experiments demonstrate that our models achieve strong performance on link prediction benchmarks,with TE-TransR outperforming existing baselines in the time prediction task.展开更多
In the international shipping industry, digital intelligence transformation has become essential, with both governments and enterprises actively working to integrate diverse datasets. The domain of maritime and shippi...In the international shipping industry, digital intelligence transformation has become essential, with both governments and enterprises actively working to integrate diverse datasets. The domain of maritime and shipping is characterized by a vast array of document types, filled with complex, large-scale, and often chaotic knowledge and relationships. Effectively managing these documents is crucial for developing a Large Language Model (LLM) in the maritime domain, enabling practitioners to access and leverage valuable information. A Knowledge Graph (KG) offers a state-of-the-art solution for enhancing knowledge retrieval, providing more accurate responses and enabling context-aware reasoning. This paper presents a framework for utilizing maritime and shipping documents to construct a knowledge graph using GraphRAG, a hybrid tool combining graph-based retrieval and generation capabilities. The extraction of entities and relationships from these documents and the KG construction process are detailed. Furthermore, the KG is integrated with an LLM to develop a Q&A system, demonstrating that the system significantly improves answer accuracy compared to traditional LLMs. Additionally, the KG construction process is up to 50% faster than conventional LLM-based approaches, underscoring the efficiency of our method. This study provides a promising approach to digital intelligence in shipping, advancing knowledge accessibility and decision-making.展开更多
Since Google introduced the concept of Knowledge Graphs(KGs)in 2012,their construction technologies have evolved into a comprehensive methodological framework encompassing knowledge acquisition,extraction,representati...Since Google introduced the concept of Knowledge Graphs(KGs)in 2012,their construction technologies have evolved into a comprehensive methodological framework encompassing knowledge acquisition,extraction,representation,modeling,fusion,computation,and storage.Within this framework,knowledge extraction,as the core component,directly determines KG quality.In military domains,traditional manual curation models face efficiency constraints due to data fragmentation,complex knowledge architectures,and confidentiality protocols.Meanwhile,crowdsourced ontology construction approaches from general domains prove non-transferable,while human-crafted ontologies struggle with generalization deficiencies.To address these challenges,this study proposes an OntologyAware LLM Methodology for Military Domain Knowledge Extraction(LLM-KE).This approach leverages the deep semantic comprehension capabilities of Large Language Models(LLMs)to simulate human experts’cognitive processes in crowdsourced ontology construction,enabling automated extraction of military textual knowledge.It concurrently enhances knowledge processing efficiency and improves KG completeness.Empirical analysis demonstrates that this method effectively resolves scalability and dynamic adaptation challenges in military KG construction,establishing a novel technological pathway for advancing military intelligence development.展开更多
Deep-time Earth research plays a pivotal role in deciphering the rates,patterns,and mechanisms of Earth's evolutionary processes throughout geological history,providing essential scientific foundations for climate...Deep-time Earth research plays a pivotal role in deciphering the rates,patterns,and mechanisms of Earth's evolutionary processes throughout geological history,providing essential scientific foundations for climate prediction,natural resource exploration,and sustainable planetary stewardship.To advance Deep-time Earth research in the era of big data and artificial intelligence,the International Union of Geological Sciences initiated the“Deeptime Digital Earth International Big Science Program”(DDE)in 2019.At the core of this ambitious program lies the development of geoscience knowledge graphs,serving as a transformative knowledge infrastructure that enables the integration,sharing,mining,and analysis of heterogeneous geoscience big data.The DDE knowledge graph initiative has made significant strides in three critical dimensions:(1)establishing a unified knowledge structure across geoscience disciplines that ensures consistent representation of geological entities and their interrelationships through standardized ontologies and semantic frameworks;(2)developing a robust and scalable software infrastructure capable of supporting both expert-driven and machine-assisted knowledge engineering for large-scale graph construction and management;(3)implementing a comprehensive three-tiered architecture encompassing basic,discipline-specific,and application-oriented knowledge graphs,spanning approximately 20 geoscience disciplines.Through its open knowledge framework and international collaborative network,this initiative has fostered multinational research collaborations,establishing a robust foundation for next-generation geoscience research while propelling the discipline toward FAIR(Findable,Accessible,Interoperable,Reusable)data practices in deep-time Earth systems research.展开更多
Tracking and analyzing data from research projects is critical for understanding research trends and supporting the development of science and technology strategies.However,the data from these projects is often comple...Tracking and analyzing data from research projects is critical for understanding research trends and supporting the development of science and technology strategies.However,the data from these projects is often complex and inadequate,making it challenging for researchers to conduct in-depth data mining to improve policies or management.To address this problem,this paper adopts a top-down approach to construct a knowledge graph(KG)for research projects.Firstly,we construct an integrated ontology by referring to the metamodel of various architectures,which is called the meta-model integration conceptual reference model.Subsequently,we use the dependency parsing method to extract knowledge from unstructured textual data and use the entity alignment method based on weakly supervised learning to classify the extracted entities,completing the construction of the KG for the research projects.In addition,a knowledge inference model based on representation learning is employed to achieve knowledge completion and improve the KG.Finally,experiments are conducted on the KG for research projects and the results demonstrate the effectiveness of the proposed method in enriching incomplete data within the KG.展开更多
Accurate prediction of drug-target interactions(DTIs)plays a pivotal role in drug discovery,facilitating optimization of lead compounds,drug repurposing and elucidation of drug side effects.However,traditional DTI pre...Accurate prediction of drug-target interactions(DTIs)plays a pivotal role in drug discovery,facilitating optimization of lead compounds,drug repurposing and elucidation of drug side effects.However,traditional DTI prediction methods are often limited by incomplete biological data and insufficient representation of protein features.In this study,we proposed KG-CNNDTI,a novel knowledge graph-enhanced framework for DTI prediction,which integrates heterogeneous biological information to improve model generalizability and predictive performance.The proposed model utilized protein embeddings derived from a biomedical knowledge graph via the Node2Vec algorithm,which were further enriched with contextualized sequence representations obtained from ProteinBERT.For compound representation,multiple molecular fingerprint schemes alongside the Uni-Mol pre-trained model were evaluated.The fused representations served as inputs to both classical machine learning models and a convolutional neural network-based predictor.Experimental evaluations across benchmark datasets demonstrated that KG-CNNDTI achieved superior performance compared to state-of-the-art methods,particularly in terms of Precision,Recall,F1-Score and area under the precision-recall curve(AUPR).Ablation analysis highlighted the substantial contribution of knowledge graph-derived features.Moreover,KG-CNNDTI was employed for virtual screening of natural products against Alzheimer's disease,resulting in 40 candidate compounds.5 were supported by literature evidence,among which 3 were further validated in vitro assays.展开更多
With the continuous development of artificial intelligence and natural language processing technologies, traditional retrieval-augmented generation (RAG) techniques face numerous challenges in document answer precisio...With the continuous development of artificial intelligence and natural language processing technologies, traditional retrieval-augmented generation (RAG) techniques face numerous challenges in document answer precision and similarity measurement. This study, set against the backdrop of the shipping industry, combines top-down and bottom-up schema design strategies to achieve precise and flexible knowledge representation. The research adopts a semi-structured approach, innovatively constructing an adaptive schema generation mechanism based on reinforcement learning, which models the knowledge graph construction process as a Markov decision process. This method begins with general concepts, defining foundational industry concepts, and then delves into abstracting core concepts specific to the maritime domain through an adaptive pattern generation mechanism that dynamically adjusts the knowledge structure. Specifically, the study designs a four-layer knowledge construction framework, including the data layer, modeling layer, technology layer, and application layer. It draws on a mutual indexing strategy, integrating large language models and traditional information extraction techniques. By leveraging self-attention mechanisms and graph attention networks, it efficiently extracts semantic relationships. The introduction of logic-form-driven solvers and symbolic decomposition techniques for reasoning significantly enhances the model’s ability to understand complex semantic relationships. Additionally, the use of open information extraction and knowledge alignment techniques further improves the efficiency and accuracy of information retrieval. Experimental results demonstrate that the proposed method not only achieves significant performance improvements in knowledge graph retrieval within the shipping domain but also holds important theoretical innovation and practical application value.展开更多
The cost and strict input format requirements of GraphRAG make it less efficient for processing large documents. This paper proposes an alternative approach for constructing a knowledge graph (KG) from a PDF document ...The cost and strict input format requirements of GraphRAG make it less efficient for processing large documents. This paper proposes an alternative approach for constructing a knowledge graph (KG) from a PDF document with a focus on simplicity and cost-effectiveness. The process involves splitting the document into chunks, extracting concepts within each chunk using a large language model (LLM), and building relationships based on the proximity of concepts in the same chunk. Unlike traditional named entity recognition (NER), which identifies entities like “Shanghai”, the proposed method identifies concepts, such as “Convenient transportation in Shanghai” which is found to be more meaningful for KG construction. Each edge in the KG represents a relationship between concepts occurring in the same text chunk. The process is computationally inexpensive, leveraging locally set up tools like Mistral 7B openorca instruct and Ollama for model inference, ensuring the entire graph generation process is cost-free. A method of assigning weights to relationships, grouping similar pairs, and summarizing multiple relationships into a single edge with associated weight and relation details is introduced. Additionally, node degrees and communities are calculated for node sizing and coloring. This approach offers a scalable, cost-effective solution for generating meaningful knowledge graphs from large documents, achieving results comparable to GraphRAG while maintaining accessibility for personal machines.展开更多
The article is based on language model,through the cue word engineering and agent thinking method,automatic knowledge extraction,with China accounting standards support to complete the corresponding knowledge map cons...The article is based on language model,through the cue word engineering and agent thinking method,automatic knowledge extraction,with China accounting standards support to complete the corresponding knowledge map construction.Through the way of extracting the accounting entities and their connections in the pattern layer,the data layer is provided for the fine-tuning and optimization of the large model.Studies found that,through the reasonable application of language model,knowledge can be realized in massive financial data neural five effective extracted tuples,and complete accounting knowledge map construction.展开更多
Multi-modal knowledge graph completion(MMKGC)aims to complete missing entities or relations in multi-modal knowledge graphs,thereby discovering more previously unknown triples.Due to the continuous growth of data and ...Multi-modal knowledge graph completion(MMKGC)aims to complete missing entities or relations in multi-modal knowledge graphs,thereby discovering more previously unknown triples.Due to the continuous growth of data and knowledge and the limitations of data sources,the visual knowledge within the knowledge graphs is generally of low quality,and some entities suffer from the issue of missing visual modality.Nevertheless,previous studies of MMKGC have primarily focused on how to facilitate modality interaction and fusion while neglecting the problems of low modality quality and modality missing.In this case,mainstream MMKGC models only use pre-trained visual encoders to extract features and transfer the semantic information to the joint embeddings through modal fusion,which inevitably suffers from problems such as error propagation and increased uncertainty.To address these problems,we propose a Multi-modal knowledge graph Completion model based on Super-resolution and Detailed Description Generation(MMCSD).Specifically,we leverage a pre-trained residual network to enhance the resolution and improve the quality of the visual modality.Moreover,we design multi-level visual semantic extraction and entity description generation,thereby further extracting entity semantics from structural triples and visual images.Meanwhile,we train a variational multi-modal auto-encoder and utilize a pre-trained multi-modal language model to complement the missing visual features.We conducted experiments on FB15K-237 and DB13K,and the results showed that MMCSD can effectively perform MMKGC and achieve state-of-the-art performance.展开更多
As large language models(LLMs)continue to demonstrate their potential in handling complex tasks,their value in knowledge-intensive industrial scenarios is becoming increasingly evident.Fault diagnosis,a critical domai...As large language models(LLMs)continue to demonstrate their potential in handling complex tasks,their value in knowledge-intensive industrial scenarios is becoming increasingly evident.Fault diagnosis,a critical domain in the industrial sector,has long faced the dual challenges of managing vast amounts of experiential knowledge and improving human-machine collaboration efficiency.Traditional fault diagnosis systems,which are primarily based on expert systems,suffer from three major limitations:(1)ineffective organization of fault diagnosis knowledge,(2)lack of adaptability between static knowledge frameworks and dynamic engineering environments,and(3)difficulties in integrating expert knowledge with real-time data streams.These systemic shortcomings restrict the ability of conventional approaches to handle uncertainty.In this study,we proposed an intelligent computer numerical control(CNC)fault diagnosis system,integrating LLMs with knowledge graph(KG).First,we constructed a comprehensive KG that consolidated multi-source data for structured representation.Second,we designed a retrievalaugmented generation(RAG)framework leveraging the KG to support multi-turn interactive fault diagnosis while incorporating real-time engineering data into the decision-making process.Finally,we introduced a learning mechanism to facilitate dynamic knowledge updates.The experimental results demonstrated that our system significantly improved fault diagnosis accuracy,outperforming engineers with two years of professional experience on our constructed benchmark datasets.By integrating LLMs and KG,our framework surpassed the limitations of traditional expert systems rooted in symbolic reasoning,offering a novel approach to addressing the cognitive paradox of unstructured knowledge modeling and dynamic environment adaptation in industrial settings.展开更多
基金supported by the National Natural Science Foundation of China(No.62267005)the Chinese Guangxi Natural Science Foundation(No.2023GXNSFAA026493)+1 种基金Guangxi Collaborative Innovation Center ofMulti-Source Information Integration and Intelligent ProcessingGuangxi Academy of Artificial Intelligence.
文摘The rapid development of information technology and accelerated digitalization have led to an explosive growth of data across various fields.As a key technology for knowledge representation and sharing,knowledge graphs play a crucial role by constructing structured networks of relationships among entities.However,data sparsity and numerous unexplored implicit relations result in the widespread incompleteness of knowledge graphs.In static knowledge graph completion,most existing methods rely on linear operations or simple interaction mechanisms for triple encoding,making it difficult to fully capture the deep semantic associations between entities and relations.Moreover,many methods focus only on the local information of individual triples,ignoring the rich semantic dependencies embedded in the neighboring nodes of entities within the graph structure,which leads to incomplete embedding representations.To address these challenges,we propose Two-Stage Mixer Embedding(TSMixerE),a static knowledge graph completion method based on entity context.In the unit semantic extraction stage,TSMixerE leveragesmulti-scale circular convolution to capture local features atmultiple granularities,enhancing the flexibility and robustness of feature interactions.A channel attention mechanism amplifies key channel responses to suppress noise and irrelevant information,thereby improving the discriminative power and semantic depth of feature representations.For contextual information fusion,a multi-layer self-attentionmechanism enables deep interactions among contextual cues,effectively integrating local details with global context.Simultaneously,type embeddings clarify the semantic identities and roles of each component,enhancing the model’s sensitivity and fusion capabilities for diverse information sources.Furthermore,TSMixerE constructs contextual unit sequences for entities,fully exploring neighborhood information within the graph structure to model complex semantic dependencies,thus improving the completeness and generalization of embedding representations.
基金funded by the Hunan Provincial Natural Science Foundation of China(Grant No.2025JJ70105)the Hunan Provincial College Students’Innovation and Entrepreneurship Training Program(Project No.S202411342056)The article processing charge(APC)was funded by the Project No.2025JJ70105.
文摘With the widespread use of social media,the propagation of health-related rumors has become a significant public health threat.Existing methods for detecting health rumors predominantly rely on external knowledge or propagation structures,with only a few recent approaches attempting causal inference;however,these have not yet effectively integrated causal discovery with domain-specific knowledge graphs for detecting health rumors.In this study,we found that the combined use of causal discovery and domain-specific knowledge graphs can effectively identify implicit pseudo-causal logic embedded within texts,holding significant potential for health rumor detection.To this end,we propose CKDG—a dual-graph fusion framework based on causal logic and medical knowledge graphs.CKDG constructs a weighted causal graph to capture the implicit causal relationships in the text and introduces a medical knowledge graph to verify semantic consistency,thereby enhancing the ability to identify the misuse of professional terminology and pseudoscientific claims.In experiments conducted on a dataset comprising 8430 health rumors,CKDG achieved an accuracy of 91.28%and an F1 score of 90.38%,representing improvements of 5.11%and 3.29%over the best baseline,respectively.Our results indicate that the integrated use of causal discovery and domainspecific knowledge graphs offers significant advantages for health rumor detection systems.This method not only improves detection performance but also enhances the transparency and credibility of model decisions by tracing causal chains and sources of knowledge conflicts.We anticipate that this work will provide key technological support for the development of trustworthy health-information filtering systems,thereby improving the reliability of public health information on social media.
基金Sichuan TCM Culture Coordinated Development Research Center Project(2023XT131)National Key Science and Technology Project of China(2023ZD0509405)National Natural Science Foundation of China(82174236).
文摘Objective To address the dual challenges of long-tail distribution and feature sparsity in traditional Chinese medicine(TCM)syndrome differentiation within real clinical settings,we propose a data-efficient learning framework enhanced by knowledge graphs.Methods We developed Agent-GNN,a three-stage decoupled learning framework,and validated it on the Traditional Chinese Medicine Syndrome Diagnosis(TCM-SD)dataset containing 54152 clinical records across 148 syndrome categories.First,we constructed a comprehensive medical knowledge graph encoding the complete TCM reasoning system.Second,we proposed a Functional Patient Profiling(FPP)method that utilizes large language models(LLMs)combined with Graph Retrieval-Augmented Generation(RAG)to extract structured symptom-etiology-pathogenesis subgraphs from medical records.Third,we employed heterogeneous graph neural networks to learn structured combination patterns explicitly.We compared our method against multiple baselines including BERT,ZY-BERT,ZY-BERT+Know,GAT,and GPT-4 Few-shot,using macro-F1 score as the primary evaluation metric.Additionally,ablation experiments were conducted to validate the contribution of each key component to model performance.Results Agent-GNN achieved an overall macro-F1 score of 72.4%,representing an 8.7 percentage points improvement over ZY-BERT+Know(63.7%),the strongest baseline among traditional methods.For long-tail syndromes with fewer than 10 samples,Agent-GNN reached a macro-F1 score of 58.6%,compared with 39.3%for ZY-BERT+Know and 41.2%for GPT-4 Few-shot,representing relative improvements of 49.2%and 42.2%,respectively.Ablation experiments confirmed that the explicit modeling of etiology-pathogenesis nodes contributed 12.4 percentage points to this enhanced long-tail syndrome performance.Conclusion This study proposes Agent-GNN,a knowledge graph-enhanced framework that effectively addresses the long-tail distribution challenge in TCM syndrome differentiation.By explicitly modeling manifestation-mechanism-essence patterns through structured knowledge graphs,our approach achieves superior performance in data-scarce scenarios while providing interpretable reasoning paths for TCM intelligent diagnosis.
基金supported byNationalNatural Science Foundation of China(GrantNos.62071098,U24B20128)Sichuan Science and Technology Program(Grant No.2022YFG0319).
文摘Knowledge graphs often suffer from sparsity and incompleteness.Knowledge graph reasoning is an effective way to address these issues.Unlike static knowledge graph reasoning,which is invariant over time,dynamic knowledge graph reasoning is more challenging due to its temporal nature.In essence,within each time step in a dynamic knowledge graph,there exists structural dependencies among entities and relations,whereas between adjacent time steps,there exists temporal continuity.Based on these structural and temporal characteristics,we propose a model named“DKGR-DR”to learn distributed representations of entities and relations by combining recurrent neural networks and graph neural networks to capture structural dependencies and temporal continuity in DKGs.In addition,we construct a static attribute graph to represent entities’inherent properties.DKGR-DR is capable of modeling both dynamic and static aspects of entities,enabling effective entity prediction and relation prediction.We conduct experiments on ICEWS05-15,ICEWS18,and ICEWS14 to demonstrate that DKGR-DR achieves competitive performance.
基金support provided by the Shandong University Education and Teaching Reform Research Project(2024Y232)the“New 20 Regulations for Universities”funding program of Jinan(202228089).
文摘To integrate traditional culture and modern technology,Shandong University’s School of Software has promoted an interdisciplinary teaching project called IYAN&ITAN,the I Ching Knowledge Graph.The project,driven by I Ching texts,guides students to practice natural language processing(NLP)and knowledge graph technology in a task-oriented curriculum,based on constructivism,situated learning,and inquiry-based pedagogy,with a progressive and task-oriented teaching model.The platform established enables the retrieval of knowledge,parsing of text,symbolic-numeric analysis,and historical commentary integration,making possible multidimensional,structured representation of I Ching knowledge,and offering an extensible reference for interdisciplinary learning in the context of New Engineering Education.
基金supported by Harbin Institute of Technology High-level Teaching Achievement Award(National Level)Cultivation Project(256709).
文摘This paper delves into effective pathways for transforming course ecosystems from resource provision to knowledge service and competency development through university-enterprise collaboration in co-building knowledge graphs and intelligent shared courses.This approach enables personalized,learning-driven teaching.Based on knowledge graphs and integrated teacher-machine-student smart teaching scenarios,it not only innovates autonomous learning environments and human-computer interaction models while optimizing teaching experiences for both instructors and students,but also effectively addresses the issues of students’“scattered,superficial,and fragmented learning”.This establishes the foundation for personalized teaching tailored to individual aptitudes.
基金supported by the State Grid Southwest Branch Project“Research on Defect Diagnosis and Early Warning Technology of Relay Protection and Safety Automation Devices Based on Multi-Source Heterogeneous Defect Data”.
文摘The reliable operation of power grid secondary equipment is an important guarantee for the safety and stability of the power system.However,various defects could be produced in the secondary equipment during longtermoperation.The complex relationship between the defect phenomenon andmulti-layer causes and the probabilistic influence of secondary equipment cannot be described through knowledge extraction and fusion technology by existing methods,which limits the real-time and accuracy of defect identification.Therefore,a defect recognition method based on the Bayesian network and knowledge graph fusion is proposed.The defect data of secondary equipment is transformed into the structured knowledge graph through knowledge extraction and fusion technology.The knowledge graph of power grid secondary equipment is mapped to the Bayesian network framework,combined with historical defect data,and introduced Noisy-OR nodes.The prior and conditional probabilities of the Bayesian network are then reasonably assigned to build a model that reflects the probability dependence between defect phenomena and potential causes in power grid secondary equipment.Defect identification of power grid secondary equipment is achieved by defect subgraph search based on the knowledge graph,and defect inference based on the Bayesian network.Practical application cases prove this method’s effectiveness in identifying secondary equipment defect causes,improving identification accuracy and efficiency.
基金supported in part by the China-Singapore International Joint Research Institute(CSIJRI)under Grant No.206-A023001the Undergraduate Teaching Reform Project of Shandong University under Grant Nos.2023Y235 and 2025Y99.
文摘Large language models(LLMs)show great potential in educational scenarios but face challenges like hallucination,knowledge gaps,and reasoning discontinuities.This study proposes a dynamic knowledge enhancement framework.By integrating local knowledge graphs and stepwise prompting mechanisms,it improves LLMs’accuracy and interpretability in solving professional domain problems.The framework has two core modules:an LLM-driven knowledge graph construction system for incremental updates and a unified reasoning engine for generating enhanced prompts.Experiments on 680 educational questions show that the method boosts accuracy by 4.5%and 4.3%for multi-step reasoning and knowledge-dependent questions respectively,and increases reasoning step completeness from 68.2%to 83.7%.It also reduces hallucination problems.Key contributions include the followings:①validation of an effective framework synergizing knowledge graphs with retrieval mechanisms to enhance LLM reliability;②a stepwise prompting strategy enforcing explicit reasoning chain generation,addressing pedagogical requirements for process interpretability;③a lightweight deployment solution for educational systems such as adaptive learning platforms.
基金support from the Scientific Funding for the Center of National Railway Intelligent Transportation System Engineering and Technology,China Academy of Railway Sciences Corporation Limited(Grant No.2023YJ354)。
文摘Knowledge graphs,which combine structured representation with semantic modeling,have shown great potential in knowledge expression,causal inference,and automated reasoning,and are widely used in fields such as intelligent question answering,decision support,and fault diagnosis.As high-speed train systems become increasingly intelligent and interconnected,fault patterns have grown more complex and dynamic.Knowledge graphs offer a promising solution to support the structured management and real-time reasoning of fault knowledge,addressing key requirements such as interpretability,accuracy,and continuous evolution in intelligent diagnostic systems.However,conventional knowledge graph construction relies heavily on domain expertise and specialized tools,resulting in high entry barriers for non-experts and limiting their practical application in frontline maintenance scenarios.To address this limitation,this paper proposes a fault knowledge modeling approach for high-speed trains that integrates structured logic diagrams with knowledge graphs.The method employs a seven-layer logic structure—comprising fault name,applicable vehicles,diagnostic logic,signal parameters,verification conditions,fault causes,and emergency measures—to transform unstructured knowledge into a visual and hierarchical representation.A semantic mapping mechanism is then used to automatically convert logic diagrams into machine-interpretable knowledge graphs,enabling dynamic reasoning and knowledge reuse.Furthermore,the proposed method establishes a three-layer architecture—logic structuring,knowledge graph transformation,and dynamic inference—to bridge human-expert logic with machinebased reasoning.Experimental validation and system implementation demonstrate that this approach not only improves knowledge interpretability and inference precision but also significantly enhances modeling efficiency and system maintainability.It provides a scalable and adaptable solution for intelligent operation and maintenance platforms in the high-speed rail domain.
基金supported by the National Natural Science Foundation of China under Grant No.72293575.
文摘Temporal knowledge graph completion(TKGC),which merges temporal information into traditional static knowledge graph completion(SKGC),has garnered increasing attention recently.Among numerous emerging approaches,translation-based embedding models constitute a prominent approach in TKGC research.However,existing translation-based methods typically incorporate timestamps into entities or relations,rather than utilizing them independently.This practice fails to fully exploit the rich semantics inherent in temporal information,thereby weakening the expressive capability of models.To address this limitation,we propose embedding timestamps,like entities and relations,in one or more dedicated semantic spaces.After projecting all embeddings into a shared space,we use the relation-timestamp pair instead of the conventional relation embedding as the translation vector between head and tail entities.Our method elevates timestamps to the same representational significance as entities and relations.Based on this strategy,we introduce two novel translation-based embedding models:TE-TransR and TE-TransT.With the independent representation of timestamps,our method not only enhances capabilities in link prediction but also facilitates a relatively underexplored task,namely time prediction.To further bolster the precision and reliability of time prediction,we introduce a granular,time unit-based timestamp setting and a relation-specific evaluation protocol.Extensive experiments demonstrate that our models achieve strong performance on link prediction benchmarks,with TE-TransR outperforming existing baselines in the time prediction task.
文摘In the international shipping industry, digital intelligence transformation has become essential, with both governments and enterprises actively working to integrate diverse datasets. The domain of maritime and shipping is characterized by a vast array of document types, filled with complex, large-scale, and often chaotic knowledge and relationships. Effectively managing these documents is crucial for developing a Large Language Model (LLM) in the maritime domain, enabling practitioners to access and leverage valuable information. A Knowledge Graph (KG) offers a state-of-the-art solution for enhancing knowledge retrieval, providing more accurate responses and enabling context-aware reasoning. This paper presents a framework for utilizing maritime and shipping documents to construct a knowledge graph using GraphRAG, a hybrid tool combining graph-based retrieval and generation capabilities. The extraction of entities and relationships from these documents and the KG construction process are detailed. Furthermore, the KG is integrated with an LLM to develop a Q&A system, demonstrating that the system significantly improves answer accuracy compared to traditional LLMs. Additionally, the KG construction process is up to 50% faster than conventional LLM-based approaches, underscoring the efficiency of our method. This study provides a promising approach to digital intelligence in shipping, advancing knowledge accessibility and decision-making.
文摘Since Google introduced the concept of Knowledge Graphs(KGs)in 2012,their construction technologies have evolved into a comprehensive methodological framework encompassing knowledge acquisition,extraction,representation,modeling,fusion,computation,and storage.Within this framework,knowledge extraction,as the core component,directly determines KG quality.In military domains,traditional manual curation models face efficiency constraints due to data fragmentation,complex knowledge architectures,and confidentiality protocols.Meanwhile,crowdsourced ontology construction approaches from general domains prove non-transferable,while human-crafted ontologies struggle with generalization deficiencies.To address these challenges,this study proposes an OntologyAware LLM Methodology for Military Domain Knowledge Extraction(LLM-KE).This approach leverages the deep semantic comprehension capabilities of Large Language Models(LLMs)to simulate human experts’cognitive processes in crowdsourced ontology construction,enabling automated extraction of military textual knowledge.It concurrently enhances knowledge processing efficiency and improves KG completeness.Empirical analysis demonstrates that this method effectively resolves scalability and dynamic adaptation challenges in military KG construction,establishing a novel technological pathway for advancing military intelligence development.
基金Strategic Priority Research Program of the Chinese Academy of Sciences,No.XDB0740000National Key Research and Development Program of China,No.2022YFB3904200,No.2022YFF0711601+1 种基金Key Project of Innovation LREIS,No.PI009National Natural Science Foundation of China,No.42471503。
文摘Deep-time Earth research plays a pivotal role in deciphering the rates,patterns,and mechanisms of Earth's evolutionary processes throughout geological history,providing essential scientific foundations for climate prediction,natural resource exploration,and sustainable planetary stewardship.To advance Deep-time Earth research in the era of big data and artificial intelligence,the International Union of Geological Sciences initiated the“Deeptime Digital Earth International Big Science Program”(DDE)in 2019.At the core of this ambitious program lies the development of geoscience knowledge graphs,serving as a transformative knowledge infrastructure that enables the integration,sharing,mining,and analysis of heterogeneous geoscience big data.The DDE knowledge graph initiative has made significant strides in three critical dimensions:(1)establishing a unified knowledge structure across geoscience disciplines that ensures consistent representation of geological entities and their interrelationships through standardized ontologies and semantic frameworks;(2)developing a robust and scalable software infrastructure capable of supporting both expert-driven and machine-assisted knowledge engineering for large-scale graph construction and management;(3)implementing a comprehensive three-tiered architecture encompassing basic,discipline-specific,and application-oriented knowledge graphs,spanning approximately 20 geoscience disciplines.Through its open knowledge framework and international collaborative network,this initiative has fostered multinational research collaborations,establishing a robust foundation for next-generation geoscience research while propelling the discipline toward FAIR(Findable,Accessible,Interoperable,Reusable)data practices in deep-time Earth systems research.
基金supported by the National Natural Science Foundation of China(72101263).
文摘Tracking and analyzing data from research projects is critical for understanding research trends and supporting the development of science and technology strategies.However,the data from these projects is often complex and inadequate,making it challenging for researchers to conduct in-depth data mining to improve policies or management.To address this problem,this paper adopts a top-down approach to construct a knowledge graph(KG)for research projects.Firstly,we construct an integrated ontology by referring to the metamodel of various architectures,which is called the meta-model integration conceptual reference model.Subsequently,we use the dependency parsing method to extract knowledge from unstructured textual data and use the entity alignment method based on weakly supervised learning to classify the extracted entities,completing the construction of the KG for the research projects.In addition,a knowledge inference model based on representation learning is employed to achieve knowledge completion and improve the KG.Finally,experiments are conducted on the KG for research projects and the results demonstrate the effectiveness of the proposed method in enriching incomplete data within the KG.
基金supported by the National Natural Science Foundation of China(Nos.82173746 and U23A20530)Shanghai Frontiers Science Center of Optogenetic Techniques for Cell Metabolism(Shanghai Municipal Education Commission)。
文摘Accurate prediction of drug-target interactions(DTIs)plays a pivotal role in drug discovery,facilitating optimization of lead compounds,drug repurposing and elucidation of drug side effects.However,traditional DTI prediction methods are often limited by incomplete biological data and insufficient representation of protein features.In this study,we proposed KG-CNNDTI,a novel knowledge graph-enhanced framework for DTI prediction,which integrates heterogeneous biological information to improve model generalizability and predictive performance.The proposed model utilized protein embeddings derived from a biomedical knowledge graph via the Node2Vec algorithm,which were further enriched with contextualized sequence representations obtained from ProteinBERT.For compound representation,multiple molecular fingerprint schemes alongside the Uni-Mol pre-trained model were evaluated.The fused representations served as inputs to both classical machine learning models and a convolutional neural network-based predictor.Experimental evaluations across benchmark datasets demonstrated that KG-CNNDTI achieved superior performance compared to state-of-the-art methods,particularly in terms of Precision,Recall,F1-Score and area under the precision-recall curve(AUPR).Ablation analysis highlighted the substantial contribution of knowledge graph-derived features.Moreover,KG-CNNDTI was employed for virtual screening of natural products against Alzheimer's disease,resulting in 40 candidate compounds.5 were supported by literature evidence,among which 3 were further validated in vitro assays.
文摘With the continuous development of artificial intelligence and natural language processing technologies, traditional retrieval-augmented generation (RAG) techniques face numerous challenges in document answer precision and similarity measurement. This study, set against the backdrop of the shipping industry, combines top-down and bottom-up schema design strategies to achieve precise and flexible knowledge representation. The research adopts a semi-structured approach, innovatively constructing an adaptive schema generation mechanism based on reinforcement learning, which models the knowledge graph construction process as a Markov decision process. This method begins with general concepts, defining foundational industry concepts, and then delves into abstracting core concepts specific to the maritime domain through an adaptive pattern generation mechanism that dynamically adjusts the knowledge structure. Specifically, the study designs a four-layer knowledge construction framework, including the data layer, modeling layer, technology layer, and application layer. It draws on a mutual indexing strategy, integrating large language models and traditional information extraction techniques. By leveraging self-attention mechanisms and graph attention networks, it efficiently extracts semantic relationships. The introduction of logic-form-driven solvers and symbolic decomposition techniques for reasoning significantly enhances the model’s ability to understand complex semantic relationships. Additionally, the use of open information extraction and knowledge alignment techniques further improves the efficiency and accuracy of information retrieval. Experimental results demonstrate that the proposed method not only achieves significant performance improvements in knowledge graph retrieval within the shipping domain but also holds important theoretical innovation and practical application value.
文摘The cost and strict input format requirements of GraphRAG make it less efficient for processing large documents. This paper proposes an alternative approach for constructing a knowledge graph (KG) from a PDF document with a focus on simplicity and cost-effectiveness. The process involves splitting the document into chunks, extracting concepts within each chunk using a large language model (LLM), and building relationships based on the proximity of concepts in the same chunk. Unlike traditional named entity recognition (NER), which identifies entities like “Shanghai”, the proposed method identifies concepts, such as “Convenient transportation in Shanghai” which is found to be more meaningful for KG construction. Each edge in the KG represents a relationship between concepts occurring in the same text chunk. The process is computationally inexpensive, leveraging locally set up tools like Mistral 7B openorca instruct and Ollama for model inference, ensuring the entire graph generation process is cost-free. A method of assigning weights to relationships, grouping similar pairs, and summarizing multiple relationships into a single edge with associated weight and relation details is introduced. Additionally, node degrees and communities are calculated for node sizing and coloring. This approach offers a scalable, cost-effective solution for generating meaningful knowledge graphs from large documents, achieving results comparable to GraphRAG while maintaining accessibility for personal machines.
文摘The article is based on language model,through the cue word engineering and agent thinking method,automatic knowledge extraction,with China accounting standards support to complete the corresponding knowledge map construction.Through the way of extracting the accounting entities and their connections in the pattern layer,the data layer is provided for the fine-tuning and optimization of the large model.Studies found that,through the reasonable application of language model,knowledge can be realized in massive financial data neural five effective extracted tuples,and complete accounting knowledge map construction.
基金funded by Research Project,grant number BHQ090003000X03。
文摘Multi-modal knowledge graph completion(MMKGC)aims to complete missing entities or relations in multi-modal knowledge graphs,thereby discovering more previously unknown triples.Due to the continuous growth of data and knowledge and the limitations of data sources,the visual knowledge within the knowledge graphs is generally of low quality,and some entities suffer from the issue of missing visual modality.Nevertheless,previous studies of MMKGC have primarily focused on how to facilitate modality interaction and fusion while neglecting the problems of low modality quality and modality missing.In this case,mainstream MMKGC models only use pre-trained visual encoders to extract features and transfer the semantic information to the joint embeddings through modal fusion,which inevitably suffers from problems such as error propagation and increased uncertainty.To address these problems,we propose a Multi-modal knowledge graph Completion model based on Super-resolution and Detailed Description Generation(MMCSD).Specifically,we leverage a pre-trained residual network to enhance the resolution and improve the quality of the visual modality.Moreover,we design multi-level visual semantic extraction and entity description generation,thereby further extracting entity semantics from structural triples and visual images.Meanwhile,we train a variational multi-modal auto-encoder and utilize a pre-trained multi-modal language model to complement the missing visual features.We conducted experiments on FB15K-237 and DB13K,and the results showed that MMCSD can effectively perform MMKGC and achieve state-of-the-art performance.
基金funded by the National Natural Science Foundation of China(72104224,L2424237,71974107,L2224059,L2124002,and 91646102)the Beijing Natural Science Foundation(9232015)+4 种基金the Beijing Social Science Foundation(24GLC058)the Construction Project of China Knowledge Center for Engineering Sciences and Technology(CKCEST-2023-1-7)the MOE(Ministry of Education in China)Project of Humanities and Social Sciences(16JDGC011)the Tsinghua University Initiative Scientific Research Program(2019Z02CAU)the Tsinghua University Project of Volvo-Supported Green Economy and Sustainable Development(20183910020)。
文摘As large language models(LLMs)continue to demonstrate their potential in handling complex tasks,their value in knowledge-intensive industrial scenarios is becoming increasingly evident.Fault diagnosis,a critical domain in the industrial sector,has long faced the dual challenges of managing vast amounts of experiential knowledge and improving human-machine collaboration efficiency.Traditional fault diagnosis systems,which are primarily based on expert systems,suffer from three major limitations:(1)ineffective organization of fault diagnosis knowledge,(2)lack of adaptability between static knowledge frameworks and dynamic engineering environments,and(3)difficulties in integrating expert knowledge with real-time data streams.These systemic shortcomings restrict the ability of conventional approaches to handle uncertainty.In this study,we proposed an intelligent computer numerical control(CNC)fault diagnosis system,integrating LLMs with knowledge graph(KG).First,we constructed a comprehensive KG that consolidated multi-source data for structured representation.Second,we designed a retrievalaugmented generation(RAG)framework leveraging the KG to support multi-turn interactive fault diagnosis while incorporating real-time engineering data into the decision-making process.Finally,we introduced a learning mechanism to facilitate dynamic knowledge updates.The experimental results demonstrated that our system significantly improved fault diagnosis accuracy,outperforming engineers with two years of professional experience on our constructed benchmark datasets.By integrating LLMs and KG,our framework surpassed the limitations of traditional expert systems rooted in symbolic reasoning,offering a novel approach to addressing the cognitive paradox of unstructured knowledge modeling and dynamic environment adaptation in industrial settings.