Since Google introduced the concept of Knowledge Graphs(KGs)in 2012,their construction technologies have evolved into a comprehensive methodological framework encompassing knowledge acquisition,extraction,representati...Since Google introduced the concept of Knowledge Graphs(KGs)in 2012,their construction technologies have evolved into a comprehensive methodological framework encompassing knowledge acquisition,extraction,representation,modeling,fusion,computation,and storage.Within this framework,knowledge extraction,as the core component,directly determines KG quality.In military domains,traditional manual curation models face efficiency constraints due to data fragmentation,complex knowledge architectures,and confidentiality protocols.Meanwhile,crowdsourced ontology construction approaches from general domains prove non-transferable,while human-crafted ontologies struggle with generalization deficiencies.To address these challenges,this study proposes an OntologyAware LLM Methodology for Military Domain Knowledge Extraction(LLM-KE).This approach leverages the deep semantic comprehension capabilities of Large Language Models(LLMs)to simulate human experts’cognitive processes in crowdsourced ontology construction,enabling automated extraction of military textual knowledge.It concurrently enhances knowledge processing efficiency and improves KG completeness.Empirical analysis demonstrates that this method effectively resolves scalability and dynamic adaptation challenges in military KG construction,establishing a novel technological pathway for advancing military intelligence development.展开更多
The rapid development of information technology and accelerated digitalization have led to an explosive growth of data across various fields.As a key technology for knowledge representation and sharing,knowledge graph...The rapid development of information technology and accelerated digitalization have led to an explosive growth of data across various fields.As a key technology for knowledge representation and sharing,knowledge graphs play a crucial role by constructing structured networks of relationships among entities.However,data sparsity and numerous unexplored implicit relations result in the widespread incompleteness of knowledge graphs.In static knowledge graph completion,most existing methods rely on linear operations or simple interaction mechanisms for triple encoding,making it difficult to fully capture the deep semantic associations between entities and relations.Moreover,many methods focus only on the local information of individual triples,ignoring the rich semantic dependencies embedded in the neighboring nodes of entities within the graph structure,which leads to incomplete embedding representations.To address these challenges,we propose Two-Stage Mixer Embedding(TSMixerE),a static knowledge graph completion method based on entity context.In the unit semantic extraction stage,TSMixerE leveragesmulti-scale circular convolution to capture local features atmultiple granularities,enhancing the flexibility and robustness of feature interactions.A channel attention mechanism amplifies key channel responses to suppress noise and irrelevant information,thereby improving the discriminative power and semantic depth of feature representations.For contextual information fusion,a multi-layer self-attentionmechanism enables deep interactions among contextual cues,effectively integrating local details with global context.Simultaneously,type embeddings clarify the semantic identities and roles of each component,enhancing the model’s sensitivity and fusion capabilities for diverse information sources.Furthermore,TSMixerE constructs contextual unit sequences for entities,fully exploring neighborhood information within the graph structure to model complex semantic dependencies,thus improving the completeness and generalization of embedding representations.展开更多
This paper delves into effective pathways for transforming course ecosystems from resource provision to knowledge service and competency development through university-enterprise collaboration in co-building knowledge...This paper delves into effective pathways for transforming course ecosystems from resource provision to knowledge service and competency development through university-enterprise collaboration in co-building knowledge graphs and intelligent shared courses.This approach enables personalized,learning-driven teaching.Based on knowledge graphs and integrated teacher-machine-student smart teaching scenarios,it not only innovates autonomous learning environments and human-computer interaction models while optimizing teaching experiences for both instructors and students,but also effectively addresses the issues of students’“scattered,superficial,and fragmented learning”.This establishes the foundation for personalized teaching tailored to individual aptitudes.展开更多
This paper proposes a deep learning-based 3D LiDAR perception framework designed for applications such as autonomous robots and vehicles.To address the high dependency on large-scale annotated data—an inherent limita...This paper proposes a deep learning-based 3D LiDAR perception framework designed for applications such as autonomous robots and vehicles.To address the high dependency on large-scale annotated data—an inherent limitation of deep learning models—this study introduces a hybrid perception architecture that incorporates expertdriven LiDAR processing techniques into the deep neural network.Traditional 3DLiDAR processingmethods typically remove ground planes and apply distance-or density-based clustering for object detection.In this work,such expert knowledge is encoded as feature-level inputs and fused with the deep network,therebymitigating the data dependency issue of conventional learning-based approaches.Specifically,the proposedmethod combines two expert algorithms—Patchwork++for ground segmentation and DBSCAN for clustering—with a PointPillars-based LiDAR detection network.We design four hybrid versions of the network depending on the stage and method of integrating expert features into the feature map of the deep model.Among these,Version 4 incorporates a modified neck structure in PointPillars and introduces a new Cluster 2D Pseudo-Map Branch that utilizes cluster-level pseudo-images generated from Patchwork++and DBSCAN.This version achieved a+3.88%improvement mean Average Precision(mAP)compared to the baseline PointPillars.The results demonstrate that embedding expert-based perception logic into deep neural architectures can effectively enhance performance and reduce dependency on extensive training datasets,offering a promising direction for robust 3D LiDAR object detection in real-world scenarios.展开更多
Wetting deformation in earth-rockfill dams is a critical factor influencingdam safety.Although numerous mathematical models have been developed to describe this phenomenon,most of them rely on empirical formulations a...Wetting deformation in earth-rockfill dams is a critical factor influencingdam safety.Although numerous mathematical models have been developed to describe this phenomenon,most of them rely on empirical formulations and lack prior knowledge of model parameters,which is essential for Bayesian parameter inversion to enhance accuracy and reduce uncertainty.This study introduces a datadriven approach to establishing prior knowledge of earth-rockfill dams.Driving factors are utilized to determine the potential range of model parameters,and settlement changes within this range are calculated.The results are iteratively compared with actual monitoring data until the calculated range encompasses the observed data,thereby providing prior knowledge of the model parameters.The proposed method is applied to the right-bank earth-rockfilldam of Danjiangkou.Employing a Gibbs sample size of 30,000,the proposed method effectively calibrates the prior knowledge of the wetting model parameters,achieving a root mean square error(RMSE)of 5.18 mm for the settlement predictions.By comparison,the use of non-informative priors with sample sizes of 30,000 and 50,000 results in significantly larger RMSE values of 11.97 mm and 16.07 mm,respectively.Furthermore,the computational efficiencyof the proposed method is demonstrated by an inversion computation time of 902 s for 30,000 samples,which is notably shorter than the 1026 s and 1558 s required for noninformative priors with 30,000 and 50,000 samples,respectively.These findingsunderscore the superior performance of the proposed approach in terms of both prediction accuracy and computational efficiency.These results demonstrate that the proposed method not only improves the predictive accuracy but also enhances the computational efficiency,enabling optimal parameter identificationwith reduced computational effort.This approach provides a robust and efficientframework for advancing dam safety assessments.展开更多
The viscosity of refining slags plays a critical role in metallurgical processes.However,obtaining accurate viscosity data remains challenging due to the complexities of high-temperature experiments,often relying on e...The viscosity of refining slags plays a critical role in metallurgical processes.However,obtaining accurate viscosity data remains challenging due to the complexities of high-temperature experiments,often relying on empirical models with limited predictive capabilities.This study focuses on the influence of optical basicity on viscosity in CaO-Al_(2)O_(3)-based refining slags,leveraging machine learning to address data scarcity and improve prediction accuracy.An automated framework for algorithm integration,parameter tuning,and evaluation ranking framework(Auto-APE)is employed to develop customized data-driven models for various slag systems,including CaO-Al_(2)O_(3)-SiO_(2),CaO-Al_(2)O_(3)-CaF_(2),CaO-Al_(2)O_(3)-SiO_(2)-MgO,and CaO-Al_(2)O_(3)-SiO_(2)-MgO-CaF_(2).By incorporating optical basicity as a key feature,the models achieve an average validation error of 8.0%to 15.1%,significantly outperforming traditional empirical models.Additionally,symbolic regression is introduced to rapidly construct domain-specific features,such as optical basicity-like descriptors,offering a potential breakthrough in performance prediction for small datasets.This work highlights the critical role of domain-specific knowledge in understanding and predicting viscosity,providing a robust machine learning-based approach for optimizing refining slag properties.展开更多
The reliable operation of power grid secondary equipment is an important guarantee for the safety and stability of the power system.However,various defects could be produced in the secondary equipment during longtermo...The reliable operation of power grid secondary equipment is an important guarantee for the safety and stability of the power system.However,various defects could be produced in the secondary equipment during longtermoperation.The complex relationship between the defect phenomenon andmulti-layer causes and the probabilistic influence of secondary equipment cannot be described through knowledge extraction and fusion technology by existing methods,which limits the real-time and accuracy of defect identification.Therefore,a defect recognition method based on the Bayesian network and knowledge graph fusion is proposed.The defect data of secondary equipment is transformed into the structured knowledge graph through knowledge extraction and fusion technology.The knowledge graph of power grid secondary equipment is mapped to the Bayesian network framework,combined with historical defect data,and introduced Noisy-OR nodes.The prior and conditional probabilities of the Bayesian network are then reasonably assigned to build a model that reflects the probability dependence between defect phenomena and potential causes in power grid secondary equipment.Defect identification of power grid secondary equipment is achieved by defect subgraph search based on the knowledge graph,and defect inference based on the Bayesian network.Practical application cases prove this method’s effectiveness in identifying secondary equipment defect causes,improving identification accuracy and efficiency.展开更多
Knowledge graphs,which combine structured representation with semantic modeling,have shown great potential in knowledge expression,causal inference,and automated reasoning,and are widely used in fields such as intelli...Knowledge graphs,which combine structured representation with semantic modeling,have shown great potential in knowledge expression,causal inference,and automated reasoning,and are widely used in fields such as intelligent question answering,decision support,and fault diagnosis.As high-speed train systems become increasingly intelligent and interconnected,fault patterns have grown more complex and dynamic.Knowledge graphs offer a promising solution to support the structured management and real-time reasoning of fault knowledge,addressing key requirements such as interpretability,accuracy,and continuous evolution in intelligent diagnostic systems.However,conventional knowledge graph construction relies heavily on domain expertise and specialized tools,resulting in high entry barriers for non-experts and limiting their practical application in frontline maintenance scenarios.To address this limitation,this paper proposes a fault knowledge modeling approach for high-speed trains that integrates structured logic diagrams with knowledge graphs.The method employs a seven-layer logic structure—comprising fault name,applicable vehicles,diagnostic logic,signal parameters,verification conditions,fault causes,and emergency measures—to transform unstructured knowledge into a visual and hierarchical representation.A semantic mapping mechanism is then used to automatically convert logic diagrams into machine-interpretable knowledge graphs,enabling dynamic reasoning and knowledge reuse.Furthermore,the proposed method establishes a three-layer architecture—logic structuring,knowledge graph transformation,and dynamic inference—to bridge human-expert logic with machinebased reasoning.Experimental validation and system implementation demonstrate that this approach not only improves knowledge interpretability and inference precision but also significantly enhances modeling efficiency and system maintainability.It provides a scalable and adaptable solution for intelligent operation and maintenance platforms in the high-speed rail domain.展开更多
Objective:To assess the knowledge and attitudes of the general population regarding hepatitis B virus(HBV)infection,their association with sociodemographic factors,and the community’s willingness to pay for hepatitis...Objective:To assess the knowledge and attitudes of the general population regarding hepatitis B virus(HBV)infection,their association with sociodemographic factors,and the community’s willingness to pay for hepatitis B vaccination.Methods:This cross-sectional study was conducted from January to March 2025 in Ho Chi Minh City,Vietnam.A self-administered questionnaire was distributed to 1098 adults who were recruited via nonprobability convenience sampling.Descriptive analysis was conducted to calculate the frequencies and percentages of the categorical variables.A Chi-square test was carried out to explore the association between the independent and categorical dependent variables.Results:A total of 926 complete questionnaires were returned and included in the analysis.Among the participants,524(56.6%)exhibited good knowledge about HBV infection and vaccination,while 651(70.3%)showed positive attitudes toward the issues of interest.The factors associated with knowledge and attitudes were household incomes,health insurance,family history of HBV infection,vaccination intention,reasons for nonvaccination,sources of HBV information,and awareness of immunization programs and vaccination sites.Conclusions:Although most participants had limited knowledge of HBV symptoms and transmission,their attitudes toward vaccination remained positive.Targeted national strategies and intervention initiatives are needed since public awareness and attitudes toward the condition greatly impact intervention success.展开更多
To integrate traditional culture and modern technology,Shandong University’s School of Software has promoted an interdisciplinary teaching project called IYAN&ITAN,the I Ching Knowledge Graph.The project,driven b...To integrate traditional culture and modern technology,Shandong University’s School of Software has promoted an interdisciplinary teaching project called IYAN&ITAN,the I Ching Knowledge Graph.The project,driven by I Ching texts,guides students to practice natural language processing(NLP)and knowledge graph technology in a task-oriented curriculum,based on constructivism,situated learning,and inquiry-based pedagogy,with a progressive and task-oriented teaching model.The platform established enables the retrieval of knowledge,parsing of text,symbolic-numeric analysis,and historical commentary integration,making possible multidimensional,structured representation of I Ching knowledge,and offering an extensible reference for interdisciplinary learning in the context of New Engineering Education.展开更多
Large language models(LLMs)show great potential in educational scenarios but face challenges like hallucination,knowledge gaps,and reasoning discontinuities.This study proposes a dynamic knowledge enhancement framewor...Large language models(LLMs)show great potential in educational scenarios but face challenges like hallucination,knowledge gaps,and reasoning discontinuities.This study proposes a dynamic knowledge enhancement framework.By integrating local knowledge graphs and stepwise prompting mechanisms,it improves LLMs’accuracy and interpretability in solving professional domain problems.The framework has two core modules:an LLM-driven knowledge graph construction system for incremental updates and a unified reasoning engine for generating enhanced prompts.Experiments on 680 educational questions show that the method boosts accuracy by 4.5%and 4.3%for multi-step reasoning and knowledge-dependent questions respectively,and increases reasoning step completeness from 68.2%to 83.7%.It also reduces hallucination problems.Key contributions include the followings:①validation of an effective framework synergizing knowledge graphs with retrieval mechanisms to enhance LLM reliability;②a stepwise prompting strategy enforcing explicit reasoning chain generation,addressing pedagogical requirements for process interpretability;③a lightweight deployment solution for educational systems such as adaptive learning platforms.展开更多
Aiming at the problem of insufficient recognition of implicit variants by existing Chinese sensitive text detection methods,this paper proposes the IPKE-MoE framework,which consists of three parts,namely,a sensitive w...Aiming at the problem of insufficient recognition of implicit variants by existing Chinese sensitive text detection methods,this paper proposes the IPKE-MoE framework,which consists of three parts,namely,a sensitive word variant extraction framework,a sensitive word variant knowledge enhancement layer and a mixture-of-experts(MoE)classification layer.First,sensitive word variants are precisely extracted through dynamic iterative prompt templates and the context-aware capabilities of Large Language Models(LLMs).Next,the extracted variants are used to construct a knowledge enhancement layer for sensitive word variants based on RoCBert models.Specifically,after locating variants via n-gram algorithms,variant types are mapped to embedding vectors and fused with original word vectors.Finally,a mixture-of-experts(MoE)classification layer is designed(sensitive word,sentiment,and semantic experts),which decouples the relationship between sensitiveword existence and text toxicity throughmultiple experts.This framework effectively combines the comprehension ability of Large Language Models(LLMs)with the discriminative ability of smaller models.Our two experiments demonstrate that the sensitive word variant extraction framework based on dynamically iterated prompt templates outperforms other baseline prompt templates.TheRoCBert models incorporating the sensitive word variant knowledge enhancement layer and a mixture-of-experts(MoE)classification layer achieve superior classification performance compared to other baselines.展开更多
With the widespread use of social media,the propagation of health-related rumors has become a significant public health threat.Existing methods for detecting health rumors predominantly rely on external knowledge or p...With the widespread use of social media,the propagation of health-related rumors has become a significant public health threat.Existing methods for detecting health rumors predominantly rely on external knowledge or propagation structures,with only a few recent approaches attempting causal inference;however,these have not yet effectively integrated causal discovery with domain-specific knowledge graphs for detecting health rumors.In this study,we found that the combined use of causal discovery and domain-specific knowledge graphs can effectively identify implicit pseudo-causal logic embedded within texts,holding significant potential for health rumor detection.To this end,we propose CKDG—a dual-graph fusion framework based on causal logic and medical knowledge graphs.CKDG constructs a weighted causal graph to capture the implicit causal relationships in the text and introduces a medical knowledge graph to verify semantic consistency,thereby enhancing the ability to identify the misuse of professional terminology and pseudoscientific claims.In experiments conducted on a dataset comprising 8430 health rumors,CKDG achieved an accuracy of 91.28%and an F1 score of 90.38%,representing improvements of 5.11%and 3.29%over the best baseline,respectively.Our results indicate that the integrated use of causal discovery and domainspecific knowledge graphs offers significant advantages for health rumor detection systems.This method not only improves detection performance but also enhances the transparency and credibility of model decisions by tracing causal chains and sources of knowledge conflicts.We anticipate that this work will provide key technological support for the development of trustworthy health-information filtering systems,thereby improving the reliability of public health information on social media.展开更多
Knowledge distillation has become a standard technique for compressing large language models into efficient student models,but existing methods often struggle to balance prediction accuracy with explanation quality.Re...Knowledge distillation has become a standard technique for compressing large language models into efficient student models,but existing methods often struggle to balance prediction accuracy with explanation quality.Recent approaches such as Distilling Step-by-Step(DSbS)introduce explanation supervision,yet they apply it in a uniform manner that may not fully exploit the different learning dynamics of prediction and explanation.In this work,we propose a task-structured curriculum learning(TSCL)framework that structures training into three sequential phases:(i)prediction-only,to establish stable feature representations;(ii)joint prediction-explanation,to align task outputs with rationale generation;and(iii)explanation-only,to refine the quality of rationales.This design provides a simple but effective modification to DSbS,requiring no architectural changes and adding negligible training cost.We justify the phase scheduling with ablation studies and convergence analysis,showing that an initial prediction-heavy stage followed by a balanced joint phase improves both stability and explanation alignment.Extensive experiments on five datasets(e-SNLI,ANLI,CommonsenseQA,SVAMP,and MedNLI)demonstrate that TSCL consistently outperforms strong baselines,achieving gains of+1.7-2.6 points in accuracy and 0.8-1.2 in ROUGE-L,corresponding to relative error reductions of up to 21%.Beyond lexical metrics,human evaluation and ERASERstyle faithfulness diagnostics confirm that TSCL produces more faithful and informative explanations.Comparative training curves further reveal faster convergence and lower variance across seeds.Efficiency analysis shows less than 3%overhead in wall-clock training time and no additional inference cost,making the approach practical for realworld deployment.This study demonstrates that a simple task-structured curriculum can significantly improve the effectiveness of knowledge distillation.By separating and sequencing objectives,TSCL achieves a better balance between accuracy,stability,and explanation quality.The framework generalizes across domains,including medical NLI,and offers a principled recipe for future applications in multimodal reasoning and reinforcement learning.展开更多
With the large-scale deployment of the Internet ofThings(IoT)devices,their weak securitymechanisms make them prime targets for malware attacks.Attackers often use Domain Generation Algorithm(DGA)to generate random dom...With the large-scale deployment of the Internet ofThings(IoT)devices,their weak securitymechanisms make them prime targets for malware attacks.Attackers often use Domain Generation Algorithm(DGA)to generate random domain names,hiding the real IP of Command and Control(C&C)servers to build botnets.Due to the randomness and dynamics of DGA,traditional methods struggle to detect them accurately,increasing the difficulty of network defense.This paper proposes a lightweight DGA detection model based on knowledge distillation for resource-constrained IoT environments.Specifically,a teacher model combining CharacterBERT,a bidirectional long short-term memory(BiLSTM)network,and attention mechanism(ATT)is constructed:it extracts character-level semantic features viaCharacterBERT,captures sequence dependencieswith the BiLSTM,and integrates theATT for key feature weighting,formingmulti-granularity feature fusion.An improved knowledge distillation approach transfers the teacher model’s learned knowledge to the simplified DistilBERT student model.Experimental results show the teacher model achieves 98.68%detection accuracy.The student modelmaintains slightly improved accuracy while significantly compressing parameters to approximately 38.4%of the teacher model’s scale,greatly reducing computational overhead for IoT deployment.展开更多
Objective To address the dual challenges of long-tail distribution and feature sparsity in traditional Chinese medicine(TCM)syndrome differentiation within real clinical settings,we propose a data-efficient learning f...Objective To address the dual challenges of long-tail distribution and feature sparsity in traditional Chinese medicine(TCM)syndrome differentiation within real clinical settings,we propose a data-efficient learning framework enhanced by knowledge graphs.Methods We developed Agent-GNN,a three-stage decoupled learning framework,and validated it on the Traditional Chinese Medicine Syndrome Diagnosis(TCM-SD)dataset containing 54152 clinical records across 148 syndrome categories.First,we constructed a comprehensive medical knowledge graph encoding the complete TCM reasoning system.Second,we proposed a Functional Patient Profiling(FPP)method that utilizes large language models(LLMs)combined with Graph Retrieval-Augmented Generation(RAG)to extract structured symptom-etiology-pathogenesis subgraphs from medical records.Third,we employed heterogeneous graph neural networks to learn structured combination patterns explicitly.We compared our method against multiple baselines including BERT,ZY-BERT,ZY-BERT+Know,GAT,and GPT-4 Few-shot,using macro-F1 score as the primary evaluation metric.Additionally,ablation experiments were conducted to validate the contribution of each key component to model performance.Results Agent-GNN achieved an overall macro-F1 score of 72.4%,representing an 8.7 percentage points improvement over ZY-BERT+Know(63.7%),the strongest baseline among traditional methods.For long-tail syndromes with fewer than 10 samples,Agent-GNN reached a macro-F1 score of 58.6%,compared with 39.3%for ZY-BERT+Know and 41.2%for GPT-4 Few-shot,representing relative improvements of 49.2%and 42.2%,respectively.Ablation experiments confirmed that the explicit modeling of etiology-pathogenesis nodes contributed 12.4 percentage points to this enhanced long-tail syndrome performance.Conclusion This study proposes Agent-GNN,a knowledge graph-enhanced framework that effectively addresses the long-tail distribution challenge in TCM syndrome differentiation.By explicitly modeling manifestation-mechanism-essence patterns through structured knowledge graphs,our approach achieves superior performance in data-scarce scenarios while providing interpretable reasoning paths for TCM intelligent diagnosis.展开更多
Amidst evolving user behavior driven by the development of the internet,enhancing the operational quality of trade publishing knowledge service platforms has become a significant challenge for publishing institutions....Amidst evolving user behavior driven by the development of the internet,enhancing the operational quality of trade publishing knowledge service platforms has become a significant challenge for publishing institutions.To address this issue,this paper employs a combined approach of theoretical analysis and case study,introducing the SICAS(Sense-Interest-Connection-Action-Share)user consumption behavior analysis model and selecting“CITIC Academy”as the case study subject.It systematically examines and summarizes the platform’s operational practices and specific strategies,aiming to offer strategic insights and practical references for the operational improvement and sustainable,high-quality development of trade publishing knowledge service platforms.展开更多
Conversational recommender systems(CRSs)focus on refining preferences and providing personalized recommendations through natural language interactions and dialogue history.Large language models(LLMs)have shown outstan...Conversational recommender systems(CRSs)focus on refining preferences and providing personalized recommendations through natural language interactions and dialogue history.Large language models(LLMs)have shown outstanding performance across various domains,thereby prompting researchers to investigate their applicability in recommendation systems.However,due to the lack of task-specific knowledge and an inefficient feature extraction process,LLMs still have suboptimal performance in recommendation tasks.Therefore,external knowledge sources,such as knowledge graphs(KGs)and knowledge bases(KBs),are often introduced to address the issue of data sparsity.Compared to KGs,KBs possess higher retrieval efficiency,making them more suitable for scenarios where LLMs serve as recommenders.To this end,we introduce a novel framework integrating LLMs with KBs for enhanced retrieval generation,namely LLMKB.LLMKB initially leverages structured knowledge to create mapping dictionaries,extracting entity-relation information from heterogeneous knowledge to construct KBs.Then,LLMKB achieves the embedding calibration between user information representations and documents in KBs through retrieval model fine-tuning.Finally,LLMKB employs retrievalaugmented generation to produce recommendations based on fused text inputs,followed by post-processing.Experiment results on two public CRS datasets demonstrate the effectiveness of our framework.Our code is publicly available at the link:https://anonymous.4open.science/r/LLMKB-6FD0.展开更多
Temporal knowledge graph completion(TKGC),which merges temporal information into traditional static knowledge graph completion(SKGC),has garnered increasing attention recently.Among numerous emerging approaches,transl...Temporal knowledge graph completion(TKGC),which merges temporal information into traditional static knowledge graph completion(SKGC),has garnered increasing attention recently.Among numerous emerging approaches,translation-based embedding models constitute a prominent approach in TKGC research.However,existing translation-based methods typically incorporate timestamps into entities or relations,rather than utilizing them independently.This practice fails to fully exploit the rich semantics inherent in temporal information,thereby weakening the expressive capability of models.To address this limitation,we propose embedding timestamps,like entities and relations,in one or more dedicated semantic spaces.After projecting all embeddings into a shared space,we use the relation-timestamp pair instead of the conventional relation embedding as the translation vector between head and tail entities.Our method elevates timestamps to the same representational significance as entities and relations.Based on this strategy,we introduce two novel translation-based embedding models:TE-TransR and TE-TransT.With the independent representation of timestamps,our method not only enhances capabilities in link prediction but also facilitates a relatively underexplored task,namely time prediction.To further bolster the precision and reliability of time prediction,we introduce a granular,time unit-based timestamp setting and a relation-specific evaluation protocol.Extensive experiments demonstrate that our models achieve strong performance on link prediction benchmarks,with TE-TransR outperforming existing baselines in the time prediction task.展开更多
Knowledge graphs often suffer from sparsity and incompleteness.Knowledge graph reasoning is an effective way to address these issues.Unlike static knowledge graph reasoning,which is invariant over time,dynamic knowled...Knowledge graphs often suffer from sparsity and incompleteness.Knowledge graph reasoning is an effective way to address these issues.Unlike static knowledge graph reasoning,which is invariant over time,dynamic knowledge graph reasoning is more challenging due to its temporal nature.In essence,within each time step in a dynamic knowledge graph,there exists structural dependencies among entities and relations,whereas between adjacent time steps,there exists temporal continuity.Based on these structural and temporal characteristics,we propose a model named“DKGR-DR”to learn distributed representations of entities and relations by combining recurrent neural networks and graph neural networks to capture structural dependencies and temporal continuity in DKGs.In addition,we construct a static attribute graph to represent entities’inherent properties.DKGR-DR is capable of modeling both dynamic and static aspects of entities,enabling effective entity prediction and relation prediction.We conduct experiments on ICEWS05-15,ICEWS18,and ICEWS14 to demonstrate that DKGR-DR achieves competitive performance.展开更多
文摘Since Google introduced the concept of Knowledge Graphs(KGs)in 2012,their construction technologies have evolved into a comprehensive methodological framework encompassing knowledge acquisition,extraction,representation,modeling,fusion,computation,and storage.Within this framework,knowledge extraction,as the core component,directly determines KG quality.In military domains,traditional manual curation models face efficiency constraints due to data fragmentation,complex knowledge architectures,and confidentiality protocols.Meanwhile,crowdsourced ontology construction approaches from general domains prove non-transferable,while human-crafted ontologies struggle with generalization deficiencies.To address these challenges,this study proposes an OntologyAware LLM Methodology for Military Domain Knowledge Extraction(LLM-KE).This approach leverages the deep semantic comprehension capabilities of Large Language Models(LLMs)to simulate human experts’cognitive processes in crowdsourced ontology construction,enabling automated extraction of military textual knowledge.It concurrently enhances knowledge processing efficiency and improves KG completeness.Empirical analysis demonstrates that this method effectively resolves scalability and dynamic adaptation challenges in military KG construction,establishing a novel technological pathway for advancing military intelligence development.
基金supported by the National Natural Science Foundation of China(No.62267005)the Chinese Guangxi Natural Science Foundation(No.2023GXNSFAA026493)+1 种基金Guangxi Collaborative Innovation Center ofMulti-Source Information Integration and Intelligent ProcessingGuangxi Academy of Artificial Intelligence.
文摘The rapid development of information technology and accelerated digitalization have led to an explosive growth of data across various fields.As a key technology for knowledge representation and sharing,knowledge graphs play a crucial role by constructing structured networks of relationships among entities.However,data sparsity and numerous unexplored implicit relations result in the widespread incompleteness of knowledge graphs.In static knowledge graph completion,most existing methods rely on linear operations or simple interaction mechanisms for triple encoding,making it difficult to fully capture the deep semantic associations between entities and relations.Moreover,many methods focus only on the local information of individual triples,ignoring the rich semantic dependencies embedded in the neighboring nodes of entities within the graph structure,which leads to incomplete embedding representations.To address these challenges,we propose Two-Stage Mixer Embedding(TSMixerE),a static knowledge graph completion method based on entity context.In the unit semantic extraction stage,TSMixerE leveragesmulti-scale circular convolution to capture local features atmultiple granularities,enhancing the flexibility and robustness of feature interactions.A channel attention mechanism amplifies key channel responses to suppress noise and irrelevant information,thereby improving the discriminative power and semantic depth of feature representations.For contextual information fusion,a multi-layer self-attentionmechanism enables deep interactions among contextual cues,effectively integrating local details with global context.Simultaneously,type embeddings clarify the semantic identities and roles of each component,enhancing the model’s sensitivity and fusion capabilities for diverse information sources.Furthermore,TSMixerE constructs contextual unit sequences for entities,fully exploring neighborhood information within the graph structure to model complex semantic dependencies,thus improving the completeness and generalization of embedding representations.
基金supported by Harbin Institute of Technology High-level Teaching Achievement Award(National Level)Cultivation Project(256709).
文摘This paper delves into effective pathways for transforming course ecosystems from resource provision to knowledge service and competency development through university-enterprise collaboration in co-building knowledge graphs and intelligent shared courses.This approach enables personalized,learning-driven teaching.Based on knowledge graphs and integrated teacher-machine-student smart teaching scenarios,it not only innovates autonomous learning environments and human-computer interaction models while optimizing teaching experiences for both instructors and students,but also effectively addresses the issues of students’“scattered,superficial,and fragmented learning”.This establishes the foundation for personalized teaching tailored to individual aptitudes.
基金supported by Basic Science Research Program through the National Research Foundation of Korea(NRF)funded by the Ministry of Education(RS-2023-00245084)by Korea Institute for Advancement of Technology(KIAT)grant funded by the Korea Government(MOTIE)(RS-2024-00415938,HRD Program for Industrial Innovation)and Soonchunhyang University.
文摘This paper proposes a deep learning-based 3D LiDAR perception framework designed for applications such as autonomous robots and vehicles.To address the high dependency on large-scale annotated data—an inherent limitation of deep learning models—this study introduces a hybrid perception architecture that incorporates expertdriven LiDAR processing techniques into the deep neural network.Traditional 3DLiDAR processingmethods typically remove ground planes and apply distance-or density-based clustering for object detection.In this work,such expert knowledge is encoded as feature-level inputs and fused with the deep network,therebymitigating the data dependency issue of conventional learning-based approaches.Specifically,the proposedmethod combines two expert algorithms—Patchwork++for ground segmentation and DBSCAN for clustering—with a PointPillars-based LiDAR detection network.We design four hybrid versions of the network depending on the stage and method of integrating expert features into the feature map of the deep model.Among these,Version 4 incorporates a modified neck structure in PointPillars and introduces a new Cluster 2D Pseudo-Map Branch that utilizes cluster-level pseudo-images generated from Patchwork++and DBSCAN.This version achieved a+3.88%improvement mean Average Precision(mAP)compared to the baseline PointPillars.The results demonstrate that embedding expert-based perception logic into deep neural architectures can effectively enhance performance and reduce dependency on extensive training datasets,offering a promising direction for robust 3D LiDAR object detection in real-world scenarios.
基金supported by the National Key R&D Program of China(Grant No.2023YFC3209504)Natural Science Foundation of Wuhan(Grant No.2024040801020271)the Fundamental Research Funds for Central Public Welfare Research Institutes(Grant No.CKSF2025718/YT).
文摘Wetting deformation in earth-rockfill dams is a critical factor influencingdam safety.Although numerous mathematical models have been developed to describe this phenomenon,most of them rely on empirical formulations and lack prior knowledge of model parameters,which is essential for Bayesian parameter inversion to enhance accuracy and reduce uncertainty.This study introduces a datadriven approach to establishing prior knowledge of earth-rockfill dams.Driving factors are utilized to determine the potential range of model parameters,and settlement changes within this range are calculated.The results are iteratively compared with actual monitoring data until the calculated range encompasses the observed data,thereby providing prior knowledge of the model parameters.The proposed method is applied to the right-bank earth-rockfilldam of Danjiangkou.Employing a Gibbs sample size of 30,000,the proposed method effectively calibrates the prior knowledge of the wetting model parameters,achieving a root mean square error(RMSE)of 5.18 mm for the settlement predictions.By comparison,the use of non-informative priors with sample sizes of 30,000 and 50,000 results in significantly larger RMSE values of 11.97 mm and 16.07 mm,respectively.Furthermore,the computational efficiencyof the proposed method is demonstrated by an inversion computation time of 902 s for 30,000 samples,which is notably shorter than the 1026 s and 1558 s required for noninformative priors with 30,000 and 50,000 samples,respectively.These findingsunderscore the superior performance of the proposed approach in terms of both prediction accuracy and computational efficiency.These results demonstrate that the proposed method not only improves the predictive accuracy but also enhances the computational efficiency,enabling optimal parameter identificationwith reduced computational effort.This approach provides a robust and efficientframework for advancing dam safety assessments.
基金supported by the National Key Research and Development Program of China(No.2023YFB3712401),the National Natural Science Foundation of China(No.52274301)the Aeronautical Science Foundation of China(No.2023Z0530S6005)the Ningbo Yongjiang Talent-Introduction Programme(No.2022A-023-C).
文摘The viscosity of refining slags plays a critical role in metallurgical processes.However,obtaining accurate viscosity data remains challenging due to the complexities of high-temperature experiments,often relying on empirical models with limited predictive capabilities.This study focuses on the influence of optical basicity on viscosity in CaO-Al_(2)O_(3)-based refining slags,leveraging machine learning to address data scarcity and improve prediction accuracy.An automated framework for algorithm integration,parameter tuning,and evaluation ranking framework(Auto-APE)is employed to develop customized data-driven models for various slag systems,including CaO-Al_(2)O_(3)-SiO_(2),CaO-Al_(2)O_(3)-CaF_(2),CaO-Al_(2)O_(3)-SiO_(2)-MgO,and CaO-Al_(2)O_(3)-SiO_(2)-MgO-CaF_(2).By incorporating optical basicity as a key feature,the models achieve an average validation error of 8.0%to 15.1%,significantly outperforming traditional empirical models.Additionally,symbolic regression is introduced to rapidly construct domain-specific features,such as optical basicity-like descriptors,offering a potential breakthrough in performance prediction for small datasets.This work highlights the critical role of domain-specific knowledge in understanding and predicting viscosity,providing a robust machine learning-based approach for optimizing refining slag properties.
基金supported by the State Grid Southwest Branch Project“Research on Defect Diagnosis and Early Warning Technology of Relay Protection and Safety Automation Devices Based on Multi-Source Heterogeneous Defect Data”.
文摘The reliable operation of power grid secondary equipment is an important guarantee for the safety and stability of the power system.However,various defects could be produced in the secondary equipment during longtermoperation.The complex relationship between the defect phenomenon andmulti-layer causes and the probabilistic influence of secondary equipment cannot be described through knowledge extraction and fusion technology by existing methods,which limits the real-time and accuracy of defect identification.Therefore,a defect recognition method based on the Bayesian network and knowledge graph fusion is proposed.The defect data of secondary equipment is transformed into the structured knowledge graph through knowledge extraction and fusion technology.The knowledge graph of power grid secondary equipment is mapped to the Bayesian network framework,combined with historical defect data,and introduced Noisy-OR nodes.The prior and conditional probabilities of the Bayesian network are then reasonably assigned to build a model that reflects the probability dependence between defect phenomena and potential causes in power grid secondary equipment.Defect identification of power grid secondary equipment is achieved by defect subgraph search based on the knowledge graph,and defect inference based on the Bayesian network.Practical application cases prove this method’s effectiveness in identifying secondary equipment defect causes,improving identification accuracy and efficiency.
基金support from the Scientific Funding for the Center of National Railway Intelligent Transportation System Engineering and Technology,China Academy of Railway Sciences Corporation Limited(Grant No.2023YJ354)。
文摘Knowledge graphs,which combine structured representation with semantic modeling,have shown great potential in knowledge expression,causal inference,and automated reasoning,and are widely used in fields such as intelligent question answering,decision support,and fault diagnosis.As high-speed train systems become increasingly intelligent and interconnected,fault patterns have grown more complex and dynamic.Knowledge graphs offer a promising solution to support the structured management and real-time reasoning of fault knowledge,addressing key requirements such as interpretability,accuracy,and continuous evolution in intelligent diagnostic systems.However,conventional knowledge graph construction relies heavily on domain expertise and specialized tools,resulting in high entry barriers for non-experts and limiting their practical application in frontline maintenance scenarios.To address this limitation,this paper proposes a fault knowledge modeling approach for high-speed trains that integrates structured logic diagrams with knowledge graphs.The method employs a seven-layer logic structure—comprising fault name,applicable vehicles,diagnostic logic,signal parameters,verification conditions,fault causes,and emergency measures—to transform unstructured knowledge into a visual and hierarchical representation.A semantic mapping mechanism is then used to automatically convert logic diagrams into machine-interpretable knowledge graphs,enabling dynamic reasoning and knowledge reuse.Furthermore,the proposed method establishes a three-layer architecture—logic structuring,knowledge graph transformation,and dynamic inference—to bridge human-expert logic with machinebased reasoning.Experimental validation and system implementation demonstrate that this approach not only improves knowledge interpretability and inference precision but also significantly enhances modeling efficiency and system maintainability.It provides a scalable and adaptable solution for intelligent operation and maintenance platforms in the high-speed rail domain.
基金funded by Vietnam National University Ho Chi Minh City(VNU-HCM)under grant number 36-2025-44-02.
文摘Objective:To assess the knowledge and attitudes of the general population regarding hepatitis B virus(HBV)infection,their association with sociodemographic factors,and the community’s willingness to pay for hepatitis B vaccination.Methods:This cross-sectional study was conducted from January to March 2025 in Ho Chi Minh City,Vietnam.A self-administered questionnaire was distributed to 1098 adults who were recruited via nonprobability convenience sampling.Descriptive analysis was conducted to calculate the frequencies and percentages of the categorical variables.A Chi-square test was carried out to explore the association between the independent and categorical dependent variables.Results:A total of 926 complete questionnaires were returned and included in the analysis.Among the participants,524(56.6%)exhibited good knowledge about HBV infection and vaccination,while 651(70.3%)showed positive attitudes toward the issues of interest.The factors associated with knowledge and attitudes were household incomes,health insurance,family history of HBV infection,vaccination intention,reasons for nonvaccination,sources of HBV information,and awareness of immunization programs and vaccination sites.Conclusions:Although most participants had limited knowledge of HBV symptoms and transmission,their attitudes toward vaccination remained positive.Targeted national strategies and intervention initiatives are needed since public awareness and attitudes toward the condition greatly impact intervention success.
基金support provided by the Shandong University Education and Teaching Reform Research Project(2024Y232)the“New 20 Regulations for Universities”funding program of Jinan(202228089).
文摘To integrate traditional culture and modern technology,Shandong University’s School of Software has promoted an interdisciplinary teaching project called IYAN&ITAN,the I Ching Knowledge Graph.The project,driven by I Ching texts,guides students to practice natural language processing(NLP)and knowledge graph technology in a task-oriented curriculum,based on constructivism,situated learning,and inquiry-based pedagogy,with a progressive and task-oriented teaching model.The platform established enables the retrieval of knowledge,parsing of text,symbolic-numeric analysis,and historical commentary integration,making possible multidimensional,structured representation of I Ching knowledge,and offering an extensible reference for interdisciplinary learning in the context of New Engineering Education.
基金supported in part by the China-Singapore International Joint Research Institute(CSIJRI)under Grant No.206-A023001the Undergraduate Teaching Reform Project of Shandong University under Grant Nos.2023Y235 and 2025Y99.
文摘Large language models(LLMs)show great potential in educational scenarios but face challenges like hallucination,knowledge gaps,and reasoning discontinuities.This study proposes a dynamic knowledge enhancement framework.By integrating local knowledge graphs and stepwise prompting mechanisms,it improves LLMs’accuracy and interpretability in solving professional domain problems.The framework has two core modules:an LLM-driven knowledge graph construction system for incremental updates and a unified reasoning engine for generating enhanced prompts.Experiments on 680 educational questions show that the method boosts accuracy by 4.5%and 4.3%for multi-step reasoning and knowledge-dependent questions respectively,and increases reasoning step completeness from 68.2%to 83.7%.It also reduces hallucination problems.Key contributions include the followings:①validation of an effective framework synergizing knowledge graphs with retrieval mechanisms to enhance LLM reliability;②a stepwise prompting strategy enforcing explicit reasoning chain generation,addressing pedagogical requirements for process interpretability;③a lightweight deployment solution for educational systems such as adaptive learning platforms.
基金funded by the National Natural Science Foundation of China(Grant No.62441212)the Major Project of the Natural Science Foundation of Inner Mongolia(Grant No.2025ZD008).
文摘Aiming at the problem of insufficient recognition of implicit variants by existing Chinese sensitive text detection methods,this paper proposes the IPKE-MoE framework,which consists of three parts,namely,a sensitive word variant extraction framework,a sensitive word variant knowledge enhancement layer and a mixture-of-experts(MoE)classification layer.First,sensitive word variants are precisely extracted through dynamic iterative prompt templates and the context-aware capabilities of Large Language Models(LLMs).Next,the extracted variants are used to construct a knowledge enhancement layer for sensitive word variants based on RoCBert models.Specifically,after locating variants via n-gram algorithms,variant types are mapped to embedding vectors and fused with original word vectors.Finally,a mixture-of-experts(MoE)classification layer is designed(sensitive word,sentiment,and semantic experts),which decouples the relationship between sensitiveword existence and text toxicity throughmultiple experts.This framework effectively combines the comprehension ability of Large Language Models(LLMs)with the discriminative ability of smaller models.Our two experiments demonstrate that the sensitive word variant extraction framework based on dynamically iterated prompt templates outperforms other baseline prompt templates.TheRoCBert models incorporating the sensitive word variant knowledge enhancement layer and a mixture-of-experts(MoE)classification layer achieve superior classification performance compared to other baselines.
基金funded by the Hunan Provincial Natural Science Foundation of China(Grant No.2025JJ70105)the Hunan Provincial College Students’Innovation and Entrepreneurship Training Program(Project No.S202411342056)The article processing charge(APC)was funded by the Project No.2025JJ70105.
文摘With the widespread use of social media,the propagation of health-related rumors has become a significant public health threat.Existing methods for detecting health rumors predominantly rely on external knowledge or propagation structures,with only a few recent approaches attempting causal inference;however,these have not yet effectively integrated causal discovery with domain-specific knowledge graphs for detecting health rumors.In this study,we found that the combined use of causal discovery and domain-specific knowledge graphs can effectively identify implicit pseudo-causal logic embedded within texts,holding significant potential for health rumor detection.To this end,we propose CKDG—a dual-graph fusion framework based on causal logic and medical knowledge graphs.CKDG constructs a weighted causal graph to capture the implicit causal relationships in the text and introduces a medical knowledge graph to verify semantic consistency,thereby enhancing the ability to identify the misuse of professional terminology and pseudoscientific claims.In experiments conducted on a dataset comprising 8430 health rumors,CKDG achieved an accuracy of 91.28%and an F1 score of 90.38%,representing improvements of 5.11%and 3.29%over the best baseline,respectively.Our results indicate that the integrated use of causal discovery and domainspecific knowledge graphs offers significant advantages for health rumor detection systems.This method not only improves detection performance but also enhances the transparency and credibility of model decisions by tracing causal chains and sources of knowledge conflicts.We anticipate that this work will provide key technological support for the development of trustworthy health-information filtering systems,thereby improving the reliability of public health information on social media.
文摘Knowledge distillation has become a standard technique for compressing large language models into efficient student models,but existing methods often struggle to balance prediction accuracy with explanation quality.Recent approaches such as Distilling Step-by-Step(DSbS)introduce explanation supervision,yet they apply it in a uniform manner that may not fully exploit the different learning dynamics of prediction and explanation.In this work,we propose a task-structured curriculum learning(TSCL)framework that structures training into three sequential phases:(i)prediction-only,to establish stable feature representations;(ii)joint prediction-explanation,to align task outputs with rationale generation;and(iii)explanation-only,to refine the quality of rationales.This design provides a simple but effective modification to DSbS,requiring no architectural changes and adding negligible training cost.We justify the phase scheduling with ablation studies and convergence analysis,showing that an initial prediction-heavy stage followed by a balanced joint phase improves both stability and explanation alignment.Extensive experiments on five datasets(e-SNLI,ANLI,CommonsenseQA,SVAMP,and MedNLI)demonstrate that TSCL consistently outperforms strong baselines,achieving gains of+1.7-2.6 points in accuracy and 0.8-1.2 in ROUGE-L,corresponding to relative error reductions of up to 21%.Beyond lexical metrics,human evaluation and ERASERstyle faithfulness diagnostics confirm that TSCL produces more faithful and informative explanations.Comparative training curves further reveal faster convergence and lower variance across seeds.Efficiency analysis shows less than 3%overhead in wall-clock training time and no additional inference cost,making the approach practical for realworld deployment.This study demonstrates that a simple task-structured curriculum can significantly improve the effectiveness of knowledge distillation.By separating and sequencing objectives,TSCL achieves a better balance between accuracy,stability,and explanation quality.The framework generalizes across domains,including medical NLI,and offers a principled recipe for future applications in multimodal reasoning and reinforcement learning.
基金supported by the following projects:National Natural Science Foundation of China(62461041)Natural Science Foundation of Jiangxi Province China(20242BAB25068).
文摘With the large-scale deployment of the Internet ofThings(IoT)devices,their weak securitymechanisms make them prime targets for malware attacks.Attackers often use Domain Generation Algorithm(DGA)to generate random domain names,hiding the real IP of Command and Control(C&C)servers to build botnets.Due to the randomness and dynamics of DGA,traditional methods struggle to detect them accurately,increasing the difficulty of network defense.This paper proposes a lightweight DGA detection model based on knowledge distillation for resource-constrained IoT environments.Specifically,a teacher model combining CharacterBERT,a bidirectional long short-term memory(BiLSTM)network,and attention mechanism(ATT)is constructed:it extracts character-level semantic features viaCharacterBERT,captures sequence dependencieswith the BiLSTM,and integrates theATT for key feature weighting,formingmulti-granularity feature fusion.An improved knowledge distillation approach transfers the teacher model’s learned knowledge to the simplified DistilBERT student model.Experimental results show the teacher model achieves 98.68%detection accuracy.The student modelmaintains slightly improved accuracy while significantly compressing parameters to approximately 38.4%of the teacher model’s scale,greatly reducing computational overhead for IoT deployment.
基金Sichuan TCM Culture Coordinated Development Research Center Project(2023XT131)National Key Science and Technology Project of China(2023ZD0509405)National Natural Science Foundation of China(82174236).
文摘Objective To address the dual challenges of long-tail distribution and feature sparsity in traditional Chinese medicine(TCM)syndrome differentiation within real clinical settings,we propose a data-efficient learning framework enhanced by knowledge graphs.Methods We developed Agent-GNN,a three-stage decoupled learning framework,and validated it on the Traditional Chinese Medicine Syndrome Diagnosis(TCM-SD)dataset containing 54152 clinical records across 148 syndrome categories.First,we constructed a comprehensive medical knowledge graph encoding the complete TCM reasoning system.Second,we proposed a Functional Patient Profiling(FPP)method that utilizes large language models(LLMs)combined with Graph Retrieval-Augmented Generation(RAG)to extract structured symptom-etiology-pathogenesis subgraphs from medical records.Third,we employed heterogeneous graph neural networks to learn structured combination patterns explicitly.We compared our method against multiple baselines including BERT,ZY-BERT,ZY-BERT+Know,GAT,and GPT-4 Few-shot,using macro-F1 score as the primary evaluation metric.Additionally,ablation experiments were conducted to validate the contribution of each key component to model performance.Results Agent-GNN achieved an overall macro-F1 score of 72.4%,representing an 8.7 percentage points improvement over ZY-BERT+Know(63.7%),the strongest baseline among traditional methods.For long-tail syndromes with fewer than 10 samples,Agent-GNN reached a macro-F1 score of 58.6%,compared with 39.3%for ZY-BERT+Know and 41.2%for GPT-4 Few-shot,representing relative improvements of 49.2%and 42.2%,respectively.Ablation experiments confirmed that the explicit modeling of etiology-pathogenesis nodes contributed 12.4 percentage points to this enhanced long-tail syndrome performance.Conclusion This study proposes Agent-GNN,a knowledge graph-enhanced framework that effectively addresses the long-tail distribution challenge in TCM syndrome differentiation.By explicitly modeling manifestation-mechanism-essence patterns through structured knowledge graphs,our approach achieves superior performance in data-scarce scenarios while providing interpretable reasoning paths for TCM intelligent diagnosis.
文摘Amidst evolving user behavior driven by the development of the internet,enhancing the operational quality of trade publishing knowledge service platforms has become a significant challenge for publishing institutions.To address this issue,this paper employs a combined approach of theoretical analysis and case study,introducing the SICAS(Sense-Interest-Connection-Action-Share)user consumption behavior analysis model and selecting“CITIC Academy”as the case study subject.It systematically examines and summarizes the platform’s operational practices and specific strategies,aiming to offer strategic insights and practical references for the operational improvement and sustainable,high-quality development of trade publishing knowledge service platforms.
文摘Conversational recommender systems(CRSs)focus on refining preferences and providing personalized recommendations through natural language interactions and dialogue history.Large language models(LLMs)have shown outstanding performance across various domains,thereby prompting researchers to investigate their applicability in recommendation systems.However,due to the lack of task-specific knowledge and an inefficient feature extraction process,LLMs still have suboptimal performance in recommendation tasks.Therefore,external knowledge sources,such as knowledge graphs(KGs)and knowledge bases(KBs),are often introduced to address the issue of data sparsity.Compared to KGs,KBs possess higher retrieval efficiency,making them more suitable for scenarios where LLMs serve as recommenders.To this end,we introduce a novel framework integrating LLMs with KBs for enhanced retrieval generation,namely LLMKB.LLMKB initially leverages structured knowledge to create mapping dictionaries,extracting entity-relation information from heterogeneous knowledge to construct KBs.Then,LLMKB achieves the embedding calibration between user information representations and documents in KBs through retrieval model fine-tuning.Finally,LLMKB employs retrievalaugmented generation to produce recommendations based on fused text inputs,followed by post-processing.Experiment results on two public CRS datasets demonstrate the effectiveness of our framework.Our code is publicly available at the link:https://anonymous.4open.science/r/LLMKB-6FD0.
基金supported by the National Natural Science Foundation of China under Grant No.72293575.
文摘Temporal knowledge graph completion(TKGC),which merges temporal information into traditional static knowledge graph completion(SKGC),has garnered increasing attention recently.Among numerous emerging approaches,translation-based embedding models constitute a prominent approach in TKGC research.However,existing translation-based methods typically incorporate timestamps into entities or relations,rather than utilizing them independently.This practice fails to fully exploit the rich semantics inherent in temporal information,thereby weakening the expressive capability of models.To address this limitation,we propose embedding timestamps,like entities and relations,in one or more dedicated semantic spaces.After projecting all embeddings into a shared space,we use the relation-timestamp pair instead of the conventional relation embedding as the translation vector between head and tail entities.Our method elevates timestamps to the same representational significance as entities and relations.Based on this strategy,we introduce two novel translation-based embedding models:TE-TransR and TE-TransT.With the independent representation of timestamps,our method not only enhances capabilities in link prediction but also facilitates a relatively underexplored task,namely time prediction.To further bolster the precision and reliability of time prediction,we introduce a granular,time unit-based timestamp setting and a relation-specific evaluation protocol.Extensive experiments demonstrate that our models achieve strong performance on link prediction benchmarks,with TE-TransR outperforming existing baselines in the time prediction task.
基金supported byNationalNatural Science Foundation of China(GrantNos.62071098,U24B20128)Sichuan Science and Technology Program(Grant No.2022YFG0319).
文摘Knowledge graphs often suffer from sparsity and incompleteness.Knowledge graph reasoning is an effective way to address these issues.Unlike static knowledge graph reasoning,which is invariant over time,dynamic knowledge graph reasoning is more challenging due to its temporal nature.In essence,within each time step in a dynamic knowledge graph,there exists structural dependencies among entities and relations,whereas between adjacent time steps,there exists temporal continuity.Based on these structural and temporal characteristics,we propose a model named“DKGR-DR”to learn distributed representations of entities and relations by combining recurrent neural networks and graph neural networks to capture structural dependencies and temporal continuity in DKGs.In addition,we construct a static attribute graph to represent entities’inherent properties.DKGR-DR is capable of modeling both dynamic and static aspects of entities,enabling effective entity prediction and relation prediction.We conduct experiments on ICEWS05-15,ICEWS18,and ICEWS14 to demonstrate that DKGR-DR achieves competitive performance.