This review presents a comprehensive and forward-looking analysis of how Large Language Models(LLMs)are transforming knowledge discovery in the rational design of advancedmicro/nano electrocatalyst materials.Electroca...This review presents a comprehensive and forward-looking analysis of how Large Language Models(LLMs)are transforming knowledge discovery in the rational design of advancedmicro/nano electrocatalyst materials.Electrocatalysis is central to sustainable energy and environmental technologies,but traditional catalyst discovery is often hindered by high complexity,fragmented knowledge,and inefficiencies.LLMs,particularly those based on Transformer architectures,offer unprecedented capabilities in extracting,synthesizing,and generating scientific knowledge from vast unstructured textual corpora.This work provides the first structured synthesis of how LLMs have been leveraged across various electrocatalysis tasks,including automated information extraction from literature,text-based property prediction,hypothesis generation,synthesis planning,and knowledge graph construction.We comparatively analyze leading LLMs and domain-specific frameworks(e.g.,CatBERTa,CataLM,CatGPT)in terms of methodology,application scope,performance metrics,and limitations.Through curated case studies across key electrocatalytic reactions—HER,OER,ORR,and CO_(2)RR—we highlight emerging trends such as the growing use of embedding-based prediction,retrieval-augmented generation,and fine-tuned scientific LLMs.The review also identifies persistent challenges,including data heterogeneity,hallucination risks,lack of standard benchmarks,and limited multimodal integration.Importantly,we articulate future research directions,such as the development of multimodal and physics-informedMatSci-LLMs,enhanced interpretability tools,and the integration of LLMswith selfdriving laboratories for autonomous discovery.By consolidating fragmented advances and outlining a unified research roadmap,this review provides valuable guidance for both materials scientists and AI practitioners seeking to accelerate catalyst innovation through large language model technologies.展开更多
Traditional Chinese medicine(TCM)serves as a treasure trove of ancient knowledge,holding a crucial position in the medical field.However,the exploration of TCM's extensive information has been hindered by challeng...Traditional Chinese medicine(TCM)serves as a treasure trove of ancient knowledge,holding a crucial position in the medical field.However,the exploration of TCM's extensive information has been hindered by challenges related to data standardization,completeness,and accuracy,primarily due to the decen-tralized distribution of TCM resources.To address these issues,we developed a platform for TCM knowledge discovery(TCMKD,https://cbcb.cdutcm.edu.cn/TCMKD/).Seven types of data,including syndromes,formulas,Chinese patent drugs(CPDs),Chinese medicinal materials(CMMs),ingredients,targets,and diseases,were manually proofread and consolidated within TCMKD.To strengthen the integration of TCM with modern medicine,TCMKD employs analytical methods such as TCM data mining,enrichment analysis,and network localization and separation.These tools help elucidate the molecular-level commonalities between TCM and contemporary scientific insights.In addition to its analytical capabilities,a quick question and answer(Q&A)system is also embedded within TCMKD to query the database efficiently,thereby improving the interactivity of the platform.The platform also provides a TCM text annotation tool,offering a simple and efficient method for TCM text mining.Overall,TCMKD not only has the potential to become a pivotal repository for TCM,delving into the pharmaco-logical foundations of TCM treatments,but its flexible embedded tools and algorithms can also be applied to the study of other traditional medical systems,extending beyond just TCM.展开更多
As a new data management paradigm,knowledge graphs can integrate multiple data sources and achieve quick responses,reasoning and better predictions in drug discovery.Characterized by powerful contagion and a high rate...As a new data management paradigm,knowledge graphs can integrate multiple data sources and achieve quick responses,reasoning and better predictions in drug discovery.Characterized by powerful contagion and a high rate of morbidity and mortality,porcine reproductive and respiratory syndrome(PRRS)is a common infectious disease in the global swine industry that causes economically great losses.Traditional Chinese medicine(TCM)has advantages in low adverse effects and a relatively affordable cost of application,and TCM is therefore conceived as a possibility to treat PRRS under the current circumstance that there is a lack of safe and effective approaches.Here,we constructed a knowledge graph containing common biomedical data from humans and Sus Scrofa as well as information from thousands of TCMs.Subsequently,we validated the effectiveness of the Sus Scrofa knowledge graph by the t-SNE algorithm and selected the optimal model(i.e.,transR)from six typical models,namely,transE,transR,DistMult,ComplEx,RESCAL and RotatE,according to five indicators,namely,MRR,MR,HITS@1,HITS@3 and HITS@10.Based on embedding vectors trained by the optimal model,anti-PRRSV TCMs were predicted by two paths,namely,VHC-Herb and VHPC-Herb,and potential anti-PRRSVTCMs were identified by retrieving the HERB database according to the phar-macological properties corresponding to symptoms of PRRS.Ultimately,Dan Shen's(Salvia miltiorrhiza Bunge)capacity to resist PRRSV infection was validated by a cell experiment in which the inhibition rate of PRRSV exceeded90%when the concentrations of Dan Shen extract were 0.004,0.008,0.016 and 0.032 mg/mL.In summary,this is the first report on the Sus Scrofa knowledge graph including TCM information,and our study reflects the important application values of deep learning on graphs in the swine industry as well as providing accessible TCM resources for PRRS.展开更多
Since Google introduced the concept of Knowledge Graphs(KGs)in 2012,their construction technologies have evolved into a comprehensive methodological framework encompassing knowledge acquisition,extraction,representati...Since Google introduced the concept of Knowledge Graphs(KGs)in 2012,their construction technologies have evolved into a comprehensive methodological framework encompassing knowledge acquisition,extraction,representation,modeling,fusion,computation,and storage.Within this framework,knowledge extraction,as the core component,directly determines KG quality.In military domains,traditional manual curation models face efficiency constraints due to data fragmentation,complex knowledge architectures,and confidentiality protocols.Meanwhile,crowdsourced ontology construction approaches from general domains prove non-transferable,while human-crafted ontologies struggle with generalization deficiencies.To address these challenges,this study proposes an OntologyAware LLM Methodology for Military Domain Knowledge Extraction(LLM-KE).This approach leverages the deep semantic comprehension capabilities of Large Language Models(LLMs)to simulate human experts’cognitive processes in crowdsourced ontology construction,enabling automated extraction of military textual knowledge.It concurrently enhances knowledge processing efficiency and improves KG completeness.Empirical analysis demonstrates that this method effectively resolves scalability and dynamic adaptation challenges in military KG construction,establishing a novel technological pathway for advancing military intelligence development.展开更多
The rapid development of information technology and accelerated digitalization have led to an explosive growth of data across various fields.As a key technology for knowledge representation and sharing,knowledge graph...The rapid development of information technology and accelerated digitalization have led to an explosive growth of data across various fields.As a key technology for knowledge representation and sharing,knowledge graphs play a crucial role by constructing structured networks of relationships among entities.However,data sparsity and numerous unexplored implicit relations result in the widespread incompleteness of knowledge graphs.In static knowledge graph completion,most existing methods rely on linear operations or simple interaction mechanisms for triple encoding,making it difficult to fully capture the deep semantic associations between entities and relations.Moreover,many methods focus only on the local information of individual triples,ignoring the rich semantic dependencies embedded in the neighboring nodes of entities within the graph structure,which leads to incomplete embedding representations.To address these challenges,we propose Two-Stage Mixer Embedding(TSMixerE),a static knowledge graph completion method based on entity context.In the unit semantic extraction stage,TSMixerE leveragesmulti-scale circular convolution to capture local features atmultiple granularities,enhancing the flexibility and robustness of feature interactions.A channel attention mechanism amplifies key channel responses to suppress noise and irrelevant information,thereby improving the discriminative power and semantic depth of feature representations.For contextual information fusion,a multi-layer self-attentionmechanism enables deep interactions among contextual cues,effectively integrating local details with global context.Simultaneously,type embeddings clarify the semantic identities and roles of each component,enhancing the model’s sensitivity and fusion capabilities for diverse information sources.Furthermore,TSMixerE constructs contextual unit sequences for entities,fully exploring neighborhood information within the graph structure to model complex semantic dependencies,thus improving the completeness and generalization of embedding representations.展开更多
Active fault-tolerant control utilizes information obtained from fault diagnosis to reconfigure the control law to compensate for faults in the wastewater treatment process. However, since the similarity of fault char...Active fault-tolerant control utilizes information obtained from fault diagnosis to reconfigure the control law to compensate for faults in the wastewater treatment process. However, since the similarity of fault characteristic in the incipient stage can result in misdiagnosis, it is a challenge for fault-tolerant control to ensure system safety and reliability. Therefore, to address this issue, a fault diagnosis and fault-tolerant control with a knowledge transfer strategy(KT-FDFTC) is proposed in this paper. First, a knowledge reasoning diagnosis strategy using multi-source transfer learning is designed to distinguish the similar characteristic of incipient faults. Then, the multi-source knowledge can assist in the diagnosis strategy to strengthen the fault information for fault-tolerant control. Second, a knowledge adaptive compensation mechanism, which makes knowledge and data coupled into the output trajectory regarded as an objective function, is employed to dynamically compute the control law. Then, KT-FDFTC can ensure the stable operation to adapt to various fault conditions. Third, the Lyapunov function is established to demonstrate the stability of KT-FDFTC. Then, the theoretical basis can offer the successful application of KTFDFTC. Finally, the proposed method is validated through a real WWTP and a simulation platform. The experimental results confirm that KT-FDFTC can provide good diagnosis performance and fault tolerance ability.展开更多
This paper delves into effective pathways for transforming course ecosystems from resource provision to knowledge service and competency development through university-enterprise collaboration in co-building knowledge...This paper delves into effective pathways for transforming course ecosystems from resource provision to knowledge service and competency development through university-enterprise collaboration in co-building knowledge graphs and intelligent shared courses.This approach enables personalized,learning-driven teaching.Based on knowledge graphs and integrated teacher-machine-student smart teaching scenarios,it not only innovates autonomous learning environments and human-computer interaction models while optimizing teaching experiences for both instructors and students,but also effectively addresses the issues of students’“scattered,superficial,and fragmented learning”.This establishes the foundation for personalized teaching tailored to individual aptitudes.展开更多
This paper proposes a deep learning-based 3D LiDAR perception framework designed for applications such as autonomous robots and vehicles.To address the high dependency on large-scale annotated data—an inherent limita...This paper proposes a deep learning-based 3D LiDAR perception framework designed for applications such as autonomous robots and vehicles.To address the high dependency on large-scale annotated data—an inherent limitation of deep learning models—this study introduces a hybrid perception architecture that incorporates expertdriven LiDAR processing techniques into the deep neural network.Traditional 3DLiDAR processingmethods typically remove ground planes and apply distance-or density-based clustering for object detection.In this work,such expert knowledge is encoded as feature-level inputs and fused with the deep network,therebymitigating the data dependency issue of conventional learning-based approaches.Specifically,the proposedmethod combines two expert algorithms—Patchwork++for ground segmentation and DBSCAN for clustering—with a PointPillars-based LiDAR detection network.We design four hybrid versions of the network depending on the stage and method of integrating expert features into the feature map of the deep model.Among these,Version 4 incorporates a modified neck structure in PointPillars and introduces a new Cluster 2D Pseudo-Map Branch that utilizes cluster-level pseudo-images generated from Patchwork++and DBSCAN.This version achieved a+3.88%improvement mean Average Precision(mAP)compared to the baseline PointPillars.The results demonstrate that embedding expert-based perception logic into deep neural architectures can effectively enhance performance and reduce dependency on extensive training datasets,offering a promising direction for robust 3D LiDAR object detection in real-world scenarios.展开更多
Wetting deformation in earth-rockfill dams is a critical factor influencingdam safety.Although numerous mathematical models have been developed to describe this phenomenon,most of them rely on empirical formulations a...Wetting deformation in earth-rockfill dams is a critical factor influencingdam safety.Although numerous mathematical models have been developed to describe this phenomenon,most of them rely on empirical formulations and lack prior knowledge of model parameters,which is essential for Bayesian parameter inversion to enhance accuracy and reduce uncertainty.This study introduces a datadriven approach to establishing prior knowledge of earth-rockfill dams.Driving factors are utilized to determine the potential range of model parameters,and settlement changes within this range are calculated.The results are iteratively compared with actual monitoring data until the calculated range encompasses the observed data,thereby providing prior knowledge of the model parameters.The proposed method is applied to the right-bank earth-rockfilldam of Danjiangkou.Employing a Gibbs sample size of 30,000,the proposed method effectively calibrates the prior knowledge of the wetting model parameters,achieving a root mean square error(RMSE)of 5.18 mm for the settlement predictions.By comparison,the use of non-informative priors with sample sizes of 30,000 and 50,000 results in significantly larger RMSE values of 11.97 mm and 16.07 mm,respectively.Furthermore,the computational efficiencyof the proposed method is demonstrated by an inversion computation time of 902 s for 30,000 samples,which is notably shorter than the 1026 s and 1558 s required for noninformative priors with 30,000 and 50,000 samples,respectively.These findingsunderscore the superior performance of the proposed approach in terms of both prediction accuracy and computational efficiency.These results demonstrate that the proposed method not only improves the predictive accuracy but also enhances the computational efficiency,enabling optimal parameter identificationwith reduced computational effort.This approach provides a robust and efficientframework for advancing dam safety assessments.展开更多
Identifying the community structure of complex networks is crucial to extracting insights and understanding network properties.Although several community detection methods have been proposed,many are unsuitable for so...Identifying the community structure of complex networks is crucial to extracting insights and understanding network properties.Although several community detection methods have been proposed,many are unsuitable for social networks due to significant limitations.Specifically,most approaches depend mainly on user-user structural links while overlooking service-centric,semantic,and multi-attribute drivers of community formation,and they also lack flexible filtering mechanisms for large-scale,service-oriented settings.Our proposed approach,called community discovery-based service(CDBS),leverages user profiles and their interactions with consulted web services.The method introduces a novel similarity measure,global similarity interaction profile(GSIP),which goes beyond typical similarity measures by unifying user and service profiles for all attributes types into a coherent representation,thereby clarifying its novelty and contribution.It applies multiple filtering criteria related to user attributes,accessed services,and interaction patterns.Experimental comparisons against Louvain,Hierarchical Agglomerative Clustering,Label Propagation and Infomap show that CDBS reveals the higher performance as it achieves 0.74 modularity,0.13 conductance,0.77 coverage,and significantly fast response time of 9.8 s,even with 10,000 users and 400 services.Moreover,community discoverybased service consistently detects a larger number of communities with distinct topics of interest,underscoring its capacity to generate detailed and efficient structures in complex networks.These results confirm both the efficiency and effectiveness of the proposed method.Beyond controlled evaluation,communities discovery based service is applicable to targeted recommendations,group-oriented marketing,access control,and service personalization,where communities are shaped not only by user links but also by service engagement.展开更多
The viscosity of refining slags plays a critical role in metallurgical processes.However,obtaining accurate viscosity data remains challenging due to the complexities of high-temperature experiments,often relying on e...The viscosity of refining slags plays a critical role in metallurgical processes.However,obtaining accurate viscosity data remains challenging due to the complexities of high-temperature experiments,often relying on empirical models with limited predictive capabilities.This study focuses on the influence of optical basicity on viscosity in CaO-Al_(2)O_(3)-based refining slags,leveraging machine learning to address data scarcity and improve prediction accuracy.An automated framework for algorithm integration,parameter tuning,and evaluation ranking framework(Auto-APE)is employed to develop customized data-driven models for various slag systems,including CaO-Al_(2)O_(3)-SiO_(2),CaO-Al_(2)O_(3)-CaF_(2),CaO-Al_(2)O_(3)-SiO_(2)-MgO,and CaO-Al_(2)O_(3)-SiO_(2)-MgO-CaF_(2).By incorporating optical basicity as a key feature,the models achieve an average validation error of 8.0%to 15.1%,significantly outperforming traditional empirical models.Additionally,symbolic regression is introduced to rapidly construct domain-specific features,such as optical basicity-like descriptors,offering a potential breakthrough in performance prediction for small datasets.This work highlights the critical role of domain-specific knowledge in understanding and predicting viscosity,providing a robust machine learning-based approach for optimizing refining slag properties.展开更多
Most Convolutional Neural Network(CNN)interpretation techniques visualize only the dominant cues that the model relies on,but there is no guarantee that these represent all the evidence the model uses for classificati...Most Convolutional Neural Network(CNN)interpretation techniques visualize only the dominant cues that the model relies on,but there is no guarantee that these represent all the evidence the model uses for classification.This limitation becomes critical when hidden secondary cues—potentially more meaningful than the visualized ones—remain undiscovered.This study introduces CasCAM(Cascaded Class Activation Mapping)to address this fundamental limitation through counterfactual reasoning.By asking“if this dominant cue were absent,what other evidence would the model use?”,CasCAM progressively masks the most salient features and systematically uncovers the hierarchy of classification evidence hidden beneath them.Experimental results demonstrate that CasCAM effectively discovers the full spectrum of reasoning evidence and can be universally applied with nine existing interpretation methods.展开更多
The reliable operation of power grid secondary equipment is an important guarantee for the safety and stability of the power system.However,various defects could be produced in the secondary equipment during longtermo...The reliable operation of power grid secondary equipment is an important guarantee for the safety and stability of the power system.However,various defects could be produced in the secondary equipment during longtermoperation.The complex relationship between the defect phenomenon andmulti-layer causes and the probabilistic influence of secondary equipment cannot be described through knowledge extraction and fusion technology by existing methods,which limits the real-time and accuracy of defect identification.Therefore,a defect recognition method based on the Bayesian network and knowledge graph fusion is proposed.The defect data of secondary equipment is transformed into the structured knowledge graph through knowledge extraction and fusion technology.The knowledge graph of power grid secondary equipment is mapped to the Bayesian network framework,combined with historical defect data,and introduced Noisy-OR nodes.The prior and conditional probabilities of the Bayesian network are then reasonably assigned to build a model that reflects the probability dependence between defect phenomena and potential causes in power grid secondary equipment.Defect identification of power grid secondary equipment is achieved by defect subgraph search based on the knowledge graph,and defect inference based on the Bayesian network.Practical application cases prove this method’s effectiveness in identifying secondary equipment defect causes,improving identification accuracy and efficiency.展开更多
Recent years have witnessed the significant breakthrough in the field of new materials discovery brought about by the artificial intelligence(AI).AI has successfully been applied for predicting the formability,reveali...Recent years have witnessed the significant breakthrough in the field of new materials discovery brought about by the artificial intelligence(AI).AI has successfully been applied for predicting the formability,revealing the properties,and guiding the experimental synthesis of materials.Rapid progress has been made in the integration of increasing database and improved computing power.Though some reviews present the development from their unique aspects,reviews from the view of how AI empowered both discovery of new materials and cognition of existing materials that covers the completed contents with two synergistical aspects are few.Here,the newest development is systematically reviewed in the field of AI empowered materials,reflecting advanced design of the intelligent systems for discovery,synthesis,prediction and validation of materials.First,background and mechanisms are briefed,after which the design for the AI systems with data,machine learning and automated laboratory included is illustrated.Next,strategies are summarized to obtain the AI systems for materials with improved performance which comprehensively cover the aspects from the in-depth cognizance of existing material and the rapid discovery of new materials,and then,the design thought for future AI systems in material science is pointed out.Finally,some perspectives are put forward.展开更多
Knowledge graphs,which combine structured representation with semantic modeling,have shown great potential in knowledge expression,causal inference,and automated reasoning,and are widely used in fields such as intelli...Knowledge graphs,which combine structured representation with semantic modeling,have shown great potential in knowledge expression,causal inference,and automated reasoning,and are widely used in fields such as intelligent question answering,decision support,and fault diagnosis.As high-speed train systems become increasingly intelligent and interconnected,fault patterns have grown more complex and dynamic.Knowledge graphs offer a promising solution to support the structured management and real-time reasoning of fault knowledge,addressing key requirements such as interpretability,accuracy,and continuous evolution in intelligent diagnostic systems.However,conventional knowledge graph construction relies heavily on domain expertise and specialized tools,resulting in high entry barriers for non-experts and limiting their practical application in frontline maintenance scenarios.To address this limitation,this paper proposes a fault knowledge modeling approach for high-speed trains that integrates structured logic diagrams with knowledge graphs.The method employs a seven-layer logic structure—comprising fault name,applicable vehicles,diagnostic logic,signal parameters,verification conditions,fault causes,and emergency measures—to transform unstructured knowledge into a visual and hierarchical representation.A semantic mapping mechanism is then used to automatically convert logic diagrams into machine-interpretable knowledge graphs,enabling dynamic reasoning and knowledge reuse.Furthermore,the proposed method establishes a three-layer architecture—logic structuring,knowledge graph transformation,and dynamic inference—to bridge human-expert logic with machinebased reasoning.Experimental validation and system implementation demonstrate that this approach not only improves knowledge interpretability and inference precision but also significantly enhances modeling efficiency and system maintainability.It provides a scalable and adaptable solution for intelligent operation and maintenance platforms in the high-speed rail domain.展开更多
Aiming at the problem of insufficient recognition of implicit variants by existing Chinese sensitive text detection methods,this paper proposes the IPKE-MoE framework,which consists of three parts,namely,a sensitive w...Aiming at the problem of insufficient recognition of implicit variants by existing Chinese sensitive text detection methods,this paper proposes the IPKE-MoE framework,which consists of three parts,namely,a sensitive word variant extraction framework,a sensitive word variant knowledge enhancement layer and a mixture-of-experts(MoE)classification layer.First,sensitive word variants are precisely extracted through dynamic iterative prompt templates and the context-aware capabilities of Large Language Models(LLMs).Next,the extracted variants are used to construct a knowledge enhancement layer for sensitive word variants based on RoCBert models.Specifically,after locating variants via n-gram algorithms,variant types are mapped to embedding vectors and fused with original word vectors.Finally,a mixture-of-experts(MoE)classification layer is designed(sensitive word,sentiment,and semantic experts),which decouples the relationship between sensitiveword existence and text toxicity throughmultiple experts.This framework effectively combines the comprehension ability of Large Language Models(LLMs)with the discriminative ability of smaller models.Our two experiments demonstrate that the sensitive word variant extraction framework based on dynamically iterated prompt templates outperforms other baseline prompt templates.TheRoCBert models incorporating the sensitive word variant knowledge enhancement layer and a mixture-of-experts(MoE)classification layer achieve superior classification performance compared to other baselines.展开更多
With the widespread use of social media,the propagation of health-related rumors has become a significant public health threat.Existing methods for detecting health rumors predominantly rely on external knowledge or p...With the widespread use of social media,the propagation of health-related rumors has become a significant public health threat.Existing methods for detecting health rumors predominantly rely on external knowledge or propagation structures,with only a few recent approaches attempting causal inference;however,these have not yet effectively integrated causal discovery with domain-specific knowledge graphs for detecting health rumors.In this study,we found that the combined use of causal discovery and domain-specific knowledge graphs can effectively identify implicit pseudo-causal logic embedded within texts,holding significant potential for health rumor detection.To this end,we propose CKDG—a dual-graph fusion framework based on causal logic and medical knowledge graphs.CKDG constructs a weighted causal graph to capture the implicit causal relationships in the text and introduces a medical knowledge graph to verify semantic consistency,thereby enhancing the ability to identify the misuse of professional terminology and pseudoscientific claims.In experiments conducted on a dataset comprising 8430 health rumors,CKDG achieved an accuracy of 91.28%and an F1 score of 90.38%,representing improvements of 5.11%and 3.29%over the best baseline,respectively.Our results indicate that the integrated use of causal discovery and domainspecific knowledge graphs offers significant advantages for health rumor detection systems.This method not only improves detection performance but also enhances the transparency and credibility of model decisions by tracing causal chains and sources of knowledge conflicts.We anticipate that this work will provide key technological support for the development of trustworthy health-information filtering systems,thereby improving the reliability of public health information on social media.展开更多
Knowledge distillation has become a standard technique for compressing large language models into efficient student models,but existing methods often struggle to balance prediction accuracy with explanation quality.Re...Knowledge distillation has become a standard technique for compressing large language models into efficient student models,but existing methods often struggle to balance prediction accuracy with explanation quality.Recent approaches such as Distilling Step-by-Step(DSbS)introduce explanation supervision,yet they apply it in a uniform manner that may not fully exploit the different learning dynamics of prediction and explanation.In this work,we propose a task-structured curriculum learning(TSCL)framework that structures training into three sequential phases:(i)prediction-only,to establish stable feature representations;(ii)joint prediction-explanation,to align task outputs with rationale generation;and(iii)explanation-only,to refine the quality of rationales.This design provides a simple but effective modification to DSbS,requiring no architectural changes and adding negligible training cost.We justify the phase scheduling with ablation studies and convergence analysis,showing that an initial prediction-heavy stage followed by a balanced joint phase improves both stability and explanation alignment.Extensive experiments on five datasets(e-SNLI,ANLI,CommonsenseQA,SVAMP,and MedNLI)demonstrate that TSCL consistently outperforms strong baselines,achieving gains of+1.7-2.6 points in accuracy and 0.8-1.2 in ROUGE-L,corresponding to relative error reductions of up to 21%.Beyond lexical metrics,human evaluation and ERASERstyle faithfulness diagnostics confirm that TSCL produces more faithful and informative explanations.Comparative training curves further reveal faster convergence and lower variance across seeds.Efficiency analysis shows less than 3%overhead in wall-clock training time and no additional inference cost,making the approach practical for realworld deployment.This study demonstrates that a simple task-structured curriculum can significantly improve the effectiveness of knowledge distillation.By separating and sequencing objectives,TSCL achieves a better balance between accuracy,stability,and explanation quality.The framework generalizes across domains,including medical NLI,and offers a principled recipe for future applications in multimodal reasoning and reinforcement learning.展开更多
With the large-scale deployment of the Internet ofThings(IoT)devices,their weak securitymechanisms make them prime targets for malware attacks.Attackers often use Domain Generation Algorithm(DGA)to generate random dom...With the large-scale deployment of the Internet ofThings(IoT)devices,their weak securitymechanisms make them prime targets for malware attacks.Attackers often use Domain Generation Algorithm(DGA)to generate random domain names,hiding the real IP of Command and Control(C&C)servers to build botnets.Due to the randomness and dynamics of DGA,traditional methods struggle to detect them accurately,increasing the difficulty of network defense.This paper proposes a lightweight DGA detection model based on knowledge distillation for resource-constrained IoT environments.Specifically,a teacher model combining CharacterBERT,a bidirectional long short-term memory(BiLSTM)network,and attention mechanism(ATT)is constructed:it extracts character-level semantic features viaCharacterBERT,captures sequence dependencieswith the BiLSTM,and integrates theATT for key feature weighting,formingmulti-granularity feature fusion.An improved knowledge distillation approach transfers the teacher model’s learned knowledge to the simplified DistilBERT student model.Experimental results show the teacher model achieves 98.68%detection accuracy.The student modelmaintains slightly improved accuracy while significantly compressing parameters to approximately 38.4%of the teacher model’s scale,greatly reducing computational overhead for IoT deployment.展开更多
文摘This review presents a comprehensive and forward-looking analysis of how Large Language Models(LLMs)are transforming knowledge discovery in the rational design of advancedmicro/nano electrocatalyst materials.Electrocatalysis is central to sustainable energy and environmental technologies,but traditional catalyst discovery is often hindered by high complexity,fragmented knowledge,and inefficiencies.LLMs,particularly those based on Transformer architectures,offer unprecedented capabilities in extracting,synthesizing,and generating scientific knowledge from vast unstructured textual corpora.This work provides the first structured synthesis of how LLMs have been leveraged across various electrocatalysis tasks,including automated information extraction from literature,text-based property prediction,hypothesis generation,synthesis planning,and knowledge graph construction.We comparatively analyze leading LLMs and domain-specific frameworks(e.g.,CatBERTa,CataLM,CatGPT)in terms of methodology,application scope,performance metrics,and limitations.Through curated case studies across key electrocatalytic reactions—HER,OER,ORR,and CO_(2)RR—we highlight emerging trends such as the growing use of embedding-based prediction,retrieval-augmented generation,and fine-tuned scientific LLMs.The review also identifies persistent challenges,including data heterogeneity,hallucination risks,lack of standard benchmarks,and limited multimodal integration.Importantly,we articulate future research directions,such as the development of multimodal and physics-informedMatSci-LLMs,enhanced interpretability tools,and the integration of LLMswith selfdriving laboratories for autonomous discovery.By consolidating fragmented advances and outlining a unified research roadmap,this review provides valuable guidance for both materials scientists and AI practitioners seeking to accelerate catalyst innovation through large language model technologies.
基金supported by Natural Science Foundation of Sichuan,China(Grant No.:2024ZDZX0019).
文摘Traditional Chinese medicine(TCM)serves as a treasure trove of ancient knowledge,holding a crucial position in the medical field.However,the exploration of TCM's extensive information has been hindered by challenges related to data standardization,completeness,and accuracy,primarily due to the decen-tralized distribution of TCM resources.To address these issues,we developed a platform for TCM knowledge discovery(TCMKD,https://cbcb.cdutcm.edu.cn/TCMKD/).Seven types of data,including syndromes,formulas,Chinese patent drugs(CPDs),Chinese medicinal materials(CMMs),ingredients,targets,and diseases,were manually proofread and consolidated within TCMKD.To strengthen the integration of TCM with modern medicine,TCMKD employs analytical methods such as TCM data mining,enrichment analysis,and network localization and separation.These tools help elucidate the molecular-level commonalities between TCM and contemporary scientific insights.In addition to its analytical capabilities,a quick question and answer(Q&A)system is also embedded within TCMKD to query the database efficiently,thereby improving the interactivity of the platform.The platform also provides a TCM text annotation tool,offering a simple and efficient method for TCM text mining.Overall,TCMKD not only has the potential to become a pivotal repository for TCM,delving into the pharmaco-logical foundations of TCM treatments,but its flexible embedded tools and algorithms can also be applied to the study of other traditional medical systems,extending beyond just TCM.
基金supported by the China Fundamental Research Funds for the Central Universities(No.2662022XXYJ001,2662022JC004,2662023XXPY005)。
文摘As a new data management paradigm,knowledge graphs can integrate multiple data sources and achieve quick responses,reasoning and better predictions in drug discovery.Characterized by powerful contagion and a high rate of morbidity and mortality,porcine reproductive and respiratory syndrome(PRRS)is a common infectious disease in the global swine industry that causes economically great losses.Traditional Chinese medicine(TCM)has advantages in low adverse effects and a relatively affordable cost of application,and TCM is therefore conceived as a possibility to treat PRRS under the current circumstance that there is a lack of safe and effective approaches.Here,we constructed a knowledge graph containing common biomedical data from humans and Sus Scrofa as well as information from thousands of TCMs.Subsequently,we validated the effectiveness of the Sus Scrofa knowledge graph by the t-SNE algorithm and selected the optimal model(i.e.,transR)from six typical models,namely,transE,transR,DistMult,ComplEx,RESCAL and RotatE,according to five indicators,namely,MRR,MR,HITS@1,HITS@3 and HITS@10.Based on embedding vectors trained by the optimal model,anti-PRRSV TCMs were predicted by two paths,namely,VHC-Herb and VHPC-Herb,and potential anti-PRRSVTCMs were identified by retrieving the HERB database according to the phar-macological properties corresponding to symptoms of PRRS.Ultimately,Dan Shen's(Salvia miltiorrhiza Bunge)capacity to resist PRRSV infection was validated by a cell experiment in which the inhibition rate of PRRSV exceeded90%when the concentrations of Dan Shen extract were 0.004,0.008,0.016 and 0.032 mg/mL.In summary,this is the first report on the Sus Scrofa knowledge graph including TCM information,and our study reflects the important application values of deep learning on graphs in the swine industry as well as providing accessible TCM resources for PRRS.
文摘Since Google introduced the concept of Knowledge Graphs(KGs)in 2012,their construction technologies have evolved into a comprehensive methodological framework encompassing knowledge acquisition,extraction,representation,modeling,fusion,computation,and storage.Within this framework,knowledge extraction,as the core component,directly determines KG quality.In military domains,traditional manual curation models face efficiency constraints due to data fragmentation,complex knowledge architectures,and confidentiality protocols.Meanwhile,crowdsourced ontology construction approaches from general domains prove non-transferable,while human-crafted ontologies struggle with generalization deficiencies.To address these challenges,this study proposes an OntologyAware LLM Methodology for Military Domain Knowledge Extraction(LLM-KE).This approach leverages the deep semantic comprehension capabilities of Large Language Models(LLMs)to simulate human experts’cognitive processes in crowdsourced ontology construction,enabling automated extraction of military textual knowledge.It concurrently enhances knowledge processing efficiency and improves KG completeness.Empirical analysis demonstrates that this method effectively resolves scalability and dynamic adaptation challenges in military KG construction,establishing a novel technological pathway for advancing military intelligence development.
基金supported by the National Natural Science Foundation of China(No.62267005)the Chinese Guangxi Natural Science Foundation(No.2023GXNSFAA026493)+1 种基金Guangxi Collaborative Innovation Center ofMulti-Source Information Integration and Intelligent ProcessingGuangxi Academy of Artificial Intelligence.
文摘The rapid development of information technology and accelerated digitalization have led to an explosive growth of data across various fields.As a key technology for knowledge representation and sharing,knowledge graphs play a crucial role by constructing structured networks of relationships among entities.However,data sparsity and numerous unexplored implicit relations result in the widespread incompleteness of knowledge graphs.In static knowledge graph completion,most existing methods rely on linear operations or simple interaction mechanisms for triple encoding,making it difficult to fully capture the deep semantic associations between entities and relations.Moreover,many methods focus only on the local information of individual triples,ignoring the rich semantic dependencies embedded in the neighboring nodes of entities within the graph structure,which leads to incomplete embedding representations.To address these challenges,we propose Two-Stage Mixer Embedding(TSMixerE),a static knowledge graph completion method based on entity context.In the unit semantic extraction stage,TSMixerE leveragesmulti-scale circular convolution to capture local features atmultiple granularities,enhancing the flexibility and robustness of feature interactions.A channel attention mechanism amplifies key channel responses to suppress noise and irrelevant information,thereby improving the discriminative power and semantic depth of feature representations.For contextual information fusion,a multi-layer self-attentionmechanism enables deep interactions among contextual cues,effectively integrating local details with global context.Simultaneously,type embeddings clarify the semantic identities and roles of each component,enhancing the model’s sensitivity and fusion capabilities for diverse information sources.Furthermore,TSMixerE constructs contextual unit sequences for entities,fully exploring neighborhood information within the graph structure to model complex semantic dependencies,thus improving the completeness and generalization of embedding representations.
基金supported by the National Natural Science Foundation of China (Grant Nos.62125301,62021003,62303024,U24A20275,62522302,62473011,92467205)the National Key Research and Development Project (Grant Nos.2022YFB3305800-5,2024YFE0212400)+2 种基金the Youth Beijing Scholars Program (Grant No.037)the Beijing Nova Program (Grant Nos.20240484694,20250484938)the Beijing Natural Science Foundation (Grant No.L253010)。
文摘Active fault-tolerant control utilizes information obtained from fault diagnosis to reconfigure the control law to compensate for faults in the wastewater treatment process. However, since the similarity of fault characteristic in the incipient stage can result in misdiagnosis, it is a challenge for fault-tolerant control to ensure system safety and reliability. Therefore, to address this issue, a fault diagnosis and fault-tolerant control with a knowledge transfer strategy(KT-FDFTC) is proposed in this paper. First, a knowledge reasoning diagnosis strategy using multi-source transfer learning is designed to distinguish the similar characteristic of incipient faults. Then, the multi-source knowledge can assist in the diagnosis strategy to strengthen the fault information for fault-tolerant control. Second, a knowledge adaptive compensation mechanism, which makes knowledge and data coupled into the output trajectory regarded as an objective function, is employed to dynamically compute the control law. Then, KT-FDFTC can ensure the stable operation to adapt to various fault conditions. Third, the Lyapunov function is established to demonstrate the stability of KT-FDFTC. Then, the theoretical basis can offer the successful application of KTFDFTC. Finally, the proposed method is validated through a real WWTP and a simulation platform. The experimental results confirm that KT-FDFTC can provide good diagnosis performance and fault tolerance ability.
基金supported by Harbin Institute of Technology High-level Teaching Achievement Award(National Level)Cultivation Project(256709).
文摘This paper delves into effective pathways for transforming course ecosystems from resource provision to knowledge service and competency development through university-enterprise collaboration in co-building knowledge graphs and intelligent shared courses.This approach enables personalized,learning-driven teaching.Based on knowledge graphs and integrated teacher-machine-student smart teaching scenarios,it not only innovates autonomous learning environments and human-computer interaction models while optimizing teaching experiences for both instructors and students,but also effectively addresses the issues of students’“scattered,superficial,and fragmented learning”.This establishes the foundation for personalized teaching tailored to individual aptitudes.
基金supported by Basic Science Research Program through the National Research Foundation of Korea(NRF)funded by the Ministry of Education(RS-2023-00245084)by Korea Institute for Advancement of Technology(KIAT)grant funded by the Korea Government(MOTIE)(RS-2024-00415938,HRD Program for Industrial Innovation)and Soonchunhyang University.
文摘This paper proposes a deep learning-based 3D LiDAR perception framework designed for applications such as autonomous robots and vehicles.To address the high dependency on large-scale annotated data—an inherent limitation of deep learning models—this study introduces a hybrid perception architecture that incorporates expertdriven LiDAR processing techniques into the deep neural network.Traditional 3DLiDAR processingmethods typically remove ground planes and apply distance-or density-based clustering for object detection.In this work,such expert knowledge is encoded as feature-level inputs and fused with the deep network,therebymitigating the data dependency issue of conventional learning-based approaches.Specifically,the proposedmethod combines two expert algorithms—Patchwork++for ground segmentation and DBSCAN for clustering—with a PointPillars-based LiDAR detection network.We design four hybrid versions of the network depending on the stage and method of integrating expert features into the feature map of the deep model.Among these,Version 4 incorporates a modified neck structure in PointPillars and introduces a new Cluster 2D Pseudo-Map Branch that utilizes cluster-level pseudo-images generated from Patchwork++and DBSCAN.This version achieved a+3.88%improvement mean Average Precision(mAP)compared to the baseline PointPillars.The results demonstrate that embedding expert-based perception logic into deep neural architectures can effectively enhance performance and reduce dependency on extensive training datasets,offering a promising direction for robust 3D LiDAR object detection in real-world scenarios.
基金supported by the National Key R&D Program of China(Grant No.2023YFC3209504)Natural Science Foundation of Wuhan(Grant No.2024040801020271)the Fundamental Research Funds for Central Public Welfare Research Institutes(Grant No.CKSF2025718/YT).
文摘Wetting deformation in earth-rockfill dams is a critical factor influencingdam safety.Although numerous mathematical models have been developed to describe this phenomenon,most of them rely on empirical formulations and lack prior knowledge of model parameters,which is essential for Bayesian parameter inversion to enhance accuracy and reduce uncertainty.This study introduces a datadriven approach to establishing prior knowledge of earth-rockfill dams.Driving factors are utilized to determine the potential range of model parameters,and settlement changes within this range are calculated.The results are iteratively compared with actual monitoring data until the calculated range encompasses the observed data,thereby providing prior knowledge of the model parameters.The proposed method is applied to the right-bank earth-rockfilldam of Danjiangkou.Employing a Gibbs sample size of 30,000,the proposed method effectively calibrates the prior knowledge of the wetting model parameters,achieving a root mean square error(RMSE)of 5.18 mm for the settlement predictions.By comparison,the use of non-informative priors with sample sizes of 30,000 and 50,000 results in significantly larger RMSE values of 11.97 mm and 16.07 mm,respectively.Furthermore,the computational efficiencyof the proposed method is demonstrated by an inversion computation time of 902 s for 30,000 samples,which is notably shorter than the 1026 s and 1558 s required for noninformative priors with 30,000 and 50,000 samples,respectively.These findingsunderscore the superior performance of the proposed approach in terms of both prediction accuracy and computational efficiency.These results demonstrate that the proposed method not only improves the predictive accuracy but also enhances the computational efficiency,enabling optimal parameter identificationwith reduced computational effort.This approach provides a robust and efficientframework for advancing dam safety assessments.
文摘Identifying the community structure of complex networks is crucial to extracting insights and understanding network properties.Although several community detection methods have been proposed,many are unsuitable for social networks due to significant limitations.Specifically,most approaches depend mainly on user-user structural links while overlooking service-centric,semantic,and multi-attribute drivers of community formation,and they also lack flexible filtering mechanisms for large-scale,service-oriented settings.Our proposed approach,called community discovery-based service(CDBS),leverages user profiles and their interactions with consulted web services.The method introduces a novel similarity measure,global similarity interaction profile(GSIP),which goes beyond typical similarity measures by unifying user and service profiles for all attributes types into a coherent representation,thereby clarifying its novelty and contribution.It applies multiple filtering criteria related to user attributes,accessed services,and interaction patterns.Experimental comparisons against Louvain,Hierarchical Agglomerative Clustering,Label Propagation and Infomap show that CDBS reveals the higher performance as it achieves 0.74 modularity,0.13 conductance,0.77 coverage,and significantly fast response time of 9.8 s,even with 10,000 users and 400 services.Moreover,community discoverybased service consistently detects a larger number of communities with distinct topics of interest,underscoring its capacity to generate detailed and efficient structures in complex networks.These results confirm both the efficiency and effectiveness of the proposed method.Beyond controlled evaluation,communities discovery based service is applicable to targeted recommendations,group-oriented marketing,access control,and service personalization,where communities are shaped not only by user links but also by service engagement.
基金supported by the National Key Research and Development Program of China(No.2023YFB3712401),the National Natural Science Foundation of China(No.52274301)the Aeronautical Science Foundation of China(No.2023Z0530S6005)the Ningbo Yongjiang Talent-Introduction Programme(No.2022A-023-C).
文摘The viscosity of refining slags plays a critical role in metallurgical processes.However,obtaining accurate viscosity data remains challenging due to the complexities of high-temperature experiments,often relying on empirical models with limited predictive capabilities.This study focuses on the influence of optical basicity on viscosity in CaO-Al_(2)O_(3)-based refining slags,leveraging machine learning to address data scarcity and improve prediction accuracy.An automated framework for algorithm integration,parameter tuning,and evaluation ranking framework(Auto-APE)is employed to develop customized data-driven models for various slag systems,including CaO-Al_(2)O_(3)-SiO_(2),CaO-Al_(2)O_(3)-CaF_(2),CaO-Al_(2)O_(3)-SiO_(2)-MgO,and CaO-Al_(2)O_(3)-SiO_(2)-MgO-CaF_(2).By incorporating optical basicity as a key feature,the models achieve an average validation error of 8.0%to 15.1%,significantly outperforming traditional empirical models.Additionally,symbolic regression is introduced to rapidly construct domain-specific features,such as optical basicity-like descriptors,offering a potential breakthrough in performance prediction for small datasets.This work highlights the critical role of domain-specific knowledge in understanding and predicting viscosity,providing a robust machine learning-based approach for optimizing refining slag properties.
基金supported by the Basic Science Research Program through the National Research Foundation of Korea(NRF),funded by the Ministry of Education(RS-2023-00249743).
文摘Most Convolutional Neural Network(CNN)interpretation techniques visualize only the dominant cues that the model relies on,but there is no guarantee that these represent all the evidence the model uses for classification.This limitation becomes critical when hidden secondary cues—potentially more meaningful than the visualized ones—remain undiscovered.This study introduces CasCAM(Cascaded Class Activation Mapping)to address this fundamental limitation through counterfactual reasoning.By asking“if this dominant cue were absent,what other evidence would the model use?”,CasCAM progressively masks the most salient features and systematically uncovers the hierarchy of classification evidence hidden beneath them.Experimental results demonstrate that CasCAM effectively discovers the full spectrum of reasoning evidence and can be universally applied with nine existing interpretation methods.
基金supported by the State Grid Southwest Branch Project“Research on Defect Diagnosis and Early Warning Technology of Relay Protection and Safety Automation Devices Based on Multi-Source Heterogeneous Defect Data”.
文摘The reliable operation of power grid secondary equipment is an important guarantee for the safety and stability of the power system.However,various defects could be produced in the secondary equipment during longtermoperation.The complex relationship between the defect phenomenon andmulti-layer causes and the probabilistic influence of secondary equipment cannot be described through knowledge extraction and fusion technology by existing methods,which limits the real-time and accuracy of defect identification.Therefore,a defect recognition method based on the Bayesian network and knowledge graph fusion is proposed.The defect data of secondary equipment is transformed into the structured knowledge graph through knowledge extraction and fusion technology.The knowledge graph of power grid secondary equipment is mapped to the Bayesian network framework,combined with historical defect data,and introduced Noisy-OR nodes.The prior and conditional probabilities of the Bayesian network are then reasonably assigned to build a model that reflects the probability dependence between defect phenomena and potential causes in power grid secondary equipment.Defect identification of power grid secondary equipment is achieved by defect subgraph search based on the knowledge graph,and defect inference based on the Bayesian network.Practical application cases prove this method’s effectiveness in identifying secondary equipment defect causes,improving identification accuracy and efficiency.
基金supported by the Hong Kong Polytechnic University(Project No.4-ZZW1,4-YWER,97D9,4-W443)。
文摘Recent years have witnessed the significant breakthrough in the field of new materials discovery brought about by the artificial intelligence(AI).AI has successfully been applied for predicting the formability,revealing the properties,and guiding the experimental synthesis of materials.Rapid progress has been made in the integration of increasing database and improved computing power.Though some reviews present the development from their unique aspects,reviews from the view of how AI empowered both discovery of new materials and cognition of existing materials that covers the completed contents with two synergistical aspects are few.Here,the newest development is systematically reviewed in the field of AI empowered materials,reflecting advanced design of the intelligent systems for discovery,synthesis,prediction and validation of materials.First,background and mechanisms are briefed,after which the design for the AI systems with data,machine learning and automated laboratory included is illustrated.Next,strategies are summarized to obtain the AI systems for materials with improved performance which comprehensively cover the aspects from the in-depth cognizance of existing material and the rapid discovery of new materials,and then,the design thought for future AI systems in material science is pointed out.Finally,some perspectives are put forward.
基金support from the Scientific Funding for the Center of National Railway Intelligent Transportation System Engineering and Technology,China Academy of Railway Sciences Corporation Limited(Grant No.2023YJ354)。
文摘Knowledge graphs,which combine structured representation with semantic modeling,have shown great potential in knowledge expression,causal inference,and automated reasoning,and are widely used in fields such as intelligent question answering,decision support,and fault diagnosis.As high-speed train systems become increasingly intelligent and interconnected,fault patterns have grown more complex and dynamic.Knowledge graphs offer a promising solution to support the structured management and real-time reasoning of fault knowledge,addressing key requirements such as interpretability,accuracy,and continuous evolution in intelligent diagnostic systems.However,conventional knowledge graph construction relies heavily on domain expertise and specialized tools,resulting in high entry barriers for non-experts and limiting their practical application in frontline maintenance scenarios.To address this limitation,this paper proposes a fault knowledge modeling approach for high-speed trains that integrates structured logic diagrams with knowledge graphs.The method employs a seven-layer logic structure—comprising fault name,applicable vehicles,diagnostic logic,signal parameters,verification conditions,fault causes,and emergency measures—to transform unstructured knowledge into a visual and hierarchical representation.A semantic mapping mechanism is then used to automatically convert logic diagrams into machine-interpretable knowledge graphs,enabling dynamic reasoning and knowledge reuse.Furthermore,the proposed method establishes a three-layer architecture—logic structuring,knowledge graph transformation,and dynamic inference—to bridge human-expert logic with machinebased reasoning.Experimental validation and system implementation demonstrate that this approach not only improves knowledge interpretability and inference precision but also significantly enhances modeling efficiency and system maintainability.It provides a scalable and adaptable solution for intelligent operation and maintenance platforms in the high-speed rail domain.
基金funded by the National Natural Science Foundation of China(Grant No.62441212)the Major Project of the Natural Science Foundation of Inner Mongolia(Grant No.2025ZD008).
文摘Aiming at the problem of insufficient recognition of implicit variants by existing Chinese sensitive text detection methods,this paper proposes the IPKE-MoE framework,which consists of three parts,namely,a sensitive word variant extraction framework,a sensitive word variant knowledge enhancement layer and a mixture-of-experts(MoE)classification layer.First,sensitive word variants are precisely extracted through dynamic iterative prompt templates and the context-aware capabilities of Large Language Models(LLMs).Next,the extracted variants are used to construct a knowledge enhancement layer for sensitive word variants based on RoCBert models.Specifically,after locating variants via n-gram algorithms,variant types are mapped to embedding vectors and fused with original word vectors.Finally,a mixture-of-experts(MoE)classification layer is designed(sensitive word,sentiment,and semantic experts),which decouples the relationship between sensitiveword existence and text toxicity throughmultiple experts.This framework effectively combines the comprehension ability of Large Language Models(LLMs)with the discriminative ability of smaller models.Our two experiments demonstrate that the sensitive word variant extraction framework based on dynamically iterated prompt templates outperforms other baseline prompt templates.TheRoCBert models incorporating the sensitive word variant knowledge enhancement layer and a mixture-of-experts(MoE)classification layer achieve superior classification performance compared to other baselines.
基金funded by the Hunan Provincial Natural Science Foundation of China(Grant No.2025JJ70105)the Hunan Provincial College Students’Innovation and Entrepreneurship Training Program(Project No.S202411342056)The article processing charge(APC)was funded by the Project No.2025JJ70105.
文摘With the widespread use of social media,the propagation of health-related rumors has become a significant public health threat.Existing methods for detecting health rumors predominantly rely on external knowledge or propagation structures,with only a few recent approaches attempting causal inference;however,these have not yet effectively integrated causal discovery with domain-specific knowledge graphs for detecting health rumors.In this study,we found that the combined use of causal discovery and domain-specific knowledge graphs can effectively identify implicit pseudo-causal logic embedded within texts,holding significant potential for health rumor detection.To this end,we propose CKDG—a dual-graph fusion framework based on causal logic and medical knowledge graphs.CKDG constructs a weighted causal graph to capture the implicit causal relationships in the text and introduces a medical knowledge graph to verify semantic consistency,thereby enhancing the ability to identify the misuse of professional terminology and pseudoscientific claims.In experiments conducted on a dataset comprising 8430 health rumors,CKDG achieved an accuracy of 91.28%and an F1 score of 90.38%,representing improvements of 5.11%and 3.29%over the best baseline,respectively.Our results indicate that the integrated use of causal discovery and domainspecific knowledge graphs offers significant advantages for health rumor detection systems.This method not only improves detection performance but also enhances the transparency and credibility of model decisions by tracing causal chains and sources of knowledge conflicts.We anticipate that this work will provide key technological support for the development of trustworthy health-information filtering systems,thereby improving the reliability of public health information on social media.
文摘Knowledge distillation has become a standard technique for compressing large language models into efficient student models,but existing methods often struggle to balance prediction accuracy with explanation quality.Recent approaches such as Distilling Step-by-Step(DSbS)introduce explanation supervision,yet they apply it in a uniform manner that may not fully exploit the different learning dynamics of prediction and explanation.In this work,we propose a task-structured curriculum learning(TSCL)framework that structures training into three sequential phases:(i)prediction-only,to establish stable feature representations;(ii)joint prediction-explanation,to align task outputs with rationale generation;and(iii)explanation-only,to refine the quality of rationales.This design provides a simple but effective modification to DSbS,requiring no architectural changes and adding negligible training cost.We justify the phase scheduling with ablation studies and convergence analysis,showing that an initial prediction-heavy stage followed by a balanced joint phase improves both stability and explanation alignment.Extensive experiments on five datasets(e-SNLI,ANLI,CommonsenseQA,SVAMP,and MedNLI)demonstrate that TSCL consistently outperforms strong baselines,achieving gains of+1.7-2.6 points in accuracy and 0.8-1.2 in ROUGE-L,corresponding to relative error reductions of up to 21%.Beyond lexical metrics,human evaluation and ERASERstyle faithfulness diagnostics confirm that TSCL produces more faithful and informative explanations.Comparative training curves further reveal faster convergence and lower variance across seeds.Efficiency analysis shows less than 3%overhead in wall-clock training time and no additional inference cost,making the approach practical for realworld deployment.This study demonstrates that a simple task-structured curriculum can significantly improve the effectiveness of knowledge distillation.By separating and sequencing objectives,TSCL achieves a better balance between accuracy,stability,and explanation quality.The framework generalizes across domains,including medical NLI,and offers a principled recipe for future applications in multimodal reasoning and reinforcement learning.
基金supported by the following projects:National Natural Science Foundation of China(62461041)Natural Science Foundation of Jiangxi Province China(20242BAB25068).
文摘With the large-scale deployment of the Internet ofThings(IoT)devices,their weak securitymechanisms make them prime targets for malware attacks.Attackers often use Domain Generation Algorithm(DGA)to generate random domain names,hiding the real IP of Command and Control(C&C)servers to build botnets.Due to the randomness and dynamics of DGA,traditional methods struggle to detect them accurately,increasing the difficulty of network defense.This paper proposes a lightweight DGA detection model based on knowledge distillation for resource-constrained IoT environments.Specifically,a teacher model combining CharacterBERT,a bidirectional long short-term memory(BiLSTM)network,and attention mechanism(ATT)is constructed:it extracts character-level semantic features viaCharacterBERT,captures sequence dependencieswith the BiLSTM,and integrates theATT for key feature weighting,formingmulti-granularity feature fusion.An improved knowledge distillation approach transfers the teacher model’s learned knowledge to the simplified DistilBERT student model.Experimental results show the teacher model achieves 98.68%detection accuracy.The student modelmaintains slightly improved accuracy while significantly compressing parameters to approximately 38.4%of the teacher model’s scale,greatly reducing computational overhead for IoT deployment.