Smart contracts are widely used on the blockchain to implement complex transactions,such as decentralized applications on Ethereum.Effective vulnerability detection of large-scale smart contracts is critical,as attack...Smart contracts are widely used on the blockchain to implement complex transactions,such as decentralized applications on Ethereum.Effective vulnerability detection of large-scale smart contracts is critical,as attacks on smart contracts often cause huge economic losses.Since it is difficult to repair and update smart contracts,it is necessary to find the vulnerabilities before they are deployed.However,code analysis,which requires traversal paths,and learning methods,which require many features to be trained,are too time-consuming to detect large-scale on-chain contracts.Learning-based methods will obtain detection models from a feature space compared to code analysis methods such as symbol execution.But the existing features lack the interpretability of the detection results and training model,even worse,the large-scale feature space also affects the efficiency of detection.This paper focuses on improving the detection efficiency by reducing the dimension of the features,combined with expert knowledge.In this paper,a feature extraction model Block-gram is proposed to form low-dimensional knowledge-based features from bytecode.First,the metadata is separated and the runtime code is converted into a sequence of opcodes,which are divided into segments based on some instructions(jumps,etc.).Then,scalable Block-gram features,including 4-dimensional block features and 8-dimensional attribute features,are mined for the learning-based model training.Finally,feature contributions are calculated from SHAP values to measure the relationship between our features and the results of the detection model.In addition,six types of vulnerability labels are made on a dataset containing 33,885 contracts,and these knowledge-based features are evaluated using seven state-of-the-art learning algorithms,which show that the average detection latency speeds up 25×to 650×,compared with the features extracted by N-gram,and also can enhance the interpretability of the detection model.展开更多
Model evaluation using benchmark datasets is an important method to measure the capability of large language models(LLMs)in specific domains,and it is mainly used to assess the knowledge and reasoning abilities of LLM...Model evaluation using benchmark datasets is an important method to measure the capability of large language models(LLMs)in specific domains,and it is mainly used to assess the knowledge and reasoning abilities of LLMs.Therefore,in order to better assess the capability of LLMs in the agricultural domain,Agri-Eval was proposed as a benchmark for assessing the knowledge and reasoning ability of LLMs in agriculture.The assessment dataset used in Agri-Eval covered seven major disciplines in the agricultural domain:crop science,horticulture,plant protection,animal husbandry,forest science,aquaculture science,and grass science,and contained a total of 2283 questions.Among domestic general-purpose LLMs,DeepSeek R1 performed best with an accuracy rate of 75.49%.In the realm of international general-purpose LLMs,Gemini 2.0 pro exp 0205 standed out as the top performer,achieving an accuracy rate of 74.28%.As an LLMs in agriculture vertical,Shennong V2.0 outperformed all the LLMs in China,and the answer accuracy rate of agricultural knowledge exceeded that of all the existing general-purpose LLMs.The launch of Agri-Eval helped the LLM developers to comprehensively evaluate the model's capability in the field of agriculture through a variety of tasks and tests to promote the development of the LLMs in the field of agriculture.展开更多
1.《工程力学》是中国科学技术协会主管、中国力学学会主办、清华大学土木工程系承办的学术期刊。2.本刊为Ei Compendex、Scopus及ISI Web of Knowledge文摘数据库收录期刊。万方数据库、中国知网、重庆维普、超星、EBSCO数据库、Euro ...1.《工程力学》是中国科学技术协会主管、中国力学学会主办、清华大学土木工程系承办的学术期刊。2.本刊为Ei Compendex、Scopus及ISI Web of Knowledge文摘数据库收录期刊。万方数据库、中国知网、重庆维普、超星、EBSCO数据库、Euro Pub数据库、JST数据库、ICI World of Journals全文检索。3.本刊于2019年和2024年两度入选“中国科技期刊卓越行动计划梯队项目”。4.本刊于2018年11月第三次荣获中国科协“精品科技期刊工程项目”资助。展开更多
The rapid development of information technology and accelerated digitalization have led to an explosive growth of data across various fields.As a key technology for knowledge representation and sharing,knowledge graph...The rapid development of information technology and accelerated digitalization have led to an explosive growth of data across various fields.As a key technology for knowledge representation and sharing,knowledge graphs play a crucial role by constructing structured networks of relationships among entities.However,data sparsity and numerous unexplored implicit relations result in the widespread incompleteness of knowledge graphs.In static knowledge graph completion,most existing methods rely on linear operations or simple interaction mechanisms for triple encoding,making it difficult to fully capture the deep semantic associations between entities and relations.Moreover,many methods focus only on the local information of individual triples,ignoring the rich semantic dependencies embedded in the neighboring nodes of entities within the graph structure,which leads to incomplete embedding representations.To address these challenges,we propose Two-Stage Mixer Embedding(TSMixerE),a static knowledge graph completion method based on entity context.In the unit semantic extraction stage,TSMixerE leveragesmulti-scale circular convolution to capture local features atmultiple granularities,enhancing the flexibility and robustness of feature interactions.A channel attention mechanism amplifies key channel responses to suppress noise and irrelevant information,thereby improving the discriminative power and semantic depth of feature representations.For contextual information fusion,a multi-layer self-attentionmechanism enables deep interactions among contextual cues,effectively integrating local details with global context.Simultaneously,type embeddings clarify the semantic identities and roles of each component,enhancing the model’s sensitivity and fusion capabilities for diverse information sources.Furthermore,TSMixerE constructs contextual unit sequences for entities,fully exploring neighborhood information within the graph structure to model complex semantic dependencies,thus improving the completeness and generalization of embedding representations.展开更多
Since Google introduced the concept of Knowledge Graphs(KGs)in 2012,their construction technologies have evolved into a comprehensive methodological framework encompassing knowledge acquisition,extraction,representati...Since Google introduced the concept of Knowledge Graphs(KGs)in 2012,their construction technologies have evolved into a comprehensive methodological framework encompassing knowledge acquisition,extraction,representation,modeling,fusion,computation,and storage.Within this framework,knowledge extraction,as the core component,directly determines KG quality.In military domains,traditional manual curation models face efficiency constraints due to data fragmentation,complex knowledge architectures,and confidentiality protocols.Meanwhile,crowdsourced ontology construction approaches from general domains prove non-transferable,while human-crafted ontologies struggle with generalization deficiencies.To address these challenges,this study proposes an OntologyAware LLM Methodology for Military Domain Knowledge Extraction(LLM-KE).This approach leverages the deep semantic comprehension capabilities of Large Language Models(LLMs)to simulate human experts’cognitive processes in crowdsourced ontology construction,enabling automated extraction of military textual knowledge.It concurrently enhances knowledge processing efficiency and improves KG completeness.Empirical analysis demonstrates that this method effectively resolves scalability and dynamic adaptation challenges in military KG construction,establishing a novel technological pathway for advancing military intelligence development.展开更多
Active fault-tolerant control utilizes information obtained from fault diagnosis to reconfigure the control law to compensate for faults in the wastewater treatment process. However, since the similarity of fault char...Active fault-tolerant control utilizes information obtained from fault diagnosis to reconfigure the control law to compensate for faults in the wastewater treatment process. However, since the similarity of fault characteristic in the incipient stage can result in misdiagnosis, it is a challenge for fault-tolerant control to ensure system safety and reliability. Therefore, to address this issue, a fault diagnosis and fault-tolerant control with a knowledge transfer strategy(KT-FDFTC) is proposed in this paper. First, a knowledge reasoning diagnosis strategy using multi-source transfer learning is designed to distinguish the similar characteristic of incipient faults. Then, the multi-source knowledge can assist in the diagnosis strategy to strengthen the fault information for fault-tolerant control. Second, a knowledge adaptive compensation mechanism, which makes knowledge and data coupled into the output trajectory regarded as an objective function, is employed to dynamically compute the control law. Then, KT-FDFTC can ensure the stable operation to adapt to various fault conditions. Third, the Lyapunov function is established to demonstrate the stability of KT-FDFTC. Then, the theoretical basis can offer the successful application of KTFDFTC. Finally, the proposed method is validated through a real WWTP and a simulation platform. The experimental results confirm that KT-FDFTC can provide good diagnosis performance and fault tolerance ability.展开更多
Wetting deformation in earth-rockfill dams is a critical factor influencingdam safety.Although numerous mathematical models have been developed to describe this phenomenon,most of them rely on empirical formulations a...Wetting deformation in earth-rockfill dams is a critical factor influencingdam safety.Although numerous mathematical models have been developed to describe this phenomenon,most of them rely on empirical formulations and lack prior knowledge of model parameters,which is essential for Bayesian parameter inversion to enhance accuracy and reduce uncertainty.This study introduces a datadriven approach to establishing prior knowledge of earth-rockfill dams.Driving factors are utilized to determine the potential range of model parameters,and settlement changes within this range are calculated.The results are iteratively compared with actual monitoring data until the calculated range encompasses the observed data,thereby providing prior knowledge of the model parameters.The proposed method is applied to the right-bank earth-rockfilldam of Danjiangkou.Employing a Gibbs sample size of 30,000,the proposed method effectively calibrates the prior knowledge of the wetting model parameters,achieving a root mean square error(RMSE)of 5.18 mm for the settlement predictions.By comparison,the use of non-informative priors with sample sizes of 30,000 and 50,000 results in significantly larger RMSE values of 11.97 mm and 16.07 mm,respectively.Furthermore,the computational efficiencyof the proposed method is demonstrated by an inversion computation time of 902 s for 30,000 samples,which is notably shorter than the 1026 s and 1558 s required for noninformative priors with 30,000 and 50,000 samples,respectively.These findingsunderscore the superior performance of the proposed approach in terms of both prediction accuracy and computational efficiency.These results demonstrate that the proposed method not only improves the predictive accuracy but also enhances the computational efficiency,enabling optimal parameter identificationwith reduced computational effort.This approach provides a robust and efficientframework for advancing dam safety assessments.展开更多
This paper proposes a deep learning-based 3D LiDAR perception framework designed for applications such as autonomous robots and vehicles.To address the high dependency on large-scale annotated data—an inherent limita...This paper proposes a deep learning-based 3D LiDAR perception framework designed for applications such as autonomous robots and vehicles.To address the high dependency on large-scale annotated data—an inherent limitation of deep learning models—this study introduces a hybrid perception architecture that incorporates expertdriven LiDAR processing techniques into the deep neural network.Traditional 3DLiDAR processingmethods typically remove ground planes and apply distance-or density-based clustering for object detection.In this work,such expert knowledge is encoded as feature-level inputs and fused with the deep network,therebymitigating the data dependency issue of conventional learning-based approaches.Specifically,the proposedmethod combines two expert algorithms—Patchwork++for ground segmentation and DBSCAN for clustering—with a PointPillars-based LiDAR detection network.We design four hybrid versions of the network depending on the stage and method of integrating expert features into the feature map of the deep model.Among these,Version 4 incorporates a modified neck structure in PointPillars and introduces a new Cluster 2D Pseudo-Map Branch that utilizes cluster-level pseudo-images generated from Patchwork++and DBSCAN.This version achieved a+3.88%improvement mean Average Precision(mAP)compared to the baseline PointPillars.The results demonstrate that embedding expert-based perception logic into deep neural architectures can effectively enhance performance and reduce dependency on extensive training datasets,offering a promising direction for robust 3D LiDAR object detection in real-world scenarios.展开更多
The viscosity of refining slags plays a critical role in metallurgical processes.However,obtaining accurate viscosity data remains challenging due to the complexities of high-temperature experiments,often relying on e...The viscosity of refining slags plays a critical role in metallurgical processes.However,obtaining accurate viscosity data remains challenging due to the complexities of high-temperature experiments,often relying on empirical models with limited predictive capabilities.This study focuses on the influence of optical basicity on viscosity in CaO-Al_(2)O_(3)-based refining slags,leveraging machine learning to address data scarcity and improve prediction accuracy.An automated framework for algorithm integration,parameter tuning,and evaluation ranking framework(Auto-APE)is employed to develop customized data-driven models for various slag systems,including CaO-Al_(2)O_(3)-SiO_(2),CaO-Al_(2)O_(3)-CaF_(2),CaO-Al_(2)O_(3)-SiO_(2)-MgO,and CaO-Al_(2)O_(3)-SiO_(2)-MgO-CaF_(2).By incorporating optical basicity as a key feature,the models achieve an average validation error of 8.0%to 15.1%,significantly outperforming traditional empirical models.Additionally,symbolic regression is introduced to rapidly construct domain-specific features,such as optical basicity-like descriptors,offering a potential breakthrough in performance prediction for small datasets.This work highlights the critical role of domain-specific knowledge in understanding and predicting viscosity,providing a robust machine learning-based approach for optimizing refining slag properties.展开更多
Knowledge distillation has become a standard technique for compressing large language models into efficient student models,but existing methods often struggle to balance prediction accuracy with explanation quality.Re...Knowledge distillation has become a standard technique for compressing large language models into efficient student models,but existing methods often struggle to balance prediction accuracy with explanation quality.Recent approaches such as Distilling Step-by-Step(DSbS)introduce explanation supervision,yet they apply it in a uniform manner that may not fully exploit the different learning dynamics of prediction and explanation.In this work,we propose a task-structured curriculum learning(TSCL)framework that structures training into three sequential phases:(i)prediction-only,to establish stable feature representations;(ii)joint prediction-explanation,to align task outputs with rationale generation;and(iii)explanation-only,to refine the quality of rationales.This design provides a simple but effective modification to DSbS,requiring no architectural changes and adding negligible training cost.We justify the phase scheduling with ablation studies and convergence analysis,showing that an initial prediction-heavy stage followed by a balanced joint phase improves both stability and explanation alignment.Extensive experiments on five datasets(e-SNLI,ANLI,CommonsenseQA,SVAMP,and MedNLI)demonstrate that TSCL consistently outperforms strong baselines,achieving gains of+1.7-2.6 points in accuracy and 0.8-1.2 in ROUGE-L,corresponding to relative error reductions of up to 21%.Beyond lexical metrics,human evaluation and ERASERstyle faithfulness diagnostics confirm that TSCL produces more faithful and informative explanations.Comparative training curves further reveal faster convergence and lower variance across seeds.Efficiency analysis shows less than 3%overhead in wall-clock training time and no additional inference cost,making the approach practical for realworld deployment.This study demonstrates that a simple task-structured curriculum can significantly improve the effectiveness of knowledge distillation.By separating and sequencing objectives,TSCL achieves a better balance between accuracy,stability,and explanation quality.The framework generalizes across domains,including medical NLI,and offers a principled recipe for future applications in multimodal reasoning and reinforcement learning.展开更多
The reliable operation of power grid secondary equipment is an important guarantee for the safety and stability of the power system.However,various defects could be produced in the secondary equipment during longtermo...The reliable operation of power grid secondary equipment is an important guarantee for the safety and stability of the power system.However,various defects could be produced in the secondary equipment during longtermoperation.The complex relationship between the defect phenomenon andmulti-layer causes and the probabilistic influence of secondary equipment cannot be described through knowledge extraction and fusion technology by existing methods,which limits the real-time and accuracy of defect identification.Therefore,a defect recognition method based on the Bayesian network and knowledge graph fusion is proposed.The defect data of secondary equipment is transformed into the structured knowledge graph through knowledge extraction and fusion technology.The knowledge graph of power grid secondary equipment is mapped to the Bayesian network framework,combined with historical defect data,and introduced Noisy-OR nodes.The prior and conditional probabilities of the Bayesian network are then reasonably assigned to build a model that reflects the probability dependence between defect phenomena and potential causes in power grid secondary equipment.Defect identification of power grid secondary equipment is achieved by defect subgraph search based on the knowledge graph,and defect inference based on the Bayesian network.Practical application cases prove this method’s effectiveness in identifying secondary equipment defect causes,improving identification accuracy and efficiency.展开更多
With the large-scale deployment of the Internet ofThings(IoT)devices,their weak securitymechanisms make them prime targets for malware attacks.Attackers often use Domain Generation Algorithm(DGA)to generate random dom...With the large-scale deployment of the Internet ofThings(IoT)devices,their weak securitymechanisms make them prime targets for malware attacks.Attackers often use Domain Generation Algorithm(DGA)to generate random domain names,hiding the real IP of Command and Control(C&C)servers to build botnets.Due to the randomness and dynamics of DGA,traditional methods struggle to detect them accurately,increasing the difficulty of network defense.This paper proposes a lightweight DGA detection model based on knowledge distillation for resource-constrained IoT environments.Specifically,a teacher model combining CharacterBERT,a bidirectional long short-term memory(BiLSTM)network,and attention mechanism(ATT)is constructed:it extracts character-level semantic features viaCharacterBERT,captures sequence dependencieswith the BiLSTM,and integrates theATT for key feature weighting,formingmulti-granularity feature fusion.An improved knowledge distillation approach transfers the teacher model’s learned knowledge to the simplified DistilBERT student model.Experimental results show the teacher model achieves 98.68%detection accuracy.The student modelmaintains slightly improved accuracy while significantly compressing parameters to approximately 38.4%of the teacher model’s scale,greatly reducing computational overhead for IoT deployment.展开更多
The micro-riblet structures have been demonstrated effective in controlling the Total Pressure Loss(TPL)of aero-engine blades.However,due to the considerable scale gap between micro-texture and an actual aero-engine b...The micro-riblet structures have been demonstrated effective in controlling the Total Pressure Loss(TPL)of aero-engine blades.However,due to the considerable scale gap between micro-texture and an actual aero-engine blade,wind tunnel tests and numerical simulations with massive grids directly describing the global flow field are costly for aerodynamic evaluation.Furthermore,the fine micro surface structure brings unavoidable manufacturing errors,and the probability prediction contributes to gaining the confidence interval of the results.Therefore,a novel relay-based probabilistic model for multi-fidelity scenarios in the TPL prediction of a compressor cascade with micro-riblet surfaces is proposed to trade off accuracy and efficiency.Combined with the low-fidelity flow data generated by an aerodynamic solution strategy using the boundary surrogate model and the high-fidelity flow data from the experiment,the relay-based modeling has been achieved through knowledge transferring,and the confidence interval can be provided by the Gaussian Process Regression(GPR)model.The TPL of compressor cascades with micro-riblet surfaces under different surface structures at March number Ma=0.64,0.74,0.84 have been evaluated using the Relay-Based Probabilistic(RBP)model.The results illustrate that the RBP model could provide higher accuracy than the Single-Fidelity-Data-Driven(SFDD)prediction model,which show the promising potential of multi-fidelity scenarios data fusion in the aerodynamic evaluation of multi-scale configurations.展开更多
Modern intelligent systems,such as autonomous vehicles and face recognition,must continuously adapt to new scenarios while preserving their ability to handle previously encountered situations.However,when neural netwo...Modern intelligent systems,such as autonomous vehicles and face recognition,must continuously adapt to new scenarios while preserving their ability to handle previously encountered situations.However,when neural networks learn new classes sequentially,they suffer from catastrophic forgetting—the tendency to lose knowledge of earlier classes.This challenge,which lies at the core of class-incremental learning,severely limits the deployment of continual learning systems in real-world applications with streaming data.Existing approaches,including rehearsalbased methods and knowledge distillation techniques,have attempted to address this issue but often struggle to effectively preserve decision boundaries and discriminative features under limited memory constraints.To overcome these limitations,we propose a support vector-guided framework for class-incremental learning.The framework integrates an enhanced feature extractor with a Support Vector Machine classifier,which generates boundary-critical support vectors to guide both replay and distillation.Building on this architecture,we design a joint feature retention strategy that combines boundary proximity with feature diversity,and a Support Vector Distillation Loss that enforces dual alignment in decision and semantic spaces.In addition,triple attention modules are incorporated into the feature extractor to enhance representation power.Extensive experiments on CIFAR-100 and Tiny-ImageNet demonstrate effective improvements.On CIFAR-100 and Tiny-ImageNet with 5 tasks,our method achieves 71.68%and 58.61%average accuracy,outperforming strong baselines by 3.34%and 2.05%.These advantages are consistently observed across different task splits,highlighting the robustness and generalization of the proposed approach.Beyond benchmark evaluations,the framework also shows potential in few-shot and resource-constrained applications such as edge computing and mobile robotics.展开更多
This paper delves into effective pathways for transforming course ecosystems from resource provision to knowledge service and competency development through university-enterprise collaboration in co-building knowledge...This paper delves into effective pathways for transforming course ecosystems from resource provision to knowledge service and competency development through university-enterprise collaboration in co-building knowledge graphs and intelligent shared courses.This approach enables personalized,learning-driven teaching.Based on knowledge graphs and integrated teacher-machine-student smart teaching scenarios,it not only innovates autonomous learning environments and human-computer interaction models while optimizing teaching experiences for both instructors and students,but also effectively addresses the issues of students’“scattered,superficial,and fragmented learning”.This establishes the foundation for personalized teaching tailored to individual aptitudes.展开更多
Aiming at the problem of insufficient recognition of implicit variants by existing Chinese sensitive text detection methods,this paper proposes the IPKE-MoE framework,which consists of three parts,namely,a sensitive w...Aiming at the problem of insufficient recognition of implicit variants by existing Chinese sensitive text detection methods,this paper proposes the IPKE-MoE framework,which consists of three parts,namely,a sensitive word variant extraction framework,a sensitive word variant knowledge enhancement layer and a mixture-of-experts(MoE)classification layer.First,sensitive word variants are precisely extracted through dynamic iterative prompt templates and the context-aware capabilities of Large Language Models(LLMs).Next,the extracted variants are used to construct a knowledge enhancement layer for sensitive word variants based on RoCBert models.Specifically,after locating variants via n-gram algorithms,variant types are mapped to embedding vectors and fused with original word vectors.Finally,a mixture-of-experts(MoE)classification layer is designed(sensitive word,sentiment,and semantic experts),which decouples the relationship between sensitiveword existence and text toxicity throughmultiple experts.This framework effectively combines the comprehension ability of Large Language Models(LLMs)with the discriminative ability of smaller models.Our two experiments demonstrate that the sensitive word variant extraction framework based on dynamically iterated prompt templates outperforms other baseline prompt templates.TheRoCBert models incorporating the sensitive word variant knowledge enhancement layer and a mixture-of-experts(MoE)classification layer achieve superior classification performance compared to other baselines.展开更多
In a wide range of engineering applications,complex constrained multi-objective optimization problems(CMOPs)present significant challenges,as the complexity of constraints often hampers algorithmic convergence and red...In a wide range of engineering applications,complex constrained multi-objective optimization problems(CMOPs)present significant challenges,as the complexity of constraints often hampers algorithmic convergence and reduces population diversity.To address these challenges,we propose a novel algorithm named Constraint IntensityDriven Evolutionary Multitasking(CIDEMT),which employs a two-stage,tri-task framework to dynamically integrates problem structure and knowledge transfer.In the first stage,three cooperative tasks are designed to explore the Constrained Pareto Front(CPF),the Unconstrained Pareto Front(UPF),and theε-relaxed constraint boundary,respectively.A CPF-UPF relationship classifier is employed to construct a problem-type-aware evolutionary strategy pool.At the end of the first stage,each task selects strategies from this strategy pool based on the specific type of problem,thereby guiding the subsequent evolutionary process.In the second stage,while each task continues to evolve,aτ-driven knowledge transfer mechanism is introduced to selectively incorporate effective solutions across tasks.enhancing the convergence and feasibility of the main task.Extensive experiments conducted on 32 benchmark problems from three test suites(LIRCMOP,DASCMOP,and DOC)demonstrate that CIDEMT achieves the best Inverted Generational Distance(IGD)values on 24 problems and the best Hypervolume values(HV)on 22 problems.Furthermore,CIDEMT significantly outperforms six state-of-the-art constrained multi-objective evolutionary algorithms(CMOEAs).These results confirm CIDEMT’s superiority in promoting convergence,diversity,and robustness in solving complex CMOPs.展开更多
With the widespread use of social media,the propagation of health-related rumors has become a significant public health threat.Existing methods for detecting health rumors predominantly rely on external knowledge or p...With the widespread use of social media,the propagation of health-related rumors has become a significant public health threat.Existing methods for detecting health rumors predominantly rely on external knowledge or propagation structures,with only a few recent approaches attempting causal inference;however,these have not yet effectively integrated causal discovery with domain-specific knowledge graphs for detecting health rumors.In this study,we found that the combined use of causal discovery and domain-specific knowledge graphs can effectively identify implicit pseudo-causal logic embedded within texts,holding significant potential for health rumor detection.To this end,we propose CKDG—a dual-graph fusion framework based on causal logic and medical knowledge graphs.CKDG constructs a weighted causal graph to capture the implicit causal relationships in the text and introduces a medical knowledge graph to verify semantic consistency,thereby enhancing the ability to identify the misuse of professional terminology and pseudoscientific claims.In experiments conducted on a dataset comprising 8430 health rumors,CKDG achieved an accuracy of 91.28%and an F1 score of 90.38%,representing improvements of 5.11%and 3.29%over the best baseline,respectively.Our results indicate that the integrated use of causal discovery and domainspecific knowledge graphs offers significant advantages for health rumor detection systems.This method not only improves detection performance but also enhances the transparency and credibility of model decisions by tracing causal chains and sources of knowledge conflicts.We anticipate that this work will provide key technological support for the development of trustworthy health-information filtering systems,thereby improving the reliability of public health information on social media.展开更多
Ensuring an information fabric safe is critical and mandatory.For its related Internet of Things(IoT)service system running on the open Internet,existing host-based monitoring methods may fail due to only inspecting s...Ensuring an information fabric safe is critical and mandatory.For its related Internet of Things(IoT)service system running on the open Internet,existing host-based monitoring methods may fail due to only inspecting software,and the physical system may not be able to be protected.In this paper,a nonintrusive virtual machine(VM)-based runtime protection framework is provided to protect the physical system with the isolated IoT services as a controlling means.Compared with existing solutions,the framework gets inconsistent and untrusted observation knowledge from multiple observation sources,and enforces property policies concurrently and incrementally in a competing-game way to avoid compositional problems.In addition,the monitoring is implemented without any modification to the protected system.Experiments are conducted to validate the proposed techniques.展开更多
Knowledge graphs,which combine structured representation with semantic modeling,have shown great potential in knowledge expression,causal inference,and automated reasoning,and are widely used in fields such as intelli...Knowledge graphs,which combine structured representation with semantic modeling,have shown great potential in knowledge expression,causal inference,and automated reasoning,and are widely used in fields such as intelligent question answering,decision support,and fault diagnosis.As high-speed train systems become increasingly intelligent and interconnected,fault patterns have grown more complex and dynamic.Knowledge graphs offer a promising solution to support the structured management and real-time reasoning of fault knowledge,addressing key requirements such as interpretability,accuracy,and continuous evolution in intelligent diagnostic systems.However,conventional knowledge graph construction relies heavily on domain expertise and specialized tools,resulting in high entry barriers for non-experts and limiting their practical application in frontline maintenance scenarios.To address this limitation,this paper proposes a fault knowledge modeling approach for high-speed trains that integrates structured logic diagrams with knowledge graphs.The method employs a seven-layer logic structure—comprising fault name,applicable vehicles,diagnostic logic,signal parameters,verification conditions,fault causes,and emergency measures—to transform unstructured knowledge into a visual and hierarchical representation.A semantic mapping mechanism is then used to automatically convert logic diagrams into machine-interpretable knowledge graphs,enabling dynamic reasoning and knowledge reuse.Furthermore,the proposed method establishes a three-layer architecture—logic structuring,knowledge graph transformation,and dynamic inference—to bridge human-expert logic with machinebased reasoning.Experimental validation and system implementation demonstrate that this approach not only improves knowledge interpretability and inference precision but also significantly enhances modeling efficiency and system maintainability.It provides a scalable and adaptable solution for intelligent operation and maintenance platforms in the high-speed rail domain.展开更多
基金partially supported by the National Natural Science Foundation (62272248)the Open Project Fund of State Key Laboratory of Computer Architecture,Institute of Computing Technology,Chinese Academy of Sciences (CARCHA202108,CARCH201905)+1 种基金the Natural Science Foundation of Tianjin (20JCZDJC00610)Sponsored by Zhejiang Lab (2021KF0AB04)。
文摘Smart contracts are widely used on the blockchain to implement complex transactions,such as decentralized applications on Ethereum.Effective vulnerability detection of large-scale smart contracts is critical,as attacks on smart contracts often cause huge economic losses.Since it is difficult to repair and update smart contracts,it is necessary to find the vulnerabilities before they are deployed.However,code analysis,which requires traversal paths,and learning methods,which require many features to be trained,are too time-consuming to detect large-scale on-chain contracts.Learning-based methods will obtain detection models from a feature space compared to code analysis methods such as symbol execution.But the existing features lack the interpretability of the detection results and training model,even worse,the large-scale feature space also affects the efficiency of detection.This paper focuses on improving the detection efficiency by reducing the dimension of the features,combined with expert knowledge.In this paper,a feature extraction model Block-gram is proposed to form low-dimensional knowledge-based features from bytecode.First,the metadata is separated and the runtime code is converted into a sequence of opcodes,which are divided into segments based on some instructions(jumps,etc.).Then,scalable Block-gram features,including 4-dimensional block features and 8-dimensional attribute features,are mined for the learning-based model training.Finally,feature contributions are calculated from SHAP values to measure the relationship between our features and the results of the detection model.In addition,six types of vulnerability labels are made on a dataset containing 33,885 contracts,and these knowledge-based features are evaluated using seven state-of-the-art learning algorithms,which show that the average detection latency speeds up 25×to 650×,compared with the features extracted by N-gram,and also can enhance the interpretability of the detection model.
文摘Model evaluation using benchmark datasets is an important method to measure the capability of large language models(LLMs)in specific domains,and it is mainly used to assess the knowledge and reasoning abilities of LLMs.Therefore,in order to better assess the capability of LLMs in the agricultural domain,Agri-Eval was proposed as a benchmark for assessing the knowledge and reasoning ability of LLMs in agriculture.The assessment dataset used in Agri-Eval covered seven major disciplines in the agricultural domain:crop science,horticulture,plant protection,animal husbandry,forest science,aquaculture science,and grass science,and contained a total of 2283 questions.Among domestic general-purpose LLMs,DeepSeek R1 performed best with an accuracy rate of 75.49%.In the realm of international general-purpose LLMs,Gemini 2.0 pro exp 0205 standed out as the top performer,achieving an accuracy rate of 74.28%.As an LLMs in agriculture vertical,Shennong V2.0 outperformed all the LLMs in China,and the answer accuracy rate of agricultural knowledge exceeded that of all the existing general-purpose LLMs.The launch of Agri-Eval helped the LLM developers to comprehensively evaluate the model's capability in the field of agriculture through a variety of tasks and tests to promote the development of the LLMs in the field of agriculture.
文摘1.《工程力学》是中国科学技术协会主管、中国力学学会主办、清华大学土木工程系承办的学术期刊。2.本刊为Ei Compendex、Scopus及ISI Web of Knowledge文摘数据库收录期刊。万方数据库、中国知网、重庆维普、超星、EBSCO数据库、Euro Pub数据库、JST数据库、ICI World of Journals全文检索。3.本刊于2019年和2024年两度入选“中国科技期刊卓越行动计划梯队项目”。4.本刊于2018年11月第三次荣获中国科协“精品科技期刊工程项目”资助。
基金supported by the National Natural Science Foundation of China(No.62267005)the Chinese Guangxi Natural Science Foundation(No.2023GXNSFAA026493)+1 种基金Guangxi Collaborative Innovation Center ofMulti-Source Information Integration and Intelligent ProcessingGuangxi Academy of Artificial Intelligence.
文摘The rapid development of information technology and accelerated digitalization have led to an explosive growth of data across various fields.As a key technology for knowledge representation and sharing,knowledge graphs play a crucial role by constructing structured networks of relationships among entities.However,data sparsity and numerous unexplored implicit relations result in the widespread incompleteness of knowledge graphs.In static knowledge graph completion,most existing methods rely on linear operations or simple interaction mechanisms for triple encoding,making it difficult to fully capture the deep semantic associations between entities and relations.Moreover,many methods focus only on the local information of individual triples,ignoring the rich semantic dependencies embedded in the neighboring nodes of entities within the graph structure,which leads to incomplete embedding representations.To address these challenges,we propose Two-Stage Mixer Embedding(TSMixerE),a static knowledge graph completion method based on entity context.In the unit semantic extraction stage,TSMixerE leveragesmulti-scale circular convolution to capture local features atmultiple granularities,enhancing the flexibility and robustness of feature interactions.A channel attention mechanism amplifies key channel responses to suppress noise and irrelevant information,thereby improving the discriminative power and semantic depth of feature representations.For contextual information fusion,a multi-layer self-attentionmechanism enables deep interactions among contextual cues,effectively integrating local details with global context.Simultaneously,type embeddings clarify the semantic identities and roles of each component,enhancing the model’s sensitivity and fusion capabilities for diverse information sources.Furthermore,TSMixerE constructs contextual unit sequences for entities,fully exploring neighborhood information within the graph structure to model complex semantic dependencies,thus improving the completeness and generalization of embedding representations.
文摘Since Google introduced the concept of Knowledge Graphs(KGs)in 2012,their construction technologies have evolved into a comprehensive methodological framework encompassing knowledge acquisition,extraction,representation,modeling,fusion,computation,and storage.Within this framework,knowledge extraction,as the core component,directly determines KG quality.In military domains,traditional manual curation models face efficiency constraints due to data fragmentation,complex knowledge architectures,and confidentiality protocols.Meanwhile,crowdsourced ontology construction approaches from general domains prove non-transferable,while human-crafted ontologies struggle with generalization deficiencies.To address these challenges,this study proposes an OntologyAware LLM Methodology for Military Domain Knowledge Extraction(LLM-KE).This approach leverages the deep semantic comprehension capabilities of Large Language Models(LLMs)to simulate human experts’cognitive processes in crowdsourced ontology construction,enabling automated extraction of military textual knowledge.It concurrently enhances knowledge processing efficiency and improves KG completeness.Empirical analysis demonstrates that this method effectively resolves scalability and dynamic adaptation challenges in military KG construction,establishing a novel technological pathway for advancing military intelligence development.
基金supported by the National Natural Science Foundation of China (Grant Nos.62125301,62021003,62303024,U24A20275,62522302,62473011,92467205)the National Key Research and Development Project (Grant Nos.2022YFB3305800-5,2024YFE0212400)+2 种基金the Youth Beijing Scholars Program (Grant No.037)the Beijing Nova Program (Grant Nos.20240484694,20250484938)the Beijing Natural Science Foundation (Grant No.L253010)。
文摘Active fault-tolerant control utilizes information obtained from fault diagnosis to reconfigure the control law to compensate for faults in the wastewater treatment process. However, since the similarity of fault characteristic in the incipient stage can result in misdiagnosis, it is a challenge for fault-tolerant control to ensure system safety and reliability. Therefore, to address this issue, a fault diagnosis and fault-tolerant control with a knowledge transfer strategy(KT-FDFTC) is proposed in this paper. First, a knowledge reasoning diagnosis strategy using multi-source transfer learning is designed to distinguish the similar characteristic of incipient faults. Then, the multi-source knowledge can assist in the diagnosis strategy to strengthen the fault information for fault-tolerant control. Second, a knowledge adaptive compensation mechanism, which makes knowledge and data coupled into the output trajectory regarded as an objective function, is employed to dynamically compute the control law. Then, KT-FDFTC can ensure the stable operation to adapt to various fault conditions. Third, the Lyapunov function is established to demonstrate the stability of KT-FDFTC. Then, the theoretical basis can offer the successful application of KTFDFTC. Finally, the proposed method is validated through a real WWTP and a simulation platform. The experimental results confirm that KT-FDFTC can provide good diagnosis performance and fault tolerance ability.
基金supported by the National Key R&D Program of China(Grant No.2023YFC3209504)Natural Science Foundation of Wuhan(Grant No.2024040801020271)the Fundamental Research Funds for Central Public Welfare Research Institutes(Grant No.CKSF2025718/YT).
文摘Wetting deformation in earth-rockfill dams is a critical factor influencingdam safety.Although numerous mathematical models have been developed to describe this phenomenon,most of them rely on empirical formulations and lack prior knowledge of model parameters,which is essential for Bayesian parameter inversion to enhance accuracy and reduce uncertainty.This study introduces a datadriven approach to establishing prior knowledge of earth-rockfill dams.Driving factors are utilized to determine the potential range of model parameters,and settlement changes within this range are calculated.The results are iteratively compared with actual monitoring data until the calculated range encompasses the observed data,thereby providing prior knowledge of the model parameters.The proposed method is applied to the right-bank earth-rockfilldam of Danjiangkou.Employing a Gibbs sample size of 30,000,the proposed method effectively calibrates the prior knowledge of the wetting model parameters,achieving a root mean square error(RMSE)of 5.18 mm for the settlement predictions.By comparison,the use of non-informative priors with sample sizes of 30,000 and 50,000 results in significantly larger RMSE values of 11.97 mm and 16.07 mm,respectively.Furthermore,the computational efficiencyof the proposed method is demonstrated by an inversion computation time of 902 s for 30,000 samples,which is notably shorter than the 1026 s and 1558 s required for noninformative priors with 30,000 and 50,000 samples,respectively.These findingsunderscore the superior performance of the proposed approach in terms of both prediction accuracy and computational efficiency.These results demonstrate that the proposed method not only improves the predictive accuracy but also enhances the computational efficiency,enabling optimal parameter identificationwith reduced computational effort.This approach provides a robust and efficientframework for advancing dam safety assessments.
基金supported by Basic Science Research Program through the National Research Foundation of Korea(NRF)funded by the Ministry of Education(RS-2023-00245084)by Korea Institute for Advancement of Technology(KIAT)grant funded by the Korea Government(MOTIE)(RS-2024-00415938,HRD Program for Industrial Innovation)and Soonchunhyang University.
文摘This paper proposes a deep learning-based 3D LiDAR perception framework designed for applications such as autonomous robots and vehicles.To address the high dependency on large-scale annotated data—an inherent limitation of deep learning models—this study introduces a hybrid perception architecture that incorporates expertdriven LiDAR processing techniques into the deep neural network.Traditional 3DLiDAR processingmethods typically remove ground planes and apply distance-or density-based clustering for object detection.In this work,such expert knowledge is encoded as feature-level inputs and fused with the deep network,therebymitigating the data dependency issue of conventional learning-based approaches.Specifically,the proposedmethod combines two expert algorithms—Patchwork++for ground segmentation and DBSCAN for clustering—with a PointPillars-based LiDAR detection network.We design four hybrid versions of the network depending on the stage and method of integrating expert features into the feature map of the deep model.Among these,Version 4 incorporates a modified neck structure in PointPillars and introduces a new Cluster 2D Pseudo-Map Branch that utilizes cluster-level pseudo-images generated from Patchwork++and DBSCAN.This version achieved a+3.88%improvement mean Average Precision(mAP)compared to the baseline PointPillars.The results demonstrate that embedding expert-based perception logic into deep neural architectures can effectively enhance performance and reduce dependency on extensive training datasets,offering a promising direction for robust 3D LiDAR object detection in real-world scenarios.
基金supported by the National Key Research and Development Program of China(No.2023YFB3712401),the National Natural Science Foundation of China(No.52274301)the Aeronautical Science Foundation of China(No.2023Z0530S6005)the Ningbo Yongjiang Talent-Introduction Programme(No.2022A-023-C).
文摘The viscosity of refining slags plays a critical role in metallurgical processes.However,obtaining accurate viscosity data remains challenging due to the complexities of high-temperature experiments,often relying on empirical models with limited predictive capabilities.This study focuses on the influence of optical basicity on viscosity in CaO-Al_(2)O_(3)-based refining slags,leveraging machine learning to address data scarcity and improve prediction accuracy.An automated framework for algorithm integration,parameter tuning,and evaluation ranking framework(Auto-APE)is employed to develop customized data-driven models for various slag systems,including CaO-Al_(2)O_(3)-SiO_(2),CaO-Al_(2)O_(3)-CaF_(2),CaO-Al_(2)O_(3)-SiO_(2)-MgO,and CaO-Al_(2)O_(3)-SiO_(2)-MgO-CaF_(2).By incorporating optical basicity as a key feature,the models achieve an average validation error of 8.0%to 15.1%,significantly outperforming traditional empirical models.Additionally,symbolic regression is introduced to rapidly construct domain-specific features,such as optical basicity-like descriptors,offering a potential breakthrough in performance prediction for small datasets.This work highlights the critical role of domain-specific knowledge in understanding and predicting viscosity,providing a robust machine learning-based approach for optimizing refining slag properties.
文摘Knowledge distillation has become a standard technique for compressing large language models into efficient student models,but existing methods often struggle to balance prediction accuracy with explanation quality.Recent approaches such as Distilling Step-by-Step(DSbS)introduce explanation supervision,yet they apply it in a uniform manner that may not fully exploit the different learning dynamics of prediction and explanation.In this work,we propose a task-structured curriculum learning(TSCL)framework that structures training into three sequential phases:(i)prediction-only,to establish stable feature representations;(ii)joint prediction-explanation,to align task outputs with rationale generation;and(iii)explanation-only,to refine the quality of rationales.This design provides a simple but effective modification to DSbS,requiring no architectural changes and adding negligible training cost.We justify the phase scheduling with ablation studies and convergence analysis,showing that an initial prediction-heavy stage followed by a balanced joint phase improves both stability and explanation alignment.Extensive experiments on five datasets(e-SNLI,ANLI,CommonsenseQA,SVAMP,and MedNLI)demonstrate that TSCL consistently outperforms strong baselines,achieving gains of+1.7-2.6 points in accuracy and 0.8-1.2 in ROUGE-L,corresponding to relative error reductions of up to 21%.Beyond lexical metrics,human evaluation and ERASERstyle faithfulness diagnostics confirm that TSCL produces more faithful and informative explanations.Comparative training curves further reveal faster convergence and lower variance across seeds.Efficiency analysis shows less than 3%overhead in wall-clock training time and no additional inference cost,making the approach practical for realworld deployment.This study demonstrates that a simple task-structured curriculum can significantly improve the effectiveness of knowledge distillation.By separating and sequencing objectives,TSCL achieves a better balance between accuracy,stability,and explanation quality.The framework generalizes across domains,including medical NLI,and offers a principled recipe for future applications in multimodal reasoning and reinforcement learning.
基金supported by the State Grid Southwest Branch Project“Research on Defect Diagnosis and Early Warning Technology of Relay Protection and Safety Automation Devices Based on Multi-Source Heterogeneous Defect Data”.
文摘The reliable operation of power grid secondary equipment is an important guarantee for the safety and stability of the power system.However,various defects could be produced in the secondary equipment during longtermoperation.The complex relationship between the defect phenomenon andmulti-layer causes and the probabilistic influence of secondary equipment cannot be described through knowledge extraction and fusion technology by existing methods,which limits the real-time and accuracy of defect identification.Therefore,a defect recognition method based on the Bayesian network and knowledge graph fusion is proposed.The defect data of secondary equipment is transformed into the structured knowledge graph through knowledge extraction and fusion technology.The knowledge graph of power grid secondary equipment is mapped to the Bayesian network framework,combined with historical defect data,and introduced Noisy-OR nodes.The prior and conditional probabilities of the Bayesian network are then reasonably assigned to build a model that reflects the probability dependence between defect phenomena and potential causes in power grid secondary equipment.Defect identification of power grid secondary equipment is achieved by defect subgraph search based on the knowledge graph,and defect inference based on the Bayesian network.Practical application cases prove this method’s effectiveness in identifying secondary equipment defect causes,improving identification accuracy and efficiency.
基金supported by the following projects:National Natural Science Foundation of China(62461041)Natural Science Foundation of Jiangxi Province China(20242BAB25068).
文摘With the large-scale deployment of the Internet ofThings(IoT)devices,their weak securitymechanisms make them prime targets for malware attacks.Attackers often use Domain Generation Algorithm(DGA)to generate random domain names,hiding the real IP of Command and Control(C&C)servers to build botnets.Due to the randomness and dynamics of DGA,traditional methods struggle to detect them accurately,increasing the difficulty of network defense.This paper proposes a lightweight DGA detection model based on knowledge distillation for resource-constrained IoT environments.Specifically,a teacher model combining CharacterBERT,a bidirectional long short-term memory(BiLSTM)network,and attention mechanism(ATT)is constructed:it extracts character-level semantic features viaCharacterBERT,captures sequence dependencieswith the BiLSTM,and integrates theATT for key feature weighting,formingmulti-granularity feature fusion.An improved knowledge distillation approach transfers the teacher model’s learned knowledge to the simplified DistilBERT student model.Experimental results show the teacher model achieves 98.68%detection accuracy.The student modelmaintains slightly improved accuracy while significantly compressing parameters to approximately 38.4%of the teacher model’s scale,greatly reducing computational overhead for IoT deployment.
基金supported by the National Natural Science Foundation of China(No.12301672)the Shanghai Science and Technology Innovation Action Plan(Yangfan Special Project),China(No.23YF1401300)。
文摘The micro-riblet structures have been demonstrated effective in controlling the Total Pressure Loss(TPL)of aero-engine blades.However,due to the considerable scale gap between micro-texture and an actual aero-engine blade,wind tunnel tests and numerical simulations with massive grids directly describing the global flow field are costly for aerodynamic evaluation.Furthermore,the fine micro surface structure brings unavoidable manufacturing errors,and the probability prediction contributes to gaining the confidence interval of the results.Therefore,a novel relay-based probabilistic model for multi-fidelity scenarios in the TPL prediction of a compressor cascade with micro-riblet surfaces is proposed to trade off accuracy and efficiency.Combined with the low-fidelity flow data generated by an aerodynamic solution strategy using the boundary surrogate model and the high-fidelity flow data from the experiment,the relay-based modeling has been achieved through knowledge transferring,and the confidence interval can be provided by the Gaussian Process Regression(GPR)model.The TPL of compressor cascades with micro-riblet surfaces under different surface structures at March number Ma=0.64,0.74,0.84 have been evaluated using the Relay-Based Probabilistic(RBP)model.The results illustrate that the RBP model could provide higher accuracy than the Single-Fidelity-Data-Driven(SFDD)prediction model,which show the promising potential of multi-fidelity scenarios data fusion in the aerodynamic evaluation of multi-scale configurations.
基金supported by the Gansu Provincial Natural Science Foundation(grant number 25JRRA074)the Gansu Provincial Key R&D Science and Technology Program(grant number 24YFGA060)the National Natural Science Foundation of China(grant number 62161019).
文摘Modern intelligent systems,such as autonomous vehicles and face recognition,must continuously adapt to new scenarios while preserving their ability to handle previously encountered situations.However,when neural networks learn new classes sequentially,they suffer from catastrophic forgetting—the tendency to lose knowledge of earlier classes.This challenge,which lies at the core of class-incremental learning,severely limits the deployment of continual learning systems in real-world applications with streaming data.Existing approaches,including rehearsalbased methods and knowledge distillation techniques,have attempted to address this issue but often struggle to effectively preserve decision boundaries and discriminative features under limited memory constraints.To overcome these limitations,we propose a support vector-guided framework for class-incremental learning.The framework integrates an enhanced feature extractor with a Support Vector Machine classifier,which generates boundary-critical support vectors to guide both replay and distillation.Building on this architecture,we design a joint feature retention strategy that combines boundary proximity with feature diversity,and a Support Vector Distillation Loss that enforces dual alignment in decision and semantic spaces.In addition,triple attention modules are incorporated into the feature extractor to enhance representation power.Extensive experiments on CIFAR-100 and Tiny-ImageNet demonstrate effective improvements.On CIFAR-100 and Tiny-ImageNet with 5 tasks,our method achieves 71.68%and 58.61%average accuracy,outperforming strong baselines by 3.34%and 2.05%.These advantages are consistently observed across different task splits,highlighting the robustness and generalization of the proposed approach.Beyond benchmark evaluations,the framework also shows potential in few-shot and resource-constrained applications such as edge computing and mobile robotics.
基金supported by Harbin Institute of Technology High-level Teaching Achievement Award(National Level)Cultivation Project(256709).
文摘This paper delves into effective pathways for transforming course ecosystems from resource provision to knowledge service and competency development through university-enterprise collaboration in co-building knowledge graphs and intelligent shared courses.This approach enables personalized,learning-driven teaching.Based on knowledge graphs and integrated teacher-machine-student smart teaching scenarios,it not only innovates autonomous learning environments and human-computer interaction models while optimizing teaching experiences for both instructors and students,but also effectively addresses the issues of students’“scattered,superficial,and fragmented learning”.This establishes the foundation for personalized teaching tailored to individual aptitudes.
基金funded by the National Natural Science Foundation of China(Grant No.62441212)the Major Project of the Natural Science Foundation of Inner Mongolia(Grant No.2025ZD008).
文摘Aiming at the problem of insufficient recognition of implicit variants by existing Chinese sensitive text detection methods,this paper proposes the IPKE-MoE framework,which consists of three parts,namely,a sensitive word variant extraction framework,a sensitive word variant knowledge enhancement layer and a mixture-of-experts(MoE)classification layer.First,sensitive word variants are precisely extracted through dynamic iterative prompt templates and the context-aware capabilities of Large Language Models(LLMs).Next,the extracted variants are used to construct a knowledge enhancement layer for sensitive word variants based on RoCBert models.Specifically,after locating variants via n-gram algorithms,variant types are mapped to embedding vectors and fused with original word vectors.Finally,a mixture-of-experts(MoE)classification layer is designed(sensitive word,sentiment,and semantic experts),which decouples the relationship between sensitiveword existence and text toxicity throughmultiple experts.This framework effectively combines the comprehension ability of Large Language Models(LLMs)with the discriminative ability of smaller models.Our two experiments demonstrate that the sensitive word variant extraction framework based on dynamically iterated prompt templates outperforms other baseline prompt templates.TheRoCBert models incorporating the sensitive word variant knowledge enhancement layer and a mixture-of-experts(MoE)classification layer achieve superior classification performance compared to other baselines.
基金supported by the National Natural Science Foundation of China under Grant No.61972040the Science and Technology Research and Development Project funded by China Railway Material Trade Group Luban Company.
文摘In a wide range of engineering applications,complex constrained multi-objective optimization problems(CMOPs)present significant challenges,as the complexity of constraints often hampers algorithmic convergence and reduces population diversity.To address these challenges,we propose a novel algorithm named Constraint IntensityDriven Evolutionary Multitasking(CIDEMT),which employs a two-stage,tri-task framework to dynamically integrates problem structure and knowledge transfer.In the first stage,three cooperative tasks are designed to explore the Constrained Pareto Front(CPF),the Unconstrained Pareto Front(UPF),and theε-relaxed constraint boundary,respectively.A CPF-UPF relationship classifier is employed to construct a problem-type-aware evolutionary strategy pool.At the end of the first stage,each task selects strategies from this strategy pool based on the specific type of problem,thereby guiding the subsequent evolutionary process.In the second stage,while each task continues to evolve,aτ-driven knowledge transfer mechanism is introduced to selectively incorporate effective solutions across tasks.enhancing the convergence and feasibility of the main task.Extensive experiments conducted on 32 benchmark problems from three test suites(LIRCMOP,DASCMOP,and DOC)demonstrate that CIDEMT achieves the best Inverted Generational Distance(IGD)values on 24 problems and the best Hypervolume values(HV)on 22 problems.Furthermore,CIDEMT significantly outperforms six state-of-the-art constrained multi-objective evolutionary algorithms(CMOEAs).These results confirm CIDEMT’s superiority in promoting convergence,diversity,and robustness in solving complex CMOPs.
基金funded by the Hunan Provincial Natural Science Foundation of China(Grant No.2025JJ70105)the Hunan Provincial College Students’Innovation and Entrepreneurship Training Program(Project No.S202411342056)The article processing charge(APC)was funded by the Project No.2025JJ70105.
文摘With the widespread use of social media,the propagation of health-related rumors has become a significant public health threat.Existing methods for detecting health rumors predominantly rely on external knowledge or propagation structures,with only a few recent approaches attempting causal inference;however,these have not yet effectively integrated causal discovery with domain-specific knowledge graphs for detecting health rumors.In this study,we found that the combined use of causal discovery and domain-specific knowledge graphs can effectively identify implicit pseudo-causal logic embedded within texts,holding significant potential for health rumor detection.To this end,we propose CKDG—a dual-graph fusion framework based on causal logic and medical knowledge graphs.CKDG constructs a weighted causal graph to capture the implicit causal relationships in the text and introduces a medical knowledge graph to verify semantic consistency,thereby enhancing the ability to identify the misuse of professional terminology and pseudoscientific claims.In experiments conducted on a dataset comprising 8430 health rumors,CKDG achieved an accuracy of 91.28%and an F1 score of 90.38%,representing improvements of 5.11%and 3.29%over the best baseline,respectively.Our results indicate that the integrated use of causal discovery and domainspecific knowledge graphs offers significant advantages for health rumor detection systems.This method not only improves detection performance but also enhances the transparency and credibility of model decisions by tracing causal chains and sources of knowledge conflicts.We anticipate that this work will provide key technological support for the development of trustworthy health-information filtering systems,thereby improving the reliability of public health information on social media.
基金supported by the National Key Research and Development Program of China under grant 2022YFF0902701the National Natural Science Foundation of China under grant U21A20468,61972043,61921003+1 种基金Zhejiang Lab under grant 2021PD0AB 02the Fundamental Research Funds for the Central Universities under grant 2020XD-A07-1.
文摘Ensuring an information fabric safe is critical and mandatory.For its related Internet of Things(IoT)service system running on the open Internet,existing host-based monitoring methods may fail due to only inspecting software,and the physical system may not be able to be protected.In this paper,a nonintrusive virtual machine(VM)-based runtime protection framework is provided to protect the physical system with the isolated IoT services as a controlling means.Compared with existing solutions,the framework gets inconsistent and untrusted observation knowledge from multiple observation sources,and enforces property policies concurrently and incrementally in a competing-game way to avoid compositional problems.In addition,the monitoring is implemented without any modification to the protected system.Experiments are conducted to validate the proposed techniques.
基金support from the Scientific Funding for the Center of National Railway Intelligent Transportation System Engineering and Technology,China Academy of Railway Sciences Corporation Limited(Grant No.2023YJ354)。
文摘Knowledge graphs,which combine structured representation with semantic modeling,have shown great potential in knowledge expression,causal inference,and automated reasoning,and are widely used in fields such as intelligent question answering,decision support,and fault diagnosis.As high-speed train systems become increasingly intelligent and interconnected,fault patterns have grown more complex and dynamic.Knowledge graphs offer a promising solution to support the structured management and real-time reasoning of fault knowledge,addressing key requirements such as interpretability,accuracy,and continuous evolution in intelligent diagnostic systems.However,conventional knowledge graph construction relies heavily on domain expertise and specialized tools,resulting in high entry barriers for non-experts and limiting their practical application in frontline maintenance scenarios.To address this limitation,this paper proposes a fault knowledge modeling approach for high-speed trains that integrates structured logic diagrams with knowledge graphs.The method employs a seven-layer logic structure—comprising fault name,applicable vehicles,diagnostic logic,signal parameters,verification conditions,fault causes,and emergency measures—to transform unstructured knowledge into a visual and hierarchical representation.A semantic mapping mechanism is then used to automatically convert logic diagrams into machine-interpretable knowledge graphs,enabling dynamic reasoning and knowledge reuse.Furthermore,the proposed method establishes a three-layer architecture—logic structuring,knowledge graph transformation,and dynamic inference—to bridge human-expert logic with machinebased reasoning.Experimental validation and system implementation demonstrate that this approach not only improves knowledge interpretability and inference precision but also significantly enhances modeling efficiency and system maintainability.It provides a scalable and adaptable solution for intelligent operation and maintenance platforms in the high-speed rail domain.