Bird's-eye-view(BEV)perception is a core technology for autonomous driving systems.However,existing solutions face the dilemma of high costs associated with multimodal methods and limited performance of vision-onl...Bird's-eye-view(BEV)perception is a core technology for autonomous driving systems.However,existing solutions face the dilemma of high costs associated with multimodal methods and limited performance of vision-only approaches.To address this issue,this paper proposes a framework named“a lightweight pure visual BEV perception method based on dual distillation of spatial-temporal knowledge”.This framework innovatively designs a lightweight vision-only student model based on Res Net,which leverages a dual distillation mechanism to learn from a powerful teacher model that integrates temporal information from both image and light detection and ranging(LiDAR)modalities.Specifically,we distill efficient multi-modal feature extraction and spatial fusion capabilities from the BEVFusion model,and distill advanced temporal information fusion and spatiotemporal attention mechanisms from the BEVFormer model.This dual distillation strategy enables the student model to achieve perception performance close to that of multi-modal models without relying on Li DAR.Experimental results on the nu Scenes dataset demonstrate that the proposed model significantly outperforms classical vision-only algorithms,achieves comparable performance to current state-of-the-art vision-only methods on the nu Scenes detection leaderboard in terms of both mean average precision(mAP)and the nu Scenes detection score(NDS)metrics,and exhibits notable advantages in inference computational efficiency.Although the proposed dual-teacher paradigm incurs higher offline training costs compared to single-model approaches,it yields a streamlined and highly efficient student model suitable for resource-constrained real-time deployment.This provides an effective pathway toward low-cost,high-performance autonomous driving perception systems.展开更多
Since Google introduced the concept of Knowledge Graphs(KGs)in 2012,their construction technologies have evolved into a comprehensive methodological framework encompassing knowledge acquisition,extraction,representati...Since Google introduced the concept of Knowledge Graphs(KGs)in 2012,their construction technologies have evolved into a comprehensive methodological framework encompassing knowledge acquisition,extraction,representation,modeling,fusion,computation,and storage.Within this framework,knowledge extraction,as the core component,directly determines KG quality.In military domains,traditional manual curation models face efficiency constraints due to data fragmentation,complex knowledge architectures,and confidentiality protocols.Meanwhile,crowdsourced ontology construction approaches from general domains prove non-transferable,while human-crafted ontologies struggle with generalization deficiencies.To address these challenges,this study proposes an OntologyAware LLM Methodology for Military Domain Knowledge Extraction(LLM-KE).This approach leverages the deep semantic comprehension capabilities of Large Language Models(LLMs)to simulate human experts’cognitive processes in crowdsourced ontology construction,enabling automated extraction of military textual knowledge.It concurrently enhances knowledge processing efficiency and improves KG completeness.Empirical analysis demonstrates that this method effectively resolves scalability and dynamic adaptation challenges in military KG construction,establishing a novel technological pathway for advancing military intelligence development.展开更多
The rapid development of information technology and accelerated digitalization have led to an explosive growth of data across various fields.As a key technology for knowledge representation and sharing,knowledge graph...The rapid development of information technology and accelerated digitalization have led to an explosive growth of data across various fields.As a key technology for knowledge representation and sharing,knowledge graphs play a crucial role by constructing structured networks of relationships among entities.However,data sparsity and numerous unexplored implicit relations result in the widespread incompleteness of knowledge graphs.In static knowledge graph completion,most existing methods rely on linear operations or simple interaction mechanisms for triple encoding,making it difficult to fully capture the deep semantic associations between entities and relations.Moreover,many methods focus only on the local information of individual triples,ignoring the rich semantic dependencies embedded in the neighboring nodes of entities within the graph structure,which leads to incomplete embedding representations.To address these challenges,we propose Two-Stage Mixer Embedding(TSMixerE),a static knowledge graph completion method based on entity context.In the unit semantic extraction stage,TSMixerE leveragesmulti-scale circular convolution to capture local features atmultiple granularities,enhancing the flexibility and robustness of feature interactions.A channel attention mechanism amplifies key channel responses to suppress noise and irrelevant information,thereby improving the discriminative power and semantic depth of feature representations.For contextual information fusion,a multi-layer self-attentionmechanism enables deep interactions among contextual cues,effectively integrating local details with global context.Simultaneously,type embeddings clarify the semantic identities and roles of each component,enhancing the model’s sensitivity and fusion capabilities for diverse information sources.Furthermore,TSMixerE constructs contextual unit sequences for entities,fully exploring neighborhood information within the graph structure to model complex semantic dependencies,thus improving the completeness and generalization of embedding representations.展开更多
Active fault-tolerant control utilizes information obtained from fault diagnosis to reconfigure the control law to compensate for faults in the wastewater treatment process. However, since the similarity of fault char...Active fault-tolerant control utilizes information obtained from fault diagnosis to reconfigure the control law to compensate for faults in the wastewater treatment process. However, since the similarity of fault characteristic in the incipient stage can result in misdiagnosis, it is a challenge for fault-tolerant control to ensure system safety and reliability. Therefore, to address this issue, a fault diagnosis and fault-tolerant control with a knowledge transfer strategy(KT-FDFTC) is proposed in this paper. First, a knowledge reasoning diagnosis strategy using multi-source transfer learning is designed to distinguish the similar characteristic of incipient faults. Then, the multi-source knowledge can assist in the diagnosis strategy to strengthen the fault information for fault-tolerant control. Second, a knowledge adaptive compensation mechanism, which makes knowledge and data coupled into the output trajectory regarded as an objective function, is employed to dynamically compute the control law. Then, KT-FDFTC can ensure the stable operation to adapt to various fault conditions. Third, the Lyapunov function is established to demonstrate the stability of KT-FDFTC. Then, the theoretical basis can offer the successful application of KTFDFTC. Finally, the proposed method is validated through a real WWTP and a simulation platform. The experimental results confirm that KT-FDFTC can provide good diagnosis performance and fault tolerance ability.展开更多
This paper delves into effective pathways for transforming course ecosystems from resource provision to knowledge service and competency development through university-enterprise collaboration in co-building knowledge...This paper delves into effective pathways for transforming course ecosystems from resource provision to knowledge service and competency development through university-enterprise collaboration in co-building knowledge graphs and intelligent shared courses.This approach enables personalized,learning-driven teaching.Based on knowledge graphs and integrated teacher-machine-student smart teaching scenarios,it not only innovates autonomous learning environments and human-computer interaction models while optimizing teaching experiences for both instructors and students,but also effectively addresses the issues of students’“scattered,superficial,and fragmented learning”.This establishes the foundation for personalized teaching tailored to individual aptitudes.展开更多
This paper proposes a deep learning-based 3D LiDAR perception framework designed for applications such as autonomous robots and vehicles.To address the high dependency on large-scale annotated data—an inherent limita...This paper proposes a deep learning-based 3D LiDAR perception framework designed for applications such as autonomous robots and vehicles.To address the high dependency on large-scale annotated data—an inherent limitation of deep learning models—this study introduces a hybrid perception architecture that incorporates expertdriven LiDAR processing techniques into the deep neural network.Traditional 3DLiDAR processingmethods typically remove ground planes and apply distance-or density-based clustering for object detection.In this work,such expert knowledge is encoded as feature-level inputs and fused with the deep network,therebymitigating the data dependency issue of conventional learning-based approaches.Specifically,the proposedmethod combines two expert algorithms—Patchwork++for ground segmentation and DBSCAN for clustering—with a PointPillars-based LiDAR detection network.We design four hybrid versions of the network depending on the stage and method of integrating expert features into the feature map of the deep model.Among these,Version 4 incorporates a modified neck structure in PointPillars and introduces a new Cluster 2D Pseudo-Map Branch that utilizes cluster-level pseudo-images generated from Patchwork++and DBSCAN.This version achieved a+3.88%improvement mean Average Precision(mAP)compared to the baseline PointPillars.The results demonstrate that embedding expert-based perception logic into deep neural architectures can effectively enhance performance and reduce dependency on extensive training datasets,offering a promising direction for robust 3D LiDAR object detection in real-world scenarios.展开更多
Wetting deformation in earth-rockfill dams is a critical factor influencingdam safety.Although numerous mathematical models have been developed to describe this phenomenon,most of them rely on empirical formulations a...Wetting deformation in earth-rockfill dams is a critical factor influencingdam safety.Although numerous mathematical models have been developed to describe this phenomenon,most of them rely on empirical formulations and lack prior knowledge of model parameters,which is essential for Bayesian parameter inversion to enhance accuracy and reduce uncertainty.This study introduces a datadriven approach to establishing prior knowledge of earth-rockfill dams.Driving factors are utilized to determine the potential range of model parameters,and settlement changes within this range are calculated.The results are iteratively compared with actual monitoring data until the calculated range encompasses the observed data,thereby providing prior knowledge of the model parameters.The proposed method is applied to the right-bank earth-rockfilldam of Danjiangkou.Employing a Gibbs sample size of 30,000,the proposed method effectively calibrates the prior knowledge of the wetting model parameters,achieving a root mean square error(RMSE)of 5.18 mm for the settlement predictions.By comparison,the use of non-informative priors with sample sizes of 30,000 and 50,000 results in significantly larger RMSE values of 11.97 mm and 16.07 mm,respectively.Furthermore,the computational efficiencyof the proposed method is demonstrated by an inversion computation time of 902 s for 30,000 samples,which is notably shorter than the 1026 s and 1558 s required for noninformative priors with 30,000 and 50,000 samples,respectively.These findingsunderscore the superior performance of the proposed approach in terms of both prediction accuracy and computational efficiency.These results demonstrate that the proposed method not only improves the predictive accuracy but also enhances the computational efficiency,enabling optimal parameter identificationwith reduced computational effort.This approach provides a robust and efficientframework for advancing dam safety assessments.展开更多
The viscosity of refining slags plays a critical role in metallurgical processes.However,obtaining accurate viscosity data remains challenging due to the complexities of high-temperature experiments,often relying on e...The viscosity of refining slags plays a critical role in metallurgical processes.However,obtaining accurate viscosity data remains challenging due to the complexities of high-temperature experiments,often relying on empirical models with limited predictive capabilities.This study focuses on the influence of optical basicity on viscosity in CaO-Al_(2)O_(3)-based refining slags,leveraging machine learning to address data scarcity and improve prediction accuracy.An automated framework for algorithm integration,parameter tuning,and evaluation ranking framework(Auto-APE)is employed to develop customized data-driven models for various slag systems,including CaO-Al_(2)O_(3)-SiO_(2),CaO-Al_(2)O_(3)-CaF_(2),CaO-Al_(2)O_(3)-SiO_(2)-MgO,and CaO-Al_(2)O_(3)-SiO_(2)-MgO-CaF_(2).By incorporating optical basicity as a key feature,the models achieve an average validation error of 8.0%to 15.1%,significantly outperforming traditional empirical models.Additionally,symbolic regression is introduced to rapidly construct domain-specific features,such as optical basicity-like descriptors,offering a potential breakthrough in performance prediction for small datasets.This work highlights the critical role of domain-specific knowledge in understanding and predicting viscosity,providing a robust machine learning-based approach for optimizing refining slag properties.展开更多
The reliable operation of power grid secondary equipment is an important guarantee for the safety and stability of the power system.However,various defects could be produced in the secondary equipment during longtermo...The reliable operation of power grid secondary equipment is an important guarantee for the safety and stability of the power system.However,various defects could be produced in the secondary equipment during longtermoperation.The complex relationship between the defect phenomenon andmulti-layer causes and the probabilistic influence of secondary equipment cannot be described through knowledge extraction and fusion technology by existing methods,which limits the real-time and accuracy of defect identification.Therefore,a defect recognition method based on the Bayesian network and knowledge graph fusion is proposed.The defect data of secondary equipment is transformed into the structured knowledge graph through knowledge extraction and fusion technology.The knowledge graph of power grid secondary equipment is mapped to the Bayesian network framework,combined with historical defect data,and introduced Noisy-OR nodes.The prior and conditional probabilities of the Bayesian network are then reasonably assigned to build a model that reflects the probability dependence between defect phenomena and potential causes in power grid secondary equipment.Defect identification of power grid secondary equipment is achieved by defect subgraph search based on the knowledge graph,and defect inference based on the Bayesian network.Practical application cases prove this method’s effectiveness in identifying secondary equipment defect causes,improving identification accuracy and efficiency.展开更多
Knowledge graphs,which combine structured representation with semantic modeling,have shown great potential in knowledge expression,causal inference,and automated reasoning,and are widely used in fields such as intelli...Knowledge graphs,which combine structured representation with semantic modeling,have shown great potential in knowledge expression,causal inference,and automated reasoning,and are widely used in fields such as intelligent question answering,decision support,and fault diagnosis.As high-speed train systems become increasingly intelligent and interconnected,fault patterns have grown more complex and dynamic.Knowledge graphs offer a promising solution to support the structured management and real-time reasoning of fault knowledge,addressing key requirements such as interpretability,accuracy,and continuous evolution in intelligent diagnostic systems.However,conventional knowledge graph construction relies heavily on domain expertise and specialized tools,resulting in high entry barriers for non-experts and limiting their practical application in frontline maintenance scenarios.To address this limitation,this paper proposes a fault knowledge modeling approach for high-speed trains that integrates structured logic diagrams with knowledge graphs.The method employs a seven-layer logic structure—comprising fault name,applicable vehicles,diagnostic logic,signal parameters,verification conditions,fault causes,and emergency measures—to transform unstructured knowledge into a visual and hierarchical representation.A semantic mapping mechanism is then used to automatically convert logic diagrams into machine-interpretable knowledge graphs,enabling dynamic reasoning and knowledge reuse.Furthermore,the proposed method establishes a three-layer architecture—logic structuring,knowledge graph transformation,and dynamic inference—to bridge human-expert logic with machinebased reasoning.Experimental validation and system implementation demonstrate that this approach not only improves knowledge interpretability and inference precision but also significantly enhances modeling efficiency and system maintainability.It provides a scalable and adaptable solution for intelligent operation and maintenance platforms in the high-speed rail domain.展开更多
Aiming at the problem of insufficient recognition of implicit variants by existing Chinese sensitive text detection methods,this paper proposes the IPKE-MoE framework,which consists of three parts,namely,a sensitive w...Aiming at the problem of insufficient recognition of implicit variants by existing Chinese sensitive text detection methods,this paper proposes the IPKE-MoE framework,which consists of three parts,namely,a sensitive word variant extraction framework,a sensitive word variant knowledge enhancement layer and a mixture-of-experts(MoE)classification layer.First,sensitive word variants are precisely extracted through dynamic iterative prompt templates and the context-aware capabilities of Large Language Models(LLMs).Next,the extracted variants are used to construct a knowledge enhancement layer for sensitive word variants based on RoCBert models.Specifically,after locating variants via n-gram algorithms,variant types are mapped to embedding vectors and fused with original word vectors.Finally,a mixture-of-experts(MoE)classification layer is designed(sensitive word,sentiment,and semantic experts),which decouples the relationship between sensitiveword existence and text toxicity throughmultiple experts.This framework effectively combines the comprehension ability of Large Language Models(LLMs)with the discriminative ability of smaller models.Our two experiments demonstrate that the sensitive word variant extraction framework based on dynamically iterated prompt templates outperforms other baseline prompt templates.TheRoCBert models incorporating the sensitive word variant knowledge enhancement layer and a mixture-of-experts(MoE)classification layer achieve superior classification performance compared to other baselines.展开更多
With the widespread use of social media,the propagation of health-related rumors has become a significant public health threat.Existing methods for detecting health rumors predominantly rely on external knowledge or p...With the widespread use of social media,the propagation of health-related rumors has become a significant public health threat.Existing methods for detecting health rumors predominantly rely on external knowledge or propagation structures,with only a few recent approaches attempting causal inference;however,these have not yet effectively integrated causal discovery with domain-specific knowledge graphs for detecting health rumors.In this study,we found that the combined use of causal discovery and domain-specific knowledge graphs can effectively identify implicit pseudo-causal logic embedded within texts,holding significant potential for health rumor detection.To this end,we propose CKDG—a dual-graph fusion framework based on causal logic and medical knowledge graphs.CKDG constructs a weighted causal graph to capture the implicit causal relationships in the text and introduces a medical knowledge graph to verify semantic consistency,thereby enhancing the ability to identify the misuse of professional terminology and pseudoscientific claims.In experiments conducted on a dataset comprising 8430 health rumors,CKDG achieved an accuracy of 91.28%and an F1 score of 90.38%,representing improvements of 5.11%and 3.29%over the best baseline,respectively.Our results indicate that the integrated use of causal discovery and domainspecific knowledge graphs offers significant advantages for health rumor detection systems.This method not only improves detection performance but also enhances the transparency and credibility of model decisions by tracing causal chains and sources of knowledge conflicts.We anticipate that this work will provide key technological support for the development of trustworthy health-information filtering systems,thereby improving the reliability of public health information on social media.展开更多
Knowledge distillation has become a standard technique for compressing large language models into efficient student models,but existing methods often struggle to balance prediction accuracy with explanation quality.Re...Knowledge distillation has become a standard technique for compressing large language models into efficient student models,but existing methods often struggle to balance prediction accuracy with explanation quality.Recent approaches such as Distilling Step-by-Step(DSbS)introduce explanation supervision,yet they apply it in a uniform manner that may not fully exploit the different learning dynamics of prediction and explanation.In this work,we propose a task-structured curriculum learning(TSCL)framework that structures training into three sequential phases:(i)prediction-only,to establish stable feature representations;(ii)joint prediction-explanation,to align task outputs with rationale generation;and(iii)explanation-only,to refine the quality of rationales.This design provides a simple but effective modification to DSbS,requiring no architectural changes and adding negligible training cost.We justify the phase scheduling with ablation studies and convergence analysis,showing that an initial prediction-heavy stage followed by a balanced joint phase improves both stability and explanation alignment.Extensive experiments on five datasets(e-SNLI,ANLI,CommonsenseQA,SVAMP,and MedNLI)demonstrate that TSCL consistently outperforms strong baselines,achieving gains of+1.7-2.6 points in accuracy and 0.8-1.2 in ROUGE-L,corresponding to relative error reductions of up to 21%.Beyond lexical metrics,human evaluation and ERASERstyle faithfulness diagnostics confirm that TSCL produces more faithful and informative explanations.Comparative training curves further reveal faster convergence and lower variance across seeds.Efficiency analysis shows less than 3%overhead in wall-clock training time and no additional inference cost,making the approach practical for realworld deployment.This study demonstrates that a simple task-structured curriculum can significantly improve the effectiveness of knowledge distillation.By separating and sequencing objectives,TSCL achieves a better balance between accuracy,stability,and explanation quality.The framework generalizes across domains,including medical NLI,and offers a principled recipe for future applications in multimodal reasoning and reinforcement learning.展开更多
With the large-scale deployment of the Internet ofThings(IoT)devices,their weak securitymechanisms make them prime targets for malware attacks.Attackers often use Domain Generation Algorithm(DGA)to generate random dom...With the large-scale deployment of the Internet ofThings(IoT)devices,their weak securitymechanisms make them prime targets for malware attacks.Attackers often use Domain Generation Algorithm(DGA)to generate random domain names,hiding the real IP of Command and Control(C&C)servers to build botnets.Due to the randomness and dynamics of DGA,traditional methods struggle to detect them accurately,increasing the difficulty of network defense.This paper proposes a lightweight DGA detection model based on knowledge distillation for resource-constrained IoT environments.Specifically,a teacher model combining CharacterBERT,a bidirectional long short-term memory(BiLSTM)network,and attention mechanism(ATT)is constructed:it extracts character-level semantic features viaCharacterBERT,captures sequence dependencieswith the BiLSTM,and integrates theATT for key feature weighting,formingmulti-granularity feature fusion.An improved knowledge distillation approach transfers the teacher model’s learned knowledge to the simplified DistilBERT student model.Experimental results show the teacher model achieves 98.68%detection accuracy.The student modelmaintains slightly improved accuracy while significantly compressing parameters to approximately 38.4%of the teacher model’s scale,greatly reducing computational overhead for IoT deployment.展开更多
Amidst evolving user behavior driven by the development of the internet,enhancing the operational quality of trade publishing knowledge service platforms has become a significant challenge for publishing institutions....Amidst evolving user behavior driven by the development of the internet,enhancing the operational quality of trade publishing knowledge service platforms has become a significant challenge for publishing institutions.To address this issue,this paper employs a combined approach of theoretical analysis and case study,introducing the SICAS(Sense-Interest-Connection-Action-Share)user consumption behavior analysis model and selecting“CITIC Academy”as the case study subject.It systematically examines and summarizes the platform’s operational practices and specific strategies,aiming to offer strategic insights and practical references for the operational improvement and sustainable,high-quality development of trade publishing knowledge service platforms.展开更多
Conversational recommender systems(CRSs)focus on refining preferences and providing personalized recommendations through natural language interactions and dialogue history.Large language models(LLMs)have shown outstan...Conversational recommender systems(CRSs)focus on refining preferences and providing personalized recommendations through natural language interactions and dialogue history.Large language models(LLMs)have shown outstanding performance across various domains,thereby prompting researchers to investigate their applicability in recommendation systems.However,due to the lack of task-specific knowledge and an inefficient feature extraction process,LLMs still have suboptimal performance in recommendation tasks.Therefore,external knowledge sources,such as knowledge graphs(KGs)and knowledge bases(KBs),are often introduced to address the issue of data sparsity.Compared to KGs,KBs possess higher retrieval efficiency,making them more suitable for scenarios where LLMs serve as recommenders.To this end,we introduce a novel framework integrating LLMs with KBs for enhanced retrieval generation,namely LLMKB.LLMKB initially leverages structured knowledge to create mapping dictionaries,extracting entity-relation information from heterogeneous knowledge to construct KBs.Then,LLMKB achieves the embedding calibration between user information representations and documents in KBs through retrieval model fine-tuning.Finally,LLMKB employs retrievalaugmented generation to produce recommendations based on fused text inputs,followed by post-processing.Experiment results on two public CRS datasets demonstrate the effectiveness of our framework.Our code is publicly available at the link:https://anonymous.4open.science/r/LLMKB-6FD0.展开更多
From the ecological viewpoint this paper discusses the urban spatial-temporal relationship. We take regional towns and cities as a complex man-land system of urban eco-community. This complex man-land system comprises...From the ecological viewpoint this paper discusses the urban spatial-temporal relationship. We take regional towns and cities as a complex man-land system of urban eco-community. This complex man-land system comprises two elements of ' man' and ' land' . Here, ' man' means organization with self-determined consciousness, and ' land' means the physical environment (niche) that ' man' depends on. The complex man-land system has three basic components. They are individual, population and community. Therefore there are six types of spatial relationship for the complex man-land system. They are individual, population,community,man-man, land-land and man-land spatial relationships. Taking the Pearl(Zhujiang) River Delta as a case study, the authors found some evidence of the urban spatial relationship from the remote sensing data. Firstly, the concentration and diffusion of the cities spatial relationship was found in the remote sensing imagery. Most of the cities concentrate in the core area of the Pearl River Delta, but the diffusion situation is also significant. Secondly, the growth behavior and succession behavior of the urban spatial relationship was found in the remote sensing images comparison with different temporal data. Thirdly, the inheritance, break, or meeting emergency behavior was observed from the remote sensing data. Fourthly, the authors found many cases of symbiosis and competition in the remote sensing data of the Pearl River Delta. Fifthly, the autoeciousness, stranglehold and invasion behavior of the urban spatial relationship was discovered from the remote sensing data.展开更多
In the international shipping industry, digital intelligence transformation has become essential, with both governments and enterprises actively working to integrate diverse datasets. The domain of maritime and shippi...In the international shipping industry, digital intelligence transformation has become essential, with both governments and enterprises actively working to integrate diverse datasets. The domain of maritime and shipping is characterized by a vast array of document types, filled with complex, large-scale, and often chaotic knowledge and relationships. Effectively managing these documents is crucial for developing a Large Language Model (LLM) in the maritime domain, enabling practitioners to access and leverage valuable information. A Knowledge Graph (KG) offers a state-of-the-art solution for enhancing knowledge retrieval, providing more accurate responses and enabling context-aware reasoning. This paper presents a framework for utilizing maritime and shipping documents to construct a knowledge graph using GraphRAG, a hybrid tool combining graph-based retrieval and generation capabilities. The extraction of entities and relationships from these documents and the KG construction process are detailed. Furthermore, the KG is integrated with an LLM to develop a Q&A system, demonstrating that the system significantly improves answer accuracy compared to traditional LLMs. Additionally, the KG construction process is up to 50% faster than conventional LLM-based approaches, underscoring the efficiency of our method. This study provides a promising approach to digital intelligence in shipping, advancing knowledge accessibility and decision-making.展开更多
As a significant city in the Yangtze River Delta regions,Hefei has experienced rapid changes in the sources of air pollution due to its high-speed economic development and urban expansion.However,there has been limite...As a significant city in the Yangtze River Delta regions,Hefei has experienced rapid changes in the sources of air pollution due to its high-speed economic development and urban expansion.However,there has been limited research in recent years on the spatial-temporal distribution and emission of its atmospheric pollutants.To address this,this study conducted mobile observations of urban roads using the Mobile-DOAS instrument from June 2021 to May 2022.The monitoring results exhibit a favourable consistent with TROPOMI satellite data and ground monitoring station data.Temporally,there were pronounced seasonal variations in air pollutants.Spatially,high concentration of HCHO and NO_(2)were closely associated with traffic congestion on roadways,while heightened SO_(2)levels were attributed to winter heating and industrial emissions.The study also revealed that with the implementation of road policies,the average vehicle speed increased by 95.4%,while the NO concentration decreased by 54.4%.In the estimation of urban NO_(x)emission flux,it was observed that in temporal terms,compared with inventory data,the emissions calculated viamobile measurements exhibitedmore distinct seasonal patterns,with the highest emission rate of 349 g/sec in winter and the lowest of 142 g/sec in summer.In spatial terms,the significant difference in emissions between the inner and outer ring roads also suggests the presence of the city’s primary NO_(x)emission sources in the area between these two rings.This study offers data support for formulating the next phase of air pollution control measures in urban areas.展开更多
基金supported by the National Natural Science Foundation of China(42476084,62203456,42276199)the Stable Support Project of National Key Laboratory(WDZC 20245250302)the National Key R&D Program of China(2024YFC2813502,2024YFC2813302)。
文摘Bird's-eye-view(BEV)perception is a core technology for autonomous driving systems.However,existing solutions face the dilemma of high costs associated with multimodal methods and limited performance of vision-only approaches.To address this issue,this paper proposes a framework named“a lightweight pure visual BEV perception method based on dual distillation of spatial-temporal knowledge”.This framework innovatively designs a lightweight vision-only student model based on Res Net,which leverages a dual distillation mechanism to learn from a powerful teacher model that integrates temporal information from both image and light detection and ranging(LiDAR)modalities.Specifically,we distill efficient multi-modal feature extraction and spatial fusion capabilities from the BEVFusion model,and distill advanced temporal information fusion and spatiotemporal attention mechanisms from the BEVFormer model.This dual distillation strategy enables the student model to achieve perception performance close to that of multi-modal models without relying on Li DAR.Experimental results on the nu Scenes dataset demonstrate that the proposed model significantly outperforms classical vision-only algorithms,achieves comparable performance to current state-of-the-art vision-only methods on the nu Scenes detection leaderboard in terms of both mean average precision(mAP)and the nu Scenes detection score(NDS)metrics,and exhibits notable advantages in inference computational efficiency.Although the proposed dual-teacher paradigm incurs higher offline training costs compared to single-model approaches,it yields a streamlined and highly efficient student model suitable for resource-constrained real-time deployment.This provides an effective pathway toward low-cost,high-performance autonomous driving perception systems.
文摘Since Google introduced the concept of Knowledge Graphs(KGs)in 2012,their construction technologies have evolved into a comprehensive methodological framework encompassing knowledge acquisition,extraction,representation,modeling,fusion,computation,and storage.Within this framework,knowledge extraction,as the core component,directly determines KG quality.In military domains,traditional manual curation models face efficiency constraints due to data fragmentation,complex knowledge architectures,and confidentiality protocols.Meanwhile,crowdsourced ontology construction approaches from general domains prove non-transferable,while human-crafted ontologies struggle with generalization deficiencies.To address these challenges,this study proposes an OntologyAware LLM Methodology for Military Domain Knowledge Extraction(LLM-KE).This approach leverages the deep semantic comprehension capabilities of Large Language Models(LLMs)to simulate human experts’cognitive processes in crowdsourced ontology construction,enabling automated extraction of military textual knowledge.It concurrently enhances knowledge processing efficiency and improves KG completeness.Empirical analysis demonstrates that this method effectively resolves scalability and dynamic adaptation challenges in military KG construction,establishing a novel technological pathway for advancing military intelligence development.
基金supported by the National Natural Science Foundation of China(No.62267005)the Chinese Guangxi Natural Science Foundation(No.2023GXNSFAA026493)+1 种基金Guangxi Collaborative Innovation Center ofMulti-Source Information Integration and Intelligent ProcessingGuangxi Academy of Artificial Intelligence.
文摘The rapid development of information technology and accelerated digitalization have led to an explosive growth of data across various fields.As a key technology for knowledge representation and sharing,knowledge graphs play a crucial role by constructing structured networks of relationships among entities.However,data sparsity and numerous unexplored implicit relations result in the widespread incompleteness of knowledge graphs.In static knowledge graph completion,most existing methods rely on linear operations or simple interaction mechanisms for triple encoding,making it difficult to fully capture the deep semantic associations between entities and relations.Moreover,many methods focus only on the local information of individual triples,ignoring the rich semantic dependencies embedded in the neighboring nodes of entities within the graph structure,which leads to incomplete embedding representations.To address these challenges,we propose Two-Stage Mixer Embedding(TSMixerE),a static knowledge graph completion method based on entity context.In the unit semantic extraction stage,TSMixerE leveragesmulti-scale circular convolution to capture local features atmultiple granularities,enhancing the flexibility and robustness of feature interactions.A channel attention mechanism amplifies key channel responses to suppress noise and irrelevant information,thereby improving the discriminative power and semantic depth of feature representations.For contextual information fusion,a multi-layer self-attentionmechanism enables deep interactions among contextual cues,effectively integrating local details with global context.Simultaneously,type embeddings clarify the semantic identities and roles of each component,enhancing the model’s sensitivity and fusion capabilities for diverse information sources.Furthermore,TSMixerE constructs contextual unit sequences for entities,fully exploring neighborhood information within the graph structure to model complex semantic dependencies,thus improving the completeness and generalization of embedding representations.
基金supported by the National Natural Science Foundation of China (Grant Nos.62125301,62021003,62303024,U24A20275,62522302,62473011,92467205)the National Key Research and Development Project (Grant Nos.2022YFB3305800-5,2024YFE0212400)+2 种基金the Youth Beijing Scholars Program (Grant No.037)the Beijing Nova Program (Grant Nos.20240484694,20250484938)the Beijing Natural Science Foundation (Grant No.L253010)。
文摘Active fault-tolerant control utilizes information obtained from fault diagnosis to reconfigure the control law to compensate for faults in the wastewater treatment process. However, since the similarity of fault characteristic in the incipient stage can result in misdiagnosis, it is a challenge for fault-tolerant control to ensure system safety and reliability. Therefore, to address this issue, a fault diagnosis and fault-tolerant control with a knowledge transfer strategy(KT-FDFTC) is proposed in this paper. First, a knowledge reasoning diagnosis strategy using multi-source transfer learning is designed to distinguish the similar characteristic of incipient faults. Then, the multi-source knowledge can assist in the diagnosis strategy to strengthen the fault information for fault-tolerant control. Second, a knowledge adaptive compensation mechanism, which makes knowledge and data coupled into the output trajectory regarded as an objective function, is employed to dynamically compute the control law. Then, KT-FDFTC can ensure the stable operation to adapt to various fault conditions. Third, the Lyapunov function is established to demonstrate the stability of KT-FDFTC. Then, the theoretical basis can offer the successful application of KTFDFTC. Finally, the proposed method is validated through a real WWTP and a simulation platform. The experimental results confirm that KT-FDFTC can provide good diagnosis performance and fault tolerance ability.
基金supported by Harbin Institute of Technology High-level Teaching Achievement Award(National Level)Cultivation Project(256709).
文摘This paper delves into effective pathways for transforming course ecosystems from resource provision to knowledge service and competency development through university-enterprise collaboration in co-building knowledge graphs and intelligent shared courses.This approach enables personalized,learning-driven teaching.Based on knowledge graphs and integrated teacher-machine-student smart teaching scenarios,it not only innovates autonomous learning environments and human-computer interaction models while optimizing teaching experiences for both instructors and students,but also effectively addresses the issues of students’“scattered,superficial,and fragmented learning”.This establishes the foundation for personalized teaching tailored to individual aptitudes.
基金supported by Basic Science Research Program through the National Research Foundation of Korea(NRF)funded by the Ministry of Education(RS-2023-00245084)by Korea Institute for Advancement of Technology(KIAT)grant funded by the Korea Government(MOTIE)(RS-2024-00415938,HRD Program for Industrial Innovation)and Soonchunhyang University.
文摘This paper proposes a deep learning-based 3D LiDAR perception framework designed for applications such as autonomous robots and vehicles.To address the high dependency on large-scale annotated data—an inherent limitation of deep learning models—this study introduces a hybrid perception architecture that incorporates expertdriven LiDAR processing techniques into the deep neural network.Traditional 3DLiDAR processingmethods typically remove ground planes and apply distance-or density-based clustering for object detection.In this work,such expert knowledge is encoded as feature-level inputs and fused with the deep network,therebymitigating the data dependency issue of conventional learning-based approaches.Specifically,the proposedmethod combines two expert algorithms—Patchwork++for ground segmentation and DBSCAN for clustering—with a PointPillars-based LiDAR detection network.We design four hybrid versions of the network depending on the stage and method of integrating expert features into the feature map of the deep model.Among these,Version 4 incorporates a modified neck structure in PointPillars and introduces a new Cluster 2D Pseudo-Map Branch that utilizes cluster-level pseudo-images generated from Patchwork++and DBSCAN.This version achieved a+3.88%improvement mean Average Precision(mAP)compared to the baseline PointPillars.The results demonstrate that embedding expert-based perception logic into deep neural architectures can effectively enhance performance and reduce dependency on extensive training datasets,offering a promising direction for robust 3D LiDAR object detection in real-world scenarios.
基金supported by the National Key R&D Program of China(Grant No.2023YFC3209504)Natural Science Foundation of Wuhan(Grant No.2024040801020271)the Fundamental Research Funds for Central Public Welfare Research Institutes(Grant No.CKSF2025718/YT).
文摘Wetting deformation in earth-rockfill dams is a critical factor influencingdam safety.Although numerous mathematical models have been developed to describe this phenomenon,most of them rely on empirical formulations and lack prior knowledge of model parameters,which is essential for Bayesian parameter inversion to enhance accuracy and reduce uncertainty.This study introduces a datadriven approach to establishing prior knowledge of earth-rockfill dams.Driving factors are utilized to determine the potential range of model parameters,and settlement changes within this range are calculated.The results are iteratively compared with actual monitoring data until the calculated range encompasses the observed data,thereby providing prior knowledge of the model parameters.The proposed method is applied to the right-bank earth-rockfilldam of Danjiangkou.Employing a Gibbs sample size of 30,000,the proposed method effectively calibrates the prior knowledge of the wetting model parameters,achieving a root mean square error(RMSE)of 5.18 mm for the settlement predictions.By comparison,the use of non-informative priors with sample sizes of 30,000 and 50,000 results in significantly larger RMSE values of 11.97 mm and 16.07 mm,respectively.Furthermore,the computational efficiencyof the proposed method is demonstrated by an inversion computation time of 902 s for 30,000 samples,which is notably shorter than the 1026 s and 1558 s required for noninformative priors with 30,000 and 50,000 samples,respectively.These findingsunderscore the superior performance of the proposed approach in terms of both prediction accuracy and computational efficiency.These results demonstrate that the proposed method not only improves the predictive accuracy but also enhances the computational efficiency,enabling optimal parameter identificationwith reduced computational effort.This approach provides a robust and efficientframework for advancing dam safety assessments.
基金supported by the National Key Research and Development Program of China(No.2023YFB3712401),the National Natural Science Foundation of China(No.52274301)the Aeronautical Science Foundation of China(No.2023Z0530S6005)the Ningbo Yongjiang Talent-Introduction Programme(No.2022A-023-C).
文摘The viscosity of refining slags plays a critical role in metallurgical processes.However,obtaining accurate viscosity data remains challenging due to the complexities of high-temperature experiments,often relying on empirical models with limited predictive capabilities.This study focuses on the influence of optical basicity on viscosity in CaO-Al_(2)O_(3)-based refining slags,leveraging machine learning to address data scarcity and improve prediction accuracy.An automated framework for algorithm integration,parameter tuning,and evaluation ranking framework(Auto-APE)is employed to develop customized data-driven models for various slag systems,including CaO-Al_(2)O_(3)-SiO_(2),CaO-Al_(2)O_(3)-CaF_(2),CaO-Al_(2)O_(3)-SiO_(2)-MgO,and CaO-Al_(2)O_(3)-SiO_(2)-MgO-CaF_(2).By incorporating optical basicity as a key feature,the models achieve an average validation error of 8.0%to 15.1%,significantly outperforming traditional empirical models.Additionally,symbolic regression is introduced to rapidly construct domain-specific features,such as optical basicity-like descriptors,offering a potential breakthrough in performance prediction for small datasets.This work highlights the critical role of domain-specific knowledge in understanding and predicting viscosity,providing a robust machine learning-based approach for optimizing refining slag properties.
基金supported by the State Grid Southwest Branch Project“Research on Defect Diagnosis and Early Warning Technology of Relay Protection and Safety Automation Devices Based on Multi-Source Heterogeneous Defect Data”.
文摘The reliable operation of power grid secondary equipment is an important guarantee for the safety and stability of the power system.However,various defects could be produced in the secondary equipment during longtermoperation.The complex relationship between the defect phenomenon andmulti-layer causes and the probabilistic influence of secondary equipment cannot be described through knowledge extraction and fusion technology by existing methods,which limits the real-time and accuracy of defect identification.Therefore,a defect recognition method based on the Bayesian network and knowledge graph fusion is proposed.The defect data of secondary equipment is transformed into the structured knowledge graph through knowledge extraction and fusion technology.The knowledge graph of power grid secondary equipment is mapped to the Bayesian network framework,combined with historical defect data,and introduced Noisy-OR nodes.The prior and conditional probabilities of the Bayesian network are then reasonably assigned to build a model that reflects the probability dependence between defect phenomena and potential causes in power grid secondary equipment.Defect identification of power grid secondary equipment is achieved by defect subgraph search based on the knowledge graph,and defect inference based on the Bayesian network.Practical application cases prove this method’s effectiveness in identifying secondary equipment defect causes,improving identification accuracy and efficiency.
基金support from the Scientific Funding for the Center of National Railway Intelligent Transportation System Engineering and Technology,China Academy of Railway Sciences Corporation Limited(Grant No.2023YJ354)。
文摘Knowledge graphs,which combine structured representation with semantic modeling,have shown great potential in knowledge expression,causal inference,and automated reasoning,and are widely used in fields such as intelligent question answering,decision support,and fault diagnosis.As high-speed train systems become increasingly intelligent and interconnected,fault patterns have grown more complex and dynamic.Knowledge graphs offer a promising solution to support the structured management and real-time reasoning of fault knowledge,addressing key requirements such as interpretability,accuracy,and continuous evolution in intelligent diagnostic systems.However,conventional knowledge graph construction relies heavily on domain expertise and specialized tools,resulting in high entry barriers for non-experts and limiting their practical application in frontline maintenance scenarios.To address this limitation,this paper proposes a fault knowledge modeling approach for high-speed trains that integrates structured logic diagrams with knowledge graphs.The method employs a seven-layer logic structure—comprising fault name,applicable vehicles,diagnostic logic,signal parameters,verification conditions,fault causes,and emergency measures—to transform unstructured knowledge into a visual and hierarchical representation.A semantic mapping mechanism is then used to automatically convert logic diagrams into machine-interpretable knowledge graphs,enabling dynamic reasoning and knowledge reuse.Furthermore,the proposed method establishes a three-layer architecture—logic structuring,knowledge graph transformation,and dynamic inference—to bridge human-expert logic with machinebased reasoning.Experimental validation and system implementation demonstrate that this approach not only improves knowledge interpretability and inference precision but also significantly enhances modeling efficiency and system maintainability.It provides a scalable and adaptable solution for intelligent operation and maintenance platforms in the high-speed rail domain.
基金funded by the National Natural Science Foundation of China(Grant No.62441212)the Major Project of the Natural Science Foundation of Inner Mongolia(Grant No.2025ZD008).
文摘Aiming at the problem of insufficient recognition of implicit variants by existing Chinese sensitive text detection methods,this paper proposes the IPKE-MoE framework,which consists of three parts,namely,a sensitive word variant extraction framework,a sensitive word variant knowledge enhancement layer and a mixture-of-experts(MoE)classification layer.First,sensitive word variants are precisely extracted through dynamic iterative prompt templates and the context-aware capabilities of Large Language Models(LLMs).Next,the extracted variants are used to construct a knowledge enhancement layer for sensitive word variants based on RoCBert models.Specifically,after locating variants via n-gram algorithms,variant types are mapped to embedding vectors and fused with original word vectors.Finally,a mixture-of-experts(MoE)classification layer is designed(sensitive word,sentiment,and semantic experts),which decouples the relationship between sensitiveword existence and text toxicity throughmultiple experts.This framework effectively combines the comprehension ability of Large Language Models(LLMs)with the discriminative ability of smaller models.Our two experiments demonstrate that the sensitive word variant extraction framework based on dynamically iterated prompt templates outperforms other baseline prompt templates.TheRoCBert models incorporating the sensitive word variant knowledge enhancement layer and a mixture-of-experts(MoE)classification layer achieve superior classification performance compared to other baselines.
基金funded by the Hunan Provincial Natural Science Foundation of China(Grant No.2025JJ70105)the Hunan Provincial College Students’Innovation and Entrepreneurship Training Program(Project No.S202411342056)The article processing charge(APC)was funded by the Project No.2025JJ70105.
文摘With the widespread use of social media,the propagation of health-related rumors has become a significant public health threat.Existing methods for detecting health rumors predominantly rely on external knowledge or propagation structures,with only a few recent approaches attempting causal inference;however,these have not yet effectively integrated causal discovery with domain-specific knowledge graphs for detecting health rumors.In this study,we found that the combined use of causal discovery and domain-specific knowledge graphs can effectively identify implicit pseudo-causal logic embedded within texts,holding significant potential for health rumor detection.To this end,we propose CKDG—a dual-graph fusion framework based on causal logic and medical knowledge graphs.CKDG constructs a weighted causal graph to capture the implicit causal relationships in the text and introduces a medical knowledge graph to verify semantic consistency,thereby enhancing the ability to identify the misuse of professional terminology and pseudoscientific claims.In experiments conducted on a dataset comprising 8430 health rumors,CKDG achieved an accuracy of 91.28%and an F1 score of 90.38%,representing improvements of 5.11%and 3.29%over the best baseline,respectively.Our results indicate that the integrated use of causal discovery and domainspecific knowledge graphs offers significant advantages for health rumor detection systems.This method not only improves detection performance but also enhances the transparency and credibility of model decisions by tracing causal chains and sources of knowledge conflicts.We anticipate that this work will provide key technological support for the development of trustworthy health-information filtering systems,thereby improving the reliability of public health information on social media.
文摘Knowledge distillation has become a standard technique for compressing large language models into efficient student models,but existing methods often struggle to balance prediction accuracy with explanation quality.Recent approaches such as Distilling Step-by-Step(DSbS)introduce explanation supervision,yet they apply it in a uniform manner that may not fully exploit the different learning dynamics of prediction and explanation.In this work,we propose a task-structured curriculum learning(TSCL)framework that structures training into three sequential phases:(i)prediction-only,to establish stable feature representations;(ii)joint prediction-explanation,to align task outputs with rationale generation;and(iii)explanation-only,to refine the quality of rationales.This design provides a simple but effective modification to DSbS,requiring no architectural changes and adding negligible training cost.We justify the phase scheduling with ablation studies and convergence analysis,showing that an initial prediction-heavy stage followed by a balanced joint phase improves both stability and explanation alignment.Extensive experiments on five datasets(e-SNLI,ANLI,CommonsenseQA,SVAMP,and MedNLI)demonstrate that TSCL consistently outperforms strong baselines,achieving gains of+1.7-2.6 points in accuracy and 0.8-1.2 in ROUGE-L,corresponding to relative error reductions of up to 21%.Beyond lexical metrics,human evaluation and ERASERstyle faithfulness diagnostics confirm that TSCL produces more faithful and informative explanations.Comparative training curves further reveal faster convergence and lower variance across seeds.Efficiency analysis shows less than 3%overhead in wall-clock training time and no additional inference cost,making the approach practical for realworld deployment.This study demonstrates that a simple task-structured curriculum can significantly improve the effectiveness of knowledge distillation.By separating and sequencing objectives,TSCL achieves a better balance between accuracy,stability,and explanation quality.The framework generalizes across domains,including medical NLI,and offers a principled recipe for future applications in multimodal reasoning and reinforcement learning.
基金supported by the following projects:National Natural Science Foundation of China(62461041)Natural Science Foundation of Jiangxi Province China(20242BAB25068).
文摘With the large-scale deployment of the Internet ofThings(IoT)devices,their weak securitymechanisms make them prime targets for malware attacks.Attackers often use Domain Generation Algorithm(DGA)to generate random domain names,hiding the real IP of Command and Control(C&C)servers to build botnets.Due to the randomness and dynamics of DGA,traditional methods struggle to detect them accurately,increasing the difficulty of network defense.This paper proposes a lightweight DGA detection model based on knowledge distillation for resource-constrained IoT environments.Specifically,a teacher model combining CharacterBERT,a bidirectional long short-term memory(BiLSTM)network,and attention mechanism(ATT)is constructed:it extracts character-level semantic features viaCharacterBERT,captures sequence dependencieswith the BiLSTM,and integrates theATT for key feature weighting,formingmulti-granularity feature fusion.An improved knowledge distillation approach transfers the teacher model’s learned knowledge to the simplified DistilBERT student model.Experimental results show the teacher model achieves 98.68%detection accuracy.The student modelmaintains slightly improved accuracy while significantly compressing parameters to approximately 38.4%of the teacher model’s scale,greatly reducing computational overhead for IoT deployment.
文摘Amidst evolving user behavior driven by the development of the internet,enhancing the operational quality of trade publishing knowledge service platforms has become a significant challenge for publishing institutions.To address this issue,this paper employs a combined approach of theoretical analysis and case study,introducing the SICAS(Sense-Interest-Connection-Action-Share)user consumption behavior analysis model and selecting“CITIC Academy”as the case study subject.It systematically examines and summarizes the platform’s operational practices and specific strategies,aiming to offer strategic insights and practical references for the operational improvement and sustainable,high-quality development of trade publishing knowledge service platforms.
文摘Conversational recommender systems(CRSs)focus on refining preferences and providing personalized recommendations through natural language interactions and dialogue history.Large language models(LLMs)have shown outstanding performance across various domains,thereby prompting researchers to investigate their applicability in recommendation systems.However,due to the lack of task-specific knowledge and an inefficient feature extraction process,LLMs still have suboptimal performance in recommendation tasks.Therefore,external knowledge sources,such as knowledge graphs(KGs)and knowledge bases(KBs),are often introduced to address the issue of data sparsity.Compared to KGs,KBs possess higher retrieval efficiency,making them more suitable for scenarios where LLMs serve as recommenders.To this end,we introduce a novel framework integrating LLMs with KBs for enhanced retrieval generation,namely LLMKB.LLMKB initially leverages structured knowledge to create mapping dictionaries,extracting entity-relation information from heterogeneous knowledge to construct KBs.Then,LLMKB achieves the embedding calibration between user information representations and documents in KBs through retrieval model fine-tuning.Finally,LLMKB employs retrievalaugmented generation to produce recommendations based on fused text inputs,followed by post-processing.Experiment results on two public CRS datasets demonstrate the effectiveness of our framework.Our code is publicly available at the link:https://anonymous.4open.science/r/LLMKB-6FD0.
基金Under the auspices of the National Natural Science Foundation of China(No.69896250-4).
文摘From the ecological viewpoint this paper discusses the urban spatial-temporal relationship. We take regional towns and cities as a complex man-land system of urban eco-community. This complex man-land system comprises two elements of ' man' and ' land' . Here, ' man' means organization with self-determined consciousness, and ' land' means the physical environment (niche) that ' man' depends on. The complex man-land system has three basic components. They are individual, population and community. Therefore there are six types of spatial relationship for the complex man-land system. They are individual, population,community,man-man, land-land and man-land spatial relationships. Taking the Pearl(Zhujiang) River Delta as a case study, the authors found some evidence of the urban spatial relationship from the remote sensing data. Firstly, the concentration and diffusion of the cities spatial relationship was found in the remote sensing imagery. Most of the cities concentrate in the core area of the Pearl River Delta, but the diffusion situation is also significant. Secondly, the growth behavior and succession behavior of the urban spatial relationship was found in the remote sensing images comparison with different temporal data. Thirdly, the inheritance, break, or meeting emergency behavior was observed from the remote sensing data. Fourthly, the authors found many cases of symbiosis and competition in the remote sensing data of the Pearl River Delta. Fifthly, the autoeciousness, stranglehold and invasion behavior of the urban spatial relationship was discovered from the remote sensing data.
文摘In the international shipping industry, digital intelligence transformation has become essential, with both governments and enterprises actively working to integrate diverse datasets. The domain of maritime and shipping is characterized by a vast array of document types, filled with complex, large-scale, and often chaotic knowledge and relationships. Effectively managing these documents is crucial for developing a Large Language Model (LLM) in the maritime domain, enabling practitioners to access and leverage valuable information. A Knowledge Graph (KG) offers a state-of-the-art solution for enhancing knowledge retrieval, providing more accurate responses and enabling context-aware reasoning. This paper presents a framework for utilizing maritime and shipping documents to construct a knowledge graph using GraphRAG, a hybrid tool combining graph-based retrieval and generation capabilities. The extraction of entities and relationships from these documents and the KG construction process are detailed. Furthermore, the KG is integrated with an LLM to develop a Q&A system, demonstrating that the system significantly improves answer accuracy compared to traditional LLMs. Additionally, the KG construction process is up to 50% faster than conventional LLM-based approaches, underscoring the efficiency of our method. This study provides a promising approach to digital intelligence in shipping, advancing knowledge accessibility and decision-making.
基金supported by the National Natural Science Foundation of China(Nos.U19A2044,42105132,42030609,41975037,and 42105133)the National Key Research and Development Program of China(No.2022YFC3703502)+1 种基金the Plan for Anhui Major Provincial Science&Technology Project(No.202203a07020003)Hefei Ecological Environment Bureau Project(No.2020BFFFD01804).
文摘As a significant city in the Yangtze River Delta regions,Hefei has experienced rapid changes in the sources of air pollution due to its high-speed economic development and urban expansion.However,there has been limited research in recent years on the spatial-temporal distribution and emission of its atmospheric pollutants.To address this,this study conducted mobile observations of urban roads using the Mobile-DOAS instrument from June 2021 to May 2022.The monitoring results exhibit a favourable consistent with TROPOMI satellite data and ground monitoring station data.Temporally,there were pronounced seasonal variations in air pollutants.Spatially,high concentration of HCHO and NO_(2)were closely associated with traffic congestion on roadways,while heightened SO_(2)levels were attributed to winter heating and industrial emissions.The study also revealed that with the implementation of road policies,the average vehicle speed increased by 95.4%,while the NO concentration decreased by 54.4%.In the estimation of urban NO_(x)emission flux,it was observed that in temporal terms,compared with inventory data,the emissions calculated viamobile measurements exhibitedmore distinct seasonal patterns,with the highest emission rate of 349 g/sec in winter and the lowest of 142 g/sec in summer.In spatial terms,the significant difference in emissions between the inner and outer ring roads also suggests the presence of the city’s primary NO_(x)emission sources in the area between these two rings.This study offers data support for formulating the next phase of air pollution control measures in urban areas.