In this paper,we propose a new privacy-aware transmission scheduling algorithm for 6G ad hoc networks.This system enables end nodes to select the optimum time and scheme to transmit private data safely.In 6G dynamic h...In this paper,we propose a new privacy-aware transmission scheduling algorithm for 6G ad hoc networks.This system enables end nodes to select the optimum time and scheme to transmit private data safely.In 6G dynamic heterogeneous infrastructures,unstable links and non-uniform hardware capabilities create critical issues regarding security and privacy.Traditional protocols are often too computationally heavy to allow 6G services to achieve their expected Quality-of-Service(QoS).As the transport network is built of ad hoc nodes,there is no guarantee about their trustworthiness or behavior,and transversal functionalities are delegated to the extreme nodes.However,while security can be guaranteed in extreme-to-extreme solutions,privacy cannot,as all intermediate nodes still have to handle the data packets they are transporting.Besides,traditional schemes for private anonymous ad hoc communications are vulnerable against modern intelligent attacks based on learning models.The proposed scheme fulfills this gap.Findings show the probability of a successful intelligent attack reduces by up to 65%compared to ad hoc networks with no privacy protection strategy when used the proposed technology.While congestion probability can remain below 0.001%,as required in 6G services.展开更多
Early and accurate detection of bone cancer and marrow cell abnormalities is critical for timely intervention and improved patient outcomes.This paper proposes a novel hybrid deep learning framework that integrates a ...Early and accurate detection of bone cancer and marrow cell abnormalities is critical for timely intervention and improved patient outcomes.This paper proposes a novel hybrid deep learning framework that integrates a Convolutional Neural Network(CNN)with a Bidirectional Long Short-Term Memory(BiLSTM)architecture,optimized using the Firefly Optimization algorithm(FO).The proposed CNN-BiLSTM-FO model is tailored for structured biomedical data,capturing both local patterns and sequential dependencies in diagnostic features,while the Firefly Algorithm fine-tunes key hyperparameters to maximize predictive performance.The approach is evaluated on two benchmark biomedical datasets:one comprising diagnostic data for bone cancer detection and another for identifying marrow cell abnormalities.Experimental results demonstrate that the proposed method outperforms standard deep learning models,including CNN,LSTM,BiLSTM,and CNN-LSTM hybrids,significantly.The CNNBiLSTM-FO model achieves an accuracy of 98.55%for bone cancer detection and 96.04%for marrow abnormality classification.The paper also presents a detailed complexity analysis of the proposed algorithm and compares its performance across multiple evaluation metrics such as precision,recall,F1-score,and AUC.The results confirm the effectiveness of the firefly-based optimization strategy in improving classification accuracy and model robustness.This work introduces a scalable and accurate diagnostic solution that holds strong potential for integration into intelligent clinical decision-support systems.展开更多
Parameter extraction of photovoltaic(PV)models is crucial for the planning,optimization,and control of PV systems.Although some methods using meta-heuristic algorithms have been proposed to determine these parameters,...Parameter extraction of photovoltaic(PV)models is crucial for the planning,optimization,and control of PV systems.Although some methods using meta-heuristic algorithms have been proposed to determine these parameters,the robustness of solutions obtained by these methods faces great challenges when the complexity of the PV model increases.The unstable results will affect the reliable operation and maintenance strategies of PV systems.In response to this challenge,an improved rime optimization algorithm with enhanced exploration and exploitation,termed TERIME,is proposed for robust and accurate parameter identification for various PV models.Specifically,the differential evolution mutation operator is integrated in the exploration phase to enhance the population diversity.Meanwhile,a new exploitation strategy incorporating randomization and neighborhood strategies simultaneously is developed to maintain the balance of exploitation width and depth.The TERIME algorithm is applied to estimate the optimal parameters of the single diode model,double diode model,and triple diode model combined with the Lambert-W function for three PV cell and module types including RTC France,Photo Watt-PWP 201 and S75.According to the statistical analysis in 100 runs,the proposed algorithm achieves more accurate and robust parameter estimations than other techniques to various PV models in varying environmental conditions.All of our source codes are publicly available at https://github.com/dirge1/TERIME.展开更多
To address the challenge of identifying the primary causes of energy consumption fluctuations and accurately assessing the influence of various factors in the converter unit of an iron and steel plant,the focus is pla...To address the challenge of identifying the primary causes of energy consumption fluctuations and accurately assessing the influence of various factors in the converter unit of an iron and steel plant,the focus is placed on the critical components of material and heat balance.Through a thorough analysis of the interactions between various components and energy consumptions,six pivotal factors have been identified—raw material composition,steel type,steel temperature,slag temperature,recycling practices,and operational parameters.Utilizing a framework based on an equivalent energy consumption model,an integrated intelligent diagnostic model has been developed that encapsulates these factors,providing a comprehensive assessment tool for converter energy consumption.Employing the K-means clustering algorithm,historical operational data from the converter have been meticulously analyzed to determine baseline values for essential variables such as energy consumption and recovery rates.Building upon this data-driven foundation,an innovative online system for the intelligent diagnosis of converter energy consumption has been crafted and implemented,enhancing the precision and efficiency of energy management.Upon implementation with energy consumption data at a steel plant in 2023,the diagnostic analysis performed by the system exposed significant variations in energy usage across different converter units.The analysis revealed that the most significant factor influencing the variation in energy consumption for both furnaces was the steel grade,with contributions of−0.550 and 0.379.展开更多
Antarctic krill(Euphausia superba),widely distributes around Antarctica,is a key species supporting the biodiversity of the Southern Ocean ecosystem.The Commission for the Conservation of Antarctic Marine Living Resou...Antarctic krill(Euphausia superba),widely distributes around Antarctica,is a key species supporting the biodiversity of the Southern Ocean ecosystem.The Commission for the Conservation of Antarctic Marine Living Resources(CCAMLR)has thus managed the krill fishery according to a precautionary way.Currently,CCAMLR is making effort to develop a refined krill fishery management approach based on more solid science,which requires accurate predictions of krill distribution.To address this need,this study investigated the effects of algorithm and spatial resolution on the performance of Antarctic krill distribution modelling.We integrated acoustic data from 4 surveys conducted in the waters adjacent to the Antarctic Peninsula with 11 environmental variables characterizing krill prey conditions,water mass properties,and seafloor topography.These data were processed at 4 spatial resolutions(5,10,15,and 20 km)to fit distribution models using 4 algorithms:Random Forests(RF),Generalized Additive Models(GAM),Extreme Gradient Boosting(XGBoost),and Artificial Neural Networks(ANN).Model performance was assessed and compared in terms of goodness-of-fit and predictive accuracy.The results showed that RF achieved the highest predictive performance at most resolutions,whereas GAM performed best at the coarsest resolution(20 km).XGBoost closely following RF in accuracy and demonstrated robustness as evidenced by the highly consistent partial dependence curves across resolutions.In contrast,ANN exhibited limitations with smaller sample sizes,resulting in comparatively poorer predictive performance.The analysis revealed a trade-off whereby reducing spatial resolution improved model fit and mitigated zero-inflation at the expense of fine-scale information and overall predictive accuracy.Ensemble models,integrating RF,GAM,and XGBoost,are proposed as potential balanced solutions to improve predictive stability,offering a more robust scientific basis for the refinement of krill management.展开更多
The estimation of the probability of informed trading(PIN)model and its extensions poses significant challenges owing to various computational problems.To address these issues,we propose a novel estimation method call...The estimation of the probability of informed trading(PIN)model and its extensions poses significant challenges owing to various computational problems.To address these issues,we propose a novel estimation method called the expectation-conditional-maximization(ECM)algorithm,which can serve as an alternative to the existing methods for estimating PIN models.Our method provides optimal estimates for the original PIN model as well as two of its extensions:the multilayer PIN model and the adjusted PIN model,along with its restricted versions.Our results indicate that estimations using the ECM algorithm are generally faster,more accurate,and more memory-efficient than the standard methods used in the literature,making it a robust alternative.More importantly,the ECM algorithm is not limited to the models discussed and can be easily adapted to estimate future extensions of the PIN model.展开更多
With the rapid adoption of artificial intelligence(AI)in domains such as power,transportation,and finance,the number of machine learning and deep learning models has grown exponentially.However,challenges such as dela...With the rapid adoption of artificial intelligence(AI)in domains such as power,transportation,and finance,the number of machine learning and deep learning models has grown exponentially.However,challenges such as delayed retraining,inconsistent version management,insufficient drift monitoring,and limited data security still hinder efficient and reliable model operations.To address these issues,this paper proposes the Intelligent Model Lifecycle Management Algorithm(IMLMA).The algorithm employs a dual-trigger mechanism based on both data volume thresholds and time intervals to automate retraining,and applies Bayesian optimization for adaptive hyperparameter tuning to improve performance.A multi-metric replacement strategy,incorporating MSE,MAE,and R2,ensures that new models replace existing ones only when performance improvements are guaranteed.A versioning and traceability database supports comparison and visualization,while real-time monitoring with stability analysis enables early warnings of latency and drift.Finally,hash-based integrity checks secure both model files and datasets.Experimental validation in a power metering operation scenario demonstrates that IMLMA reduces model update delays,enhances predictive accuracy and stability,and maintains low latency under high concurrency.This work provides a practical,reusable,and scalable solution for intelligent model lifecycle management,with broad applicability to complex systems such as smart grids.展开更多
Accurate and reliable photovoltaic(PV)modeling is crucial for the performance evaluation,control,and optimization of PV systems.However,existing methods for PV parameter identification often suffer from limitations in...Accurate and reliable photovoltaic(PV)modeling is crucial for the performance evaluation,control,and optimization of PV systems.However,existing methods for PV parameter identification often suffer from limitations in accuracy and efficiency.To address these challenges,we propose an adaptive multi-learning cooperation search algorithm(AMLCSA)for efficient identification of unknown parameters in PV models.AMLCSA is a novel algorithm inspired by teamwork behaviors in modern enterprises.It enhances the original cooperation search algorithm in two key aspects:(i)an adaptive multi-learning strategy that dynamically adjusts search ranges using adaptive weights,allowing better individuals to focus on local exploitation while guiding poorer individuals toward global exploration;and(ii)a chaotic grouping reflection strategy that introduces chaotic sequences to enhance population diversity and improve search performance.The effectiveness of AMLCSA is demonstrated on single-diode,double-diode,and three PV-module models.Simulation results show that AMLCSA offers significant advantages in convergence,accuracy,and stability compared to existing state-of-the-art algorithms.展开更多
When dealing with expensive multiobjective optimization problems,majority of existing surrogate-assisted evolutionary algorithms(SAEAs)generate solutions in decision space and screen candidate solutions mostly by usin...When dealing with expensive multiobjective optimization problems,majority of existing surrogate-assisted evolutionary algorithms(SAEAs)generate solutions in decision space and screen candidate solutions mostly by using designed surrogate models.The generated solutions exhibit excessive randomness,which tends to reduce the likelihood of generating good-quality solutions and cause a long evolution to the optima.To improve SAEAs greatly,this work proposes an evolutionary algorithm based on surrogate and inverse surrogate models by 1)Employing a surrogate model in lieu of expensive(true)function evaluations;and 2)Proposing and using an inverse surrogate model to generate new solutions.By using the same training data but with its inputs and outputs being reversed,the latter is simple to train.It is then used to generate new vectors in objective space,which are mapped into decision space to obtain their corresponding solutions.Using a particular example,this work shows its advantages over existing SAEAs.The results of comparing it with state-of-the-art algorithms on expensive optimization problems show that it is highly competitive in both solution performance and efficiency.展开更多
“Algorithm Design and Analysis”is not only one of the important courses in the undergraduate teaching of computer science and technology but also a key part of computer professional skills.In recent years,with the r...“Algorithm Design and Analysis”is not only one of the important courses in the undergraduate teaching of computer science and technology but also a key part of computer professional skills.In recent years,with the rise and widespread application of big language models,many teaching reform plans have been produced to promote the quality and efficiency of teaching.This paper studies how to refer to software development professional skills standards,investigates the knowledge points of“Algorithm Design and Analysis”courses in other educational institutions,uses cutting-edge core technology big language models to drive the improvement of teaching evaluation methods,improves teaching efficiency,and carries out reforms and practices in teaching content for undergraduate students in computer science.展开更多
Domain Generation Algorithms(DGAs)continue to pose a significant threat inmodernmalware infrastructures by enabling resilient and evasive communication with Command and Control(C&C)servers.Traditional detection me...Domain Generation Algorithms(DGAs)continue to pose a significant threat inmodernmalware infrastructures by enabling resilient and evasive communication with Command and Control(C&C)servers.Traditional detection methods-rooted in statistical heuristics,feature engineering,and shallow machine learning-struggle to adapt to the increasing sophistication,linguistic mimicry,and adversarial variability of DGA variants.The emergence of Large Language Models(LLMs)marks a transformative shift in this landscape.Leveraging deep contextual understanding,semantic generalization,and few-shot learning capabilities,LLMs such as BERT,GPT,and T5 have shown promising results in detecting both character-based and dictionary-based DGAs,including previously unseen(zeroday)variants.This paper provides a comprehensive and critical review of LLM-driven DGA detection,introducing a structured taxonomy of LLM architectures,evaluating the linguistic and behavioral properties of benchmark datasets,and comparing recent detection frameworks across accuracy,latency,robustness,and multilingual performance.We also highlight key limitations,including challenges in adversarial resilience,model interpretability,deployment scalability,and privacy risks.To address these gaps,we present a forward-looking research roadmap encompassing adversarial training,model compression,cross-lingual benchmarking,and real-time integration with SIEM/SOAR platforms.This survey aims to serve as a foundational resource for advancing the development of scalable,explainable,and operationally viable LLM-based DGA detection systems.展开更多
Vulnerability assessment is a systematic process to identify security gaps in the design and evaluation of physical protection systems.Adversarial path planning is a widely used method for identifying potential vulner...Vulnerability assessment is a systematic process to identify security gaps in the design and evaluation of physical protection systems.Adversarial path planning is a widely used method for identifying potential vulnerabilities and threats to the security and resilience of critical infrastructures.However,achieving efficient path optimization in complex large-scale three-dimensional(3D)scenes remains a significant challenge for vulnerability assessment.This paper introduces a novel A^(*)-algorithmic framework for 3D security modeling and vulnerability assessment.Within this framework,the 3D facility models were first developed in 3ds Max and then incorporated into Unity for A^(*)heuristic pathfinding.The A^(*)-heuristic pathfinding algorithm was implemented with a geometric probability model to refine the detection and distance fields and achieve a rational approximation of the cost to reach the goal.An admissible heuristic is ensured by incorporating the minimum probability of detection(P_(D)^(min))and diagonal distance to estimate the heuristic function.The 3D A^(*)heuristic search was demonstrated using a hypothetical laboratory facility,where a comparison was also carried out between the A^(*)and Dijkstra algorithms for optimal path identification.Comparative results indicate that the proposed A^(*)-heuristic algorithm effectively identifies the most vulnerable adversarial pathfinding with high efficiency.Finally,the paper discusses hidden phenomena and open issues in efficient 3D pathfinding for security applications.展开更多
The distillation process is an important chemical process,and the application of data-driven modelling approach has the potential to reduce model complexity compared to mechanistic modelling,thus improving the efficie...The distillation process is an important chemical process,and the application of data-driven modelling approach has the potential to reduce model complexity compared to mechanistic modelling,thus improving the efficiency of process optimization or monitoring studies.However,the distillation process is highly nonlinear and has multiple uncertainty perturbation intervals,which brings challenges to accurate data-driven modelling of distillation processes.This paper proposes a systematic data-driven modelling framework to solve these problems.Firstly,data segment variance was introduced into the K-means algorithm to form K-means data interval(KMDI)clustering in order to cluster the data into perturbed and steady state intervals for steady-state data extraction.Secondly,maximal information coefficient(MIC)was employed to calculate the nonlinear correlation between variables for removing redundant features.Finally,extreme gradient boosting(XGBoost)was integrated as the basic learner into adaptive boosting(AdaBoost)with the error threshold(ET)set to improve weights update strategy to construct the new integrated learning algorithm,XGBoost-AdaBoost-ET.The superiority of the proposed framework is verified by applying this data-driven modelling framework to a real industrial process of propylene distillation.展开更多
BACKGROUND Difficulty of colonoscopy insertion(DCI)significantly affects colonoscopy effectiveness and serves as a key quality indicator.Predicting and evaluating DCI risk preoperatively is crucial for optimizing intr...BACKGROUND Difficulty of colonoscopy insertion(DCI)significantly affects colonoscopy effectiveness and serves as a key quality indicator.Predicting and evaluating DCI risk preoperatively is crucial for optimizing intraoperative strategies.AIM To evaluate the predictive performance of machine learning(ML)algorithms for DCI by comparing three modeling approaches,identify factors influencing DCI,and develop a preoperative prediction model using ML algorithms to enhance colonoscopy quality and efficiency.METHODS This cross-sectional study enrolled 712 patients who underwent colonoscopy at a tertiary hospital between June 2020 and May 2021.Demographic data,past medical history,medication use,and psychological status were collected.The endoscopist assessed DCI using the visual analogue scale.After univariate screening,predictive models were developed using multivariable logistic regression,least absolute shrinkage and selection operator(LASSO)regression,and random forest(RF)algorithms.Model performance was evaluated based on discrimination,calibration,and decision curve analysis(DCA),and results were visualized using nomograms.RESULTS A total of 712 patients(53.8%male;mean age 54.5 years±12.9 years)were included.Logistic regression analysis identified constipation[odds ratio(OR)=2.254,95%confidence interval(CI):1.289-3.931],abdominal circumference(AC)(77.5–91.9 cm,OR=1.895,95%CI:1.065-3.350;AC≥92 cm,OR=1.271,95%CI:0.730-2.188),and anxiety(OR=1.071,95%CI:1.044-1.100)as predictive factors for DCI,validated by LASSO and RF methods.Model performance revealed training/validation sensitivities of 0.826/0.925,0.924/0.868,and 1.000/0.981;specificities of 0.602/0.511,0.510/0.562,and 0.977/0.526;and corresponding area under the receiver operating characteristic curves(AUCs)of 0.780(0.737-0.823)/0.726(0.654-0.799),0.754(0.710-0.798)/0.723(0.656-0.791),and 1.000(1.000-1.000)/0.754(0.688-0.820),respectively.DCA indicated optimal net benefit within probability thresholds of 0-0.9 and 0.05-0.37.The RF model demonstrated superior diagnostic accuracy,reflected by perfect training sensitivity(1.000)and highest validation AUC(0.754),outperforming other methods in clinical applicability.CONCLUSION The RF-based model exhibited superior predictive accuracy for DCI compared to multivariable logistic and LASSO regression models.This approach supports individualized preoperative optimization,enhancing colonoscopy quality through targeted risk stratification.展开更多
Challenges in stratigraphic modeling arise from underground uncertainty.While borehole exploration is reliable,it remains sparse due to economic and site constraints.Electrical resistivity tomography(ERT)as a cost-eff...Challenges in stratigraphic modeling arise from underground uncertainty.While borehole exploration is reliable,it remains sparse due to economic and site constraints.Electrical resistivity tomography(ERT)as a cost-effective geophysical technique can acquire high-density data;however,uncertainty and nonuniqueness inherent in ERT impede its usage for stratigraphy identification.This paper integrates ERT and onsite observations for the first time to propose a novel method for characterizing stratigraphic profiles.The method consists of two steps:(1)ERT for prior knowledge:ERT data are processed by soft clustering using the Gaussian mixture model,followed by probability smoothing to quantify its depthdependent uncertainty;and(2)Observations for calibration:a spatial sequential Bayesian updating(SSBU)algorithm is developed to update the prior knowledge based on likelihoods derived from onsite observations,namely topsoil and boreholes.The effectiveness of the proposed method is validated through its application to a real slope site in Foshan,China.Comparative analysis with advanced borehole-driven methods highlights the superiority of incorporating ERT data in stratigraphic modeling,in terms of prediction accuracy at borehole locations and sensitivity to borehole data.Informed by ERT,reduced sensitivity to boreholes provides a fundamental solution to the longstanding challenge of sparse measurements.The paper further discusses the impact of ERT uncertainty on the proposed model using time-lapse measurements,the impact of model resolution,and applicability in engineering projects.This study,as a breakthrough in stratigraphic modeling,bridges gaps in combining geophysical and geotechnical data to address measurement sparsity and paves the way for more economical geotechnical exploration.展开更多
The work proposes a distributed Kalman filtering(KF)algorithm to track a time-varying unknown signal process for a stochastic regression model over network systems in a cooperative way.We provide the stability analysi...The work proposes a distributed Kalman filtering(KF)algorithm to track a time-varying unknown signal process for a stochastic regression model over network systems in a cooperative way.We provide the stability analysis of the proposed distributed KF algorithm without independent and stationary signal assumptions,which implies that the theoretical results are able to be applied to stochastic feedback systems.Note that the main difficulty of stability analysis lies in analyzing the properties of the product of non-independent and non-stationary random matrices involved in the error equation.We employ analysis techniques such as stochastic Lyapunov function,stability theory of stochastic systems,and algebraic graph theory to deal with the above issue.The stochastic spatio-temporal cooperative information condition shows the cooperative property of multiple sensors that even though any local sensor cannot track the time-varying unknown signal,the distributed KF algorithm can be utilized to finish the filtering task in a cooperative way.At last,we illustrate the property of the proposed distributed KF algorithm by a simulation example.展开更多
With the rapid development of transportation infrastructure,ensuring road safety through timely and accurate highway inspection has become increasingly critical.Traditional manual inspection methods are not only time-...With the rapid development of transportation infrastructure,ensuring road safety through timely and accurate highway inspection has become increasingly critical.Traditional manual inspection methods are not only time-consuming and labor-intensive,but they also struggle to provide consistent,high-precision detection and realtime monitoring of pavement surface defects.To overcome these limitations,we propose an Automatic Recognition of PavementDefect(ARPD)algorithm,which leverages unmanned aerial vehicle(UAV)-based aerial imagery to automate the inspection process.The ARPD framework incorporates a backbone network based on the Selective State Space Model(S3M),which is designed to capture long-range temporal dependencies.This enables effective modeling of dynamic correlations among redundant and often repetitive structures commonly found in road imagery.Furthermore,a neck structure based on Semantics and Detail Infusion(SDI)is introduced to guide cross-scale feature fusion.The SDI module enhances the integration of low-level spatial details with high-level semantic cues,thereby improving feature expressiveness and defect localization accuracy.Experimental evaluations demonstrate that theARPDalgorithm achieves a mean average precision(mAP)of 86.1%on a custom-labeled pavement defect dataset,outperforming the state-of-the-art YOLOv11 segmentation model.The algorithm also maintains strong generalization ability on public datasets.These results confirm that ARPD is well-suited for diverse real-world applications in intelligent,large-scale highway defect monitoring and maintenance planning.展开更多
BACKGROUND:This study aims to develop and validate a machine learning-based in-hospital mortality predictive model for acute aortic syndrome(AAS)in the emergency department(ED)and to derive a simplifi ed version suita...BACKGROUND:This study aims to develop and validate a machine learning-based in-hospital mortality predictive model for acute aortic syndrome(AAS)in the emergency department(ED)and to derive a simplifi ed version suitable for rapid clinical application.METHODS:In this multi-center retrospective cohort study,AAS patient data from three hospitals were analyzed.The modeling cohort included data from the First Affiliated Hospital of Zhengzhou University and the People’s Hospital of Xinjiang Uygur Autonomous Region,with Peking University Third Hospital data serving as the external test set.Four machine learning algorithms—logistic regression(LR),multilayer perceptron(MLP),Gaussian naive Bayes(GNB),and random forest(RF)—were used to develop predictive models based on 34 early-accessible clinical variables.A simplifi ed model was then derived based on fi ve key variables(Stanford type,pericardial eff usion,asymmetric peripheral arterial pulsation,decreased bowel sounds,and dyspnea)via Least Absolute Shrinkage and Selection Operator(LASSO)regression to improve ED applicability.RESULTS:A total of 929 patients were included in the modeling cohort,and 210 were included in the external test set.Four machine learning models based on 34 clinical variables were developed,achieving internal and external validation AUCs of 0.85-0.90 and 0.73-0.85,respectively.The simplifi ed model incorporating fi ve key variables demonstrated internal and external validation AUCs of 0.71-0.86 and 0.75-0.78,respectively.Both models showed robust calibration and predictive stability across datasets.CONCLUSION:Both kinds of models were built based on machine learning tools,and proved to have certain prediction performance and extrapolation.展开更多
Accurate prediction of flood events is important for flood control and risk management.Machine learning techniques contributed greatly to advances in flood predictions,and existing studies mainly focused on predicting...Accurate prediction of flood events is important for flood control and risk management.Machine learning techniques contributed greatly to advances in flood predictions,and existing studies mainly focused on predicting flood resource variables using single or hybrid machine learning techniques.However,class-based flood predictions have rarely been investigated,which can aid in quickly diagnosing comprehensive flood characteristics and proposing targeted management strategies.This study proposed a prediction approach of flood regime metrics and event classes coupling machine learning algorithms with clustering-deduced membership degrees.Five algorithms were adopted for this exploration.Results showed that the class membership degrees accurately determined event classes with class hit rates up to 100%,compared with the four classes clustered from nine regime metrics.The nonlinear algorithms(Multiple Linear Regression,Random Forest,and least squares-Support Vector Machine)outperformed the linear techniques(Multiple Linear Regression and Stepwise Regression)in predicting flood regime metrics.The proposed approach well predicted flood event classes with average class hit rates of 66.0%-85.4%and 47.2%-76.0%in calibration and validation periods,respectively,particularly for the slow and late flood events.The predictive capability of the proposed prediction approach for flood regime metrics and classes was considerably stronger than that of hydrological modeling approach.展开更多
Gas-bearing volcanic reservoirs have been found in the deep Songliao Basin, China. Choosing proper interpretation parameters for log evaluation is difficult due to complicated mineral compositions and variable mineral...Gas-bearing volcanic reservoirs have been found in the deep Songliao Basin, China. Choosing proper interpretation parameters for log evaluation is difficult due to complicated mineral compositions and variable mineral contents. Based on the QAPF classification scheme given by IUGS, we propose a method to determine the mineral contents of volcanic rocks using log data and a genetic algorithm. According to the QAPF scheme, minerals in volcanic rocks are divided into five groups: Q(quartz), A (Alkaline feldspar), P (plagioclase), M (mafic) and F (feldspathoid). We propose a model called QAPM including porosity for the volumetric analysis of reservoirs. The log response equations for density, apparent neutron porosity, transit time, gamma ray and volume photoelectrical cross section index were first established with the mineral parameters obtained from the Schlumberger handbook of log mineral parameters. Then the volumes of the four minerals in the matrix were calculated using the genetic algorithm (GA). The calculated porosity, based on the interpretation parameters, can be compared with core porosity, and the rock names given in the paper based on QAPF classification according to the four mineral contents are compatible with those from the chemical analysis of the core samples.展开更多
基金funding from the European Commission by the Ruralities project(grant agreement no.101060876).
文摘In this paper,we propose a new privacy-aware transmission scheduling algorithm for 6G ad hoc networks.This system enables end nodes to select the optimum time and scheme to transmit private data safely.In 6G dynamic heterogeneous infrastructures,unstable links and non-uniform hardware capabilities create critical issues regarding security and privacy.Traditional protocols are often too computationally heavy to allow 6G services to achieve their expected Quality-of-Service(QoS).As the transport network is built of ad hoc nodes,there is no guarantee about their trustworthiness or behavior,and transversal functionalities are delegated to the extreme nodes.However,while security can be guaranteed in extreme-to-extreme solutions,privacy cannot,as all intermediate nodes still have to handle the data packets they are transporting.Besides,traditional schemes for private anonymous ad hoc communications are vulnerable against modern intelligent attacks based on learning models.The proposed scheme fulfills this gap.Findings show the probability of a successful intelligent attack reduces by up to 65%compared to ad hoc networks with no privacy protection strategy when used the proposed technology.While congestion probability can remain below 0.001%,as required in 6G services.
文摘Early and accurate detection of bone cancer and marrow cell abnormalities is critical for timely intervention and improved patient outcomes.This paper proposes a novel hybrid deep learning framework that integrates a Convolutional Neural Network(CNN)with a Bidirectional Long Short-Term Memory(BiLSTM)architecture,optimized using the Firefly Optimization algorithm(FO).The proposed CNN-BiLSTM-FO model is tailored for structured biomedical data,capturing both local patterns and sequential dependencies in diagnostic features,while the Firefly Algorithm fine-tunes key hyperparameters to maximize predictive performance.The approach is evaluated on two benchmark biomedical datasets:one comprising diagnostic data for bone cancer detection and another for identifying marrow cell abnormalities.Experimental results demonstrate that the proposed method outperforms standard deep learning models,including CNN,LSTM,BiLSTM,and CNN-LSTM hybrids,significantly.The CNNBiLSTM-FO model achieves an accuracy of 98.55%for bone cancer detection and 96.04%for marrow abnormality classification.The paper also presents a detailed complexity analysis of the proposed algorithm and compares its performance across multiple evaluation metrics such as precision,recall,F1-score,and AUC.The results confirm the effectiveness of the firefly-based optimization strategy in improving classification accuracy and model robustness.This work introduces a scalable and accurate diagnostic solution that holds strong potential for integration into intelligent clinical decision-support systems.
基金supported by the National Natural Science Foundation of China[grant number 51775020]the Science Challenge Project[grant number.TZ2018007]+2 种基金the National Natural Science Foundation of China[grant number 62073009]the Postdoctoral Fellowship Program of CPSF[grant number GZC20233365]the Fundamental Research Funds for Central Universities[grant number JKF-20240559].
文摘Parameter extraction of photovoltaic(PV)models is crucial for the planning,optimization,and control of PV systems.Although some methods using meta-heuristic algorithms have been proposed to determine these parameters,the robustness of solutions obtained by these methods faces great challenges when the complexity of the PV model increases.The unstable results will affect the reliable operation and maintenance strategies of PV systems.In response to this challenge,an improved rime optimization algorithm with enhanced exploration and exploitation,termed TERIME,is proposed for robust and accurate parameter identification for various PV models.Specifically,the differential evolution mutation operator is integrated in the exploration phase to enhance the population diversity.Meanwhile,a new exploitation strategy incorporating randomization and neighborhood strategies simultaneously is developed to maintain the balance of exploitation width and depth.The TERIME algorithm is applied to estimate the optimal parameters of the single diode model,double diode model,and triple diode model combined with the Lambert-W function for three PV cell and module types including RTC France,Photo Watt-PWP 201 and S75.According to the statistical analysis in 100 runs,the proposed algorithm achieves more accurate and robust parameter estimations than other techniques to various PV models in varying environmental conditions.All of our source codes are publicly available at https://github.com/dirge1/TERIME.
基金financial support from the National Key R&D Program of China(Grant No.2020YFB1711100).
文摘To address the challenge of identifying the primary causes of energy consumption fluctuations and accurately assessing the influence of various factors in the converter unit of an iron and steel plant,the focus is placed on the critical components of material and heat balance.Through a thorough analysis of the interactions between various components and energy consumptions,six pivotal factors have been identified—raw material composition,steel type,steel temperature,slag temperature,recycling practices,and operational parameters.Utilizing a framework based on an equivalent energy consumption model,an integrated intelligent diagnostic model has been developed that encapsulates these factors,providing a comprehensive assessment tool for converter energy consumption.Employing the K-means clustering algorithm,historical operational data from the converter have been meticulously analyzed to determine baseline values for essential variables such as energy consumption and recovery rates.Building upon this data-driven foundation,an innovative online system for the intelligent diagnosis of converter energy consumption has been crafted and implemented,enhancing the precision and efficiency of energy management.Upon implementation with energy consumption data at a steel plant in 2023,the diagnostic analysis performed by the system exposed significant variations in energy usage across different converter units.The analysis revealed that the most significant factor influencing the variation in energy consumption for both furnaces was the steel grade,with contributions of−0.550 and 0.379.
基金funded by the National Key R&D Program of China(Grant no.2022YFC2807504)the Marine S&T Fund of Shandong Province for Qingdao Marine Science and Technology Center(Grant no.2022QNLM030002-1)the Central Public-interest Scientific Institution Basal Research(Grant no.2023TD02).
文摘Antarctic krill(Euphausia superba),widely distributes around Antarctica,is a key species supporting the biodiversity of the Southern Ocean ecosystem.The Commission for the Conservation of Antarctic Marine Living Resources(CCAMLR)has thus managed the krill fishery according to a precautionary way.Currently,CCAMLR is making effort to develop a refined krill fishery management approach based on more solid science,which requires accurate predictions of krill distribution.To address this need,this study investigated the effects of algorithm and spatial resolution on the performance of Antarctic krill distribution modelling.We integrated acoustic data from 4 surveys conducted in the waters adjacent to the Antarctic Peninsula with 11 environmental variables characterizing krill prey conditions,water mass properties,and seafloor topography.These data were processed at 4 spatial resolutions(5,10,15,and 20 km)to fit distribution models using 4 algorithms:Random Forests(RF),Generalized Additive Models(GAM),Extreme Gradient Boosting(XGBoost),and Artificial Neural Networks(ANN).Model performance was assessed and compared in terms of goodness-of-fit and predictive accuracy.The results showed that RF achieved the highest predictive performance at most resolutions,whereas GAM performed best at the coarsest resolution(20 km).XGBoost closely following RF in accuracy and demonstrated robustness as evidenced by the highly consistent partial dependence curves across resolutions.In contrast,ANN exhibited limitations with smaller sample sizes,resulting in comparatively poorer predictive performance.The analysis revealed a trade-off whereby reducing spatial resolution improved model fit and mitigated zero-inflation at the expense of fine-scale information and overall predictive accuracy.Ensemble models,integrating RF,GAM,and XGBoost,are proposed as potential balanced solutions to improve predictive stability,offering a more robust scientific basis for the refinement of krill management.
基金supported by the Scientific and Technological Research Council of Turkey(TUBITAK)[grant no 122K637].
文摘The estimation of the probability of informed trading(PIN)model and its extensions poses significant challenges owing to various computational problems.To address these issues,we propose a novel estimation method called the expectation-conditional-maximization(ECM)algorithm,which can serve as an alternative to the existing methods for estimating PIN models.Our method provides optimal estimates for the original PIN model as well as two of its extensions:the multilayer PIN model and the adjusted PIN model,along with its restricted versions.Our results indicate that estimations using the ECM algorithm are generally faster,more accurate,and more memory-efficient than the standard methods used in the literature,making it a robust alternative.More importantly,the ECM algorithm is not limited to the models discussed and can be easily adapted to estimate future extensions of the PIN model.
基金funded by Anhui NARI ZT Electric Co.,Ltd.,entitled“Research on the Shared Operation and Maintenance Service Model for Metering Equipment and Platform Development for the Modern Industrial Chain”(Grant No.524636250005).
文摘With the rapid adoption of artificial intelligence(AI)in domains such as power,transportation,and finance,the number of machine learning and deep learning models has grown exponentially.However,challenges such as delayed retraining,inconsistent version management,insufficient drift monitoring,and limited data security still hinder efficient and reliable model operations.To address these issues,this paper proposes the Intelligent Model Lifecycle Management Algorithm(IMLMA).The algorithm employs a dual-trigger mechanism based on both data volume thresholds and time intervals to automate retraining,and applies Bayesian optimization for adaptive hyperparameter tuning to improve performance.A multi-metric replacement strategy,incorporating MSE,MAE,and R2,ensures that new models replace existing ones only when performance improvements are guaranteed.A versioning and traceability database supports comparison and visualization,while real-time monitoring with stability analysis enables early warnings of latency and drift.Finally,hash-based integrity checks secure both model files and datasets.Experimental validation in a power metering operation scenario demonstrates that IMLMA reduces model update delays,enhances predictive accuracy and stability,and maintains low latency under high concurrency.This work provides a practical,reusable,and scalable solution for intelligent model lifecycle management,with broad applicability to complex systems such as smart grids.
基金supported by the National Natural Science Foundation of China(Grant Nos.62303197,62273214)the Natural Science Foundation of Shandong Province(ZR2024MFO18).
文摘Accurate and reliable photovoltaic(PV)modeling is crucial for the performance evaluation,control,and optimization of PV systems.However,existing methods for PV parameter identification often suffer from limitations in accuracy and efficiency.To address these challenges,we propose an adaptive multi-learning cooperation search algorithm(AMLCSA)for efficient identification of unknown parameters in PV models.AMLCSA is a novel algorithm inspired by teamwork behaviors in modern enterprises.It enhances the original cooperation search algorithm in two key aspects:(i)an adaptive multi-learning strategy that dynamically adjusts search ranges using adaptive weights,allowing better individuals to focus on local exploitation while guiding poorer individuals toward global exploration;and(ii)a chaotic grouping reflection strategy that introduces chaotic sequences to enhance population diversity and improve search performance.The effectiveness of AMLCSA is demonstrated on single-diode,double-diode,and three PV-module models.Simulation results show that AMLCSA offers significant advantages in convergence,accuracy,and stability compared to existing state-of-the-art algorithms.
基金supported in part by the National Natural Science Foundation of China(51775385)the Natural Science Foundation of Shanghai(23ZR1466000)+2 种基金the Shanghai Industrial Collaborative Science and Technology Innovation Project(2021-cyxt2-kj10)the Innovation Program of Shanghai Municipal Education Commission(202101070007E00098)Fundo para o Desenvolvimento das Ciencias e da Tecnologia(FDCT)(0147/2024/AFJ).
文摘When dealing with expensive multiobjective optimization problems,majority of existing surrogate-assisted evolutionary algorithms(SAEAs)generate solutions in decision space and screen candidate solutions mostly by using designed surrogate models.The generated solutions exhibit excessive randomness,which tends to reduce the likelihood of generating good-quality solutions and cause a long evolution to the optima.To improve SAEAs greatly,this work proposes an evolutionary algorithm based on surrogate and inverse surrogate models by 1)Employing a surrogate model in lieu of expensive(true)function evaluations;and 2)Proposing and using an inverse surrogate model to generate new solutions.By using the same training data but with its inputs and outputs being reversed,the latter is simple to train.It is then used to generate new vectors in objective space,which are mapped into decision space to obtain their corresponding solutions.Using a particular example,this work shows its advantages over existing SAEAs.The results of comparing it with state-of-the-art algorithms on expensive optimization problems show that it is highly competitive in both solution performance and efficiency.
基金supported by Harbin Engineering University’s 2021 Education Reform Project“How to Make Computer Theory Teaching Serve Employment”(Grant No.JG2021B0609).
文摘“Algorithm Design and Analysis”is not only one of the important courses in the undergraduate teaching of computer science and technology but also a key part of computer professional skills.In recent years,with the rise and widespread application of big language models,many teaching reform plans have been produced to promote the quality and efficiency of teaching.This paper studies how to refer to software development professional skills standards,investigates the knowledge points of“Algorithm Design and Analysis”courses in other educational institutions,uses cutting-edge core technology big language models to drive the improvement of teaching evaluation methods,improves teaching efficiency,and carries out reforms and practices in teaching content for undergraduate students in computer science.
基金the Deanship of Scientific Research at King Khalid University for funding this work through large group under grant number(GRP.2/663/46).
文摘Domain Generation Algorithms(DGAs)continue to pose a significant threat inmodernmalware infrastructures by enabling resilient and evasive communication with Command and Control(C&C)servers.Traditional detection methods-rooted in statistical heuristics,feature engineering,and shallow machine learning-struggle to adapt to the increasing sophistication,linguistic mimicry,and adversarial variability of DGA variants.The emergence of Large Language Models(LLMs)marks a transformative shift in this landscape.Leveraging deep contextual understanding,semantic generalization,and few-shot learning capabilities,LLMs such as BERT,GPT,and T5 have shown promising results in detecting both character-based and dictionary-based DGAs,including previously unseen(zeroday)variants.This paper provides a comprehensive and critical review of LLM-driven DGA detection,introducing a structured taxonomy of LLM architectures,evaluating the linguistic and behavioral properties of benchmark datasets,and comparing recent detection frameworks across accuracy,latency,robustness,and multilingual performance.We also highlight key limitations,including challenges in adversarial resilience,model interpretability,deployment scalability,and privacy risks.To address these gaps,we present a forward-looking research roadmap encompassing adversarial training,model compression,cross-lingual benchmarking,and real-time integration with SIEM/SOAR platforms.This survey aims to serve as a foundational resource for advancing the development of scalable,explainable,and operationally viable LLM-based DGA detection systems.
基金supported by the fundings from 2024 Young Talents Program for Science and Technology Thinking Tanks(No.XMSB20240711041)2024 Student Research Program on Dynamic Simulation and Force-on-Force Exercise of Nuclear Security in 3D Interactive Environment Using Reinforcement Learning,Natural Science Foundation of Top Talent of SZTU(No.GDRC202407)+2 种基金Shenzhen Science and Technology Program(No.KCXFZ20240903092603005)Shenzhen Science and Technology Program(No.JCYJ20241202124703004)Shenzhen Science and Technology Program(No.KJZD20230923114117032)。
文摘Vulnerability assessment is a systematic process to identify security gaps in the design and evaluation of physical protection systems.Adversarial path planning is a widely used method for identifying potential vulnerabilities and threats to the security and resilience of critical infrastructures.However,achieving efficient path optimization in complex large-scale three-dimensional(3D)scenes remains a significant challenge for vulnerability assessment.This paper introduces a novel A^(*)-algorithmic framework for 3D security modeling and vulnerability assessment.Within this framework,the 3D facility models were first developed in 3ds Max and then incorporated into Unity for A^(*)heuristic pathfinding.The A^(*)-heuristic pathfinding algorithm was implemented with a geometric probability model to refine the detection and distance fields and achieve a rational approximation of the cost to reach the goal.An admissible heuristic is ensured by incorporating the minimum probability of detection(P_(D)^(min))and diagonal distance to estimate the heuristic function.The 3D A^(*)heuristic search was demonstrated using a hypothetical laboratory facility,where a comparison was also carried out between the A^(*)and Dijkstra algorithms for optimal path identification.Comparative results indicate that the proposed A^(*)-heuristic algorithm effectively identifies the most vulnerable adversarial pathfinding with high efficiency.Finally,the paper discusses hidden phenomena and open issues in efficient 3D pathfinding for security applications.
基金supported by the National Key Research and Development Program of China(2023YFB3307801)the National Natural Science Foundation of China(62394343,62373155,62073142)+3 种基金Major Science and Technology Project of Xinjiang(No.2022A01006-4)the Programme of Introducing Talents of Discipline to Universities(the 111 Project)under Grant B17017the Fundamental Research Funds for the Central Universities,Science Foundation of China University of Petroleum,Beijing(No.2462024YJRC011)the Open Research Project of the State Key Laboratory of Industrial Control Technology,China(Grant No.ICT2024B70).
文摘The distillation process is an important chemical process,and the application of data-driven modelling approach has the potential to reduce model complexity compared to mechanistic modelling,thus improving the efficiency of process optimization or monitoring studies.However,the distillation process is highly nonlinear and has multiple uncertainty perturbation intervals,which brings challenges to accurate data-driven modelling of distillation processes.This paper proposes a systematic data-driven modelling framework to solve these problems.Firstly,data segment variance was introduced into the K-means algorithm to form K-means data interval(KMDI)clustering in order to cluster the data into perturbed and steady state intervals for steady-state data extraction.Secondly,maximal information coefficient(MIC)was employed to calculate the nonlinear correlation between variables for removing redundant features.Finally,extreme gradient boosting(XGBoost)was integrated as the basic learner into adaptive boosting(AdaBoost)with the error threshold(ET)set to improve weights update strategy to construct the new integrated learning algorithm,XGBoost-AdaBoost-ET.The superiority of the proposed framework is verified by applying this data-driven modelling framework to a real industrial process of propylene distillation.
基金the Chinese Clinical Trial Registry(No.ChiCTR2000040109)approved by the Hospital Ethics Committee(No.20210130017).
文摘BACKGROUND Difficulty of colonoscopy insertion(DCI)significantly affects colonoscopy effectiveness and serves as a key quality indicator.Predicting and evaluating DCI risk preoperatively is crucial for optimizing intraoperative strategies.AIM To evaluate the predictive performance of machine learning(ML)algorithms for DCI by comparing three modeling approaches,identify factors influencing DCI,and develop a preoperative prediction model using ML algorithms to enhance colonoscopy quality and efficiency.METHODS This cross-sectional study enrolled 712 patients who underwent colonoscopy at a tertiary hospital between June 2020 and May 2021.Demographic data,past medical history,medication use,and psychological status were collected.The endoscopist assessed DCI using the visual analogue scale.After univariate screening,predictive models were developed using multivariable logistic regression,least absolute shrinkage and selection operator(LASSO)regression,and random forest(RF)algorithms.Model performance was evaluated based on discrimination,calibration,and decision curve analysis(DCA),and results were visualized using nomograms.RESULTS A total of 712 patients(53.8%male;mean age 54.5 years±12.9 years)were included.Logistic regression analysis identified constipation[odds ratio(OR)=2.254,95%confidence interval(CI):1.289-3.931],abdominal circumference(AC)(77.5–91.9 cm,OR=1.895,95%CI:1.065-3.350;AC≥92 cm,OR=1.271,95%CI:0.730-2.188),and anxiety(OR=1.071,95%CI:1.044-1.100)as predictive factors for DCI,validated by LASSO and RF methods.Model performance revealed training/validation sensitivities of 0.826/0.925,0.924/0.868,and 1.000/0.981;specificities of 0.602/0.511,0.510/0.562,and 0.977/0.526;and corresponding area under the receiver operating characteristic curves(AUCs)of 0.780(0.737-0.823)/0.726(0.654-0.799),0.754(0.710-0.798)/0.723(0.656-0.791),and 1.000(1.000-1.000)/0.754(0.688-0.820),respectively.DCA indicated optimal net benefit within probability thresholds of 0-0.9 and 0.05-0.37.The RF model demonstrated superior diagnostic accuracy,reflected by perfect training sensitivity(1.000)and highest validation AUC(0.754),outperforming other methods in clinical applicability.CONCLUSION The RF-based model exhibited superior predictive accuracy for DCI compared to multivariable logistic and LASSO regression models.This approach supports individualized preoperative optimization,enhancing colonoscopy quality through targeted risk stratification.
基金the financial support from the National Key R&D Program of China(Grant No.2021YFC3001003)Science and Technology Development Fund,Macao SAR(File No.0056/2023/RIB2)Guangdong Provincial Department of Science and Technology(Grant No.2022A0505030019).
文摘Challenges in stratigraphic modeling arise from underground uncertainty.While borehole exploration is reliable,it remains sparse due to economic and site constraints.Electrical resistivity tomography(ERT)as a cost-effective geophysical technique can acquire high-density data;however,uncertainty and nonuniqueness inherent in ERT impede its usage for stratigraphy identification.This paper integrates ERT and onsite observations for the first time to propose a novel method for characterizing stratigraphic profiles.The method consists of two steps:(1)ERT for prior knowledge:ERT data are processed by soft clustering using the Gaussian mixture model,followed by probability smoothing to quantify its depthdependent uncertainty;and(2)Observations for calibration:a spatial sequential Bayesian updating(SSBU)algorithm is developed to update the prior knowledge based on likelihoods derived from onsite observations,namely topsoil and boreholes.The effectiveness of the proposed method is validated through its application to a real slope site in Foshan,China.Comparative analysis with advanced borehole-driven methods highlights the superiority of incorporating ERT data in stratigraphic modeling,in terms of prediction accuracy at borehole locations and sensitivity to borehole data.Informed by ERT,reduced sensitivity to boreholes provides a fundamental solution to the longstanding challenge of sparse measurements.The paper further discusses the impact of ERT uncertainty on the proposed model using time-lapse measurements,the impact of model resolution,and applicability in engineering projects.This study,as a breakthrough in stratigraphic modeling,bridges gaps in combining geophysical and geotechnical data to address measurement sparsity and paves the way for more economical geotechnical exploration.
基金supported in part by Sichuan Science and Technology Program under Grant No.2025ZNSFSC151in part by the Strategic Priority Research Program of Chinese Academy of Sciences under Grant No.XDA27030201+1 种基金the Natural Science Foundation of China under Grant No.U21B6001in part by the Natural Science Foundation of Tianjin under Grant No.24JCQNJC01930.
文摘The work proposes a distributed Kalman filtering(KF)algorithm to track a time-varying unknown signal process for a stochastic regression model over network systems in a cooperative way.We provide the stability analysis of the proposed distributed KF algorithm without independent and stationary signal assumptions,which implies that the theoretical results are able to be applied to stochastic feedback systems.Note that the main difficulty of stability analysis lies in analyzing the properties of the product of non-independent and non-stationary random matrices involved in the error equation.We employ analysis techniques such as stochastic Lyapunov function,stability theory of stochastic systems,and algebraic graph theory to deal with the above issue.The stochastic spatio-temporal cooperative information condition shows the cooperative property of multiple sensors that even though any local sensor cannot track the time-varying unknown signal,the distributed KF algorithm can be utilized to finish the filtering task in a cooperative way.At last,we illustrate the property of the proposed distributed KF algorithm by a simulation example.
基金supported in part by the Technical Service for the Development and Application of an Intelligent Visual Management Platformfor Expressway Construction Progress Based on BIM Technology(grant NO.JKYZLX-2023-09)in partby the Technical Service for the Development of an Early Warning Model in the Research and Application of Key Technologies for Tunnel Operation Safety Monitoring and Early Warning Based on Digital Twin(grant NO.JK-S02-ZNGS-202412-JISHU-FA-0035)sponsored by Yunnan Transportation Science Research Institute Co.,Ltd.
文摘With the rapid development of transportation infrastructure,ensuring road safety through timely and accurate highway inspection has become increasingly critical.Traditional manual inspection methods are not only time-consuming and labor-intensive,but they also struggle to provide consistent,high-precision detection and realtime monitoring of pavement surface defects.To overcome these limitations,we propose an Automatic Recognition of PavementDefect(ARPD)algorithm,which leverages unmanned aerial vehicle(UAV)-based aerial imagery to automate the inspection process.The ARPD framework incorporates a backbone network based on the Selective State Space Model(S3M),which is designed to capture long-range temporal dependencies.This enables effective modeling of dynamic correlations among redundant and often repetitive structures commonly found in road imagery.Furthermore,a neck structure based on Semantics and Detail Infusion(SDI)is introduced to guide cross-scale feature fusion.The SDI module enhances the integration of low-level spatial details with high-level semantic cues,thereby improving feature expressiveness and defect localization accuracy.Experimental evaluations demonstrate that theARPDalgorithm achieves a mean average precision(mAP)of 86.1%on a custom-labeled pavement defect dataset,outperforming the state-of-the-art YOLOv11 segmentation model.The algorithm also maintains strong generalization ability on public datasets.These results confirm that ARPD is well-suited for diverse real-world applications in intelligent,large-scale highway defect monitoring and maintenance planning.
基金supported by the special fund of the National Clinical Key Specialty Construction Program[(2022)301-2305].
文摘BACKGROUND:This study aims to develop and validate a machine learning-based in-hospital mortality predictive model for acute aortic syndrome(AAS)in the emergency department(ED)and to derive a simplifi ed version suitable for rapid clinical application.METHODS:In this multi-center retrospective cohort study,AAS patient data from three hospitals were analyzed.The modeling cohort included data from the First Affiliated Hospital of Zhengzhou University and the People’s Hospital of Xinjiang Uygur Autonomous Region,with Peking University Third Hospital data serving as the external test set.Four machine learning algorithms—logistic regression(LR),multilayer perceptron(MLP),Gaussian naive Bayes(GNB),and random forest(RF)—were used to develop predictive models based on 34 early-accessible clinical variables.A simplifi ed model was then derived based on fi ve key variables(Stanford type,pericardial eff usion,asymmetric peripheral arterial pulsation,decreased bowel sounds,and dyspnea)via Least Absolute Shrinkage and Selection Operator(LASSO)regression to improve ED applicability.RESULTS:A total of 929 patients were included in the modeling cohort,and 210 were included in the external test set.Four machine learning models based on 34 clinical variables were developed,achieving internal and external validation AUCs of 0.85-0.90 and 0.73-0.85,respectively.The simplifi ed model incorporating fi ve key variables demonstrated internal and external validation AUCs of 0.71-0.86 and 0.75-0.78,respectively.Both models showed robust calibration and predictive stability across datasets.CONCLUSION:Both kinds of models were built based on machine learning tools,and proved to have certain prediction performance and extrapolation.
基金National Key Research and Development Program of China,No.2023YFC3006704National Natural Science Foundation of China,No.42171047CAS-CSIRO Partnership Joint Project of 2024,No.177GJHZ2023097MI。
文摘Accurate prediction of flood events is important for flood control and risk management.Machine learning techniques contributed greatly to advances in flood predictions,and existing studies mainly focused on predicting flood resource variables using single or hybrid machine learning techniques.However,class-based flood predictions have rarely been investigated,which can aid in quickly diagnosing comprehensive flood characteristics and proposing targeted management strategies.This study proposed a prediction approach of flood regime metrics and event classes coupling machine learning algorithms with clustering-deduced membership degrees.Five algorithms were adopted for this exploration.Results showed that the class membership degrees accurately determined event classes with class hit rates up to 100%,compared with the four classes clustered from nine regime metrics.The nonlinear algorithms(Multiple Linear Regression,Random Forest,and least squares-Support Vector Machine)outperformed the linear techniques(Multiple Linear Regression and Stepwise Regression)in predicting flood regime metrics.The proposed approach well predicted flood event classes with average class hit rates of 66.0%-85.4%and 47.2%-76.0%in calibration and validation periods,respectively,particularly for the slow and late flood events.The predictive capability of the proposed prediction approach for flood regime metrics and classes was considerably stronger than that of hydrological modeling approach.
基金National Natural Science Foundation of China (No. 49894194-4)
文摘Gas-bearing volcanic reservoirs have been found in the deep Songliao Basin, China. Choosing proper interpretation parameters for log evaluation is difficult due to complicated mineral compositions and variable mineral contents. Based on the QAPF classification scheme given by IUGS, we propose a method to determine the mineral contents of volcanic rocks using log data and a genetic algorithm. According to the QAPF scheme, minerals in volcanic rocks are divided into five groups: Q(quartz), A (Alkaline feldspar), P (plagioclase), M (mafic) and F (feldspathoid). We propose a model called QAPM including porosity for the volumetric analysis of reservoirs. The log response equations for density, apparent neutron porosity, transit time, gamma ray and volume photoelectrical cross section index were first established with the mineral parameters obtained from the Schlumberger handbook of log mineral parameters. Then the volumes of the four minerals in the matrix were calculated using the genetic algorithm (GA). The calculated porosity, based on the interpretation parameters, can be compared with core porosity, and the rock names given in the paper based on QAPF classification according to the four mineral contents are compatible with those from the chemical analysis of the core samples.