Concrete-filled steel tubes(CFST)are widely utilized in civil engineering due to their superior load-bearing capacity,ductility,and seismic resistance.However,existing design codes,such as AISC and Eurocode 4,tend to ...Concrete-filled steel tubes(CFST)are widely utilized in civil engineering due to their superior load-bearing capacity,ductility,and seismic resistance.However,existing design codes,such as AISC and Eurocode 4,tend to be excessively conservative as they fail to account for the composite action between the steel tube and the concrete core.To address this limitation,this study proposes a hybrid model that integrates XGBoost with the Pied Kingfisher Optimizer(PKO),a nature-inspired algorithm,to enhance the accuracy of shear strength prediction for CFST columns.Additionally,quantile regression is employed to construct prediction intervals for the ultimate shear force,while the Asymmetric Squared Error Loss(ASEL)function is incorporated to mitigate overestimation errors.The computational results demonstrate that the PKO-XGBoost model delivers superior predictive accuracy,achieving a Mean Absolute Percentage Error(MAPE)of 4.431%and R2 of 0.9925 on the test set.Furthermore,the ASEL-PKO-XGBoost model substantially reduces overestimation errors to 28.26%,with negligible impact on predictive performance.Additionally,based on the Genetic Algorithm(GA)and existing equation models,a strength equation model is developed,achieving markedly higher accuracy than existing models(R^(2)=0.934).Lastly,web-based Graphical User Interfaces(GUIs)were developed to enable real-time prediction.展开更多
This paper presents an Eulerian-Lagrangian algorithm for direct numerical simulation(DNS)of particle-laden flows.The algorithm is applicable to perform simulations of dilute suspensions of small inertial particles in ...This paper presents an Eulerian-Lagrangian algorithm for direct numerical simulation(DNS)of particle-laden flows.The algorithm is applicable to perform simulations of dilute suspensions of small inertial particles in turbulent carrier flow.The Eulerian framework numerically resolves turbulent carrier flow using a parallelized,finite-volume DNS solver on a staggered Cartesian grid.Particles are tracked using a point-particle method utilizing a Lagrangian particle tracking(LPT)algorithm.The proposed Eulerian-Lagrangian algorithm is validated using an inertial particle-laden turbulent channel flow for different Stokes number cases.The particle concentration profiles and higher-order statistics of the carrier and dispersed phases agree well with the benchmark results.We investigated the effect of fluid velocity interpolation and numerical integration schemes of particle tracking algorithms on particle dispersion statistics.The suitability of fluid velocity interpolation schemes for predicting the particle dispersion statistics is discussed in the framework of the particle tracking algorithm coupled to the finite-volume solver.In addition,we present parallelization strategies implemented in the algorithm and evaluate their parallel performance.展开更多
The original intention of the algorithmic recommender system is to grapple with the negative impacts caused by information overload,but the system also can be used as"hypernudge",a new form of online manipul...The original intention of the algorithmic recommender system is to grapple with the negative impacts caused by information overload,but the system also can be used as"hypernudge",a new form of online manipulation,to inten⁃tionally exploit people's cognitive and decision-making gaps to influence their decisions in practice,which is particu⁃larly detrimental to the sustainable development of the digital market.Limiting harmful algorithmic online manipula⁃tion in digital markets has become a challenging task.Globally,both the EU and China have responded to this issue,and the differences between them are so evident that their governance measures can serve as the typical case.The EU focuses on improving citizens'digital literacy and their ability to integrate into digital social life to independently ad⁃dress this issue,and expects to address harmful manipulation behavior through binding and applicable hard law,which is part of the digital strategy.By comparison,although there exist certain legal norms that have made relevant stipula⁃tions on manipulation issues,China continues to issue specific departmental regulations to regulate algorithmic recom⁃mender services,and pays more attention to addressing collective harm caused by algorithmic online manipulation through a multiple co-governance approach led by the government or industry associations to implement supervision.展开更多
The study conducts a bibliometric review of artificial intelligence applications in two areas:the entrepreneurial finance literature,and the corporate finance literature with implications for entrepreneurship.A rigoro...The study conducts a bibliometric review of artificial intelligence applications in two areas:the entrepreneurial finance literature,and the corporate finance literature with implications for entrepreneurship.A rigorous search and screening of the web of science core collection identified 1,890 journal articles for analysis.The bibliometrics provide a detailed view of the knowledge field,indicating underdeveloped research directions.An important contribution comes from insights through artificial intelligence methods in entrepreneurship.The results demonstrate a high representation of artificial neural networks,deep neural networks,and support vector machines across almost all identified topic niches.In contrast,applications of topic modeling,fuzzy neural networks,and growing hierarchical self-organizing maps are rare.Additionally,we take a broader view by addressing the problem of applying artificial intelligence in economic science.Specifically,we present the foundational paradigm and a bespoke demonstration of the Monte Carlo randomized algorithm.展开更多
Aiming to solve the steering instability and hysteresis of agricultural robots in the process of movement,a fusion PID control method of particle swarm optimization(PSO)and genetic algorithm(GA)was proposed.The fusion...Aiming to solve the steering instability and hysteresis of agricultural robots in the process of movement,a fusion PID control method of particle swarm optimization(PSO)and genetic algorithm(GA)was proposed.The fusion algorithm took advantage of the fast optimization ability of PSO to optimize the population screening link of GA.The Simulink simulation results showed that the convergence of the fitness function of the fusion algorithm was accelerated,the system response adjustment time was reduced,and the overshoot was almost zero.Then the algorithm was applied to the steering test of agricultural robot in various scenes.After modeling the steering system of agricultural robot,the steering test results in the unloaded suspended state showed that the PID control based on fusion algorithm reduced the rise time,response adjustment time and overshoot of the system,and improved the response speed and stability of the system,compared with the artificial trial and error PID control and the PID control based on GA.The actual road steering test results showed that the PID control response rise time based on the fusion algorithm was the shortest,about 4.43 s.When the target pulse number was set to 100,the actual mean value in the steady-state regulation stage was about 102.9,which was the closest to the target value among the three control methods,and the overshoot was reduced at the same time.The steering test results under various scene states showed that the PID control based on the proposed fusion algorithm had good anti-interference ability,it can adapt to the changes of environment and load and improve the performance of the control system.It was effective in the steering control of agricultural robot.This method can provide a reference for the precise steering control of other robots.展开更多
Efficient elastic wave focusing is crucial in materials and physical engineering.Elastic coding metasurfaces,which are innovative planar artificial structures,show great potential for use in the field of wave focusing...Efficient elastic wave focusing is crucial in materials and physical engineering.Elastic coding metasurfaces,which are innovative planar artificial structures,show great potential for use in the field of wave focusing.However,elastic coding lenses(ECLs)still suffer from low focusing performance,thickness comparable to wavelength,and frequency sensitivity.Here,we consider both the structural and material properties of the coding unit,thus realizing further compression of the thickness of the ECL.We chose the simplest ECL,which consists of only two encoding units.The coding unit 0 is a straight structure constructed using a carbon fiber reinforced composite material,and the coding unit 1 is a zigzag structure constructed using an aluminum material,and the thickness of the ECL constructed using them is only 1/8 of the wavelength.Based on the theoretical design,the arrangement of coding units is further optimized using genetic algorithms,which significantly improves the focusing performance of the lens at different focus and frequencies.This study provides a more effective way to control vibration and noise in advanced structures.展开更多
We explored the effects of algorithmic opacity on employees’playing dumb and evasive hiding rather than rationalized hiding.We examined the mediating role of job insecurity and the moderating role of employee-AI coll...We explored the effects of algorithmic opacity on employees’playing dumb and evasive hiding rather than rationalized hiding.We examined the mediating role of job insecurity and the moderating role of employee-AI collaboration.Participants were 421 full-time employees(female=46.32%,junior employees=31.83%)from a variety of organizations and industries that interact with AI.Employees filled out data on algorithm opacity,job insecurity,knowledge hiding,employee-AI collaboration,and control variables.The results of the structural equation modeling indicated that algorithm opacity exacerbated employees’job insecurity,and job insecurity mediated between algorithm opacity and playing dumb and evasive hiding rather than rationalized hiding.The relationship between algorithmic opacity and playing dumb and evasive hiding was more positive when the level of employee-AI collaboration was higher.These findings suggest that employee-AI collaboration reinforces the indirect relationship between algorithmic opacity and playing dumb and evasive hiding.Our study contributes to research on human and AI collaboration by exploring the dark side of employee-AI collaboration.展开更多
This paper investigates the optimization of data sampling and target labeling techniques to enhance algorithmic trading strategies in cryptocurrency markets,focusing on Bitcoin(BTC)and Ethereum(ETH).Traditional data s...This paper investigates the optimization of data sampling and target labeling techniques to enhance algorithmic trading strategies in cryptocurrency markets,focusing on Bitcoin(BTC)and Ethereum(ETH).Traditional data sampling methods,such as time bars,often fail to capture the nuances of the continuously active and highly volatile cryptocurrency market and force traders to wait for arbitrary points in time.To address this,we propose an alternative approach using information-driven sampling methods,including the CUSUM filter,range bars,volume bars,and dollar bars,and evaluate their performance using tick-level data from January 2018 to June 2023.Additionally,we introduce the Triple Barrier method for target labeling,which offers a solution tailored for algorithmic trading as opposed to the widely used next-bar prediction.We empirically assess the effectiveness of these data sampling and labeling methods to craft profitable trading strategies.The results demonstrate that the innovative combination of CUSUM-filtered data with Triple Barrier labeling outperforms traditional time bars and next-bar prediction,achieving consistently positive trading performance even after accounting for transaction costs.Moreover,our system enables making trading decisions at any point in time on the basis of market conditions,providing an advantage over traditional methods that rely on fixed time intervals.Furthermore,the paper contributes to the ongoing debate on the applicability of Transformer models to time series classification in the context of algorithmic trading by evaluating various Transformer architectures—including the vanilla Transformer encoder,FEDformer,and Autoformer—alongside other deep learning architectures and classical machine learning models,revealing insights into their relative performance.展开更多
With the rapid advancement of medical artificial intelligence(AI)technology,particularly the widespread adoption of AI diagnostic systems,ethical challenges in medical decision-making have garnered increasing attentio...With the rapid advancement of medical artificial intelligence(AI)technology,particularly the widespread adoption of AI diagnostic systems,ethical challenges in medical decision-making have garnered increasing attention.This paper analyzes the limitations of algorithmic ethics in medical decision-making and explores accountability mechanisms,aiming to provide theoretical support for ethically informed medical practices.The study highlights how the opacity of AI algorithms complicates the definition of decision-making responsibility,undermines doctor-patient trust,and affects informed consent.By thoroughly investigating issues such as the algorithmic“black box”problem and data privacy protection,we develop accountability assessment models to address ethical concerns related to medical resource allocation.Furthermore,this research examines the effective implementation of AI diagnostic systems through case studies of both successful and unsuccessful applications,extracting lessons on accountability mechanisms and response strategies.Finally,we emphasize that establishing a transparent accountability framework is crucial for enhancing the ethical standards of medical AI systems and protecting patients’rights and interests.展开更多
This paper presents a novel method for reconstructing a highly accurate 3D nose model of the human from 2D images and pre-marked landmarks based on algorithmic methods.The study focuses on the reconstruction of a 3D n...This paper presents a novel method for reconstructing a highly accurate 3D nose model of the human from 2D images and pre-marked landmarks based on algorithmic methods.The study focuses on the reconstruction of a 3D nose model tailored for applications in healthcare and cosmetic surgery.The approach leverages advanced image processing techniques,3D Morphable Models(3DMM),and deformation techniques to overcome the limita-tions of deep learning models,particularly addressing the interpretability issues commonly encountered in medical applications.The proposed method estimates the 3D coordinates of landmark points using a 3D structure estimation algorithm.Sub-landmarks are extracted through image processing techniques and interpolation.The initial surface is generated using a 3DMM,though its accuracy remains limited.To enhance precision,deformation techniques are applied,utilizing the coordinates of 76 identified landmarks and sub-landmarks.The resulting 3D nose model is constructed based on algorithmic methods and pre-marked landmarks.Evaluation of the 3D model is conducted by comparing landmark distances and shape similarity with expert-determined ground truth on 30 Vietnamese volunteers aged 18 to 47,all of whom were either preparing for or required nasal surgery.Experimental results demonstrate a strong agreement between the reconstructed 3D model and the ground truth.The method achieved a mean landmark distance error of 0.631 mm and a shape error of 1.738 mm,demonstrating its potential for medical applications.展开更多
Domain Generation Algorithms(DGAs)continue to pose a significant threat inmodernmalware infrastructures by enabling resilient and evasive communication with Command and Control(C&C)servers.Traditional detection me...Domain Generation Algorithms(DGAs)continue to pose a significant threat inmodernmalware infrastructures by enabling resilient and evasive communication with Command and Control(C&C)servers.Traditional detection methods-rooted in statistical heuristics,feature engineering,and shallow machine learning-struggle to adapt to the increasing sophistication,linguistic mimicry,and adversarial variability of DGA variants.The emergence of Large Language Models(LLMs)marks a transformative shift in this landscape.Leveraging deep contextual understanding,semantic generalization,and few-shot learning capabilities,LLMs such as BERT,GPT,and T5 have shown promising results in detecting both character-based and dictionary-based DGAs,including previously unseen(zeroday)variants.This paper provides a comprehensive and critical review of LLM-driven DGA detection,introducing a structured taxonomy of LLM architectures,evaluating the linguistic and behavioral properties of benchmark datasets,and comparing recent detection frameworks across accuracy,latency,robustness,and multilingual performance.We also highlight key limitations,including challenges in adversarial resilience,model interpretability,deployment scalability,and privacy risks.To address these gaps,we present a forward-looking research roadmap encompassing adversarial training,model compression,cross-lingual benchmarking,and real-time integration with SIEM/SOAR platforms.This survey aims to serve as a foundational resource for advancing the development of scalable,explainable,and operationally viable LLM-based DGA detection systems.展开更多
Optimization is the key to obtaining efficient utilization of resources in structural design.Due to the complex nature of truss systems,this study presents a method based on metaheuristic modelling that minimises stru...Optimization is the key to obtaining efficient utilization of resources in structural design.Due to the complex nature of truss systems,this study presents a method based on metaheuristic modelling that minimises structural weight under stress and frequency constraints.Two new algorithms,the Red Kite Optimization Algorithm(ROA)and Secretary Bird Optimization Algorithm(SBOA),are utilized on five benchmark trusses with 10,18,37,72,and 200-bar trusses.Both algorithms are evaluated against benchmarks in the literature.The results indicate that SBOA always reaches a lighter optimal.Designs with reducing structural weight ranging from 0.02%to 0.15%compared to ROA,and up to 6%–8%as compared to conventional algorithms.In addition,SBOA can achieve 15%–20%faster convergence speed and 10%–18%reduction in computational time with a smaller standard deviation over independent runs,which demonstrates its robustness and reliability.It is indicated that the adaptive exploration mechanism of SBOA,especially its Levy flight–based search strategy,can obviously improve optimization performance for low-and high-dimensional trusses.The research has implications in the context of promoting bio-inspired optimization techniques by demonstrating the viability of SBOA,a reliable model for large-scale structural design that provides significant enhancements in performance and convergence behavior.展开更多
This study investigates how artificial intelligence(AI)algorithms enable mainstream media to achieve precise emotional matching and improve communication efficiency through reconstructed communication logic.As digital...This study investigates how artificial intelligence(AI)algorithms enable mainstream media to achieve precise emotional matching and improve communication efficiency through reconstructed communication logic.As digital intelligence technology rapidly evolves,mainstream media organizations are increasingly leveraging AI-driven empathy algorithms to enhance audience engagement and optimize content delivery.This research employs a mixed-methods approach,combining quantitative analysis of algorithmic performance metrics with qualitative examination of media communication patterns.Through systematic review of 150 academic papers and analysis of data from 12 major media platforms,this study reveals that algorithmic empathy systems can improve emotional resonance by 34.7%and increase audience engagement by 28.3%compared to traditional communication methods.The findings demonstrate that AI algorithms reconstruct media communication logic through three primary pathways:emotional pattern recognition,personalized content curation,and real-time sentiment adaptation.However,the study also identifies significant challenges including algorithmic bias,emotional authenticity concerns,and ethical implications of automated empathy.The research contributes to understanding how mainstream media can leverage AI technology to build high-quality empathetic communication while maintaining journalistic integrity and social responsibility.展开更多
This paper introduces a novel nature-inspired metaheuristic algorithm called the Gekko japonicus algorithm.The algo-rithm draws inspiration mainly from the predation strategies and survival behaviors of the Gekko japo...This paper introduces a novel nature-inspired metaheuristic algorithm called the Gekko japonicus algorithm.The algo-rithm draws inspiration mainly from the predation strategies and survival behaviors of the Gekko japonicus.The math-ematical model is developed by simulating various biological behaviors of the Gekko japonicus,such as hybrid loco-motion patterns,directional olfactory guidance,implicit group advantage tendencies,and the tail autotomy mechanism.By integrating multi-stage mutual constraints and dynamically adjusting parameters,GJA maintains an optimal balance between global exploration and local exploitation,thereby effectively solving complex optimization problems.To assess the performance of GJA,comparative analyses were performed against fourteen state-of-the-art metaheuristic algorithms using the CEC2017 and CEC2022 benchmark test sets.Additionally,a Friedman test was performed on the experimen-tal results to assess the statistical significance of differences between various algorithms.And GJA was evaluated using multiple qualitative indicators,further confirming its superiority in exploration and exploitation.Finally,GJA was utilized to solve four engineering optimization problems and further implemented in robotic path planning to verify its practical applicability.Experimental results indicate that,compared to other high-performance algorithms,GJA demonstrates excep-tional performance as a powerful optimization algorithm in complex optimization problems.We make the code publicly available at:https://github.com/zhy1109/Gekko-japonicusalgorithm.展开更多
The Intrusion Detection System(IDS)is a security mechanism developed to observe network traffic and recognize suspicious or malicious activities.Clustering algorithms are often incorporated into IDS;however,convention...The Intrusion Detection System(IDS)is a security mechanism developed to observe network traffic and recognize suspicious or malicious activities.Clustering algorithms are often incorporated into IDS;however,conventional clustering-based methods face notable drawbacks,including poor scalability in handling high-dimensional datasets and a strong dependence of outcomes on initial conditions.To overcome the performance limitations of existing methods,this study proposes a novel quantum-inspired clustering algorithm that relies on a similarity coefficient-based quantum genetic algorithm(SC-QGA)and an improved quantum artificial bee colony algorithm hybrid K-means(IQABC-K).First,the SC-QGA algorithmis constructed based on quantum computing and integrates similarity coefficient theory to strengthen genetic diversity and feature extraction capabilities.For the subsequent clustering phase,the process based on the IQABC-K algorithm is enhanced with the core improvement of adaptive rotation gate and movement exploitation strategies to balance the exploration capabilities of global search and the exploitation capabilities of local search.Simultaneously,the acceleration of convergence toward the global optimum and a reduction in computational complexity are facilitated by means of the global optimum bootstrap strategy and a linear population reduction strategy.Through experimental evaluation with multiple algorithms and diverse performance metrics,the proposed algorithm confirms reliable accuracy on three datasets:KDD CUP99,NSL_KDD,and UNSW_NB15,achieving accuracy of 98.57%,98.81%,and 98.32%,respectively.These results affirm its potential as an effective solution for practical clustering applications.展开更多
In this paper,we propose a new privacy-aware transmission scheduling algorithm for 6G ad hoc networks.This system enables end nodes to select the optimum time and scheme to transmit private data safely.In 6G dynamic h...In this paper,we propose a new privacy-aware transmission scheduling algorithm for 6G ad hoc networks.This system enables end nodes to select the optimum time and scheme to transmit private data safely.In 6G dynamic heterogeneous infrastructures,unstable links and non-uniform hardware capabilities create critical issues regarding security and privacy.Traditional protocols are often too computationally heavy to allow 6G services to achieve their expected Quality-of-Service(QoS).As the transport network is built of ad hoc nodes,there is no guarantee about their trustworthiness or behavior,and transversal functionalities are delegated to the extreme nodes.However,while security can be guaranteed in extreme-to-extreme solutions,privacy cannot,as all intermediate nodes still have to handle the data packets they are transporting.Besides,traditional schemes for private anonymous ad hoc communications are vulnerable against modern intelligent attacks based on learning models.The proposed scheme fulfills this gap.Findings show the probability of a successful intelligent attack reduces by up to 65%compared to ad hoc networks with no privacy protection strategy when used the proposed technology.While congestion probability can remain below 0.001%,as required in 6G services.展开更多
ThePigeon-InspiredOptimization(PIO)algorithmconstitutes ametaheuristic method derived fromthe homing behaviour of pigeons.Initially formulated for three-dimensional path planning in unmanned aerial vehicles(UAVs),the ...ThePigeon-InspiredOptimization(PIO)algorithmconstitutes ametaheuristic method derived fromthe homing behaviour of pigeons.Initially formulated for three-dimensional path planning in unmanned aerial vehicles(UAVs),the algorithmhas attracted considerable academic and industrial interest owing to its effective balance between exploration and exploitation,coupled with advantages in real-time performance and robustness.Nevertheless,as applications have diversified,limitations in convergence precision and a tendency toward premature convergence have become increasingly evident,highlighting a need for improvement.This reviewsystematically outlines the developmental trajectory of the PIO algorithm,with a particular focus on its core applications in UAV navigation,multi-objective formulations,and a spectrum of variantmodels that have emerged in recent years.It offers a structured analysis of the foundational principles underlying the PIO.It conducts a comparative assessment of various performance-enhanced versions,including hybrid models that integrate mechanisms from other optimization paradigms.Additionally,the strengths andweaknesses of distinct PIOvariants are critically examined frommultiple perspectives,including intrinsic algorithmic characteristics,suitability for specific application scenarios,objective function design,and the rigor of the statistical evaluation methodologies employed in empirical studies.Finally,this paper identifies principal challenges within current PIO research and proposes several prospective research directions.Future work should focus on mitigating premature convergence by refining the two-phase search structure and adjusting the exponential decrease of individual numbers during the landmark operator.Enhancing parameter adaptation strategies,potentially using reinforcement learning for dynamic tuning,and advancing theoretical analyses on convergence and complexity are also critical.Further applications should be explored in constrained path planning,Neural Architecture Search(NAS),and other real-worldmulti-objective problems.For Multi-objective PIO(MPIO),key improvements include controlling the growth of the external archive and designing more effective selection mechanisms to maintain convergence efficiency.These efforts are expected to strengthen both the theoretical foundation and practical versatility of PIO and its variants.展开更多
With the rapid development of transportation infrastructure,ensuring road safety through timely and accurate highway inspection has become increasingly critical.Traditional manual inspection methods are not only time-...With the rapid development of transportation infrastructure,ensuring road safety through timely and accurate highway inspection has become increasingly critical.Traditional manual inspection methods are not only time-consuming and labor-intensive,but they also struggle to provide consistent,high-precision detection and realtime monitoring of pavement surface defects.To overcome these limitations,we propose an Automatic Recognition of PavementDefect(ARPD)algorithm,which leverages unmanned aerial vehicle(UAV)-based aerial imagery to automate the inspection process.The ARPD framework incorporates a backbone network based on the Selective State Space Model(S3M),which is designed to capture long-range temporal dependencies.This enables effective modeling of dynamic correlations among redundant and often repetitive structures commonly found in road imagery.Furthermore,a neck structure based on Semantics and Detail Infusion(SDI)is introduced to guide cross-scale feature fusion.The SDI module enhances the integration of low-level spatial details with high-level semantic cues,thereby improving feature expressiveness and defect localization accuracy.Experimental evaluations demonstrate that theARPDalgorithm achieves a mean average precision(mAP)of 86.1%on a custom-labeled pavement defect dataset,outperforming the state-of-the-art YOLOv11 segmentation model.The algorithm also maintains strong generalization ability on public datasets.These results confirm that ARPD is well-suited for diverse real-world applications in intelligent,large-scale highway defect monitoring and maintenance planning.展开更多
Accurate prediction of flood events is important for flood control and risk management.Machine learning techniques contributed greatly to advances in flood predictions,and existing studies mainly focused on predicting...Accurate prediction of flood events is important for flood control and risk management.Machine learning techniques contributed greatly to advances in flood predictions,and existing studies mainly focused on predicting flood resource variables using single or hybrid machine learning techniques.However,class-based flood predictions have rarely been investigated,which can aid in quickly diagnosing comprehensive flood characteristics and proposing targeted management strategies.This study proposed a prediction approach of flood regime metrics and event classes coupling machine learning algorithms with clustering-deduced membership degrees.Five algorithms were adopted for this exploration.Results showed that the class membership degrees accurately determined event classes with class hit rates up to 100%,compared with the four classes clustered from nine regime metrics.The nonlinear algorithms(Multiple Linear Regression,Random Forest,and least squares-Support Vector Machine)outperformed the linear techniques(Multiple Linear Regression and Stepwise Regression)in predicting flood regime metrics.The proposed approach well predicted flood event classes with average class hit rates of 66.0%-85.4%and 47.2%-76.0%in calibration and validation periods,respectively,particularly for the slow and late flood events.The predictive capability of the proposed prediction approach for flood regime metrics and classes was considerably stronger than that of hydrological modeling approach.展开更多
In reliability analyses,the absence of a priori information on the most probable point of failure(MPP)may result in overlooking critical points,thereby leading to biased assessment outcomes.Moreover,second-order relia...In reliability analyses,the absence of a priori information on the most probable point of failure(MPP)may result in overlooking critical points,thereby leading to biased assessment outcomes.Moreover,second-order reliability methods exhibit limited accuracy in highly nonlinear scenarios.To overcome these challenges,a novel reliability analysis strategy based on a multimodal differential evolution algorithm and a hypersphere integration method is proposed.Initially,the penalty function method is employed to reformulate the MPP search problem as a conditionally constrained optimization task.Subsequently,a differential evolution algorithm incorporating a population delineation strategy is utilized to identify all MPPs.Finally,a paraboloid equation is constructed based on the curvature of the limit-state function at the MPPs,and the failure probability of the structure is calculated by using the hypersphere integration method.The localization effectiveness of the MPPs is compared through multiple numerical cases and two engineering examples,with accuracy comparisons of failure probabilities against the first-order reliability method(FORM)and the secondorder reliability method(SORM).The results indicate that the method effectively identifies existing MPPs and achieves higher solution precision.展开更多
基金funded by United Arab Emirates University(UAEU)under the UAEU-AUA grant number G00004577(12N145)with the corresponding grant at Universiti Malaya(UM)under grant number IF019-2024.
文摘Concrete-filled steel tubes(CFST)are widely utilized in civil engineering due to their superior load-bearing capacity,ductility,and seismic resistance.However,existing design codes,such as AISC and Eurocode 4,tend to be excessively conservative as they fail to account for the composite action between the steel tube and the concrete core.To address this limitation,this study proposes a hybrid model that integrates XGBoost with the Pied Kingfisher Optimizer(PKO),a nature-inspired algorithm,to enhance the accuracy of shear strength prediction for CFST columns.Additionally,quantile regression is employed to construct prediction intervals for the ultimate shear force,while the Asymmetric Squared Error Loss(ASEL)function is incorporated to mitigate overestimation errors.The computational results demonstrate that the PKO-XGBoost model delivers superior predictive accuracy,achieving a Mean Absolute Percentage Error(MAPE)of 4.431%and R2 of 0.9925 on the test set.Furthermore,the ASEL-PKO-XGBoost model substantially reduces overestimation errors to 28.26%,with negligible impact on predictive performance.Additionally,based on the Genetic Algorithm(GA)and existing equation models,a strength equation model is developed,achieving markedly higher accuracy than existing models(R^(2)=0.934).Lastly,web-based Graphical User Interfaces(GUIs)were developed to enable real-time prediction.
基金supported by the P.G.Senapathy Center for Computing Resources at IIT Madrasfunding provided by the Ministry of Education,Government of Indiasupported by the National Natural Science Foundation of China(Grant Nos.12388101,12472224 and 92252104).
文摘This paper presents an Eulerian-Lagrangian algorithm for direct numerical simulation(DNS)of particle-laden flows.The algorithm is applicable to perform simulations of dilute suspensions of small inertial particles in turbulent carrier flow.The Eulerian framework numerically resolves turbulent carrier flow using a parallelized,finite-volume DNS solver on a staggered Cartesian grid.Particles are tracked using a point-particle method utilizing a Lagrangian particle tracking(LPT)algorithm.The proposed Eulerian-Lagrangian algorithm is validated using an inertial particle-laden turbulent channel flow for different Stokes number cases.The particle concentration profiles and higher-order statistics of the carrier and dispersed phases agree well with the benchmark results.We investigated the effect of fluid velocity interpolation and numerical integration schemes of particle tracking algorithms on particle dispersion statistics.The suitability of fluid velocity interpolation schemes for predicting the particle dispersion statistics is discussed in the framework of the particle tracking algorithm coupled to the finite-volume solver.In addition,we present parallelization strategies implemented in the algorithm and evaluate their parallel performance.
文摘The original intention of the algorithmic recommender system is to grapple with the negative impacts caused by information overload,but the system also can be used as"hypernudge",a new form of online manipulation,to inten⁃tionally exploit people's cognitive and decision-making gaps to influence their decisions in practice,which is particu⁃larly detrimental to the sustainable development of the digital market.Limiting harmful algorithmic online manipula⁃tion in digital markets has become a challenging task.Globally,both the EU and China have responded to this issue,and the differences between them are so evident that their governance measures can serve as the typical case.The EU focuses on improving citizens'digital literacy and their ability to integrate into digital social life to independently ad⁃dress this issue,and expects to address harmful manipulation behavior through binding and applicable hard law,which is part of the digital strategy.By comparison,although there exist certain legal norms that have made relevant stipula⁃tions on manipulation issues,China continues to issue specific departmental regulations to regulate algorithmic recom⁃mender services,and pays more attention to addressing collective harm caused by algorithmic online manipulation through a multiple co-governance approach led by the government or industry associations to implement supervision.
文摘The study conducts a bibliometric review of artificial intelligence applications in two areas:the entrepreneurial finance literature,and the corporate finance literature with implications for entrepreneurship.A rigorous search and screening of the web of science core collection identified 1,890 journal articles for analysis.The bibliometrics provide a detailed view of the knowledge field,indicating underdeveloped research directions.An important contribution comes from insights through artificial intelligence methods in entrepreneurship.The results demonstrate a high representation of artificial neural networks,deep neural networks,and support vector machines across almost all identified topic niches.In contrast,applications of topic modeling,fuzzy neural networks,and growing hierarchical self-organizing maps are rare.Additionally,we take a broader view by addressing the problem of applying artificial intelligence in economic science.Specifically,we present the foundational paradigm and a bespoke demonstration of the Monte Carlo randomized algorithm.
文摘Aiming to solve the steering instability and hysteresis of agricultural robots in the process of movement,a fusion PID control method of particle swarm optimization(PSO)and genetic algorithm(GA)was proposed.The fusion algorithm took advantage of the fast optimization ability of PSO to optimize the population screening link of GA.The Simulink simulation results showed that the convergence of the fitness function of the fusion algorithm was accelerated,the system response adjustment time was reduced,and the overshoot was almost zero.Then the algorithm was applied to the steering test of agricultural robot in various scenes.After modeling the steering system of agricultural robot,the steering test results in the unloaded suspended state showed that the PID control based on fusion algorithm reduced the rise time,response adjustment time and overshoot of the system,and improved the response speed and stability of the system,compared with the artificial trial and error PID control and the PID control based on GA.The actual road steering test results showed that the PID control response rise time based on the fusion algorithm was the shortest,about 4.43 s.When the target pulse number was set to 100,the actual mean value in the steady-state regulation stage was about 102.9,which was the closest to the target value among the three control methods,and the overshoot was reduced at the same time.The steering test results under various scene states showed that the PID control based on the proposed fusion algorithm had good anti-interference ability,it can adapt to the changes of environment and load and improve the performance of the control system.It was effective in the steering control of agricultural robot.This method can provide a reference for the precise steering control of other robots.
基金Project supported by the National Natural Science Foundation of China(Grant No.12404531)the Natural Science Foundation of the Higher Education Institutions of Jiangsu Province,China(Grant No.23KJB140011)。
文摘Efficient elastic wave focusing is crucial in materials and physical engineering.Elastic coding metasurfaces,which are innovative planar artificial structures,show great potential for use in the field of wave focusing.However,elastic coding lenses(ECLs)still suffer from low focusing performance,thickness comparable to wavelength,and frequency sensitivity.Here,we consider both the structural and material properties of the coding unit,thus realizing further compression of the thickness of the ECL.We chose the simplest ECL,which consists of only two encoding units.The coding unit 0 is a straight structure constructed using a carbon fiber reinforced composite material,and the coding unit 1 is a zigzag structure constructed using an aluminum material,and the thickness of the ECL constructed using them is only 1/8 of the wavelength.Based on the theoretical design,the arrangement of coding units is further optimized using genetic algorithms,which significantly improves the focusing performance of the lens at different focus and frequencies.This study provides a more effective way to control vibration and noise in advanced structures.
基金supported by the Social Science Foundation of Liaoning Province(L23BJY022).
文摘We explored the effects of algorithmic opacity on employees’playing dumb and evasive hiding rather than rationalized hiding.We examined the mediating role of job insecurity and the moderating role of employee-AI collaboration.Participants were 421 full-time employees(female=46.32%,junior employees=31.83%)from a variety of organizations and industries that interact with AI.Employees filled out data on algorithm opacity,job insecurity,knowledge hiding,employee-AI collaboration,and control variables.The results of the structural equation modeling indicated that algorithm opacity exacerbated employees’job insecurity,and job insecurity mediated between algorithm opacity and playing dumb and evasive hiding rather than rationalized hiding.The relationship between algorithmic opacity and playing dumb and evasive hiding was more positive when the level of employee-AI collaboration was higher.These findings suggest that employee-AI collaboration reinforces the indirect relationship between algorithmic opacity and playing dumb and evasive hiding.Our study contributes to research on human and AI collaboration by exploring the dark side of employee-AI collaboration.
基金support of the University of Warsaw under’New Ideas 3B’competition in POB Ⅲ implemented under the’Excellence Initiative-Research University’Programme.
文摘This paper investigates the optimization of data sampling and target labeling techniques to enhance algorithmic trading strategies in cryptocurrency markets,focusing on Bitcoin(BTC)and Ethereum(ETH).Traditional data sampling methods,such as time bars,often fail to capture the nuances of the continuously active and highly volatile cryptocurrency market and force traders to wait for arbitrary points in time.To address this,we propose an alternative approach using information-driven sampling methods,including the CUSUM filter,range bars,volume bars,and dollar bars,and evaluate their performance using tick-level data from January 2018 to June 2023.Additionally,we introduce the Triple Barrier method for target labeling,which offers a solution tailored for algorithmic trading as opposed to the widely used next-bar prediction.We empirically assess the effectiveness of these data sampling and labeling methods to craft profitable trading strategies.The results demonstrate that the innovative combination of CUSUM-filtered data with Triple Barrier labeling outperforms traditional time bars and next-bar prediction,achieving consistently positive trading performance even after accounting for transaction costs.Moreover,our system enables making trading decisions at any point in time on the basis of market conditions,providing an advantage over traditional methods that rely on fixed time intervals.Furthermore,the paper contributes to the ongoing debate on the applicability of Transformer models to time series classification in the context of algorithmic trading by evaluating various Transformer architectures—including the vanilla Transformer encoder,FEDformer,and Autoformer—alongside other deep learning architectures and classical machine learning models,revealing insights into their relative performance.
文摘With the rapid advancement of medical artificial intelligence(AI)technology,particularly the widespread adoption of AI diagnostic systems,ethical challenges in medical decision-making have garnered increasing attention.This paper analyzes the limitations of algorithmic ethics in medical decision-making and explores accountability mechanisms,aiming to provide theoretical support for ethically informed medical practices.The study highlights how the opacity of AI algorithms complicates the definition of decision-making responsibility,undermines doctor-patient trust,and affects informed consent.By thoroughly investigating issues such as the algorithmic“black box”problem and data privacy protection,we develop accountability assessment models to address ethical concerns related to medical resource allocation.Furthermore,this research examines the effective implementation of AI diagnostic systems through case studies of both successful and unsuccessful applications,extracting lessons on accountability mechanisms and response strategies.Finally,we emphasize that establishing a transparent accountability framework is crucial for enhancing the ethical standards of medical AI systems and protecting patients’rights and interests.
文摘This paper presents a novel method for reconstructing a highly accurate 3D nose model of the human from 2D images and pre-marked landmarks based on algorithmic methods.The study focuses on the reconstruction of a 3D nose model tailored for applications in healthcare and cosmetic surgery.The approach leverages advanced image processing techniques,3D Morphable Models(3DMM),and deformation techniques to overcome the limita-tions of deep learning models,particularly addressing the interpretability issues commonly encountered in medical applications.The proposed method estimates the 3D coordinates of landmark points using a 3D structure estimation algorithm.Sub-landmarks are extracted through image processing techniques and interpolation.The initial surface is generated using a 3DMM,though its accuracy remains limited.To enhance precision,deformation techniques are applied,utilizing the coordinates of 76 identified landmarks and sub-landmarks.The resulting 3D nose model is constructed based on algorithmic methods and pre-marked landmarks.Evaluation of the 3D model is conducted by comparing landmark distances and shape similarity with expert-determined ground truth on 30 Vietnamese volunteers aged 18 to 47,all of whom were either preparing for or required nasal surgery.Experimental results demonstrate a strong agreement between the reconstructed 3D model and the ground truth.The method achieved a mean landmark distance error of 0.631 mm and a shape error of 1.738 mm,demonstrating its potential for medical applications.
基金the Deanship of Scientific Research at King Khalid University for funding this work through large group under grant number(GRP.2/663/46).
文摘Domain Generation Algorithms(DGAs)continue to pose a significant threat inmodernmalware infrastructures by enabling resilient and evasive communication with Command and Control(C&C)servers.Traditional detection methods-rooted in statistical heuristics,feature engineering,and shallow machine learning-struggle to adapt to the increasing sophistication,linguistic mimicry,and adversarial variability of DGA variants.The emergence of Large Language Models(LLMs)marks a transformative shift in this landscape.Leveraging deep contextual understanding,semantic generalization,and few-shot learning capabilities,LLMs such as BERT,GPT,and T5 have shown promising results in detecting both character-based and dictionary-based DGAs,including previously unseen(zeroday)variants.This paper provides a comprehensive and critical review of LLM-driven DGA detection,introducing a structured taxonomy of LLM architectures,evaluating the linguistic and behavioral properties of benchmark datasets,and comparing recent detection frameworks across accuracy,latency,robustness,and multilingual performance.We also highlight key limitations,including challenges in adversarial resilience,model interpretability,deployment scalability,and privacy risks.To address these gaps,we present a forward-looking research roadmap encompassing adversarial training,model compression,cross-lingual benchmarking,and real-time integration with SIEM/SOAR platforms.This survey aims to serve as a foundational resource for advancing the development of scalable,explainable,and operationally viable LLM-based DGA detection systems.
文摘Optimization is the key to obtaining efficient utilization of resources in structural design.Due to the complex nature of truss systems,this study presents a method based on metaheuristic modelling that minimises structural weight under stress and frequency constraints.Two new algorithms,the Red Kite Optimization Algorithm(ROA)and Secretary Bird Optimization Algorithm(SBOA),are utilized on five benchmark trusses with 10,18,37,72,and 200-bar trusses.Both algorithms are evaluated against benchmarks in the literature.The results indicate that SBOA always reaches a lighter optimal.Designs with reducing structural weight ranging from 0.02%to 0.15%compared to ROA,and up to 6%–8%as compared to conventional algorithms.In addition,SBOA can achieve 15%–20%faster convergence speed and 10%–18%reduction in computational time with a smaller standard deviation over independent runs,which demonstrates its robustness and reliability.It is indicated that the adaptive exploration mechanism of SBOA,especially its Levy flight–based search strategy,can obviously improve optimization performance for low-and high-dimensional trusses.The research has implications in the context of promoting bio-inspired optimization techniques by demonstrating the viability of SBOA,a reliable model for large-scale structural design that provides significant enhancements in performance and convergence behavior.
文摘This study investigates how artificial intelligence(AI)algorithms enable mainstream media to achieve precise emotional matching and improve communication efficiency through reconstructed communication logic.As digital intelligence technology rapidly evolves,mainstream media organizations are increasingly leveraging AI-driven empathy algorithms to enhance audience engagement and optimize content delivery.This research employs a mixed-methods approach,combining quantitative analysis of algorithmic performance metrics with qualitative examination of media communication patterns.Through systematic review of 150 academic papers and analysis of data from 12 major media platforms,this study reveals that algorithmic empathy systems can improve emotional resonance by 34.7%and increase audience engagement by 28.3%compared to traditional communication methods.The findings demonstrate that AI algorithms reconstruct media communication logic through three primary pathways:emotional pattern recognition,personalized content curation,and real-time sentiment adaptation.However,the study also identifies significant challenges including algorithmic bias,emotional authenticity concerns,and ethical implications of automated empathy.The research contributes to understanding how mainstream media can leverage AI technology to build high-quality empathetic communication while maintaining journalistic integrity and social responsibility.
基金CHINA POSTDOCTORAL SCIENCE FOUNDATION(Grant No.2025M771925)Young Scientists Fund(C Class)(Grant No.32501636)Special Fund of Fundamental Scientific Research Business Expense for Higher School of Central Government(Grant No.2572025JT04).
文摘This paper introduces a novel nature-inspired metaheuristic algorithm called the Gekko japonicus algorithm.The algo-rithm draws inspiration mainly from the predation strategies and survival behaviors of the Gekko japonicus.The math-ematical model is developed by simulating various biological behaviors of the Gekko japonicus,such as hybrid loco-motion patterns,directional olfactory guidance,implicit group advantage tendencies,and the tail autotomy mechanism.By integrating multi-stage mutual constraints and dynamically adjusting parameters,GJA maintains an optimal balance between global exploration and local exploitation,thereby effectively solving complex optimization problems.To assess the performance of GJA,comparative analyses were performed against fourteen state-of-the-art metaheuristic algorithms using the CEC2017 and CEC2022 benchmark test sets.Additionally,a Friedman test was performed on the experimen-tal results to assess the statistical significance of differences between various algorithms.And GJA was evaluated using multiple qualitative indicators,further confirming its superiority in exploration and exploitation.Finally,GJA was utilized to solve four engineering optimization problems and further implemented in robotic path planning to verify its practical applicability.Experimental results indicate that,compared to other high-performance algorithms,GJA demonstrates excep-tional performance as a powerful optimization algorithm in complex optimization problems.We make the code publicly available at:https://github.com/zhy1109/Gekko-japonicusalgorithm.
基金supported by the NSFC(Grant Nos.62176273,62271070,62441212)The Open Foundation of State Key Laboratory of Networking and Switching Technology(Beijing University of Posts and Telecommunications)under Grant SKLNST-2024-1-062025Major Project of the Natural Science Foundation of Inner Mongolia(2025ZD008).
文摘The Intrusion Detection System(IDS)is a security mechanism developed to observe network traffic and recognize suspicious or malicious activities.Clustering algorithms are often incorporated into IDS;however,conventional clustering-based methods face notable drawbacks,including poor scalability in handling high-dimensional datasets and a strong dependence of outcomes on initial conditions.To overcome the performance limitations of existing methods,this study proposes a novel quantum-inspired clustering algorithm that relies on a similarity coefficient-based quantum genetic algorithm(SC-QGA)and an improved quantum artificial bee colony algorithm hybrid K-means(IQABC-K).First,the SC-QGA algorithmis constructed based on quantum computing and integrates similarity coefficient theory to strengthen genetic diversity and feature extraction capabilities.For the subsequent clustering phase,the process based on the IQABC-K algorithm is enhanced with the core improvement of adaptive rotation gate and movement exploitation strategies to balance the exploration capabilities of global search and the exploitation capabilities of local search.Simultaneously,the acceleration of convergence toward the global optimum and a reduction in computational complexity are facilitated by means of the global optimum bootstrap strategy and a linear population reduction strategy.Through experimental evaluation with multiple algorithms and diverse performance metrics,the proposed algorithm confirms reliable accuracy on three datasets:KDD CUP99,NSL_KDD,and UNSW_NB15,achieving accuracy of 98.57%,98.81%,and 98.32%,respectively.These results affirm its potential as an effective solution for practical clustering applications.
基金funding from the European Commission by the Ruralities project(grant agreement no.101060876).
文摘In this paper,we propose a new privacy-aware transmission scheduling algorithm for 6G ad hoc networks.This system enables end nodes to select the optimum time and scheme to transmit private data safely.In 6G dynamic heterogeneous infrastructures,unstable links and non-uniform hardware capabilities create critical issues regarding security and privacy.Traditional protocols are often too computationally heavy to allow 6G services to achieve their expected Quality-of-Service(QoS).As the transport network is built of ad hoc nodes,there is no guarantee about their trustworthiness or behavior,and transversal functionalities are delegated to the extreme nodes.However,while security can be guaranteed in extreme-to-extreme solutions,privacy cannot,as all intermediate nodes still have to handle the data packets they are transporting.Besides,traditional schemes for private anonymous ad hoc communications are vulnerable against modern intelligent attacks based on learning models.The proposed scheme fulfills this gap.Findings show the probability of a successful intelligent attack reduces by up to 65%compared to ad hoc networks with no privacy protection strategy when used the proposed technology.While congestion probability can remain below 0.001%,as required in 6G services.
基金supported by the National Natural Science Foundation of China under grant number 62066016the Natural Science Foundation of Hunan Province of China under grant number 2024JJ7395+2 种基金International and Regional Science and Technology Cooperation and Exchange Program of the Hunan Association for Science and Technology under grant number 025SKX-KJ-04Hunan Provincial Postgraduate Research Innovation Project under grant numberCX20251611Liye Qin Bamboo Slips Research Special Project of JishouUniversity 25LYY03.
文摘ThePigeon-InspiredOptimization(PIO)algorithmconstitutes ametaheuristic method derived fromthe homing behaviour of pigeons.Initially formulated for three-dimensional path planning in unmanned aerial vehicles(UAVs),the algorithmhas attracted considerable academic and industrial interest owing to its effective balance between exploration and exploitation,coupled with advantages in real-time performance and robustness.Nevertheless,as applications have diversified,limitations in convergence precision and a tendency toward premature convergence have become increasingly evident,highlighting a need for improvement.This reviewsystematically outlines the developmental trajectory of the PIO algorithm,with a particular focus on its core applications in UAV navigation,multi-objective formulations,and a spectrum of variantmodels that have emerged in recent years.It offers a structured analysis of the foundational principles underlying the PIO.It conducts a comparative assessment of various performance-enhanced versions,including hybrid models that integrate mechanisms from other optimization paradigms.Additionally,the strengths andweaknesses of distinct PIOvariants are critically examined frommultiple perspectives,including intrinsic algorithmic characteristics,suitability for specific application scenarios,objective function design,and the rigor of the statistical evaluation methodologies employed in empirical studies.Finally,this paper identifies principal challenges within current PIO research and proposes several prospective research directions.Future work should focus on mitigating premature convergence by refining the two-phase search structure and adjusting the exponential decrease of individual numbers during the landmark operator.Enhancing parameter adaptation strategies,potentially using reinforcement learning for dynamic tuning,and advancing theoretical analyses on convergence and complexity are also critical.Further applications should be explored in constrained path planning,Neural Architecture Search(NAS),and other real-worldmulti-objective problems.For Multi-objective PIO(MPIO),key improvements include controlling the growth of the external archive and designing more effective selection mechanisms to maintain convergence efficiency.These efforts are expected to strengthen both the theoretical foundation and practical versatility of PIO and its variants.
基金supported in part by the Technical Service for the Development and Application of an Intelligent Visual Management Platformfor Expressway Construction Progress Based on BIM Technology(grant NO.JKYZLX-2023-09)in partby the Technical Service for the Development of an Early Warning Model in the Research and Application of Key Technologies for Tunnel Operation Safety Monitoring and Early Warning Based on Digital Twin(grant NO.JK-S02-ZNGS-202412-JISHU-FA-0035)sponsored by Yunnan Transportation Science Research Institute Co.,Ltd.
文摘With the rapid development of transportation infrastructure,ensuring road safety through timely and accurate highway inspection has become increasingly critical.Traditional manual inspection methods are not only time-consuming and labor-intensive,but they also struggle to provide consistent,high-precision detection and realtime monitoring of pavement surface defects.To overcome these limitations,we propose an Automatic Recognition of PavementDefect(ARPD)algorithm,which leverages unmanned aerial vehicle(UAV)-based aerial imagery to automate the inspection process.The ARPD framework incorporates a backbone network based on the Selective State Space Model(S3M),which is designed to capture long-range temporal dependencies.This enables effective modeling of dynamic correlations among redundant and often repetitive structures commonly found in road imagery.Furthermore,a neck structure based on Semantics and Detail Infusion(SDI)is introduced to guide cross-scale feature fusion.The SDI module enhances the integration of low-level spatial details with high-level semantic cues,thereby improving feature expressiveness and defect localization accuracy.Experimental evaluations demonstrate that theARPDalgorithm achieves a mean average precision(mAP)of 86.1%on a custom-labeled pavement defect dataset,outperforming the state-of-the-art YOLOv11 segmentation model.The algorithm also maintains strong generalization ability on public datasets.These results confirm that ARPD is well-suited for diverse real-world applications in intelligent,large-scale highway defect monitoring and maintenance planning.
基金National Key Research and Development Program of China,No.2023YFC3006704National Natural Science Foundation of China,No.42171047CAS-CSIRO Partnership Joint Project of 2024,No.177GJHZ2023097MI。
文摘Accurate prediction of flood events is important for flood control and risk management.Machine learning techniques contributed greatly to advances in flood predictions,and existing studies mainly focused on predicting flood resource variables using single or hybrid machine learning techniques.However,class-based flood predictions have rarely been investigated,which can aid in quickly diagnosing comprehensive flood characteristics and proposing targeted management strategies.This study proposed a prediction approach of flood regime metrics and event classes coupling machine learning algorithms with clustering-deduced membership degrees.Five algorithms were adopted for this exploration.Results showed that the class membership degrees accurately determined event classes with class hit rates up to 100%,compared with the four classes clustered from nine regime metrics.The nonlinear algorithms(Multiple Linear Regression,Random Forest,and least squares-Support Vector Machine)outperformed the linear techniques(Multiple Linear Regression and Stepwise Regression)in predicting flood regime metrics.The proposed approach well predicted flood event classes with average class hit rates of 66.0%-85.4%and 47.2%-76.0%in calibration and validation periods,respectively,particularly for the slow and late flood events.The predictive capability of the proposed prediction approach for flood regime metrics and classes was considerably stronger than that of hydrological modeling approach.
基金National Natural Science Foundation of China(No.52375236)Fundamental Research Funds for the Central Universities of China(No.23D110316)。
文摘In reliability analyses,the absence of a priori information on the most probable point of failure(MPP)may result in overlooking critical points,thereby leading to biased assessment outcomes.Moreover,second-order reliability methods exhibit limited accuracy in highly nonlinear scenarios.To overcome these challenges,a novel reliability analysis strategy based on a multimodal differential evolution algorithm and a hypersphere integration method is proposed.Initially,the penalty function method is employed to reformulate the MPP search problem as a conditionally constrained optimization task.Subsequently,a differential evolution algorithm incorporating a population delineation strategy is utilized to identify all MPPs.Finally,a paraboloid equation is constructed based on the curvature of the limit-state function at the MPPs,and the failure probability of the structure is calculated by using the hypersphere integration method.The localization effectiveness of the MPPs is compared through multiple numerical cases and two engineering examples,with accuracy comparisons of failure probabilities against the first-order reliability method(FORM)and the secondorder reliability method(SORM).The results indicate that the method effectively identifies existing MPPs and achieves higher solution precision.