Open caissons are widely used in foundation engineering because of their load-bearing efficiency and adaptability in diverse soil conditions.However,accurately predicting their undrained bearing capacity in layered so...Open caissons are widely used in foundation engineering because of their load-bearing efficiency and adaptability in diverse soil conditions.However,accurately predicting their undrained bearing capacity in layered soils remains a complex challenge.This study presents a novel application of five ensemble machine(ML)algorithms-random forest(RF),gradient boosting machine(GBM),extreme gradient boosting(XGBoost),adaptive boosting(AdaBoost),and categorical boosting(CatBoost)-to predict the undrained bearing capacity factor(Nc)of circular open caissons embedded in two-layered clay on the basis of results from finite element limit analysis(FELA).The input dataset consists of 1188 numerical simulations using the Tresca failure criterion,varying in geometrical and soil parameters.The FELA was performed via OptumG2 software with adaptive meshing techniques and verified against existing benchmark studies.The ML models were trained on 70% of the dataset and tested on the remaining 30%.Their performance was evaluated using six statistical metrics:coefficient of determination(R²),mean absolute error(MAE),root mean squared error(RMSE),index of scatter(IOS),RMSE-to-standard deviation ratio(RSR),and variance explained factor(VAF).The results indicate that all the models achieved high accuracy,with R²values exceeding 97.6%and RMSE values below 0.02.Among them,AdaBoost and CatBoost consistently outperformed the other methods across both the training and testing datasets,demonstrating superior generalizability and robustness.The proposed ML framework offers an efficient,accurate,and data-driven alternative to traditional methods for estimating caisson capacity in stratified soils.This approach can aid in reducing computational costs while improving reliability in the early stages of foundation design.展开更多
The Cross-domain Heuristic Search Challenge(CHeSC)is a competition focused on creating efficient search algorithms adaptable to diverse problem domains.Selection hyper-heuristics are a class of algorithms that dynamic...The Cross-domain Heuristic Search Challenge(CHeSC)is a competition focused on creating efficient search algorithms adaptable to diverse problem domains.Selection hyper-heuristics are a class of algorithms that dynamically choose heuristics during the search process.Numerous selection hyper-heuristics have different imple-mentation strategies.However,comparisons between them are lacking in the literature,and previous works have not highlighted the beneficial and detrimental implementation methods of different components.The question is how to effectively employ them to produce an efficient search heuristic.Furthermore,the algorithms that competed in the inaugural CHeSC have not been collectively reviewed.This work conducts a review analysis of the top twenty competitors from this competition to identify effective and ineffective strategies influencing algorithmic performance.A summary of the main characteristics and classification of the algorithms is presented.The analysis underlines efficient and inefficient methods in eight key components,including search points,search phases,heuristic selection,move acceptance,feedback,Tabu mechanism,restart mechanism,and low-level heuristic parameter control.This review analyzes the components referencing the competition’s final leaderboard and discusses future research directions for these components.The effective approaches,identified as having the highest quality index,are mixed search point,iterated search phases,relay hybridization selection,threshold acceptance,mixed learning,Tabu heuristics,stochastic restart,and dynamic parameters.Findings are also compared with recent trends in hyper-heuristics.This work enhances the understanding of selection hyper-heuristics,offering valuable insights for researchers and practitioners aiming to develop effective search algorithms for diverse problem domains.展开更多
Heuristic optimization algorithms have been widely used in solving complex optimization problems in various fields such as engineering,economics,and computer science.These algorithms are designed to find high-quality ...Heuristic optimization algorithms have been widely used in solving complex optimization problems in various fields such as engineering,economics,and computer science.These algorithms are designed to find high-quality solutions efficiently by balancing exploration of the search space and exploitation of promising solutions.While heuristic optimization algorithms vary in their specific details,they often exhibit common patterns that are essential to their effectiveness.This paper aims to analyze and explore common patterns in heuristic optimization algorithms.Through a comprehensive review of the literature,we identify the patterns that are commonly observed in these algorithms,including initialization,local search,diversity maintenance,adaptation,and stochasticity.For each pattern,we describe the motivation behind it,its implementation,and its impact on the search process.To demonstrate the utility of our analysis,we identify these patterns in multiple heuristic optimization algorithms.For each case study,we analyze how the patterns are implemented in the algorithm and how they contribute to its performance.Through these case studies,we show how our analysis can be used to understand the behavior of heuristic optimization algorithms and guide the design of new algorithms.Our analysis reveals that patterns in heuristic optimization algorithms are essential to their effectiveness.By understanding and incorporating these patterns into the design of new algorithms,researchers can develop more efficient and effective optimization algorithms.展开更多
The rapid advancement of 6G communication technologies and generative artificial intelligence(AI)is catalyzing a new wave of innovation at the intersection of networking and intelligent computing.On the one hand,6G en...The rapid advancement of 6G communication technologies and generative artificial intelligence(AI)is catalyzing a new wave of innovation at the intersection of networking and intelligent computing.On the one hand,6G envisions a hyper-connected environment that supports ubiquitous intelligence through ultra-low latency,high throughput,massive device connectivity,and integrated sensing and communication.On the other hand,generative AI,powered by large foundation models,has emerged as a powerful paradigm capable of creating.展开更多
Neuromorphic computing extends beyond sequential processing modalities and outperforms traditional von Neumann architectures in implementing more complicated tasks,e.g.,pattern processing,image recognition,and decisio...Neuromorphic computing extends beyond sequential processing modalities and outperforms traditional von Neumann architectures in implementing more complicated tasks,e.g.,pattern processing,image recognition,and decision making.It features parallel interconnected neural networks,high fault tolerance,robustness,autonomous learning capability,and ultralow energy dissipation.The algorithms of artificial neural network(ANN)have also been widely used because of their facile self-organization and self-learning capabilities,which mimic those of the human brain.To some extent,ANN reflects several basic functions of the human brain and can be efficiently integrated into neuromorphic devices to perform neuromorphic computations.This review highlights recent advances in neuromorphic devices assisted by machine learning algorithms.First,the basic structure of simple neuron models inspired by biological neurons and the information processing in simple neural networks are particularly discussed.Second,the fabrication and research progress of neuromorphic devices are presented regarding to materials and structures.Furthermore,the fabrication of neuromorphic devices,including stand-alone neuromorphic devices,neuromorphic device arrays,and integrated neuromorphic systems,is discussed and demonstrated with reference to some respective studies.The applications of neuromorphic devices assisted by machine learning algorithms in different fields are categorized and investigated.Finally,perspectives,suggestions,and potential solutions to the current challenges of neuromorphic devices are provided.展开更多
Wind energy has emerged as a potential replacement for fossil fuel-based energy sources.To harness maximum wind energy,a crucial decision in the development of an efficient wind farm is the optimal layout design.This ...Wind energy has emerged as a potential replacement for fossil fuel-based energy sources.To harness maximum wind energy,a crucial decision in the development of an efficient wind farm is the optimal layout design.This layout defines the specific locations of the turbines within the wind farm.The process of finding the optimal locations of turbines,in the presence of various technical and technological constraints,makes the wind farm layout design problem a complex optimization problem.This problem has traditionally been solved with nature-inspired algorithms with promising results.The performance and convergence of nature-inspired algorithms depend on several parameters,among which the algorithm termination criterion plays a crucial role.Timely convergence is an important aspect of efficient algorithm design because an inefficient algorithm results in wasted computational resources,unwarranted electricity consumption,and hardware stress.This study provides an in-depth analysis of several termination criteria while using the genetic algorithm as a test bench,with its application to the wind farm layout design problem while considering various wind scenarios.The performance of six termination criteria is empirically evaluated with respect to the quality of solutions produced and the execution time involved.Due to the conflicting nature of these two attributes,fuzzy logic-based multi-attribute decision-making is employed in the decision process.Results for the fuzzy decision approach indicate that among the various criteria tested,the criterion Phi achieves an improvement in the range of 2.44%to 32.93%for wind scenario 1.For scenario 2,Best-worst termination criterion performed well compared to the other criteria evaluated,with an improvement in the range of 1.2%to 9.64%.For scenario 3,Hitting bound was the best performer with an improvement of 1.16%to 20.93%.展开更多
This study addresses the critical challenge of reconfiguration in unbalanced power distribution networks(UPDNs),focusing on the complex 123-Bus test system.Three scenarios are investigated:(1)simultaneous power loss r...This study addresses the critical challenge of reconfiguration in unbalanced power distribution networks(UPDNs),focusing on the complex 123-Bus test system.Three scenarios are investigated:(1)simultaneous power loss reduction and voltage profile improvement,(2)minimization of voltage and current unbalance indices under various operational cases,and(3)multi-objective optimization using Pareto front analysis to concurrently optimize voltage unbalance index,active power loss,and current unbalance index.Unlike previous research that oftensimplified system components,this work maintains all equipment,including capacitor banks,transformers,and voltage regulators,to ensure realistic results.The study evaluates twelve metaheuristic algorithms to solve the reconfiguration problem(RecPrb)in UPDNs.A comprehensive statistical analysis is conducted to identify the most efficient algorithm for solving the RecPrb in the 123-Bus UPDN,employing multiple performance metrics and comparative techniques.The Artificial Hummingbird Algorithm emerges as the top-performing algorithm and is subsequently applied to address a multi-objective optimization challenge in the 123-Bus UPDN.This research contributes valuable insights for network operators and researchers in selecting suitable algorithms for specific reconfiguration scenarios,advancing the field of UPDN optimization and management.展开更多
The word“spatial”fundamentally relates to human existence,evolution,and activity in terrestrial and even celestial spaces.After reviewing the spatial features of many areas,the paper describes basics of high level m...The word“spatial”fundamentally relates to human existence,evolution,and activity in terrestrial and even celestial spaces.After reviewing the spatial features of many areas,the paper describes basics of high level model and technology called Spatial Grasp for dealing with large distributed systems,which can provide spatial vision,awareness,management,control,and even consciousness.The technology description includes its key Spatial Grasp Language(SGL),self-evolution of recursive SGL scenarios,and implementation of SGL interpreter converting distributed networked systems into powerful spatial engines.Examples of typical spatial scenarios in SGL include finding shortest path tree and shortest path between network nodes,collecting proper information throughout the whole world,elimination of multiple targets by intelligent teams of chasers,and withstanding cyber attacks in distributed networked systems.Also this paper compares Spatial Grasp model with traditional algorithms,confirming universality of the former for any spatial systems,while the latter just tools for concrete applications.展开更多
Natural Language Processing(NLP)has become essential in text classification,sentiment analysis,machine translation,and speech recognition applications.As these tasks become complex,traditionalmachine learning and deep...Natural Language Processing(NLP)has become essential in text classification,sentiment analysis,machine translation,and speech recognition applications.As these tasks become complex,traditionalmachine learning and deep learning models encounter challenges with optimization,parameter tuning,and handling large-scale,highdimensional data.Bio-inspired algorithms,which mimic natural processes,offer robust optimization capabilities that can enhance NLP performance by improving feature selection,optimizing model parameters,and integrating adaptive learning mechanisms.This review explores the state-of-the-art applications of bio-inspired algorithms—such as Genetic Algorithms(GA),Particle Swarm Optimization(PSO),and Ant Colony Optimization(ACO)—across core NLP tasks.We analyze their comparative advantages,discuss their integration with neural network models,and address computational and scalability limitations.Through a synthesis of existing research,this paper highlights the unique strengths and current challenges of bio-inspired approaches in NLP,offering insights into hybrid models and lightweight,resource-efficient adaptations for real-time processing.Finally,we outline future research directions that emphasize the development of scalable,effective bio-inspired methods adaptable to evolving data environments.展开更多
Background:In recent years,there has been a growing trend in the utilization of observational studies that make use of routinely collected healthcare data(RCD).These studies rely on algorithms to identify specific hea...Background:In recent years,there has been a growing trend in the utilization of observational studies that make use of routinely collected healthcare data(RCD).These studies rely on algorithms to identify specific health conditions(e.g.,diabetes or sepsis)for statistical analyses.However,there has been substantial variation in the algorithm development and validation,leading to frequently suboptimal performance and posing a significant threat to the validity of study findings.Unfortunately,these issues are often overlooked.Methods:We systematically developed guidance for the development,validation,and evaluation of algorithms designed to identify health status(DEVELOP-RCD).Our initial efforts involved conducting both a narrative review and a systematic review of published studies on the concepts and methodological issues related to algorithm development,validation,and evaluation.Subsequently,we conducted an empirical study on an algorithm for identifying sepsis.Based on these findings,we formulated specific workflow and recommendations for algorithm development,validation,and evaluation within the guidance.Finally,the guidance underwent independent review by a panel of 20 external experts who then convened a consensus meeting to finalize it.Results:A standardized workflow for algorithm development,validation,and evaluation was established.Guided by specific health status considerations,the workflow comprises four integrated steps:assessing an existing algorithm’s suitability for the target health status;developing a new algorithm using recommended methods;validating the algorithm using prescribed performance measures;and evaluating the impact of the algorithm on study results.Additionally,13 good practice recommendations were formulated with detailed explanations.Furthermore,a practical study on sepsis identification was included to demonstrate the application of this guidance.Conclusions:The establishment of guidance is intended to aid researchers and clinicians in the appropriate and accurate development and application of algorithms for identifying health status from RCD.This guidance has the potential to enhance the credibility of findings from observational studies involving RCD.展开更多
The distribution of shear-wave velocities in the subsurface is generally used to assess the potential forseismic liquefaction and soil amplification effects and to classify seismic sites. Newly developeddistributed ac...The distribution of shear-wave velocities in the subsurface is generally used to assess the potential forseismic liquefaction and soil amplification effects and to classify seismic sites. Newly developeddistributed acoustic sensing (DAS) technology enables estimation of the shear-wave distribution as ahigh-density seismic observation system. This technology is characterized by low maintenance costs,high-resolution outputs, and real-time data transmission capabilities, albeit with the challenge ofmanaging massive data generation. Rapid and efficient interpretation of data is the key to advancingapplication of the DAS technology. In this study, field tests were carried out to record ambient noise overa short period using DAS technology, from which the surface-wave dispersion curves were extracted. Inorder to reduce the influence of directional effects on the results, an unsupervised clustering method isused to select appropriate clusters to extract the Green's function. A combination of a genetic algorithmand Monte Carlo (GA-MC) simulation is proposed to invert the subsurface velocity structure. Thestratigraphic profiles obtained by the GA-MC method are in agreement with the borehole profiles.Compared to other methods, the proposed optimization method not only improves the solution qualitybut also reduces the solution time.展开更多
In recent years,feature selection(FS)optimization of high-dimensional gene expression data has become one of the most promising approaches for cancer prediction and classification.This work reviews FS and classificati...In recent years,feature selection(FS)optimization of high-dimensional gene expression data has become one of the most promising approaches for cancer prediction and classification.This work reviews FS and classification methods that utilize evolutionary algorithms(EAs)for gene expression profiles in cancer or medical applications based on research motivations,challenges,and recommendations.Relevant studies were retrieved from four major academic databases-IEEE,Scopus,Springer,and ScienceDirect-using the keywords‘cancer classification’,‘optimization’,‘FS’,and‘gene expression profile’.A total of 67 papers were finally selected with key advancements identified as follows:(1)The majority of papers(44.8%)focused on developing algorithms and models for FS and classification.(2)The second category encompassed studies on biomarker identification by EAs,including 20 papers(30%).(3)The third category comprised works that applied FS to cancer data for decision support system purposes,addressing high-dimensional data and the formulation of chromosome length.These studies accounted for 12%of the total number of studies.(4)The remaining three papers(4.5%)were reviews and surveys focusing on models and developments in prediction and classification optimization for cancer classification under current technical conditions.This review highlights the importance of optimizing FS in EAs to manage high-dimensional data effectively.Despite recent advancements,significant limitations remain:the dynamic formulation of chromosome length remains an underexplored area.Thus,further research is needed on dynamic-length chromosome techniques for more sophisticated biomarker gene selection techniques.The findings suggest that further advancements in dynamic chromosome length formulations and adaptive algorithms could enhance cancer classification accuracy and efficiency.展开更多
Quantum computing offers unprecedented computational power, enabling simultaneous computations beyond traditional computers. Quantum computers differ significantly from classical computers, necessitating a distinct ap...Quantum computing offers unprecedented computational power, enabling simultaneous computations beyond traditional computers. Quantum computers differ significantly from classical computers, necessitating a distinct approach to algorithm design, which involves taming quantum mechanical phenomena. This paper extends the numbering of computable programs to be applied in the quantum computing context. Numbering computable programs is a theoretical computer science concept that assigns unique numbers to individual programs or algorithms. Common methods include Gödel numbering which encodes programs as strings of symbols or characters, often used in formal systems and mathematical logic. Based on the proposed numbering approach, this paper presents a mechanism to explore the set of possible quantum algorithms. The proposed approach is able to construct useful circuits such as Quantum Key Distribution BB84 protocol, which enables sender and receiver to establish a secure cryptographic key via a quantum channel. The proposed approach facilitates the process of exploring and constructing quantum algorithms.展开更多
Face recognition has emerged as one of the most prominent applications of image analysis and under-standing,gaining considerable attention in recent years.This growing interest is driven by two key factors:its extensi...Face recognition has emerged as one of the most prominent applications of image analysis and under-standing,gaining considerable attention in recent years.This growing interest is driven by two key factors:its extensive applications in law enforcement and the commercial domain,and the rapid advancement of practical technologies.Despite the significant advancements,modern recognition algorithms still struggle in real-world conditions such as varying lighting conditions,occlusion,and diverse facial postures.In such scenarios,human perception is still well above the capabilities of present technology.Using the systematic mapping study,this paper presents an in-depth review of face detection algorithms and face recognition algorithms,presenting a detailed survey of advancements made between 2015 and 2024.We analyze key methodologies,highlighting their strengths and restrictions in the application context.Additionally,we examine various datasets used for face detection/recognition datasets focusing on the task-specific applications,size,diversity,and complexity.By analyzing these algorithms and datasets,this survey works as a valuable resource for researchers,identifying the research gap in the field of face detection and recognition and outlining potential directions for future research.展开更多
The advent of microgrids in modern energy systems heralds a promising era of resilience,sustainability,and efficiency.Within the realm of grid-tied microgrids,the selection of an optimal optimization algorithm is crit...The advent of microgrids in modern energy systems heralds a promising era of resilience,sustainability,and efficiency.Within the realm of grid-tied microgrids,the selection of an optimal optimization algorithm is critical for effective energy management,particularly in economic dispatching.This study compares the performance of Particle Swarm Optimization(PSO)and Genetic Algorithms(GA)in microgrid energy management systems,implemented using MATLAB tools.Through a comprehensive review of the literature and sim-ulations conducted in MATLAB,the study analyzes performance metrics,convergence speed,and the overall efficacy of GA and PSO,with a focus on economic dispatching tasks.Notably,a significant distinction emerges between the cost curves generated by the two algo-rithms for microgrid operation,with the PSO algorithm consistently resulting in lower costs due to its effective economic dispatching capabilities.Specifically,the utilization of the PSO approach could potentially lead to substantial savings on the power bill,amounting to approximately$15.30 in this evaluation.Thefindings provide insights into the strengths and limitations of each algorithm within the complex dynamics of grid-tied microgrids,thereby assisting stakeholders and researchers in arriving at informed decisions.This study contributes to the discourse on sustainable energy management by offering actionable guidance for the advancement of grid-tied micro-grid technologies through MATLAB-implemented optimization algorithms.展开更多
This study presents a novel hybrid topology optimization and mold design framework that integrates process fitting,runner system optimization,and structural analysis to significantly enhance the performance of injecti...This study presents a novel hybrid topology optimization and mold design framework that integrates process fitting,runner system optimization,and structural analysis to significantly enhance the performance of injection-molded parts.At its core,the framework employs a greedy algorithm that generates runner systems based on adjacency and shortest path principles,leading to improvements in both mechanical strength and material efficiency.The design optimization is validated through a series of rigorous experimental tests,including three-point bending and torsion tests performed on key-socket frames,ensuring that the optimized designs meet practical performance requirements.A critical innovation of the framework is the development of the Adjacent Element Temperature-Driven Prestress Algorithm(AETDPA),which refines the prediction of mechanical failure and strength fitting.This algorithm has been shown to deliver mesh-independent accuracy,thereby enhancing the reliability of simulation results across various design iterations.The framework’s adaptability is further demonstrated by its ability to adjust optimization methods based on the unique geometry of each part,thus accelerating the overall design process while ensuring struc-tural integrity.In addition to its immediate applications in injection molding,the study explores the potential extension of this framework to metal additive manufacturing,opening new avenues for its use in advanced manufacturing technologies.Numerical simulations,including finite element analysis,support the experimental findings and confirm that the optimized designs provide a balanced combination of strength,durability,and efficiency.Furthermore,the integration challenges with existing injection molding practices are addressed,underscoring the framework’s scalability and industrial relevance.Overall,this hybrid topology optimization framework offers a computationally efficient and robust solution for advanced manufacturing applications,promising significant improvements in design efficiency,cost-effectiveness,and product performance.Future work will focus on further enhancing algorithm robustness and exploring additional applications across diverse manufacturing processes.展开更多
In the field of calculating the attack area of air-to-air missiles in modern air combat scenarios,the limitations of existing research,including real-time calculation,accuracy efficiency trade-off,and the absence of t...In the field of calculating the attack area of air-to-air missiles in modern air combat scenarios,the limitations of existing research,including real-time calculation,accuracy efficiency trade-off,and the absence of the three-dimensional attack area model,restrict their practical applications.To address these issues,an improved backtracking algorithm is proposed to improve calculation efficiency.A significant reduction in solution time and maintenance of accuracy in the three-dimensional attack area are achieved by using the proposed algorithm.Furthermore,the age-layered population structure genetic programming(ALPS-GP)algorithm is introduced to determine an analytical polynomial model of the three-dimensional attack area,considering real-time requirements.The accuracy of the polynomial model is enhanced through the coefficient correction using an improved gradient descent algorithm.The study reveals a remarkable combination of high accuracy and efficient real-time computation,with a mean error of 91.89 m using the analytical polynomial model of the three-dimensional attack area solved in just 10^(-4)s,thus meeting the requirements of real-time combat scenarios.展开更多
This study proposes a novel time-synchronization protocol inspired by stochastic gradient algorithms.The clock model of each network node in this synchronizer is configured as a generic adaptive filter where different...This study proposes a novel time-synchronization protocol inspired by stochastic gradient algorithms.The clock model of each network node in this synchronizer is configured as a generic adaptive filter where different stochastic gradient algorithms can be adopted for adaptive clock frequency adjustments.The study analyzes the pairwise synchronization behavior of the protocol and proves the generalized convergence of the synchronization error and clock frequency.A novel closed-form expression is also derived for a generalized asymptotic error variance steady state.Steady and convergence analyses are then presented for the synchronization,with frequency adaptations done using least mean square(LMS),the Newton search,the gradient descent(GraDes),the normalized LMS(N-LMS),and the Sign-Data LMS algorithms.Results obtained from real-time experiments showed a better performance of our protocols as compared to the Average Proportional-Integral Synchronization Protocol(AvgPISync)regarding the impact of quantization error on synchronization accuracy,precision,and convergence time.This generalized approach to time synchronization allows flexibility in selecting a suitable protocol for different wireless sensor network applications.展开更多
In recent years,the development of new types of nuclear reactors,such as transportable,marine,and space reactors,has presented new challenges for the optimization of reactor radiation-shielding design.Shielding struct...In recent years,the development of new types of nuclear reactors,such as transportable,marine,and space reactors,has presented new challenges for the optimization of reactor radiation-shielding design.Shielding structures typically need to be lightweight,miniaturized,and radiation-protected,which is a multi-parameter and multi-objective optimization problem.The conventional multi-objective(two or three objectives)optimization method for radiation-shielding design exhibits limitations for a number of optimization objectives and variable parameters,as well as a deficiency in achieving a global optimal solution,thereby failing to meet the requirements of shielding optimization for newly developed reactors.In this study,genetic and artificial bee-colony algorithms are combined with a reference-point-selection strategy and applied to the many-objective(having four or more objectives)optimal design of reactor radiation shielding.To validate the reliability of the methods,an optimization simulation is conducted on three-dimensional shielding structures and another complicated shielding-optimization problem.The numerical results demonstrate that the proposed algorithms outperform conventional shielding-design methods in terms of optimization performance,and they exhibit their reliability in practical engineering problems.The many-objective optimization algorithms developed in this study are proven to efficiently and consistently search for Pareto-front shielding schemes.Therefore,the algorithms proposed in this study offer novel insights into improving the shielding-design performance and shielding quality of new reactor types.展开更多
This paper studies polygon simplification algorithms for 3D models,focuses on the optimization algorithm of quadratic error metric(QEM),explores the impacts of different methods on the simplification of different mode...This paper studies polygon simplification algorithms for 3D models,focuses on the optimization algorithm of quadratic error metric(QEM),explores the impacts of different methods on the simplification of different models,and develops a web-based visualization application.Metrics such as the Hausdorff distance are used to evaluate the balance between the degree of simplification and the retention of model details.展开更多
文摘Open caissons are widely used in foundation engineering because of their load-bearing efficiency and adaptability in diverse soil conditions.However,accurately predicting their undrained bearing capacity in layered soils remains a complex challenge.This study presents a novel application of five ensemble machine(ML)algorithms-random forest(RF),gradient boosting machine(GBM),extreme gradient boosting(XGBoost),adaptive boosting(AdaBoost),and categorical boosting(CatBoost)-to predict the undrained bearing capacity factor(Nc)of circular open caissons embedded in two-layered clay on the basis of results from finite element limit analysis(FELA).The input dataset consists of 1188 numerical simulations using the Tresca failure criterion,varying in geometrical and soil parameters.The FELA was performed via OptumG2 software with adaptive meshing techniques and verified against existing benchmark studies.The ML models were trained on 70% of the dataset and tested on the remaining 30%.Their performance was evaluated using six statistical metrics:coefficient of determination(R²),mean absolute error(MAE),root mean squared error(RMSE),index of scatter(IOS),RMSE-to-standard deviation ratio(RSR),and variance explained factor(VAF).The results indicate that all the models achieved high accuracy,with R²values exceeding 97.6%and RMSE values below 0.02.Among them,AdaBoost and CatBoost consistently outperformed the other methods across both the training and testing datasets,demonstrating superior generalizability and robustness.The proposed ML framework offers an efficient,accurate,and data-driven alternative to traditional methods for estimating caisson capacity in stratified soils.This approach can aid in reducing computational costs while improving reliability in the early stages of foundation design.
基金funded by Ministry of Higher Education(MoHE)Malaysia,under Transdisciplinary Research Grant Scheme(TRGS/1/2019/UKM/01/4/2).
文摘The Cross-domain Heuristic Search Challenge(CHeSC)is a competition focused on creating efficient search algorithms adaptable to diverse problem domains.Selection hyper-heuristics are a class of algorithms that dynamically choose heuristics during the search process.Numerous selection hyper-heuristics have different imple-mentation strategies.However,comparisons between them are lacking in the literature,and previous works have not highlighted the beneficial and detrimental implementation methods of different components.The question is how to effectively employ them to produce an efficient search heuristic.Furthermore,the algorithms that competed in the inaugural CHeSC have not been collectively reviewed.This work conducts a review analysis of the top twenty competitors from this competition to identify effective and ineffective strategies influencing algorithmic performance.A summary of the main characteristics and classification of the algorithms is presented.The analysis underlines efficient and inefficient methods in eight key components,including search points,search phases,heuristic selection,move acceptance,feedback,Tabu mechanism,restart mechanism,and low-level heuristic parameter control.This review analyzes the components referencing the competition’s final leaderboard and discusses future research directions for these components.The effective approaches,identified as having the highest quality index,are mixed search point,iterated search phases,relay hybridization selection,threshold acceptance,mixed learning,Tabu heuristics,stochastic restart,and dynamic parameters.Findings are also compared with recent trends in hyper-heuristics.This work enhances the understanding of selection hyper-heuristics,offering valuable insights for researchers and practitioners aiming to develop effective search algorithms for diverse problem domains.
文摘Heuristic optimization algorithms have been widely used in solving complex optimization problems in various fields such as engineering,economics,and computer science.These algorithms are designed to find high-quality solutions efficiently by balancing exploration of the search space and exploitation of promising solutions.While heuristic optimization algorithms vary in their specific details,they often exhibit common patterns that are essential to their effectiveness.This paper aims to analyze and explore common patterns in heuristic optimization algorithms.Through a comprehensive review of the literature,we identify the patterns that are commonly observed in these algorithms,including initialization,local search,diversity maintenance,adaptation,and stochasticity.For each pattern,we describe the motivation behind it,its implementation,and its impact on the search process.To demonstrate the utility of our analysis,we identify these patterns in multiple heuristic optimization algorithms.For each case study,we analyze how the patterns are implemented in the algorithm and how they contribute to its performance.Through these case studies,we show how our analysis can be used to understand the behavior of heuristic optimization algorithms and guide the design of new algorithms.Our analysis reveals that patterns in heuristic optimization algorithms are essential to their effectiveness.By understanding and incorporating these patterns into the design of new algorithms,researchers can develop more efficient and effective optimization algorithms.
文摘The rapid advancement of 6G communication technologies and generative artificial intelligence(AI)is catalyzing a new wave of innovation at the intersection of networking and intelligent computing.On the one hand,6G envisions a hyper-connected environment that supports ubiquitous intelligence through ultra-low latency,high throughput,massive device connectivity,and integrated sensing and communication.On the other hand,generative AI,powered by large foundation models,has emerged as a powerful paradigm capable of creating.
基金financially supported by the National Natural Science Foundation of China(No.52073031)the National Key Research and Development Program of China(Nos.2023YFB3208102,2021YFB3200304)+4 种基金the China National Postdoctoral Program for Innovative Talents(No.BX2021302)the Beijing Nova Program(Nos.Z191100001119047,Z211100002121148)the Fundamental Research Funds for the Central Universities(No.E0EG6801X2)the‘Hundred Talents Program’of the Chinese Academy of Sciencesthe BrainLink program funded by the MSIT through the NRF of Korea(No.RS-2023-00237308).
文摘Neuromorphic computing extends beyond sequential processing modalities and outperforms traditional von Neumann architectures in implementing more complicated tasks,e.g.,pattern processing,image recognition,and decision making.It features parallel interconnected neural networks,high fault tolerance,robustness,autonomous learning capability,and ultralow energy dissipation.The algorithms of artificial neural network(ANN)have also been widely used because of their facile self-organization and self-learning capabilities,which mimic those of the human brain.To some extent,ANN reflects several basic functions of the human brain and can be efficiently integrated into neuromorphic devices to perform neuromorphic computations.This review highlights recent advances in neuromorphic devices assisted by machine learning algorithms.First,the basic structure of simple neuron models inspired by biological neurons and the information processing in simple neural networks are particularly discussed.Second,the fabrication and research progress of neuromorphic devices are presented regarding to materials and structures.Furthermore,the fabrication of neuromorphic devices,including stand-alone neuromorphic devices,neuromorphic device arrays,and integrated neuromorphic systems,is discussed and demonstrated with reference to some respective studies.The applications of neuromorphic devices assisted by machine learning algorithms in different fields are categorized and investigated.Finally,perspectives,suggestions,and potential solutions to the current challenges of neuromorphic devices are provided.
基金funded by King Fahd University of Petroleum&Minerals,Saudi Arabia under IRC-SES grant#INRE 2217.
文摘Wind energy has emerged as a potential replacement for fossil fuel-based energy sources.To harness maximum wind energy,a crucial decision in the development of an efficient wind farm is the optimal layout design.This layout defines the specific locations of the turbines within the wind farm.The process of finding the optimal locations of turbines,in the presence of various technical and technological constraints,makes the wind farm layout design problem a complex optimization problem.This problem has traditionally been solved with nature-inspired algorithms with promising results.The performance and convergence of nature-inspired algorithms depend on several parameters,among which the algorithm termination criterion plays a crucial role.Timely convergence is an important aspect of efficient algorithm design because an inefficient algorithm results in wasted computational resources,unwarranted electricity consumption,and hardware stress.This study provides an in-depth analysis of several termination criteria while using the genetic algorithm as a test bench,with its application to the wind farm layout design problem while considering various wind scenarios.The performance of six termination criteria is empirically evaluated with respect to the quality of solutions produced and the execution time involved.Due to the conflicting nature of these two attributes,fuzzy logic-based multi-attribute decision-making is employed in the decision process.Results for the fuzzy decision approach indicate that among the various criteria tested,the criterion Phi achieves an improvement in the range of 2.44%to 32.93%for wind scenario 1.For scenario 2,Best-worst termination criterion performed well compared to the other criteria evaluated,with an improvement in the range of 1.2%to 9.64%.For scenario 3,Hitting bound was the best performer with an improvement of 1.16%to 20.93%.
基金supported by the Scientific and Technological Research Council of Turkey(TUBITAK)under Grant No.124E002(1001-Project).
文摘This study addresses the critical challenge of reconfiguration in unbalanced power distribution networks(UPDNs),focusing on the complex 123-Bus test system.Three scenarios are investigated:(1)simultaneous power loss reduction and voltage profile improvement,(2)minimization of voltage and current unbalance indices under various operational cases,and(3)multi-objective optimization using Pareto front analysis to concurrently optimize voltage unbalance index,active power loss,and current unbalance index.Unlike previous research that oftensimplified system components,this work maintains all equipment,including capacitor banks,transformers,and voltage regulators,to ensure realistic results.The study evaluates twelve metaheuristic algorithms to solve the reconfiguration problem(RecPrb)in UPDNs.A comprehensive statistical analysis is conducted to identify the most efficient algorithm for solving the RecPrb in the 123-Bus UPDN,employing multiple performance metrics and comparative techniques.The Artificial Hummingbird Algorithm emerges as the top-performing algorithm and is subsequently applied to address a multi-objective optimization challenge in the 123-Bus UPDN.This research contributes valuable insights for network operators and researchers in selecting suitable algorithms for specific reconfiguration scenarios,advancing the field of UPDN optimization and management.
文摘The word“spatial”fundamentally relates to human existence,evolution,and activity in terrestrial and even celestial spaces.After reviewing the spatial features of many areas,the paper describes basics of high level model and technology called Spatial Grasp for dealing with large distributed systems,which can provide spatial vision,awareness,management,control,and even consciousness.The technology description includes its key Spatial Grasp Language(SGL),self-evolution of recursive SGL scenarios,and implementation of SGL interpreter converting distributed networked systems into powerful spatial engines.Examples of typical spatial scenarios in SGL include finding shortest path tree and shortest path between network nodes,collecting proper information throughout the whole world,elimination of multiple targets by intelligent teams of chasers,and withstanding cyber attacks in distributed networked systems.Also this paper compares Spatial Grasp model with traditional algorithms,confirming universality of the former for any spatial systems,while the latter just tools for concrete applications.
基金supported by AIT Laboratory,FPT University,Danang Campus,Vietnam,2024.
文摘Natural Language Processing(NLP)has become essential in text classification,sentiment analysis,machine translation,and speech recognition applications.As these tasks become complex,traditionalmachine learning and deep learning models encounter challenges with optimization,parameter tuning,and handling large-scale,highdimensional data.Bio-inspired algorithms,which mimic natural processes,offer robust optimization capabilities that can enhance NLP performance by improving feature selection,optimizing model parameters,and integrating adaptive learning mechanisms.This review explores the state-of-the-art applications of bio-inspired algorithms—such as Genetic Algorithms(GA),Particle Swarm Optimization(PSO),and Ant Colony Optimization(ACO)—across core NLP tasks.We analyze their comparative advantages,discuss their integration with neural network models,and address computational and scalability limitations.Through a synthesis of existing research,this paper highlights the unique strengths and current challenges of bio-inspired approaches in NLP,offering insights into hybrid models and lightweight,resource-efficient adaptations for real-time processing.Finally,we outline future research directions that emphasize the development of scalable,effective bio-inspired methods adaptable to evolving data environments.
基金supported by the National Natural Science Foundation of China(82225049,72104155)the Sichuan Provincial Central Government Guides Local Science and Technology Development Special Project(2022ZYD0127)the 1·3·5 Project for Disciplines of Excellence,West China Hospital,Sichuan University(ZYGD23004).
文摘Background:In recent years,there has been a growing trend in the utilization of observational studies that make use of routinely collected healthcare data(RCD).These studies rely on algorithms to identify specific health conditions(e.g.,diabetes or sepsis)for statistical analyses.However,there has been substantial variation in the algorithm development and validation,leading to frequently suboptimal performance and posing a significant threat to the validity of study findings.Unfortunately,these issues are often overlooked.Methods:We systematically developed guidance for the development,validation,and evaluation of algorithms designed to identify health status(DEVELOP-RCD).Our initial efforts involved conducting both a narrative review and a systematic review of published studies on the concepts and methodological issues related to algorithm development,validation,and evaluation.Subsequently,we conducted an empirical study on an algorithm for identifying sepsis.Based on these findings,we formulated specific workflow and recommendations for algorithm development,validation,and evaluation within the guidance.Finally,the guidance underwent independent review by a panel of 20 external experts who then convened a consensus meeting to finalize it.Results:A standardized workflow for algorithm development,validation,and evaluation was established.Guided by specific health status considerations,the workflow comprises four integrated steps:assessing an existing algorithm’s suitability for the target health status;developing a new algorithm using recommended methods;validating the algorithm using prescribed performance measures;and evaluating the impact of the algorithm on study results.Additionally,13 good practice recommendations were formulated with detailed explanations.Furthermore,a practical study on sepsis identification was included to demonstrate the application of this guidance.Conclusions:The establishment of guidance is intended to aid researchers and clinicians in the appropriate and accurate development and application of algorithms for identifying health status from RCD.This guidance has the potential to enhance the credibility of findings from observational studies involving RCD.
基金supported by the National Natural Science Foundation of China(Grant Nos.42225702 and 42077235)the Natural Science Foundation of Jiangsu Province(Grant No.BK20211086)the open fund of the Key Laboratory of Earth Fissures Geological Disaster,Ministry of Natural Resources.
文摘The distribution of shear-wave velocities in the subsurface is generally used to assess the potential forseismic liquefaction and soil amplification effects and to classify seismic sites. Newly developeddistributed acoustic sensing (DAS) technology enables estimation of the shear-wave distribution as ahigh-density seismic observation system. This technology is characterized by low maintenance costs,high-resolution outputs, and real-time data transmission capabilities, albeit with the challenge ofmanaging massive data generation. Rapid and efficient interpretation of data is the key to advancingapplication of the DAS technology. In this study, field tests were carried out to record ambient noise overa short period using DAS technology, from which the surface-wave dispersion curves were extracted. Inorder to reduce the influence of directional effects on the results, an unsupervised clustering method isused to select appropriate clusters to extract the Green's function. A combination of a genetic algorithmand Monte Carlo (GA-MC) simulation is proposed to invert the subsurface velocity structure. Thestratigraphic profiles obtained by the GA-MC method are in agreement with the borehole profiles.Compared to other methods, the proposed optimization method not only improves the solution qualitybut also reduces the solution time.
基金funded by the Ministry of Higher Education of Malaysia,grant number FRGS/1/2022/ICT02/UPSI/02/1.
文摘In recent years,feature selection(FS)optimization of high-dimensional gene expression data has become one of the most promising approaches for cancer prediction and classification.This work reviews FS and classification methods that utilize evolutionary algorithms(EAs)for gene expression profiles in cancer or medical applications based on research motivations,challenges,and recommendations.Relevant studies were retrieved from four major academic databases-IEEE,Scopus,Springer,and ScienceDirect-using the keywords‘cancer classification’,‘optimization’,‘FS’,and‘gene expression profile’.A total of 67 papers were finally selected with key advancements identified as follows:(1)The majority of papers(44.8%)focused on developing algorithms and models for FS and classification.(2)The second category encompassed studies on biomarker identification by EAs,including 20 papers(30%).(3)The third category comprised works that applied FS to cancer data for decision support system purposes,addressing high-dimensional data and the formulation of chromosome length.These studies accounted for 12%of the total number of studies.(4)The remaining three papers(4.5%)were reviews and surveys focusing on models and developments in prediction and classification optimization for cancer classification under current technical conditions.This review highlights the importance of optimizing FS in EAs to manage high-dimensional data effectively.Despite recent advancements,significant limitations remain:the dynamic formulation of chromosome length remains an underexplored area.Thus,further research is needed on dynamic-length chromosome techniques for more sophisticated biomarker gene selection techniques.The findings suggest that further advancements in dynamic chromosome length formulations and adaptive algorithms could enhance cancer classification accuracy and efficiency.
文摘Quantum computing offers unprecedented computational power, enabling simultaneous computations beyond traditional computers. Quantum computers differ significantly from classical computers, necessitating a distinct approach to algorithm design, which involves taming quantum mechanical phenomena. This paper extends the numbering of computable programs to be applied in the quantum computing context. Numbering computable programs is a theoretical computer science concept that assigns unique numbers to individual programs or algorithms. Common methods include Gödel numbering which encodes programs as strings of symbols or characters, often used in formal systems and mathematical logic. Based on the proposed numbering approach, this paper presents a mechanism to explore the set of possible quantum algorithms. The proposed approach is able to construct useful circuits such as Quantum Key Distribution BB84 protocol, which enables sender and receiver to establish a secure cryptographic key via a quantum channel. The proposed approach facilitates the process of exploring and constructing quantum algorithms.
文摘Face recognition has emerged as one of the most prominent applications of image analysis and under-standing,gaining considerable attention in recent years.This growing interest is driven by two key factors:its extensive applications in law enforcement and the commercial domain,and the rapid advancement of practical technologies.Despite the significant advancements,modern recognition algorithms still struggle in real-world conditions such as varying lighting conditions,occlusion,and diverse facial postures.In such scenarios,human perception is still well above the capabilities of present technology.Using the systematic mapping study,this paper presents an in-depth review of face detection algorithms and face recognition algorithms,presenting a detailed survey of advancements made between 2015 and 2024.We analyze key methodologies,highlighting their strengths and restrictions in the application context.Additionally,we examine various datasets used for face detection/recognition datasets focusing on the task-specific applications,size,diversity,and complexity.By analyzing these algorithms and datasets,this survey works as a valuable resource for researchers,identifying the research gap in the field of face detection and recognition and outlining potential directions for future research.
文摘The advent of microgrids in modern energy systems heralds a promising era of resilience,sustainability,and efficiency.Within the realm of grid-tied microgrids,the selection of an optimal optimization algorithm is critical for effective energy management,particularly in economic dispatching.This study compares the performance of Particle Swarm Optimization(PSO)and Genetic Algorithms(GA)in microgrid energy management systems,implemented using MATLAB tools.Through a comprehensive review of the literature and sim-ulations conducted in MATLAB,the study analyzes performance metrics,convergence speed,and the overall efficacy of GA and PSO,with a focus on economic dispatching tasks.Notably,a significant distinction emerges between the cost curves generated by the two algo-rithms for microgrid operation,with the PSO algorithm consistently resulting in lower costs due to its effective economic dispatching capabilities.Specifically,the utilization of the PSO approach could potentially lead to substantial savings on the power bill,amounting to approximately$15.30 in this evaluation.Thefindings provide insights into the strengths and limitations of each algorithm within the complex dynamics of grid-tied microgrids,thereby assisting stakeholders and researchers in arriving at informed decisions.This study contributes to the discourse on sustainable energy management by offering actionable guidance for the advancement of grid-tied micro-grid technologies through MATLAB-implemented optimization algorithms.
文摘This study presents a novel hybrid topology optimization and mold design framework that integrates process fitting,runner system optimization,and structural analysis to significantly enhance the performance of injection-molded parts.At its core,the framework employs a greedy algorithm that generates runner systems based on adjacency and shortest path principles,leading to improvements in both mechanical strength and material efficiency.The design optimization is validated through a series of rigorous experimental tests,including three-point bending and torsion tests performed on key-socket frames,ensuring that the optimized designs meet practical performance requirements.A critical innovation of the framework is the development of the Adjacent Element Temperature-Driven Prestress Algorithm(AETDPA),which refines the prediction of mechanical failure and strength fitting.This algorithm has been shown to deliver mesh-independent accuracy,thereby enhancing the reliability of simulation results across various design iterations.The framework’s adaptability is further demonstrated by its ability to adjust optimization methods based on the unique geometry of each part,thus accelerating the overall design process while ensuring struc-tural integrity.In addition to its immediate applications in injection molding,the study explores the potential extension of this framework to metal additive manufacturing,opening new avenues for its use in advanced manufacturing technologies.Numerical simulations,including finite element analysis,support the experimental findings and confirm that the optimized designs provide a balanced combination of strength,durability,and efficiency.Furthermore,the integration challenges with existing injection molding practices are addressed,underscoring the framework’s scalability and industrial relevance.Overall,this hybrid topology optimization framework offers a computationally efficient and robust solution for advanced manufacturing applications,promising significant improvements in design efficiency,cost-effectiveness,and product performance.Future work will focus on further enhancing algorithm robustness and exploring additional applications across diverse manufacturing processes.
基金National Natural Science Foundation of China(62373187)Forward-looking Layout Special Projects(ILA220591A22)。
文摘In the field of calculating the attack area of air-to-air missiles in modern air combat scenarios,the limitations of existing research,including real-time calculation,accuracy efficiency trade-off,and the absence of the three-dimensional attack area model,restrict their practical applications.To address these issues,an improved backtracking algorithm is proposed to improve calculation efficiency.A significant reduction in solution time and maintenance of accuracy in the three-dimensional attack area are achieved by using the proposed algorithm.Furthermore,the age-layered population structure genetic programming(ALPS-GP)algorithm is introduced to determine an analytical polynomial model of the three-dimensional attack area,considering real-time requirements.The accuracy of the polynomial model is enhanced through the coefficient correction using an improved gradient descent algorithm.The study reveals a remarkable combination of high accuracy and efficient real-time computation,with a mean error of 91.89 m using the analytical polynomial model of the three-dimensional attack area solved in just 10^(-4)s,thus meeting the requirements of real-time combat scenarios.
基金funded by Universiti Putra Malaysia under a Geran Putra Inisiatif(GPI)research grant with reference to GP-GPI/2023/9762100.
文摘This study proposes a novel time-synchronization protocol inspired by stochastic gradient algorithms.The clock model of each network node in this synchronizer is configured as a generic adaptive filter where different stochastic gradient algorithms can be adopted for adaptive clock frequency adjustments.The study analyzes the pairwise synchronization behavior of the protocol and proves the generalized convergence of the synchronization error and clock frequency.A novel closed-form expression is also derived for a generalized asymptotic error variance steady state.Steady and convergence analyses are then presented for the synchronization,with frequency adaptations done using least mean square(LMS),the Newton search,the gradient descent(GraDes),the normalized LMS(N-LMS),and the Sign-Data LMS algorithms.Results obtained from real-time experiments showed a better performance of our protocols as compared to the Average Proportional-Integral Synchronization Protocol(AvgPISync)regarding the impact of quantization error on synchronization accuracy,precision,and convergence time.This generalized approach to time synchronization allows flexibility in selecting a suitable protocol for different wireless sensor network applications.
基金supported by the National Natural Science Foundation of China(Nos.12475174 and 12175101)Yue Lu Shan Center Industrial Innovation(No.2024YCII0108)。
文摘In recent years,the development of new types of nuclear reactors,such as transportable,marine,and space reactors,has presented new challenges for the optimization of reactor radiation-shielding design.Shielding structures typically need to be lightweight,miniaturized,and radiation-protected,which is a multi-parameter and multi-objective optimization problem.The conventional multi-objective(two or three objectives)optimization method for radiation-shielding design exhibits limitations for a number of optimization objectives and variable parameters,as well as a deficiency in achieving a global optimal solution,thereby failing to meet the requirements of shielding optimization for newly developed reactors.In this study,genetic and artificial bee-colony algorithms are combined with a reference-point-selection strategy and applied to the many-objective(having four or more objectives)optimal design of reactor radiation shielding.To validate the reliability of the methods,an optimization simulation is conducted on three-dimensional shielding structures and another complicated shielding-optimization problem.The numerical results demonstrate that the proposed algorithms outperform conventional shielding-design methods in terms of optimization performance,and they exhibit their reliability in practical engineering problems.The many-objective optimization algorithms developed in this study are proven to efficiently and consistently search for Pareto-front shielding schemes.Therefore,the algorithms proposed in this study offer novel insights into improving the shielding-design performance and shielding quality of new reactor types.
文摘This paper studies polygon simplification algorithms for 3D models,focuses on the optimization algorithm of quadratic error metric(QEM),explores the impacts of different methods on the simplification of different models,and develops a web-based visualization application.Metrics such as the Hausdorff distance are used to evaluate the balance between the degree of simplification and the retention of model details.