Open caissons are widely used in foundation engineering because of their load-bearing efficiency and adaptability in diverse soil conditions.However,accurately predicting their undrained bearing capacity in layered so...Open caissons are widely used in foundation engineering because of their load-bearing efficiency and adaptability in diverse soil conditions.However,accurately predicting their undrained bearing capacity in layered soils remains a complex challenge.This study presents a novel application of five ensemble machine(ML)algorithms-random forest(RF),gradient boosting machine(GBM),extreme gradient boosting(XGBoost),adaptive boosting(AdaBoost),and categorical boosting(CatBoost)-to predict the undrained bearing capacity factor(Nc)of circular open caissons embedded in two-layered clay on the basis of results from finite element limit analysis(FELA).The input dataset consists of 1188 numerical simulations using the Tresca failure criterion,varying in geometrical and soil parameters.The FELA was performed via OptumG2 software with adaptive meshing techniques and verified against existing benchmark studies.The ML models were trained on 70% of the dataset and tested on the remaining 30%.Their performance was evaluated using six statistical metrics:coefficient of determination(R²),mean absolute error(MAE),root mean squared error(RMSE),index of scatter(IOS),RMSE-to-standard deviation ratio(RSR),and variance explained factor(VAF).The results indicate that all the models achieved high accuracy,with R²values exceeding 97.6%and RMSE values below 0.02.Among them,AdaBoost and CatBoost consistently outperformed the other methods across both the training and testing datasets,demonstrating superior generalizability and robustness.The proposed ML framework offers an efficient,accurate,and data-driven alternative to traditional methods for estimating caisson capacity in stratified soils.This approach can aid in reducing computational costs while improving reliability in the early stages of foundation design.展开更多
Quantum computing offers unprecedented computational power, enabling simultaneous computations beyond traditional computers. Quantum computers differ significantly from classical computers, necessitating a distinct ap...Quantum computing offers unprecedented computational power, enabling simultaneous computations beyond traditional computers. Quantum computers differ significantly from classical computers, necessitating a distinct approach to algorithm design, which involves taming quantum mechanical phenomena. This paper extends the numbering of computable programs to be applied in the quantum computing context. Numbering computable programs is a theoretical computer science concept that assigns unique numbers to individual programs or algorithms. Common methods include Gödel numbering which encodes programs as strings of symbols or characters, often used in formal systems and mathematical logic. Based on the proposed numbering approach, this paper presents a mechanism to explore the set of possible quantum algorithms. The proposed approach is able to construct useful circuits such as Quantum Key Distribution BB84 protocol, which enables sender and receiver to establish a secure cryptographic key via a quantum channel. The proposed approach facilitates the process of exploring and constructing quantum algorithms.展开更多
The advent of microgrids in modern energy systems heralds a promising era of resilience,sustainability,and efficiency.Within the realm of grid-tied microgrids,the selection of an optimal optimization algorithm is crit...The advent of microgrids in modern energy systems heralds a promising era of resilience,sustainability,and efficiency.Within the realm of grid-tied microgrids,the selection of an optimal optimization algorithm is critical for effective energy management,particularly in economic dispatching.This study compares the performance of Particle Swarm Optimization(PSO)and Genetic Algorithms(GA)in microgrid energy management systems,implemented using MATLAB tools.Through a comprehensive review of the literature and sim-ulations conducted in MATLAB,the study analyzes performance metrics,convergence speed,and the overall efficacy of GA and PSO,with a focus on economic dispatching tasks.Notably,a significant distinction emerges between the cost curves generated by the two algo-rithms for microgrid operation,with the PSO algorithm consistently resulting in lower costs due to its effective economic dispatching capabilities.Specifically,the utilization of the PSO approach could potentially lead to substantial savings on the power bill,amounting to approximately$15.30 in this evaluation.Thefindings provide insights into the strengths and limitations of each algorithm within the complex dynamics of grid-tied microgrids,thereby assisting stakeholders and researchers in arriving at informed decisions.This study contributes to the discourse on sustainable energy management by offering actionable guidance for the advancement of grid-tied micro-grid technologies through MATLAB-implemented optimization algorithms.展开更多
In the field of calculating the attack area of air-to-air missiles in modern air combat scenarios,the limitations of existing research,including real-time calculation,accuracy efficiency trade-off,and the absence of t...In the field of calculating the attack area of air-to-air missiles in modern air combat scenarios,the limitations of existing research,including real-time calculation,accuracy efficiency trade-off,and the absence of the three-dimensional attack area model,restrict their practical applications.To address these issues,an improved backtracking algorithm is proposed to improve calculation efficiency.A significant reduction in solution time and maintenance of accuracy in the three-dimensional attack area are achieved by using the proposed algorithm.Furthermore,the age-layered population structure genetic programming(ALPS-GP)algorithm is introduced to determine an analytical polynomial model of the three-dimensional attack area,considering real-time requirements.The accuracy of the polynomial model is enhanced through the coefficient correction using an improved gradient descent algorithm.The study reveals a remarkable combination of high accuracy and efficient real-time computation,with a mean error of 91.89 m using the analytical polynomial model of the three-dimensional attack area solved in just 10^(-4)s,thus meeting the requirements of real-time combat scenarios.展开更多
This study presents a novel hybrid topology optimization and mold design framework that integrates process fitting,runner system optimization,and structural analysis to significantly enhance the performance of injecti...This study presents a novel hybrid topology optimization and mold design framework that integrates process fitting,runner system optimization,and structural analysis to significantly enhance the performance of injection-molded parts.At its core,the framework employs a greedy algorithm that generates runner systems based on adjacency and shortest path principles,leading to improvements in both mechanical strength and material efficiency.The design optimization is validated through a series of rigorous experimental tests,including three-point bending and torsion tests performed on key-socket frames,ensuring that the optimized designs meet practical performance requirements.A critical innovation of the framework is the development of the Adjacent Element Temperature-Driven Prestress Algorithm(AETDPA),which refines the prediction of mechanical failure and strength fitting.This algorithm has been shown to deliver mesh-independent accuracy,thereby enhancing the reliability of simulation results across various design iterations.The framework’s adaptability is further demonstrated by its ability to adjust optimization methods based on the unique geometry of each part,thus accelerating the overall design process while ensuring struc-tural integrity.In addition to its immediate applications in injection molding,the study explores the potential extension of this framework to metal additive manufacturing,opening new avenues for its use in advanced manufacturing technologies.Numerical simulations,including finite element analysis,support the experimental findings and confirm that the optimized designs provide a balanced combination of strength,durability,and efficiency.Furthermore,the integration challenges with existing injection molding practices are addressed,underscoring the framework’s scalability and industrial relevance.Overall,this hybrid topology optimization framework offers a computationally efficient and robust solution for advanced manufacturing applications,promising significant improvements in design efficiency,cost-effectiveness,and product performance.Future work will focus on further enhancing algorithm robustness and exploring additional applications across diverse manufacturing processes.展开更多
The application of machine learning was investigated for predicting end-point temperature in the basic oxygen furnace steelmaking process,addressing gaps in the field,particularly large-scale dataset sizes and the und...The application of machine learning was investigated for predicting end-point temperature in the basic oxygen furnace steelmaking process,addressing gaps in the field,particularly large-scale dataset sizes and the underutilization of boosting algorithms.Utilizing a substantial dataset containing over 20,000 heats,significantly bigger than those in previous studies,a comprehensive evaluation of five advanced machine learning models was conducted.These include four ensemble learning algorithms:XGBoost,LightGBM,CatBoost(three boosting algorithms),along with random forest(a bagging algorithm),as well as a neural network model,namely the multilayer perceptron.Our comparative analysis reveals that Bayesian-optimized boosting models demonstrate exceptional robustness and accuracy,achieving the highest R-squared values,the lowest root mean square error,and lowest mean absolute error,along with the best hit ratio.CatBoost exhibited superior performance,with its test R-squared improving by 4.2%compared to that of the random forest and by 0.8%compared to that of the multilayer perceptron.This highlights the efficacy of boosting algorithms in refining complex industrial processes.Additionally,our investigation into the impact of varying dataset sizes,ranging from 500 to 20,000 heats,on model accuracy underscores the importance of leveraging larger-scale datasets to improve the accuracy and stability of predictive models.展开更多
This study proposes a novel time-synchronization protocol inspired by stochastic gradient algorithms.The clock model of each network node in this synchronizer is configured as a generic adaptive filter where different...This study proposes a novel time-synchronization protocol inspired by stochastic gradient algorithms.The clock model of each network node in this synchronizer is configured as a generic adaptive filter where different stochastic gradient algorithms can be adopted for adaptive clock frequency adjustments.The study analyzes the pairwise synchronization behavior of the protocol and proves the generalized convergence of the synchronization error and clock frequency.A novel closed-form expression is also derived for a generalized asymptotic error variance steady state.Steady and convergence analyses are then presented for the synchronization,with frequency adaptations done using least mean square(LMS),the Newton search,the gradient descent(GraDes),the normalized LMS(N-LMS),and the Sign-Data LMS algorithms.Results obtained from real-time experiments showed a better performance of our protocols as compared to the Average Proportional-Integral Synchronization Protocol(AvgPISync)regarding the impact of quantization error on synchronization accuracy,precision,and convergence time.This generalized approach to time synchronization allows flexibility in selecting a suitable protocol for different wireless sensor network applications.展开更多
In recent years,the development of new types of nuclear reactors,such as transportable,marine,and space reactors,has presented new challenges for the optimization of reactor radiation-shielding design.Shielding struct...In recent years,the development of new types of nuclear reactors,such as transportable,marine,and space reactors,has presented new challenges for the optimization of reactor radiation-shielding design.Shielding structures typically need to be lightweight,miniaturized,and radiation-protected,which is a multi-parameter and multi-objective optimization problem.The conventional multi-objective(two or three objectives)optimization method for radiation-shielding design exhibits limitations for a number of optimization objectives and variable parameters,as well as a deficiency in achieving a global optimal solution,thereby failing to meet the requirements of shielding optimization for newly developed reactors.In this study,genetic and artificial bee-colony algorithms are combined with a reference-point-selection strategy and applied to the many-objective(having four or more objectives)optimal design of reactor radiation shielding.To validate the reliability of the methods,an optimization simulation is conducted on three-dimensional shielding structures and another complicated shielding-optimization problem.The numerical results demonstrate that the proposed algorithms outperform conventional shielding-design methods in terms of optimization performance,and they exhibit their reliability in practical engineering problems.The many-objective optimization algorithms developed in this study are proven to efficiently and consistently search for Pareto-front shielding schemes.Therefore,the algorithms proposed in this study offer novel insights into improving the shielding-design performance and shielding quality of new reactor types.展开更多
This paper studies polygon simplification algorithms for 3D models,focuses on the optimization algorithm of quadratic error metric(QEM),explores the impacts of different methods on the simplification of different mode...This paper studies polygon simplification algorithms for 3D models,focuses on the optimization algorithm of quadratic error metric(QEM),explores the impacts of different methods on the simplification of different models,and develops a web-based visualization application.Metrics such as the Hausdorff distance are used to evaluate the balance between the degree of simplification and the retention of model details.展开更多
Based on the Google Earth Engine cloud computing data platform,this study employed three algorithms including Support Vector Machine,Random Forest,and Classification and Regression Tree to classify the current status ...Based on the Google Earth Engine cloud computing data platform,this study employed three algorithms including Support Vector Machine,Random Forest,and Classification and Regression Tree to classify the current status of land covers in Hung Yen province of Vietnam using Landsat 8 OLI satellite images,a free data source with reasonable spatial and temporal resolution.The results of the study show that all three algorithms presented good classification for five basic types of land cover including Rice land,Water bodies,Perennial vegetation,Annual vegetation,Built-up areas as their overall accuracy and Kappa coefficient were greater than 80%and 0.8,respectively.Among the three algorithms,SVM achieved the highest accuracy as its overall accuracy was 86%and the Kappa coefficient was 0.88.Land cover classification based on the SVM algorithm shows that Built-up areas cover the largest area with nearly 31,495 ha,accounting for more than 33.8%of the total natural area,followed by Rice land and Perennial vegetation which cover an area of over 30,767 ha(33%)and 15,637 ha(16.8%),respectively.Water bodies and Annual vegetation cover the smallest areas with 8,820(9.5%)ha and 6,302 ha(6.8%),respectively.The results of this study can be used for land use management and planning as well as other natural resource and environmental management purposes in the province.展开更多
Neuromorphic computing extends beyond sequential processing modalities and outperforms traditional von Neumann architectures in implementing more complicated tasks,e.g.,pattern processing,image recognition,and decisio...Neuromorphic computing extends beyond sequential processing modalities and outperforms traditional von Neumann architectures in implementing more complicated tasks,e.g.,pattern processing,image recognition,and decision making.It features parallel interconnected neural networks,high fault tolerance,robustness,autonomous learning capability,and ultralow energy dissipation.The algorithms of artificial neural network(ANN)have also been widely used because of their facile self-organization and self-learning capabilities,which mimic those of the human brain.To some extent,ANN reflects several basic functions of the human brain and can be efficiently integrated into neuromorphic devices to perform neuromorphic computations.This review highlights recent advances in neuromorphic devices assisted by machine learning algorithms.First,the basic structure of simple neuron models inspired by biological neurons and the information processing in simple neural networks are particularly discussed.Second,the fabrication and research progress of neuromorphic devices are presented regarding to materials and structures.Furthermore,the fabrication of neuromorphic devices,including stand-alone neuromorphic devices,neuromorphic device arrays,and integrated neuromorphic systems,is discussed and demonstrated with reference to some respective studies.The applications of neuromorphic devices assisted by machine learning algorithms in different fields are categorized and investigated.Finally,perspectives,suggestions,and potential solutions to the current challenges of neuromorphic devices are provided.展开更多
The optimization of reaction processes is crucial for the green, efficient, and sustainable development of the chemical industry. However, how to address the problems posed by multiple variables, nonlinearities, and u...The optimization of reaction processes is crucial for the green, efficient, and sustainable development of the chemical industry. However, how to address the problems posed by multiple variables, nonlinearities, and uncertainties during optimization remains a formidable challenge. In this study, a strategy combining interpretable machine learning with metaheuristic optimization algorithms is employed to optimize the reaction process. First, experimental data from a biodiesel production process are collected to establish a database. These data are then used to construct a predictive model based on artificial neural network (ANN) models. Subsequently, interpretable machine learning techniques are applied for quantitative analysis and verification of the model. Finally, four metaheuristic optimization algorithms are coupled with the ANN model to achieve the desired optimization. The research results show that the methanol: palm fatty acid distillate (PFAD) molar ratio contributes the most to the reaction outcome, accounting for 41%. The ANN-simulated annealing (SA) hybrid method is more suitable for this optimization, and the optimal process parameters are a catalyst concentration of 3.00% (mass), a methanol: PFAD molar ratio of 8.67, and a reaction time of 30 min. This study provides deeper insights into reaction process optimization, which will facilitate future applications in various reaction optimization processes.展开更多
Background:In recent years,there has been a growing trend in the utilization of observational studies that make use of routinely collected healthcare data(RCD).These studies rely on algorithms to identify specific hea...Background:In recent years,there has been a growing trend in the utilization of observational studies that make use of routinely collected healthcare data(RCD).These studies rely on algorithms to identify specific health conditions(e.g.,diabetes or sepsis)for statistical analyses.However,there has been substantial variation in the algorithm development and validation,leading to frequently suboptimal performance and posing a significant threat to the validity of study findings.Unfortunately,these issues are often overlooked.Methods:We systematically developed guidance for the development,validation,and evaluation of algorithms designed to identify health status(DEVELOP-RCD).Our initial efforts involved conducting both a narrative review and a systematic review of published studies on the concepts and methodological issues related to algorithm development,validation,and evaluation.Subsequently,we conducted an empirical study on an algorithm for identifying sepsis.Based on these findings,we formulated specific workflow and recommendations for algorithm development,validation,and evaluation within the guidance.Finally,the guidance underwent independent review by a panel of 20 external experts who then convened a consensus meeting to finalize it.Results:A standardized workflow for algorithm development,validation,and evaluation was established.Guided by specific health status considerations,the workflow comprises four integrated steps:assessing an existing algorithm’s suitability for the target health status;developing a new algorithm using recommended methods;validating the algorithm using prescribed performance measures;and evaluating the impact of the algorithm on study results.Additionally,13 good practice recommendations were formulated with detailed explanations.Furthermore,a practical study on sepsis identification was included to demonstrate the application of this guidance.Conclusions:The establishment of guidance is intended to aid researchers and clinicians in the appropriate and accurate development and application of algorithms for identifying health status from RCD.This guidance has the potential to enhance the credibility of findings from observational studies involving RCD.展开更多
The Cross-domain Heuristic Search Challenge(CHeSC)is a competition focused on creating efficient search algorithms adaptable to diverse problem domains.Selection hyper-heuristics are a class of algorithms that dynamic...The Cross-domain Heuristic Search Challenge(CHeSC)is a competition focused on creating efficient search algorithms adaptable to diverse problem domains.Selection hyper-heuristics are a class of algorithms that dynamically choose heuristics during the search process.Numerous selection hyper-heuristics have different imple-mentation strategies.However,comparisons between them are lacking in the literature,and previous works have not highlighted the beneficial and detrimental implementation methods of different components.The question is how to effectively employ them to produce an efficient search heuristic.Furthermore,the algorithms that competed in the inaugural CHeSC have not been collectively reviewed.This work conducts a review analysis of the top twenty competitors from this competition to identify effective and ineffective strategies influencing algorithmic performance.A summary of the main characteristics and classification of the algorithms is presented.The analysis underlines efficient and inefficient methods in eight key components,including search points,search phases,heuristic selection,move acceptance,feedback,Tabu mechanism,restart mechanism,and low-level heuristic parameter control.This review analyzes the components referencing the competition’s final leaderboard and discusses future research directions for these components.The effective approaches,identified as having the highest quality index,are mixed search point,iterated search phases,relay hybridization selection,threshold acceptance,mixed learning,Tabu heuristics,stochastic restart,and dynamic parameters.Findings are also compared with recent trends in hyper-heuristics.This work enhances the understanding of selection hyper-heuristics,offering valuable insights for researchers and practitioners aiming to develop effective search algorithms for diverse problem domains.展开更多
This study examines the bidirectional shaping mechanism between short-video algorithms and film narratives within the attention economy.It investigates how algorithmic logic influences cinematic storytelling and how f...This study examines the bidirectional shaping mechanism between short-video algorithms and film narratives within the attention economy.It investigates how algorithmic logic influences cinematic storytelling and how films,in turn,contribute to the aesthetic enhancement of short-video content.Drawing on Communication Accommodation Theory and Berry’s Acculturation Theory,along with case analyses and industry data,this research demonstrates that algorithms push films toward high-stimulus,fast-paced narrative patterns—characterized by increased shot density and structural fragmentation—to capture and retain viewer attention.Conversely,films counter this influence by supplying narratively deep and artistically refined content that elevates short-video aesthetics and encourages critical audience engagement.This dynamic reflects a process of mutual adaptation rather than one-sided dominance.The study concludes that such interaction signifies a broader restructuring of cultural production logic,facilitating cross-media convergence while simultaneously posing risks to cultural diversity due to the prioritization of high-traffic content.Balancing this relationship will require policy support,algorithmic transparency,and strengthened industry self-regulation to preserve artistic integrity and cultural ecosystem diversity.展开更多
Natural Language Processing(NLP)has become essential in text classification,sentiment analysis,machine translation,and speech recognition applications.As these tasks become complex,traditionalmachine learning and deep...Natural Language Processing(NLP)has become essential in text classification,sentiment analysis,machine translation,and speech recognition applications.As these tasks become complex,traditionalmachine learning and deep learning models encounter challenges with optimization,parameter tuning,and handling large-scale,highdimensional data.Bio-inspired algorithms,which mimic natural processes,offer robust optimization capabilities that can enhance NLP performance by improving feature selection,optimizing model parameters,and integrating adaptive learning mechanisms.This review explores the state-of-the-art applications of bio-inspired algorithms—such as Genetic Algorithms(GA),Particle Swarm Optimization(PSO),and Ant Colony Optimization(ACO)—across core NLP tasks.We analyze their comparative advantages,discuss their integration with neural network models,and address computational and scalability limitations.Through a synthesis of existing research,this paper highlights the unique strengths and current challenges of bio-inspired approaches in NLP,offering insights into hybrid models and lightweight,resource-efficient adaptations for real-time processing.Finally,we outline future research directions that emphasize the development of scalable,effective bio-inspired methods adaptable to evolving data environments.展开更多
Quantum computing is a promising technology that has the potential to revolutionize many areas of science and technology,including communication.In this review,we discuss the current state of quantum computing in comm...Quantum computing is a promising technology that has the potential to revolutionize many areas of science and technology,including communication.In this review,we discuss the current state of quantum computing in communication and its potential applications in various areas such as network optimization,signal processing,and machine learning for communication.First,the basic principle of quantum computing,quantum physics systems,and quantum algorithms are analyzed.Then,based on the classification of quantum algorithms,several important basic quantum algorithms,quantum optimization algorithms,and quantum machine learning algorithms are discussed in detail.Finally,the basic ideas and feasibility of introducing quantum algorithms into communications are emphatically analyzed,which provides a reference to address computational bottlenecks in communication networks.展开更多
The excessive use of artificial intelligence(AI)algorithms has caused the problem of errors in AI algorithms,which has challenged the fairness of decision-making,and has intensified people’s inequality.Therefore,it i...The excessive use of artificial intelligence(AI)algorithms has caused the problem of errors in AI algorithms,which has challenged the fairness of decision-making,and has intensified people’s inequality.Therefore,it is necessary to conduct in-depth research and propose corresponding error detection and error elimination methods.This paper first proposes the root causes and threats of bias in AI algorithms,then summarizes the existing bias detection and error elimination methods,and proposes a bias processing framework in three-level dimensions of data,models,and conclusions,aiming to provide a framework for a comprehensive solution to errors in algorithms.At the same time,it also summarizes the problems and challenges in existing research and makes a prospect for future research trends.It is hoped that it will be helpful for us to build fairer AI.展开更多
Face recognition has emerged as one of the most prominent applications of image analysis and under-standing,gaining considerable attention in recent years.This growing interest is driven by two key factors:its extensi...Face recognition has emerged as one of the most prominent applications of image analysis and under-standing,gaining considerable attention in recent years.This growing interest is driven by two key factors:its extensive applications in law enforcement and the commercial domain,and the rapid advancement of practical technologies.Despite the significant advancements,modern recognition algorithms still struggle in real-world conditions such as varying lighting conditions,occlusion,and diverse facial postures.In such scenarios,human perception is still well above the capabilities of present technology.Using the systematic mapping study,this paper presents an in-depth review of face detection algorithms and face recognition algorithms,presenting a detailed survey of advancements made between 2015 and 2024.We analyze key methodologies,highlighting their strengths and restrictions in the application context.Additionally,we examine various datasets used for face detection/recognition datasets focusing on the task-specific applications,size,diversity,and complexity.By analyzing these algorithms and datasets,this survey works as a valuable resource for researchers,identifying the research gap in the field of face detection and recognition and outlining potential directions for future research.展开更多
Wind energy has emerged as a potential replacement for fossil fuel-based energy sources.To harness maximum wind energy,a crucial decision in the development of an efficient wind farm is the optimal layout design.This ...Wind energy has emerged as a potential replacement for fossil fuel-based energy sources.To harness maximum wind energy,a crucial decision in the development of an efficient wind farm is the optimal layout design.This layout defines the specific locations of the turbines within the wind farm.The process of finding the optimal locations of turbines,in the presence of various technical and technological constraints,makes the wind farm layout design problem a complex optimization problem.This problem has traditionally been solved with nature-inspired algorithms with promising results.The performance and convergence of nature-inspired algorithms depend on several parameters,among which the algorithm termination criterion plays a crucial role.Timely convergence is an important aspect of efficient algorithm design because an inefficient algorithm results in wasted computational resources,unwarranted electricity consumption,and hardware stress.This study provides an in-depth analysis of several termination criteria while using the genetic algorithm as a test bench,with its application to the wind farm layout design problem while considering various wind scenarios.The performance of six termination criteria is empirically evaluated with respect to the quality of solutions produced and the execution time involved.Due to the conflicting nature of these two attributes,fuzzy logic-based multi-attribute decision-making is employed in the decision process.Results for the fuzzy decision approach indicate that among the various criteria tested,the criterion Phi achieves an improvement in the range of 2.44%to 32.93%for wind scenario 1.For scenario 2,Best-worst termination criterion performed well compared to the other criteria evaluated,with an improvement in the range of 1.2%to 9.64%.For scenario 3,Hitting bound was the best performer with an improvement of 1.16%to 20.93%.展开更多
文摘Open caissons are widely used in foundation engineering because of their load-bearing efficiency and adaptability in diverse soil conditions.However,accurately predicting their undrained bearing capacity in layered soils remains a complex challenge.This study presents a novel application of five ensemble machine(ML)algorithms-random forest(RF),gradient boosting machine(GBM),extreme gradient boosting(XGBoost),adaptive boosting(AdaBoost),and categorical boosting(CatBoost)-to predict the undrained bearing capacity factor(Nc)of circular open caissons embedded in two-layered clay on the basis of results from finite element limit analysis(FELA).The input dataset consists of 1188 numerical simulations using the Tresca failure criterion,varying in geometrical and soil parameters.The FELA was performed via OptumG2 software with adaptive meshing techniques and verified against existing benchmark studies.The ML models were trained on 70% of the dataset and tested on the remaining 30%.Their performance was evaluated using six statistical metrics:coefficient of determination(R²),mean absolute error(MAE),root mean squared error(RMSE),index of scatter(IOS),RMSE-to-standard deviation ratio(RSR),and variance explained factor(VAF).The results indicate that all the models achieved high accuracy,with R²values exceeding 97.6%and RMSE values below 0.02.Among them,AdaBoost and CatBoost consistently outperformed the other methods across both the training and testing datasets,demonstrating superior generalizability and robustness.The proposed ML framework offers an efficient,accurate,and data-driven alternative to traditional methods for estimating caisson capacity in stratified soils.This approach can aid in reducing computational costs while improving reliability in the early stages of foundation design.
文摘Quantum computing offers unprecedented computational power, enabling simultaneous computations beyond traditional computers. Quantum computers differ significantly from classical computers, necessitating a distinct approach to algorithm design, which involves taming quantum mechanical phenomena. This paper extends the numbering of computable programs to be applied in the quantum computing context. Numbering computable programs is a theoretical computer science concept that assigns unique numbers to individual programs or algorithms. Common methods include Gödel numbering which encodes programs as strings of symbols or characters, often used in formal systems and mathematical logic. Based on the proposed numbering approach, this paper presents a mechanism to explore the set of possible quantum algorithms. The proposed approach is able to construct useful circuits such as Quantum Key Distribution BB84 protocol, which enables sender and receiver to establish a secure cryptographic key via a quantum channel. The proposed approach facilitates the process of exploring and constructing quantum algorithms.
文摘The advent of microgrids in modern energy systems heralds a promising era of resilience,sustainability,and efficiency.Within the realm of grid-tied microgrids,the selection of an optimal optimization algorithm is critical for effective energy management,particularly in economic dispatching.This study compares the performance of Particle Swarm Optimization(PSO)and Genetic Algorithms(GA)in microgrid energy management systems,implemented using MATLAB tools.Through a comprehensive review of the literature and sim-ulations conducted in MATLAB,the study analyzes performance metrics,convergence speed,and the overall efficacy of GA and PSO,with a focus on economic dispatching tasks.Notably,a significant distinction emerges between the cost curves generated by the two algo-rithms for microgrid operation,with the PSO algorithm consistently resulting in lower costs due to its effective economic dispatching capabilities.Specifically,the utilization of the PSO approach could potentially lead to substantial savings on the power bill,amounting to approximately$15.30 in this evaluation.Thefindings provide insights into the strengths and limitations of each algorithm within the complex dynamics of grid-tied microgrids,thereby assisting stakeholders and researchers in arriving at informed decisions.This study contributes to the discourse on sustainable energy management by offering actionable guidance for the advancement of grid-tied micro-grid technologies through MATLAB-implemented optimization algorithms.
基金National Natural Science Foundation of China(62373187)Forward-looking Layout Special Projects(ILA220591A22)。
文摘In the field of calculating the attack area of air-to-air missiles in modern air combat scenarios,the limitations of existing research,including real-time calculation,accuracy efficiency trade-off,and the absence of the three-dimensional attack area model,restrict their practical applications.To address these issues,an improved backtracking algorithm is proposed to improve calculation efficiency.A significant reduction in solution time and maintenance of accuracy in the three-dimensional attack area are achieved by using the proposed algorithm.Furthermore,the age-layered population structure genetic programming(ALPS-GP)algorithm is introduced to determine an analytical polynomial model of the three-dimensional attack area,considering real-time requirements.The accuracy of the polynomial model is enhanced through the coefficient correction using an improved gradient descent algorithm.The study reveals a remarkable combination of high accuracy and efficient real-time computation,with a mean error of 91.89 m using the analytical polynomial model of the three-dimensional attack area solved in just 10^(-4)s,thus meeting the requirements of real-time combat scenarios.
文摘This study presents a novel hybrid topology optimization and mold design framework that integrates process fitting,runner system optimization,and structural analysis to significantly enhance the performance of injection-molded parts.At its core,the framework employs a greedy algorithm that generates runner systems based on adjacency and shortest path principles,leading to improvements in both mechanical strength and material efficiency.The design optimization is validated through a series of rigorous experimental tests,including three-point bending and torsion tests performed on key-socket frames,ensuring that the optimized designs meet practical performance requirements.A critical innovation of the framework is the development of the Adjacent Element Temperature-Driven Prestress Algorithm(AETDPA),which refines the prediction of mechanical failure and strength fitting.This algorithm has been shown to deliver mesh-independent accuracy,thereby enhancing the reliability of simulation results across various design iterations.The framework’s adaptability is further demonstrated by its ability to adjust optimization methods based on the unique geometry of each part,thus accelerating the overall design process while ensuring struc-tural integrity.In addition to its immediate applications in injection molding,the study explores the potential extension of this framework to metal additive manufacturing,opening new avenues for its use in advanced manufacturing technologies.Numerical simulations,including finite element analysis,support the experimental findings and confirm that the optimized designs provide a balanced combination of strength,durability,and efficiency.Furthermore,the integration challenges with existing injection molding practices are addressed,underscoring the framework’s scalability and industrial relevance.Overall,this hybrid topology optimization framework offers a computationally efficient and robust solution for advanced manufacturing applications,promising significant improvements in design efficiency,cost-effectiveness,and product performance.Future work will focus on further enhancing algorithm robustness and exploring additional applications across diverse manufacturing processes.
文摘The application of machine learning was investigated for predicting end-point temperature in the basic oxygen furnace steelmaking process,addressing gaps in the field,particularly large-scale dataset sizes and the underutilization of boosting algorithms.Utilizing a substantial dataset containing over 20,000 heats,significantly bigger than those in previous studies,a comprehensive evaluation of five advanced machine learning models was conducted.These include four ensemble learning algorithms:XGBoost,LightGBM,CatBoost(three boosting algorithms),along with random forest(a bagging algorithm),as well as a neural network model,namely the multilayer perceptron.Our comparative analysis reveals that Bayesian-optimized boosting models demonstrate exceptional robustness and accuracy,achieving the highest R-squared values,the lowest root mean square error,and lowest mean absolute error,along with the best hit ratio.CatBoost exhibited superior performance,with its test R-squared improving by 4.2%compared to that of the random forest and by 0.8%compared to that of the multilayer perceptron.This highlights the efficacy of boosting algorithms in refining complex industrial processes.Additionally,our investigation into the impact of varying dataset sizes,ranging from 500 to 20,000 heats,on model accuracy underscores the importance of leveraging larger-scale datasets to improve the accuracy and stability of predictive models.
基金funded by Universiti Putra Malaysia under a Geran Putra Inisiatif(GPI)research grant with reference to GP-GPI/2023/9762100.
文摘This study proposes a novel time-synchronization protocol inspired by stochastic gradient algorithms.The clock model of each network node in this synchronizer is configured as a generic adaptive filter where different stochastic gradient algorithms can be adopted for adaptive clock frequency adjustments.The study analyzes the pairwise synchronization behavior of the protocol and proves the generalized convergence of the synchronization error and clock frequency.A novel closed-form expression is also derived for a generalized asymptotic error variance steady state.Steady and convergence analyses are then presented for the synchronization,with frequency adaptations done using least mean square(LMS),the Newton search,the gradient descent(GraDes),the normalized LMS(N-LMS),and the Sign-Data LMS algorithms.Results obtained from real-time experiments showed a better performance of our protocols as compared to the Average Proportional-Integral Synchronization Protocol(AvgPISync)regarding the impact of quantization error on synchronization accuracy,precision,and convergence time.This generalized approach to time synchronization allows flexibility in selecting a suitable protocol for different wireless sensor network applications.
基金supported by the National Natural Science Foundation of China(Nos.12475174 and 12175101)Yue Lu Shan Center Industrial Innovation(No.2024YCII0108)。
文摘In recent years,the development of new types of nuclear reactors,such as transportable,marine,and space reactors,has presented new challenges for the optimization of reactor radiation-shielding design.Shielding structures typically need to be lightweight,miniaturized,and radiation-protected,which is a multi-parameter and multi-objective optimization problem.The conventional multi-objective(two or three objectives)optimization method for radiation-shielding design exhibits limitations for a number of optimization objectives and variable parameters,as well as a deficiency in achieving a global optimal solution,thereby failing to meet the requirements of shielding optimization for newly developed reactors.In this study,genetic and artificial bee-colony algorithms are combined with a reference-point-selection strategy and applied to the many-objective(having four or more objectives)optimal design of reactor radiation shielding.To validate the reliability of the methods,an optimization simulation is conducted on three-dimensional shielding structures and another complicated shielding-optimization problem.The numerical results demonstrate that the proposed algorithms outperform conventional shielding-design methods in terms of optimization performance,and they exhibit their reliability in practical engineering problems.The many-objective optimization algorithms developed in this study are proven to efficiently and consistently search for Pareto-front shielding schemes.Therefore,the algorithms proposed in this study offer novel insights into improving the shielding-design performance and shielding quality of new reactor types.
文摘This paper studies polygon simplification algorithms for 3D models,focuses on the optimization algorithm of quadratic error metric(QEM),explores the impacts of different methods on the simplification of different models,and develops a web-based visualization application.Metrics such as the Hausdorff distance are used to evaluate the balance between the degree of simplification and the retention of model details.
文摘Based on the Google Earth Engine cloud computing data platform,this study employed three algorithms including Support Vector Machine,Random Forest,and Classification and Regression Tree to classify the current status of land covers in Hung Yen province of Vietnam using Landsat 8 OLI satellite images,a free data source with reasonable spatial and temporal resolution.The results of the study show that all three algorithms presented good classification for five basic types of land cover including Rice land,Water bodies,Perennial vegetation,Annual vegetation,Built-up areas as their overall accuracy and Kappa coefficient were greater than 80%and 0.8,respectively.Among the three algorithms,SVM achieved the highest accuracy as its overall accuracy was 86%and the Kappa coefficient was 0.88.Land cover classification based on the SVM algorithm shows that Built-up areas cover the largest area with nearly 31,495 ha,accounting for more than 33.8%of the total natural area,followed by Rice land and Perennial vegetation which cover an area of over 30,767 ha(33%)and 15,637 ha(16.8%),respectively.Water bodies and Annual vegetation cover the smallest areas with 8,820(9.5%)ha and 6,302 ha(6.8%),respectively.The results of this study can be used for land use management and planning as well as other natural resource and environmental management purposes in the province.
基金financially supported by the National Natural Science Foundation of China(No.52073031)the National Key Research and Development Program of China(Nos.2023YFB3208102,2021YFB3200304)+4 种基金the China National Postdoctoral Program for Innovative Talents(No.BX2021302)the Beijing Nova Program(Nos.Z191100001119047,Z211100002121148)the Fundamental Research Funds for the Central Universities(No.E0EG6801X2)the‘Hundred Talents Program’of the Chinese Academy of Sciencesthe BrainLink program funded by the MSIT through the NRF of Korea(No.RS-2023-00237308).
文摘Neuromorphic computing extends beyond sequential processing modalities and outperforms traditional von Neumann architectures in implementing more complicated tasks,e.g.,pattern processing,image recognition,and decision making.It features parallel interconnected neural networks,high fault tolerance,robustness,autonomous learning capability,and ultralow energy dissipation.The algorithms of artificial neural network(ANN)have also been widely used because of their facile self-organization and self-learning capabilities,which mimic those of the human brain.To some extent,ANN reflects several basic functions of the human brain and can be efficiently integrated into neuromorphic devices to perform neuromorphic computations.This review highlights recent advances in neuromorphic devices assisted by machine learning algorithms.First,the basic structure of simple neuron models inspired by biological neurons and the information processing in simple neural networks are particularly discussed.Second,the fabrication and research progress of neuromorphic devices are presented regarding to materials and structures.Furthermore,the fabrication of neuromorphic devices,including stand-alone neuromorphic devices,neuromorphic device arrays,and integrated neuromorphic systems,is discussed and demonstrated with reference to some respective studies.The applications of neuromorphic devices assisted by machine learning algorithms in different fields are categorized and investigated.Finally,perspectives,suggestions,and potential solutions to the current challenges of neuromorphic devices are provided.
基金supported by the National Natural Science Foundation of China(22408227,22238005)the Postdoctoral Research Foundation of China(GZC20231576).
文摘The optimization of reaction processes is crucial for the green, efficient, and sustainable development of the chemical industry. However, how to address the problems posed by multiple variables, nonlinearities, and uncertainties during optimization remains a formidable challenge. In this study, a strategy combining interpretable machine learning with metaheuristic optimization algorithms is employed to optimize the reaction process. First, experimental data from a biodiesel production process are collected to establish a database. These data are then used to construct a predictive model based on artificial neural network (ANN) models. Subsequently, interpretable machine learning techniques are applied for quantitative analysis and verification of the model. Finally, four metaheuristic optimization algorithms are coupled with the ANN model to achieve the desired optimization. The research results show that the methanol: palm fatty acid distillate (PFAD) molar ratio contributes the most to the reaction outcome, accounting for 41%. The ANN-simulated annealing (SA) hybrid method is more suitable for this optimization, and the optimal process parameters are a catalyst concentration of 3.00% (mass), a methanol: PFAD molar ratio of 8.67, and a reaction time of 30 min. This study provides deeper insights into reaction process optimization, which will facilitate future applications in various reaction optimization processes.
基金supported by the National Natural Science Foundation of China(82225049,72104155)the Sichuan Provincial Central Government Guides Local Science and Technology Development Special Project(2022ZYD0127)the 1·3·5 Project for Disciplines of Excellence,West China Hospital,Sichuan University(ZYGD23004).
文摘Background:In recent years,there has been a growing trend in the utilization of observational studies that make use of routinely collected healthcare data(RCD).These studies rely on algorithms to identify specific health conditions(e.g.,diabetes or sepsis)for statistical analyses.However,there has been substantial variation in the algorithm development and validation,leading to frequently suboptimal performance and posing a significant threat to the validity of study findings.Unfortunately,these issues are often overlooked.Methods:We systematically developed guidance for the development,validation,and evaluation of algorithms designed to identify health status(DEVELOP-RCD).Our initial efforts involved conducting both a narrative review and a systematic review of published studies on the concepts and methodological issues related to algorithm development,validation,and evaluation.Subsequently,we conducted an empirical study on an algorithm for identifying sepsis.Based on these findings,we formulated specific workflow and recommendations for algorithm development,validation,and evaluation within the guidance.Finally,the guidance underwent independent review by a panel of 20 external experts who then convened a consensus meeting to finalize it.Results:A standardized workflow for algorithm development,validation,and evaluation was established.Guided by specific health status considerations,the workflow comprises four integrated steps:assessing an existing algorithm’s suitability for the target health status;developing a new algorithm using recommended methods;validating the algorithm using prescribed performance measures;and evaluating the impact of the algorithm on study results.Additionally,13 good practice recommendations were formulated with detailed explanations.Furthermore,a practical study on sepsis identification was included to demonstrate the application of this guidance.Conclusions:The establishment of guidance is intended to aid researchers and clinicians in the appropriate and accurate development and application of algorithms for identifying health status from RCD.This guidance has the potential to enhance the credibility of findings from observational studies involving RCD.
基金funded by Ministry of Higher Education(MoHE)Malaysia,under Transdisciplinary Research Grant Scheme(TRGS/1/2019/UKM/01/4/2).
文摘The Cross-domain Heuristic Search Challenge(CHeSC)is a competition focused on creating efficient search algorithms adaptable to diverse problem domains.Selection hyper-heuristics are a class of algorithms that dynamically choose heuristics during the search process.Numerous selection hyper-heuristics have different imple-mentation strategies.However,comparisons between them are lacking in the literature,and previous works have not highlighted the beneficial and detrimental implementation methods of different components.The question is how to effectively employ them to produce an efficient search heuristic.Furthermore,the algorithms that competed in the inaugural CHeSC have not been collectively reviewed.This work conducts a review analysis of the top twenty competitors from this competition to identify effective and ineffective strategies influencing algorithmic performance.A summary of the main characteristics and classification of the algorithms is presented.The analysis underlines efficient and inefficient methods in eight key components,including search points,search phases,heuristic selection,move acceptance,feedback,Tabu mechanism,restart mechanism,and low-level heuristic parameter control.This review analyzes the components referencing the competition’s final leaderboard and discusses future research directions for these components.The effective approaches,identified as having the highest quality index,are mixed search point,iterated search phases,relay hybridization selection,threshold acceptance,mixed learning,Tabu heuristics,stochastic restart,and dynamic parameters.Findings are also compared with recent trends in hyper-heuristics.This work enhances the understanding of selection hyper-heuristics,offering valuable insights for researchers and practitioners aiming to develop effective search algorithms for diverse problem domains.
文摘This study examines the bidirectional shaping mechanism between short-video algorithms and film narratives within the attention economy.It investigates how algorithmic logic influences cinematic storytelling and how films,in turn,contribute to the aesthetic enhancement of short-video content.Drawing on Communication Accommodation Theory and Berry’s Acculturation Theory,along with case analyses and industry data,this research demonstrates that algorithms push films toward high-stimulus,fast-paced narrative patterns—characterized by increased shot density and structural fragmentation—to capture and retain viewer attention.Conversely,films counter this influence by supplying narratively deep and artistically refined content that elevates short-video aesthetics and encourages critical audience engagement.This dynamic reflects a process of mutual adaptation rather than one-sided dominance.The study concludes that such interaction signifies a broader restructuring of cultural production logic,facilitating cross-media convergence while simultaneously posing risks to cultural diversity due to the prioritization of high-traffic content.Balancing this relationship will require policy support,algorithmic transparency,and strengthened industry self-regulation to preserve artistic integrity and cultural ecosystem diversity.
基金supported by AIT Laboratory,FPT University,Danang Campus,Vietnam,2024.
文摘Natural Language Processing(NLP)has become essential in text classification,sentiment analysis,machine translation,and speech recognition applications.As these tasks become complex,traditionalmachine learning and deep learning models encounter challenges with optimization,parameter tuning,and handling large-scale,highdimensional data.Bio-inspired algorithms,which mimic natural processes,offer robust optimization capabilities that can enhance NLP performance by improving feature selection,optimizing model parameters,and integrating adaptive learning mechanisms.This review explores the state-of-the-art applications of bio-inspired algorithms—such as Genetic Algorithms(GA),Particle Swarm Optimization(PSO),and Ant Colony Optimization(ACO)—across core NLP tasks.We analyze their comparative advantages,discuss their integration with neural network models,and address computational and scalability limitations.Through a synthesis of existing research,this paper highlights the unique strengths and current challenges of bio-inspired approaches in NLP,offering insights into hybrid models and lightweight,resource-efficient adaptations for real-time processing.Finally,we outline future research directions that emphasize the development of scalable,effective bio-inspired methods adaptable to evolving data environments.
文摘Quantum computing is a promising technology that has the potential to revolutionize many areas of science and technology,including communication.In this review,we discuss the current state of quantum computing in communication and its potential applications in various areas such as network optimization,signal processing,and machine learning for communication.First,the basic principle of quantum computing,quantum physics systems,and quantum algorithms are analyzed.Then,based on the classification of quantum algorithms,several important basic quantum algorithms,quantum optimization algorithms,and quantum machine learning algorithms are discussed in detail.Finally,the basic ideas and feasibility of introducing quantum algorithms into communications are emphatically analyzed,which provides a reference to address computational bottlenecks in communication networks.
文摘The excessive use of artificial intelligence(AI)algorithms has caused the problem of errors in AI algorithms,which has challenged the fairness of decision-making,and has intensified people’s inequality.Therefore,it is necessary to conduct in-depth research and propose corresponding error detection and error elimination methods.This paper first proposes the root causes and threats of bias in AI algorithms,then summarizes the existing bias detection and error elimination methods,and proposes a bias processing framework in three-level dimensions of data,models,and conclusions,aiming to provide a framework for a comprehensive solution to errors in algorithms.At the same time,it also summarizes the problems and challenges in existing research and makes a prospect for future research trends.It is hoped that it will be helpful for us to build fairer AI.
文摘Face recognition has emerged as one of the most prominent applications of image analysis and under-standing,gaining considerable attention in recent years.This growing interest is driven by two key factors:its extensive applications in law enforcement and the commercial domain,and the rapid advancement of practical technologies.Despite the significant advancements,modern recognition algorithms still struggle in real-world conditions such as varying lighting conditions,occlusion,and diverse facial postures.In such scenarios,human perception is still well above the capabilities of present technology.Using the systematic mapping study,this paper presents an in-depth review of face detection algorithms and face recognition algorithms,presenting a detailed survey of advancements made between 2015 and 2024.We analyze key methodologies,highlighting their strengths and restrictions in the application context.Additionally,we examine various datasets used for face detection/recognition datasets focusing on the task-specific applications,size,diversity,and complexity.By analyzing these algorithms and datasets,this survey works as a valuable resource for researchers,identifying the research gap in the field of face detection and recognition and outlining potential directions for future research.
基金funded by King Fahd University of Petroleum&Minerals,Saudi Arabia under IRC-SES grant#INRE 2217.
文摘Wind energy has emerged as a potential replacement for fossil fuel-based energy sources.To harness maximum wind energy,a crucial decision in the development of an efficient wind farm is the optimal layout design.This layout defines the specific locations of the turbines within the wind farm.The process of finding the optimal locations of turbines,in the presence of various technical and technological constraints,makes the wind farm layout design problem a complex optimization problem.This problem has traditionally been solved with nature-inspired algorithms with promising results.The performance and convergence of nature-inspired algorithms depend on several parameters,among which the algorithm termination criterion plays a crucial role.Timely convergence is an important aspect of efficient algorithm design because an inefficient algorithm results in wasted computational resources,unwarranted electricity consumption,and hardware stress.This study provides an in-depth analysis of several termination criteria while using the genetic algorithm as a test bench,with its application to the wind farm layout design problem while considering various wind scenarios.The performance of six termination criteria is empirically evaluated with respect to the quality of solutions produced and the execution time involved.Due to the conflicting nature of these two attributes,fuzzy logic-based multi-attribute decision-making is employed in the decision process.Results for the fuzzy decision approach indicate that among the various criteria tested,the criterion Phi achieves an improvement in the range of 2.44%to 32.93%for wind scenario 1.For scenario 2,Best-worst termination criterion performed well compared to the other criteria evaluated,with an improvement in the range of 1.2%to 9.64%.For scenario 3,Hitting bound was the best performer with an improvement of 1.16%to 20.93%.