The problem of collision avoidance for non-cooperative targets has received significant attention from researchers in recent years.Non-cooperative targets exhibit uncertain states and unpredictable behaviors,making co...The problem of collision avoidance for non-cooperative targets has received significant attention from researchers in recent years.Non-cooperative targets exhibit uncertain states and unpredictable behaviors,making collision avoidance significantly more challenging than that for space debris.Much existing research focuses on the continuous thrust model,whereas the impulsive maneuver model is more appropriate for long-duration and long-distance avoidance missions.Additionally,it is important to minimize the impact on the original mission while avoiding noncooperative targets.On the other hand,the existing avoidance algorithms are computationally complex and time-consuming especially with the limited computing capability of the on-board computer,posing challenges for practical engineering applications.To conquer these difficulties,this paper makes the following key contributions:(A)a turn-based(sequential decision-making)limited-area impulsive collision avoidance model considering the time delay of precision orbit determination is established for the first time;(B)a novel Selection Probability Learning Adaptive Search-depth Search Tree(SPL-ASST)algorithm is proposed for non-cooperative target avoidance,which improves the decision-making efficiency by introducing an adaptive-search-depth mechanism and a neural network into the traditional Monte Carlo Tree Search(MCTS).Numerical simulations confirm the effectiveness and efficiency of the proposed method.展开更多
The Bat algorithm,a metaheuristic optimization technique inspired by the foraging behaviour of bats,has been employed to tackle optimization problems.Known for its ease of implementation,parameter tunability,and stron...The Bat algorithm,a metaheuristic optimization technique inspired by the foraging behaviour of bats,has been employed to tackle optimization problems.Known for its ease of implementation,parameter tunability,and strong global search capabilities,this algorithm finds application across diverse optimization problem domains.However,in the face of increasingly complex optimization challenges,the Bat algorithm encounters certain limitations,such as slow convergence and sensitivity to initial solutions.In order to tackle these challenges,the present study incorporates a range of optimization compo-nents into the Bat algorithm,thereby proposing a variant called PKEBA.A projection screening strategy is implemented to mitigate its sensitivity to initial solutions,thereby enhancing the quality of the initial solution set.A kinetic adaptation strategy reforms exploration patterns,while an elite communication strategy enhances group interaction,to avoid algorithm from local optima.Subsequently,the effectiveness of the proposed PKEBA is rigorously evaluated.Testing encompasses 30 benchmark functions from IEEE CEC2014,featuring ablation experiments and comparative assessments against classical algorithms and their variants.Moreover,real-world engineering problems are employed as further validation.The results conclusively demonstrate that PKEBA ex-hibits superior convergence and precision compared to existing algorithms.展开更多
The Internet of Things(IoT)has emerged as an important future technology.IoT-Fog is a new computing paradigm that processes IoT data on servers close to the source of the data.In IoT-Fog computing,resource allocation ...The Internet of Things(IoT)has emerged as an important future technology.IoT-Fog is a new computing paradigm that processes IoT data on servers close to the source of the data.In IoT-Fog computing,resource allocation and independent task scheduling aim to deliver short response time services demanded by the IoT devices and performed by fog servers.The heterogeneity of the IoT-Fog resources and the huge amount of data that needs to be processed by the IoT-Fog tasks make scheduling fog computing tasks a challenging problem.This study proposes an Adaptive Firefly Algorithm(AFA)for dependent task scheduling in IoT-Fog computing.The proposed AFA is a modified version of the standard Firefly Algorithm(FA),considering the execution times of the submitted tasks,the impact of synchronization requirements,and the communication time between dependent tasks.As IoT-Fog computing depends mainly on distributed fog node servers that receive tasks in a dynamic manner,tackling the communications and synchronization issues between dependent tasks is becoming a challenging problem.The proposed AFA aims to address the dynamic nature of IoT-Fog computing environments.The proposed AFA mechanism considers a dynamic light absorption coefficient to control the decrease in attractiveness over iterations.The proposed AFA mechanism performance was benchmarked against the standard Firefly Algorithm(FA),Puma Optimizer(PO),Genetic Algorithm(GA),and Ant Colony Optimization(ACO)through simulations under light,typical,and heavy workload scenarios.In heavy workloads,the proposed AFA mechanism obtained the shortest average execution time,968.98 ms compared to 970.96,1352.87,1247.28,and 1773.62 of FA,PO,GA,and ACO,respectively.The simulation results demonstrate the proposed AFA’s ability to rapidly converge to optimal solutions,emphasizing its adaptability and efficiency in typical and heavy workloads.展开更多
Data clustering is an essential technique for analyzing complex datasets and continues to be a central research topic in data analysis.Traditional clustering algorithms,such as K-means,are widely used due to their sim...Data clustering is an essential technique for analyzing complex datasets and continues to be a central research topic in data analysis.Traditional clustering algorithms,such as K-means,are widely used due to their simplicity and efficiency.This paper proposes a novel Spiral Mechanism-Optimized Phasmatodea Population Evolution Algorithm(SPPE)to improve clustering performance.The SPPE algorithm introduces several enhancements to the standard Phasmatodea Population Evolution(PPE)algorithm.Firstly,a Variable Neighborhood Search(VNS)factor is incorporated to strengthen the local search capability and foster population diversity.Secondly,a position update model,incorporating a spiral mechanism,is designed to improve the algorithm’s global exploration and convergence speed.Finally,a dynamic balancing factor,guided by fitness values,adjusts the search process to balance exploration and exploitation effectively.The performance of SPPE is first validated on CEC2013 benchmark functions,where it demonstrates excellent convergence speed and superior optimization results compared to several state-of-the-art metaheuristic algorithms.To further verify its practical applicability,SPPE is combined with the K-means algorithm for data clustering and tested on seven datasets.Experimental results show that SPPE-K-means improves clustering accuracy,reduces dependency on initialization,and outperforms other clustering approaches.This study highlights SPPE’s robustness and efficiency in solving both optimization and clustering challenges,making it a promising tool for complex data analysis tasks.展开更多
The bat algorithm(BA)is a metaheuristic algorithm for global optimisation that simulates the echolocation behaviour of bats with varying pulse rates of emission and loudness,which can be used to find the globally opti...The bat algorithm(BA)is a metaheuristic algorithm for global optimisation that simulates the echolocation behaviour of bats with varying pulse rates of emission and loudness,which can be used to find the globally optimal solutions for various optimisation problems.Knowing the recent criticises of the originality of equations,the principle of BA is concise and easy to implement,and its mathematical structure can be seen as a hybrid particle swarm with simulated annealing.In this research,the authors focus on the performance optimisation of BA as a solver rather than discussing its originality issues.In terms of operation effect,BA has an acceptable convergence speed.However,due to the low proportion of time used to explore the search space,it is easy to converge prematurely and fall into the local optima.The authors propose an adaptive multi-stage bat algorithm(AMSBA).By tuning the algorithm's focus at three different stages of the search process,AMSBA can achieve a better balance between exploration and exploitation and improve its exploration ability by enhancing its performance in escaping local optima as well as maintaining a certain convergence speed.Therefore,AMSBA can achieve solutions with better quality.A convergence analysis was conducted to demonstrate the global convergence of AMSBA.The authors also perform simulation experiments on 30 benchmark functions from IEEE CEC 2017 as the objective functions and compare AMSBA with some original and improved swarm-based algorithms.The results verify the effectiveness and superiority of AMSBA.AMSBA is also compared with eight representative optimisation algorithms on 10 benchmark functions derived from IEEE CEC 2020,while this experiment is carried out on five different dimensions of the objective functions respectively.A balance and diversity analysis was performed on AMSBA to demonstrate its improvement over the original BA in terms of balance.AMSBA was also applied to the multi-threshold image segmentation of Citrus Macular disease,which is a bacterial infection that causes lesions on citrus trees.The segmentation results were analysed by comparing each comparative algorithm's peak signal-to-noise ratio,structural similarity index and feature similarity index.The results show that the proposed BA-based algorithm has apparent advantages,and it can effectively segment the disease spots from citrus leaves when the segmentation threshold is at a low level.Based on a comprehensive study,the authors think the proposed optimiser has mitigated the main drawbacks of the BA,and it can be utilised as an effective optimisation tool.展开更多
Cloud computing has become an essential technology for the management and processing of large datasets,offering scalability,high availability,and fault tolerance.However,optimizing data replication across multiple dat...Cloud computing has become an essential technology for the management and processing of large datasets,offering scalability,high availability,and fault tolerance.However,optimizing data replication across multiple data centers poses a significant challenge,especially when balancing opposing goals such as latency,storage costs,energy consumption,and network efficiency.This study introduces a novel Dynamic Optimization Algorithm called Dynamic Multi-Objective Gannet Optimization(DMGO),designed to enhance data replication efficiency in cloud environments.Unlike traditional static replication systems,DMGO adapts dynamically to variations in network conditions,system demand,and resource availability.The approach utilizes multi-objective optimization approaches to efficiently balance data access latency,storage efficiency,and operational costs.DMGO consistently evaluates data center performance and adjusts replication algorithms in real time to guarantee optimal system efficiency.Experimental evaluations conducted in a simulated cloud environment demonstrate that DMGO significantly outperforms conventional static algorithms,achieving faster data access,lower storage overhead,reduced energy consumption,and improved scalability.The proposed methodology offers a robust and adaptable solution for modern cloud systems,ensuring efficient resource consumption while maintaining high performance.展开更多
This paper examines the impact of algorithmic recommendations and data-driven marketing on consumer engagement and business performance.By leveraging large volumes of user data,businesses can deliver personalized cont...This paper examines the impact of algorithmic recommendations and data-driven marketing on consumer engagement and business performance.By leveraging large volumes of user data,businesses can deliver personalized content that enhances user experiences and increases conversion rates.However,the growing reliance on these technologies introduces significant risks,including privacy violations,algorithmic bias,and ethical concerns.This paper explores these challenges and provides recommendations for businesses to mitigate associated risks while optimizing marketing strategies.It highlights the importance of transparency,fairness,and user control in ensuring responsible and effective data-driven marketing.展开更多
The application of machine learning was investigated for predicting end-point temperature in the basic oxygen furnace steelmaking process,addressing gaps in the field,particularly large-scale dataset sizes and the und...The application of machine learning was investigated for predicting end-point temperature in the basic oxygen furnace steelmaking process,addressing gaps in the field,particularly large-scale dataset sizes and the underutilization of boosting algorithms.Utilizing a substantial dataset containing over 20,000 heats,significantly bigger than those in previous studies,a comprehensive evaluation of five advanced machine learning models was conducted.These include four ensemble learning algorithms:XGBoost,LightGBM,CatBoost(three boosting algorithms),along with random forest(a bagging algorithm),as well as a neural network model,namely the multilayer perceptron.Our comparative analysis reveals that Bayesian-optimized boosting models demonstrate exceptional robustness and accuracy,achieving the highest R-squared values,the lowest root mean square error,and lowest mean absolute error,along with the best hit ratio.CatBoost exhibited superior performance,with its test R-squared improving by 4.2%compared to that of the random forest and by 0.8%compared to that of the multilayer perceptron.This highlights the efficacy of boosting algorithms in refining complex industrial processes.Additionally,our investigation into the impact of varying dataset sizes,ranging from 500 to 20,000 heats,on model accuracy underscores the importance of leveraging larger-scale datasets to improve the accuracy and stability of predictive models.展开更多
When dealing with imbalanced datasets,the traditional support vectormachine(SVM)tends to produce a classification hyperplane that is biased towards the majority class,which exhibits poor robustness.This paper proposes...When dealing with imbalanced datasets,the traditional support vectormachine(SVM)tends to produce a classification hyperplane that is biased towards the majority class,which exhibits poor robustness.This paper proposes a high-performance classification algorithm specifically designed for imbalanced datasets.The proposed method first uses a biased second-order cone programming support vectormachine(B-SOCP-SVM)to identify the support vectors(SVs)and non-support vectors(NSVs)in the imbalanced data.Then,it applies the synthetic minority over-sampling technique(SV-SMOTE)to oversample the support vectors of the minority class and uses the random under-sampling technique(NSV-RUS)multiple times to undersample the non-support vectors of the majority class.Combining the above-obtained minority class data set withmultiple majority class datasets can obtainmultiple new balanced data sets.Finally,SOCP-SVM is used to classify each data set,and the final result is obtained through the integrated algorithm.Experimental results demonstrate that the proposed method performs excellently on imbalanced datasets.展开更多
This study proposes a novel time-synchronization protocol inspired by stochastic gradient algorithms.The clock model of each network node in this synchronizer is configured as a generic adaptive filter where different...This study proposes a novel time-synchronization protocol inspired by stochastic gradient algorithms.The clock model of each network node in this synchronizer is configured as a generic adaptive filter where different stochastic gradient algorithms can be adopted for adaptive clock frequency adjustments.The study analyzes the pairwise synchronization behavior of the protocol and proves the generalized convergence of the synchronization error and clock frequency.A novel closed-form expression is also derived for a generalized asymptotic error variance steady state.Steady and convergence analyses are then presented for the synchronization,with frequency adaptations done using least mean square(LMS),the Newton search,the gradient descent(GraDes),the normalized LMS(N-LMS),and the Sign-Data LMS algorithms.Results obtained from real-time experiments showed a better performance of our protocols as compared to the Average Proportional-Integral Synchronization Protocol(AvgPISync)regarding the impact of quantization error on synchronization accuracy,precision,and convergence time.This generalized approach to time synchronization allows flexibility in selecting a suitable protocol for different wireless sensor network applications.展开更多
DNA microarrays, a cornerstone in biomedicine, measure gene expression across thousands to tens of thousands of genes. Identifying the genes vital for accurate cancer classification is a key challenge. Here, we presen...DNA microarrays, a cornerstone in biomedicine, measure gene expression across thousands to tens of thousands of genes. Identifying the genes vital for accurate cancer classification is a key challenge. Here, we present Fs-LSA (F-score based Learning Search Algorithm), a novel gene selection algorithm designed to enhance the precision and efficiency of target gene identification from microarray data for cancer classification. This algorithm is divided into two phases: the first leverages F-score values to prioritize and select feature genes with the most significant differential expression;the second phase introduces our Learning Search Algorithm (LSA), which harnesses swarm intelligence to identify the optimal subset among the remaining genes. Inspired by human social learning, LSA integrates historical data and collective intelligence for a thorough search, with a dynamic control mechanism that balances exploration and refinement, thereby enhancing the gene selection process. We conducted a rigorous validation of Fs-LSA’s performance using eight publicly available cancer microarray expression datasets. Fs-LSA achieved accuracy, precision, sensitivity, and F1-score values of 0.9932, 0.9923, 0.9962, and 0.994, respectively. Comparative analyses with state-of-the-art algorithms revealed Fs-LSA’s superior performance in terms of simplicity and efficiency. Additionally, we validated the algorithm’s efficacy independently using glioblastoma data from GEO and TCGA databases. It was significantly superior to those of the comparison algorithms. Importantly, the driver genes identified by Fs-LSA were instrumental in developing a predictive model as an independent prognostic indicator for glioblastoma, underscoring Fs-LSA’s transformative potential in genomics and personalized medicine.展开更多
With the rapid advancement of medical artificial intelligence(AI)technology,particularly the widespread adoption of AI diagnostic systems,ethical challenges in medical decision-making have garnered increasing attentio...With the rapid advancement of medical artificial intelligence(AI)technology,particularly the widespread adoption of AI diagnostic systems,ethical challenges in medical decision-making have garnered increasing attention.This paper analyzes the limitations of algorithmic ethics in medical decision-making and explores accountability mechanisms,aiming to provide theoretical support for ethically informed medical practices.The study highlights how the opacity of AI algorithms complicates the definition of decision-making responsibility,undermines doctor-patient trust,and affects informed consent.By thoroughly investigating issues such as the algorithmic“black box”problem and data privacy protection,we develop accountability assessment models to address ethical concerns related to medical resource allocation.Furthermore,this research examines the effective implementation of AI diagnostic systems through case studies of both successful and unsuccessful applications,extracting lessons on accountability mechanisms and response strategies.Finally,we emphasize that establishing a transparent accountability framework is crucial for enhancing the ethical standards of medical AI systems and protecting patients’rights and interests.展开更多
Background:In recent years,there has been a growing trend in the utilization of observational studies that make use of routinely collected healthcare data(RCD).These studies rely on algorithms to identify specific hea...Background:In recent years,there has been a growing trend in the utilization of observational studies that make use of routinely collected healthcare data(RCD).These studies rely on algorithms to identify specific health conditions(e.g.,diabetes or sepsis)for statistical analyses.However,there has been substantial variation in the algorithm development and validation,leading to frequently suboptimal performance and posing a significant threat to the validity of study findings.Unfortunately,these issues are often overlooked.Methods:We systematically developed guidance for the development,validation,and evaluation of algorithms designed to identify health status(DEVELOP-RCD).Our initial efforts involved conducting both a narrative review and a systematic review of published studies on the concepts and methodological issues related to algorithm development,validation,and evaluation.Subsequently,we conducted an empirical study on an algorithm for identifying sepsis.Based on these findings,we formulated specific workflow and recommendations for algorithm development,validation,and evaluation within the guidance.Finally,the guidance underwent independent review by a panel of 20 external experts who then convened a consensus meeting to finalize it.Results:A standardized workflow for algorithm development,validation,and evaluation was established.Guided by specific health status considerations,the workflow comprises four integrated steps:assessing an existing algorithm’s suitability for the target health status;developing a new algorithm using recommended methods;validating the algorithm using prescribed performance measures;and evaluating the impact of the algorithm on study results.Additionally,13 good practice recommendations were formulated with detailed explanations.Furthermore,a practical study on sepsis identification was included to demonstrate the application of this guidance.Conclusions:The establishment of guidance is intended to aid researchers and clinicians in the appropriate and accurate development and application of algorithms for identifying health status from RCD.This guidance has the potential to enhance the credibility of findings from observational studies involving RCD.展开更多
Autism Spectrum Disorder(ASD)is a complex neurodevelopmental condition that causes multiple challenges in behavioral and communication activities.In the medical field,the data related to ASD,the security measures are ...Autism Spectrum Disorder(ASD)is a complex neurodevelopmental condition that causes multiple challenges in behavioral and communication activities.In the medical field,the data related to ASD,the security measures are integrated in this research responsibly and effectively to develop the Mobile Neuron Attention Stage-by-Stage Network(MNASNet)model,which is the integration of both Mobile Network(MobileNet)and Neuron Attention Stage-by-Stage.The steps followed to detect ASD with privacy-preserved data are data normalization,data augmentation,and K-Anonymization.The clinical data of individuals are taken initially and preprocessed using the Z-score Normalization.Then,data augmentation is performed using the oversampling technique.Subsequently,K-Anonymization is effectuated by utilizing the Black-winged Kite Algorithm to ensure the privacy of medical data,where the best fitness solution is based on data utility and privacy.Finally,after improving the data privacy,the developed approach MNASNet is implemented for ASD detection,which achieves highly accurate results compared to traditional methods to detect autism behavior.Hence,the final results illustrate that the proposed MNASNet achieves an accuracy of 92.9%,TPR of 95.9%,and TNR of 90.9%at the k-samples of 8.展开更多
The implicit partition algorithm used to solve fluid–structure coupling problems has high accuracy,but it requires a long computation time.In this paper,a semi-implicit fluid–structure coupling algorithm based on mo...The implicit partition algorithm used to solve fluid–structure coupling problems has high accuracy,but it requires a long computation time.In this paper,a semi-implicit fluid–structure coupling algorithm based on modal force prediction-correction is proposed to improve the computational efficiency.In the pre-processing stage,the fluid domain is assumed to be a pseudo-elastic solid and merged with the solid domain to form a holistic system,and the normalized modal information of the holistic system is calculated and stored.During the sub-step cycle,the modal superposition method is used to obtain the response of the holistic system with the predicted modal force as the load,so that the deformation of the structure and the updating of the fluid mesh can be achieved simultaneously.After solving the Reynolds-averaged Navier-Stokes equations in the fluid domain,the predicted modal force is corrected and a new sub-step cycle is started until the converged result is obtained.In this method,the computation of the fluid equations and the updating of the dynamic mesh are done implicitly,while the deformation of the structure is done explicitly.Two numerical cases,vortex induced oscillation of an elastic beam and fluid–structure interaction of a final stage blade,are used to verify the efficiency and accuracy of the proposed algorithm.The results show that the proposed method achieves the same accuracy as the implicit method while the computational time is reduced.In the case of the vortex-induced oscillation problem,the computational time can be reduced to 18.6%.In the case of the final stage blade vibration,the computational time can be reduced to 53.8%.展开更多
It is evident that complex optimization problems are becoming increasingly prominent,metaheuristic algorithms have demonstrated unique advantages in solving high-dimensional,nonlinear problems.However,the traditional ...It is evident that complex optimization problems are becoming increasingly prominent,metaheuristic algorithms have demonstrated unique advantages in solving high-dimensional,nonlinear problems.However,the traditional Sparrow Search Algorithm(SSA)suffers from limited global search capability,insufficient population diversity,and slow convergence,which often leads to premature stagnation in local optima.Despite the proposal of various enhanced versions,the effective balancing of exploration and exploitation remains an unsolved challenge.To address the previously mentioned problems,this study proposes a multi-strategy collaborative improved SSA,which systematically integrates four complementary strategies:(1)the Northern Goshawk Optimization(NGO)mechanism enhances global exploration through guided prey-attacking dynamics;(2)an adaptive t-distribution mutation strategy balances the transition between exploration and exploitation via dynamic adjustment of the degrees of freedom;(3)a dual chaotic initialization method(Bernoulli and Sinusoidal maps)increases population diversity and distribution uniformity;and(4)an elite retention strategy maintains solution quality and prevents degradation during iterations.These strategies cooperate synergistically,forming a tightly coupled optimization framework that significantly improves search efficiency and robustness.Therefore,this paper names it NTSSA:A Novel Multi-Strategy Enhanced Sparrow Search Algorithm with Northern Goshawk Optimization and Adaptive t-Distribution for Global Optimization.Extensive experiments on the CEC2005 benchmark set demonstrate that NTSSA achieves theoretical optimal accuracy on unimodal functions and significantly enhances global optimum discovery for multimodal functions by 2–5 orders of magnitude.Compared with SSA,GWO,ISSA,and CSSOA,NTSSA improves solution accuracy by up to 14.3%(F8)and 99.8%(F12),while accelerating convergence by approximately 1.5–2×.The Wilcoxon rank-sum test(p<0.05)indicates that NTSSA demonstrates a statistically substantial performance advantage.Theoretical analysis demonstrates that the collaborative synergy among adaptive mutation,chaos-based diversification,and elite preservation ensures both high convergence accuracy and global stability.This work bridges a key research gap in SSA by realizing a coordinated optimization mechanism between exploration and exploitation,offering a robust and efficient solution framework for complex high-dimensional problems in intelligent computation and engineering design.展开更多
The adaptive filtering algorithm with a fixed projection order is unable to adjust its performance in response to changes in the external environment of airborne radars.To overcome this limitation,a new approach is in...The adaptive filtering algorithm with a fixed projection order is unable to adjust its performance in response to changes in the external environment of airborne radars.To overcome this limitation,a new approach is introduced,which is the variable projection order Ekblom norm-promoted adaptive algorithm(VPO-EPAA).The method begins by examining the mean squared deviation(MSD)of the EPAA,deriving a formula for its MSD.Next,it compares the MSD of EPAA at two different projection orders and selects the one that minimizes the MSD as the parameter for the current iteration.Furthermore,the algorithm’s computational complexity is analyzed theoretically.Simulation results from system identification and self-interference cancellation show that the proposed algorithm performs exceptionally well in airborne radar signal self-interference cancellation,even under various noise intensities and types of interference.展开更多
With the widespread adoption of hydraulic fracturing technology in oil and gas resource development,improving the accuracy and efficiency of fracturing simulations has become a critical research focus.This paper propo...With the widespread adoption of hydraulic fracturing technology in oil and gas resource development,improving the accuracy and efficiency of fracturing simulations has become a critical research focus.This paper proposes an improved fluid flow algorithm,aiming to enhance the computational efficiency of hydraulic fracturing simulations while ensuring computational accuracy.The algorithm optimizes the aperture law and iteration criteria,focusing on improving the domain volume and crack pressure update strategy,thereby enabling precise capture of dynamic borehole pressure variations during injection tests.The effectiveness of the algorithm is verified through three flow-solid coupling cases.The study also analyzes the effects of borehole size,domain volume,and crack pressure update strategy on fracturing behavior.Furthermore,the performance of the improved algorithm in terms of crack propagation rate,micro-crack formation,and fluid pressure distribution was further evaluated.The results indicate that while large-size boreholes delay crack initiation,the cracks propagate more rapidly once formed.Additionally,the optimized domain volume calculation and crack pressure update strategy significantly shorten the pressure propagation stage,promote crack propagation,and improve computational efficiency.展开更多
Face recognition has emerged as one of the most prominent applications of image analysis and under-standing,gaining considerable attention in recent years.This growing interest is driven by two key factors:its extensi...Face recognition has emerged as one of the most prominent applications of image analysis and under-standing,gaining considerable attention in recent years.This growing interest is driven by two key factors:its extensive applications in law enforcement and the commercial domain,and the rapid advancement of practical technologies.Despite the significant advancements,modern recognition algorithms still struggle in real-world conditions such as varying lighting conditions,occlusion,and diverse facial postures.In such scenarios,human perception is still well above the capabilities of present technology.Using the systematic mapping study,this paper presents an in-depth review of face detection algorithms and face recognition algorithms,presenting a detailed survey of advancements made between 2015 and 2024.We analyze key methodologies,highlighting their strengths and restrictions in the application context.Additionally,we examine various datasets used for face detection/recognition datasets focusing on the task-specific applications,size,diversity,and complexity.By analyzing these algorithms and datasets,this survey works as a valuable resource for researchers,identifying the research gap in the field of face detection and recognition and outlining potential directions for future research.展开更多
Accurate and reliable photovoltaic(PV)modeling is crucial for the performance evaluation,control,and optimization of PV systems.However,existing methods for PV parameter identification often suffer from limitations in...Accurate and reliable photovoltaic(PV)modeling is crucial for the performance evaluation,control,and optimization of PV systems.However,existing methods for PV parameter identification often suffer from limitations in accuracy and efficiency.To address these challenges,we propose an adaptive multi-learning cooperation search algorithm(AMLCSA)for efficient identification of unknown parameters in PV models.AMLCSA is a novel algorithm inspired by teamwork behaviors in modern enterprises.It enhances the original cooperation search algorithm in two key aspects:(i)an adaptive multi-learning strategy that dynamically adjusts search ranges using adaptive weights,allowing better individuals to focus on local exploitation while guiding poorer individuals toward global exploration;and(ii)a chaotic grouping reflection strategy that introduces chaotic sequences to enhance population diversity and improve search performance.The effectiveness of AMLCSA is demonstrated on single-diode,double-diode,and three PV-module models.Simulation results show that AMLCSA offers significant advantages in convergence,accuracy,and stability compared to existing state-of-the-art algorithms.展开更多
基金co-supported by the Foundation of Shanghai Astronautics Science and Technology Innovation,China(No.SAST2022-114)the National Natural Science Foundation of China(No.62303378),the National Natural Science Foundation of China(Nos.124B2031,12202281)the Foundation of China National Key Laboratory of Science and Technology on Test Physics&Numerical Mathematics,China(No.08-YY-2023-R11)。
文摘The problem of collision avoidance for non-cooperative targets has received significant attention from researchers in recent years.Non-cooperative targets exhibit uncertain states and unpredictable behaviors,making collision avoidance significantly more challenging than that for space debris.Much existing research focuses on the continuous thrust model,whereas the impulsive maneuver model is more appropriate for long-duration and long-distance avoidance missions.Additionally,it is important to minimize the impact on the original mission while avoiding noncooperative targets.On the other hand,the existing avoidance algorithms are computationally complex and time-consuming especially with the limited computing capability of the on-board computer,posing challenges for practical engineering applications.To conquer these difficulties,this paper makes the following key contributions:(A)a turn-based(sequential decision-making)limited-area impulsive collision avoidance model considering the time delay of precision orbit determination is established for the first time;(B)a novel Selection Probability Learning Adaptive Search-depth Search Tree(SPL-ASST)algorithm is proposed for non-cooperative target avoidance,which improves the decision-making efficiency by introducing an adaptive-search-depth mechanism and a neural network into the traditional Monte Carlo Tree Search(MCTS).Numerical simulations confirm the effectiveness and efficiency of the proposed method.
基金partially supported by MRC(MC_PC_17171)Royal Society(RP202G0230)+8 种基金BHF(AA/18/3/34220)Hope Foundation for Cancer Research(RM60G0680)GCRF(20P2PF11)Sino-UK Industrial Fund(RP202G0289)LIAS(20P2ED10,20P2RE969)Data Science Enhancement Fund(20P2RE237)Fight for Sight(24NN201)Sino-UK Education Fund(OP202006)BBSRC(RM32G0178B8).
文摘The Bat algorithm,a metaheuristic optimization technique inspired by the foraging behaviour of bats,has been employed to tackle optimization problems.Known for its ease of implementation,parameter tunability,and strong global search capabilities,this algorithm finds application across diverse optimization problem domains.However,in the face of increasingly complex optimization challenges,the Bat algorithm encounters certain limitations,such as slow convergence and sensitivity to initial solutions.In order to tackle these challenges,the present study incorporates a range of optimization compo-nents into the Bat algorithm,thereby proposing a variant called PKEBA.A projection screening strategy is implemented to mitigate its sensitivity to initial solutions,thereby enhancing the quality of the initial solution set.A kinetic adaptation strategy reforms exploration patterns,while an elite communication strategy enhances group interaction,to avoid algorithm from local optima.Subsequently,the effectiveness of the proposed PKEBA is rigorously evaluated.Testing encompasses 30 benchmark functions from IEEE CEC2014,featuring ablation experiments and comparative assessments against classical algorithms and their variants.Moreover,real-world engineering problems are employed as further validation.The results conclusively demonstrate that PKEBA ex-hibits superior convergence and precision compared to existing algorithms.
基金the Deanship of Graduate Studies and Scientific Research at Najran University for funding this work under the Easy Funding Program grant code(NU/EFP/SERC/13/166).
文摘The Internet of Things(IoT)has emerged as an important future technology.IoT-Fog is a new computing paradigm that processes IoT data on servers close to the source of the data.In IoT-Fog computing,resource allocation and independent task scheduling aim to deliver short response time services demanded by the IoT devices and performed by fog servers.The heterogeneity of the IoT-Fog resources and the huge amount of data that needs to be processed by the IoT-Fog tasks make scheduling fog computing tasks a challenging problem.This study proposes an Adaptive Firefly Algorithm(AFA)for dependent task scheduling in IoT-Fog computing.The proposed AFA is a modified version of the standard Firefly Algorithm(FA),considering the execution times of the submitted tasks,the impact of synchronization requirements,and the communication time between dependent tasks.As IoT-Fog computing depends mainly on distributed fog node servers that receive tasks in a dynamic manner,tackling the communications and synchronization issues between dependent tasks is becoming a challenging problem.The proposed AFA aims to address the dynamic nature of IoT-Fog computing environments.The proposed AFA mechanism considers a dynamic light absorption coefficient to control the decrease in attractiveness over iterations.The proposed AFA mechanism performance was benchmarked against the standard Firefly Algorithm(FA),Puma Optimizer(PO),Genetic Algorithm(GA),and Ant Colony Optimization(ACO)through simulations under light,typical,and heavy workload scenarios.In heavy workloads,the proposed AFA mechanism obtained the shortest average execution time,968.98 ms compared to 970.96,1352.87,1247.28,and 1773.62 of FA,PO,GA,and ACO,respectively.The simulation results demonstrate the proposed AFA’s ability to rapidly converge to optimal solutions,emphasizing its adaptability and efficiency in typical and heavy workloads.
文摘Data clustering is an essential technique for analyzing complex datasets and continues to be a central research topic in data analysis.Traditional clustering algorithms,such as K-means,are widely used due to their simplicity and efficiency.This paper proposes a novel Spiral Mechanism-Optimized Phasmatodea Population Evolution Algorithm(SPPE)to improve clustering performance.The SPPE algorithm introduces several enhancements to the standard Phasmatodea Population Evolution(PPE)algorithm.Firstly,a Variable Neighborhood Search(VNS)factor is incorporated to strengthen the local search capability and foster population diversity.Secondly,a position update model,incorporating a spiral mechanism,is designed to improve the algorithm’s global exploration and convergence speed.Finally,a dynamic balancing factor,guided by fitness values,adjusts the search process to balance exploration and exploitation effectively.The performance of SPPE is first validated on CEC2013 benchmark functions,where it demonstrates excellent convergence speed and superior optimization results compared to several state-of-the-art metaheuristic algorithms.To further verify its practical applicability,SPPE is combined with the K-means algorithm for data clustering and tested on seven datasets.Experimental results show that SPPE-K-means improves clustering accuracy,reduces dependency on initialization,and outperforms other clustering approaches.This study highlights SPPE’s robustness and efficiency in solving both optimization and clustering challenges,making it a promising tool for complex data analysis tasks.
基金BBSRC,Grant/Award Number:RM32G0178B8National Natural Science Foundation of China,Grant/Award Numbers:U19A2061,U1809209,62076185+11 种基金Science and Technology Development Project of Jilin Province,Grant/Award Number:20190301024NYJilin Provincial Industrial Innovation Special Fund Project,Grant/Award Number:2018C039-3MRC,Grant/Award Number:MC_PC_17171Royal Society,Grant/Award Number:RP202G0230BHF,Grant/Award Number:AA/18/3/34220Hope Foundation for Cancer Research,Grant/Award Number:RM60G0680GCRF,Grant/Award Number:P202PF11Sino-UK Industrial Fund,Grant/Award Number:RP202G0289LIAS,Grant/Award Numbers:P202ED10,P202RE969Data Science Enhancement Fund,Grant/Award Number:P202RE237Fight for Sight,Grant/Award Number:24NN201Sino-UK Education Fund,Grant/Award Number:OP202006。
文摘The bat algorithm(BA)is a metaheuristic algorithm for global optimisation that simulates the echolocation behaviour of bats with varying pulse rates of emission and loudness,which can be used to find the globally optimal solutions for various optimisation problems.Knowing the recent criticises of the originality of equations,the principle of BA is concise and easy to implement,and its mathematical structure can be seen as a hybrid particle swarm with simulated annealing.In this research,the authors focus on the performance optimisation of BA as a solver rather than discussing its originality issues.In terms of operation effect,BA has an acceptable convergence speed.However,due to the low proportion of time used to explore the search space,it is easy to converge prematurely and fall into the local optima.The authors propose an adaptive multi-stage bat algorithm(AMSBA).By tuning the algorithm's focus at three different stages of the search process,AMSBA can achieve a better balance between exploration and exploitation and improve its exploration ability by enhancing its performance in escaping local optima as well as maintaining a certain convergence speed.Therefore,AMSBA can achieve solutions with better quality.A convergence analysis was conducted to demonstrate the global convergence of AMSBA.The authors also perform simulation experiments on 30 benchmark functions from IEEE CEC 2017 as the objective functions and compare AMSBA with some original and improved swarm-based algorithms.The results verify the effectiveness and superiority of AMSBA.AMSBA is also compared with eight representative optimisation algorithms on 10 benchmark functions derived from IEEE CEC 2020,while this experiment is carried out on five different dimensions of the objective functions respectively.A balance and diversity analysis was performed on AMSBA to demonstrate its improvement over the original BA in terms of balance.AMSBA was also applied to the multi-threshold image segmentation of Citrus Macular disease,which is a bacterial infection that causes lesions on citrus trees.The segmentation results were analysed by comparing each comparative algorithm's peak signal-to-noise ratio,structural similarity index and feature similarity index.The results show that the proposed BA-based algorithm has apparent advantages,and it can effectively segment the disease spots from citrus leaves when the segmentation threshold is at a low level.Based on a comprehensive study,the authors think the proposed optimiser has mitigated the main drawbacks of the BA,and it can be utilised as an effective optimisation tool.
文摘Cloud computing has become an essential technology for the management and processing of large datasets,offering scalability,high availability,and fault tolerance.However,optimizing data replication across multiple data centers poses a significant challenge,especially when balancing opposing goals such as latency,storage costs,energy consumption,and network efficiency.This study introduces a novel Dynamic Optimization Algorithm called Dynamic Multi-Objective Gannet Optimization(DMGO),designed to enhance data replication efficiency in cloud environments.Unlike traditional static replication systems,DMGO adapts dynamically to variations in network conditions,system demand,and resource availability.The approach utilizes multi-objective optimization approaches to efficiently balance data access latency,storage efficiency,and operational costs.DMGO consistently evaluates data center performance and adjusts replication algorithms in real time to guarantee optimal system efficiency.Experimental evaluations conducted in a simulated cloud environment demonstrate that DMGO significantly outperforms conventional static algorithms,achieving faster data access,lower storage overhead,reduced energy consumption,and improved scalability.The proposed methodology offers a robust and adaptable solution for modern cloud systems,ensuring efficient resource consumption while maintaining high performance.
文摘This paper examines the impact of algorithmic recommendations and data-driven marketing on consumer engagement and business performance.By leveraging large volumes of user data,businesses can deliver personalized content that enhances user experiences and increases conversion rates.However,the growing reliance on these technologies introduces significant risks,including privacy violations,algorithmic bias,and ethical concerns.This paper explores these challenges and provides recommendations for businesses to mitigate associated risks while optimizing marketing strategies.It highlights the importance of transparency,fairness,and user control in ensuring responsible and effective data-driven marketing.
文摘The application of machine learning was investigated for predicting end-point temperature in the basic oxygen furnace steelmaking process,addressing gaps in the field,particularly large-scale dataset sizes and the underutilization of boosting algorithms.Utilizing a substantial dataset containing over 20,000 heats,significantly bigger than those in previous studies,a comprehensive evaluation of five advanced machine learning models was conducted.These include four ensemble learning algorithms:XGBoost,LightGBM,CatBoost(three boosting algorithms),along with random forest(a bagging algorithm),as well as a neural network model,namely the multilayer perceptron.Our comparative analysis reveals that Bayesian-optimized boosting models demonstrate exceptional robustness and accuracy,achieving the highest R-squared values,the lowest root mean square error,and lowest mean absolute error,along with the best hit ratio.CatBoost exhibited superior performance,with its test R-squared improving by 4.2%compared to that of the random forest and by 0.8%compared to that of the multilayer perceptron.This highlights the efficacy of boosting algorithms in refining complex industrial processes.Additionally,our investigation into the impact of varying dataset sizes,ranging from 500 to 20,000 heats,on model accuracy underscores the importance of leveraging larger-scale datasets to improve the accuracy and stability of predictive models.
基金supported by the Natural Science Basic Research Program of Shaanxi(Program No.2024JC-YBMS-026).
文摘When dealing with imbalanced datasets,the traditional support vectormachine(SVM)tends to produce a classification hyperplane that is biased towards the majority class,which exhibits poor robustness.This paper proposes a high-performance classification algorithm specifically designed for imbalanced datasets.The proposed method first uses a biased second-order cone programming support vectormachine(B-SOCP-SVM)to identify the support vectors(SVs)and non-support vectors(NSVs)in the imbalanced data.Then,it applies the synthetic minority over-sampling technique(SV-SMOTE)to oversample the support vectors of the minority class and uses the random under-sampling technique(NSV-RUS)multiple times to undersample the non-support vectors of the majority class.Combining the above-obtained minority class data set withmultiple majority class datasets can obtainmultiple new balanced data sets.Finally,SOCP-SVM is used to classify each data set,and the final result is obtained through the integrated algorithm.Experimental results demonstrate that the proposed method performs excellently on imbalanced datasets.
基金funded by Universiti Putra Malaysia under a Geran Putra Inisiatif(GPI)research grant with reference to GP-GPI/2023/9762100.
文摘This study proposes a novel time-synchronization protocol inspired by stochastic gradient algorithms.The clock model of each network node in this synchronizer is configured as a generic adaptive filter where different stochastic gradient algorithms can be adopted for adaptive clock frequency adjustments.The study analyzes the pairwise synchronization behavior of the protocol and proves the generalized convergence of the synchronization error and clock frequency.A novel closed-form expression is also derived for a generalized asymptotic error variance steady state.Steady and convergence analyses are then presented for the synchronization,with frequency adaptations done using least mean square(LMS),the Newton search,the gradient descent(GraDes),the normalized LMS(N-LMS),and the Sign-Data LMS algorithms.Results obtained from real-time experiments showed a better performance of our protocols as compared to the Average Proportional-Integral Synchronization Protocol(AvgPISync)regarding the impact of quantization error on synchronization accuracy,precision,and convergence time.This generalized approach to time synchronization allows flexibility in selecting a suitable protocol for different wireless sensor network applications.
基金supported by the National Natural Science Foundation of China(Grant Number 62341210)Natural Science Foundation of Guangxi Province(Grant Number:2025GXNSFHA069267)Science and Technology Development Plan for Baise City(Grant Number 20233654).
文摘DNA microarrays, a cornerstone in biomedicine, measure gene expression across thousands to tens of thousands of genes. Identifying the genes vital for accurate cancer classification is a key challenge. Here, we present Fs-LSA (F-score based Learning Search Algorithm), a novel gene selection algorithm designed to enhance the precision and efficiency of target gene identification from microarray data for cancer classification. This algorithm is divided into two phases: the first leverages F-score values to prioritize and select feature genes with the most significant differential expression;the second phase introduces our Learning Search Algorithm (LSA), which harnesses swarm intelligence to identify the optimal subset among the remaining genes. Inspired by human social learning, LSA integrates historical data and collective intelligence for a thorough search, with a dynamic control mechanism that balances exploration and refinement, thereby enhancing the gene selection process. We conducted a rigorous validation of Fs-LSA’s performance using eight publicly available cancer microarray expression datasets. Fs-LSA achieved accuracy, precision, sensitivity, and F1-score values of 0.9932, 0.9923, 0.9962, and 0.994, respectively. Comparative analyses with state-of-the-art algorithms revealed Fs-LSA’s superior performance in terms of simplicity and efficiency. Additionally, we validated the algorithm’s efficacy independently using glioblastoma data from GEO and TCGA databases. It was significantly superior to those of the comparison algorithms. Importantly, the driver genes identified by Fs-LSA were instrumental in developing a predictive model as an independent prognostic indicator for glioblastoma, underscoring Fs-LSA’s transformative potential in genomics and personalized medicine.
文摘With the rapid advancement of medical artificial intelligence(AI)technology,particularly the widespread adoption of AI diagnostic systems,ethical challenges in medical decision-making have garnered increasing attention.This paper analyzes the limitations of algorithmic ethics in medical decision-making and explores accountability mechanisms,aiming to provide theoretical support for ethically informed medical practices.The study highlights how the opacity of AI algorithms complicates the definition of decision-making responsibility,undermines doctor-patient trust,and affects informed consent.By thoroughly investigating issues such as the algorithmic“black box”problem and data privacy protection,we develop accountability assessment models to address ethical concerns related to medical resource allocation.Furthermore,this research examines the effective implementation of AI diagnostic systems through case studies of both successful and unsuccessful applications,extracting lessons on accountability mechanisms and response strategies.Finally,we emphasize that establishing a transparent accountability framework is crucial for enhancing the ethical standards of medical AI systems and protecting patients’rights and interests.
基金supported by the National Natural Science Foundation of China(82225049,72104155)the Sichuan Provincial Central Government Guides Local Science and Technology Development Special Project(2022ZYD0127)the 1·3·5 Project for Disciplines of Excellence,West China Hospital,Sichuan University(ZYGD23004).
文摘Background:In recent years,there has been a growing trend in the utilization of observational studies that make use of routinely collected healthcare data(RCD).These studies rely on algorithms to identify specific health conditions(e.g.,diabetes or sepsis)for statistical analyses.However,there has been substantial variation in the algorithm development and validation,leading to frequently suboptimal performance and posing a significant threat to the validity of study findings.Unfortunately,these issues are often overlooked.Methods:We systematically developed guidance for the development,validation,and evaluation of algorithms designed to identify health status(DEVELOP-RCD).Our initial efforts involved conducting both a narrative review and a systematic review of published studies on the concepts and methodological issues related to algorithm development,validation,and evaluation.Subsequently,we conducted an empirical study on an algorithm for identifying sepsis.Based on these findings,we formulated specific workflow and recommendations for algorithm development,validation,and evaluation within the guidance.Finally,the guidance underwent independent review by a panel of 20 external experts who then convened a consensus meeting to finalize it.Results:A standardized workflow for algorithm development,validation,and evaluation was established.Guided by specific health status considerations,the workflow comprises four integrated steps:assessing an existing algorithm’s suitability for the target health status;developing a new algorithm using recommended methods;validating the algorithm using prescribed performance measures;and evaluating the impact of the algorithm on study results.Additionally,13 good practice recommendations were formulated with detailed explanations.Furthermore,a practical study on sepsis identification was included to demonstrate the application of this guidance.Conclusions:The establishment of guidance is intended to aid researchers and clinicians in the appropriate and accurate development and application of algorithms for identifying health status from RCD.This guidance has the potential to enhance the credibility of findings from observational studies involving RCD.
文摘Autism Spectrum Disorder(ASD)is a complex neurodevelopmental condition that causes multiple challenges in behavioral and communication activities.In the medical field,the data related to ASD,the security measures are integrated in this research responsibly and effectively to develop the Mobile Neuron Attention Stage-by-Stage Network(MNASNet)model,which is the integration of both Mobile Network(MobileNet)and Neuron Attention Stage-by-Stage.The steps followed to detect ASD with privacy-preserved data are data normalization,data augmentation,and K-Anonymization.The clinical data of individuals are taken initially and preprocessed using the Z-score Normalization.Then,data augmentation is performed using the oversampling technique.Subsequently,K-Anonymization is effectuated by utilizing the Black-winged Kite Algorithm to ensure the privacy of medical data,where the best fitness solution is based on data utility and privacy.Finally,after improving the data privacy,the developed approach MNASNet is implemented for ASD detection,which achieves highly accurate results compared to traditional methods to detect autism behavior.Hence,the final results illustrate that the proposed MNASNet achieves an accuracy of 92.9%,TPR of 95.9%,and TNR of 90.9%at the k-samples of 8.
基金support of the National Natural Science Foundation of China(No.51675406)the Basic Research Project Group,China(No.514010106-205)。
文摘The implicit partition algorithm used to solve fluid–structure coupling problems has high accuracy,but it requires a long computation time.In this paper,a semi-implicit fluid–structure coupling algorithm based on modal force prediction-correction is proposed to improve the computational efficiency.In the pre-processing stage,the fluid domain is assumed to be a pseudo-elastic solid and merged with the solid domain to form a holistic system,and the normalized modal information of the holistic system is calculated and stored.During the sub-step cycle,the modal superposition method is used to obtain the response of the holistic system with the predicted modal force as the load,so that the deformation of the structure and the updating of the fluid mesh can be achieved simultaneously.After solving the Reynolds-averaged Navier-Stokes equations in the fluid domain,the predicted modal force is corrected and a new sub-step cycle is started until the converged result is obtained.In this method,the computation of the fluid equations and the updating of the dynamic mesh are done implicitly,while the deformation of the structure is done explicitly.Two numerical cases,vortex induced oscillation of an elastic beam and fluid–structure interaction of a final stage blade,are used to verify the efficiency and accuracy of the proposed algorithm.The results show that the proposed method achieves the same accuracy as the implicit method while the computational time is reduced.In the case of the vortex-induced oscillation problem,the computational time can be reduced to 18.6%.In the case of the final stage blade vibration,the computational time can be reduced to 53.8%.
文摘It is evident that complex optimization problems are becoming increasingly prominent,metaheuristic algorithms have demonstrated unique advantages in solving high-dimensional,nonlinear problems.However,the traditional Sparrow Search Algorithm(SSA)suffers from limited global search capability,insufficient population diversity,and slow convergence,which often leads to premature stagnation in local optima.Despite the proposal of various enhanced versions,the effective balancing of exploration and exploitation remains an unsolved challenge.To address the previously mentioned problems,this study proposes a multi-strategy collaborative improved SSA,which systematically integrates four complementary strategies:(1)the Northern Goshawk Optimization(NGO)mechanism enhances global exploration through guided prey-attacking dynamics;(2)an adaptive t-distribution mutation strategy balances the transition between exploration and exploitation via dynamic adjustment of the degrees of freedom;(3)a dual chaotic initialization method(Bernoulli and Sinusoidal maps)increases population diversity and distribution uniformity;and(4)an elite retention strategy maintains solution quality and prevents degradation during iterations.These strategies cooperate synergistically,forming a tightly coupled optimization framework that significantly improves search efficiency and robustness.Therefore,this paper names it NTSSA:A Novel Multi-Strategy Enhanced Sparrow Search Algorithm with Northern Goshawk Optimization and Adaptive t-Distribution for Global Optimization.Extensive experiments on the CEC2005 benchmark set demonstrate that NTSSA achieves theoretical optimal accuracy on unimodal functions and significantly enhances global optimum discovery for multimodal functions by 2–5 orders of magnitude.Compared with SSA,GWO,ISSA,and CSSOA,NTSSA improves solution accuracy by up to 14.3%(F8)and 99.8%(F12),while accelerating convergence by approximately 1.5–2×.The Wilcoxon rank-sum test(p<0.05)indicates that NTSSA demonstrates a statistically substantial performance advantage.Theoretical analysis demonstrates that the collaborative synergy among adaptive mutation,chaos-based diversification,and elite preservation ensures both high convergence accuracy and global stability.This work bridges a key research gap in SSA by realizing a coordinated optimization mechanism between exploration and exploitation,offering a robust and efficient solution framework for complex high-dimensional problems in intelligent computation and engineering design.
基金supported by the Shan⁃dong Provincial Natural Science Foundation(No.ZR2022MF314).
文摘The adaptive filtering algorithm with a fixed projection order is unable to adjust its performance in response to changes in the external environment of airborne radars.To overcome this limitation,a new approach is introduced,which is the variable projection order Ekblom norm-promoted adaptive algorithm(VPO-EPAA).The method begins by examining the mean squared deviation(MSD)of the EPAA,deriving a formula for its MSD.Next,it compares the MSD of EPAA at two different projection orders and selects the one that minimizes the MSD as the parameter for the current iteration.Furthermore,the algorithm’s computational complexity is analyzed theoretically.Simulation results from system identification and self-interference cancellation show that the proposed algorithm performs exceptionally well in airborne radar signal self-interference cancellation,even under various noise intensities and types of interference.
基金supported by the National Natural Science Foundation of China(Nos.52164001,52064006,52004072 and 52364004)the Science and Technology Support Project of Guizhou(Nos.[2020]4Y044,[2021]N404 and[2021]N511)+1 种基金the Guizhou Provincial Science and Technology Foundation(No.GCC[2022]005-1),Talents of Guizhou University(No.201901)the Special Research Funds of Guizhou University(Nos.201903,202011,and 202012).
文摘With the widespread adoption of hydraulic fracturing technology in oil and gas resource development,improving the accuracy and efficiency of fracturing simulations has become a critical research focus.This paper proposes an improved fluid flow algorithm,aiming to enhance the computational efficiency of hydraulic fracturing simulations while ensuring computational accuracy.The algorithm optimizes the aperture law and iteration criteria,focusing on improving the domain volume and crack pressure update strategy,thereby enabling precise capture of dynamic borehole pressure variations during injection tests.The effectiveness of the algorithm is verified through three flow-solid coupling cases.The study also analyzes the effects of borehole size,domain volume,and crack pressure update strategy on fracturing behavior.Furthermore,the performance of the improved algorithm in terms of crack propagation rate,micro-crack formation,and fluid pressure distribution was further evaluated.The results indicate that while large-size boreholes delay crack initiation,the cracks propagate more rapidly once formed.Additionally,the optimized domain volume calculation and crack pressure update strategy significantly shorten the pressure propagation stage,promote crack propagation,and improve computational efficiency.
文摘Face recognition has emerged as one of the most prominent applications of image analysis and under-standing,gaining considerable attention in recent years.This growing interest is driven by two key factors:its extensive applications in law enforcement and the commercial domain,and the rapid advancement of practical technologies.Despite the significant advancements,modern recognition algorithms still struggle in real-world conditions such as varying lighting conditions,occlusion,and diverse facial postures.In such scenarios,human perception is still well above the capabilities of present technology.Using the systematic mapping study,this paper presents an in-depth review of face detection algorithms and face recognition algorithms,presenting a detailed survey of advancements made between 2015 and 2024.We analyze key methodologies,highlighting their strengths and restrictions in the application context.Additionally,we examine various datasets used for face detection/recognition datasets focusing on the task-specific applications,size,diversity,and complexity.By analyzing these algorithms and datasets,this survey works as a valuable resource for researchers,identifying the research gap in the field of face detection and recognition and outlining potential directions for future research.
基金supported by the National Natural Science Foundation of China(Grant Nos.62303197,62273214)the Natural Science Foundation of Shandong Province(ZR2024MFO18).
文摘Accurate and reliable photovoltaic(PV)modeling is crucial for the performance evaluation,control,and optimization of PV systems.However,existing methods for PV parameter identification often suffer from limitations in accuracy and efficiency.To address these challenges,we propose an adaptive multi-learning cooperation search algorithm(AMLCSA)for efficient identification of unknown parameters in PV models.AMLCSA is a novel algorithm inspired by teamwork behaviors in modern enterprises.It enhances the original cooperation search algorithm in two key aspects:(i)an adaptive multi-learning strategy that dynamically adjusts search ranges using adaptive weights,allowing better individuals to focus on local exploitation while guiding poorer individuals toward global exploration;and(ii)a chaotic grouping reflection strategy that introduces chaotic sequences to enhance population diversity and improve search performance.The effectiveness of AMLCSA is demonstrated on single-diode,double-diode,and three PV-module models.Simulation results show that AMLCSA offers significant advantages in convergence,accuracy,and stability compared to existing state-of-the-art algorithms.