Optimization is the key to obtaining efficient utilization of resources in structural design.Due to the complex nature of truss systems,this study presents a method based on metaheuristic modelling that minimises stru...Optimization is the key to obtaining efficient utilization of resources in structural design.Due to the complex nature of truss systems,this study presents a method based on metaheuristic modelling that minimises structural weight under stress and frequency constraints.Two new algorithms,the Red Kite Optimization Algorithm(ROA)and Secretary Bird Optimization Algorithm(SBOA),are utilized on five benchmark trusses with 10,18,37,72,and 200-bar trusses.Both algorithms are evaluated against benchmarks in the literature.The results indicate that SBOA always reaches a lighter optimal.Designs with reducing structural weight ranging from 0.02%to 0.15%compared to ROA,and up to 6%–8%as compared to conventional algorithms.In addition,SBOA can achieve 15%–20%faster convergence speed and 10%–18%reduction in computational time with a smaller standard deviation over independent runs,which demonstrates its robustness and reliability.It is indicated that the adaptive exploration mechanism of SBOA,especially its Levy flight–based search strategy,can obviously improve optimization performance for low-and high-dimensional trusses.The research has implications in the context of promoting bio-inspired optimization techniques by demonstrating the viability of SBOA,a reliable model for large-scale structural design that provides significant enhancements in performance and convergence behavior.展开更多
Aiming to solve the steering instability and hysteresis of agricultural robots in the process of movement,a fusion PID control method of particle swarm optimization(PSO)and genetic algorithm(GA)was proposed.The fusion...Aiming to solve the steering instability and hysteresis of agricultural robots in the process of movement,a fusion PID control method of particle swarm optimization(PSO)and genetic algorithm(GA)was proposed.The fusion algorithm took advantage of the fast optimization ability of PSO to optimize the population screening link of GA.The Simulink simulation results showed that the convergence of the fitness function of the fusion algorithm was accelerated,the system response adjustment time was reduced,and the overshoot was almost zero.Then the algorithm was applied to the steering test of agricultural robot in various scenes.After modeling the steering system of agricultural robot,the steering test results in the unloaded suspended state showed that the PID control based on fusion algorithm reduced the rise time,response adjustment time and overshoot of the system,and improved the response speed and stability of the system,compared with the artificial trial and error PID control and the PID control based on GA.The actual road steering test results showed that the PID control response rise time based on the fusion algorithm was the shortest,about 4.43 s.When the target pulse number was set to 100,the actual mean value in the steady-state regulation stage was about 102.9,which was the closest to the target value among the three control methods,and the overshoot was reduced at the same time.The steering test results under various scene states showed that the PID control based on the proposed fusion algorithm had good anti-interference ability,it can adapt to the changes of environment and load and improve the performance of the control system.It was effective in the steering control of agricultural robot.This method can provide a reference for the precise steering control of other robots.展开更多
This paper presents an Eulerian-Lagrangian algorithm for direct numerical simulation(DNS)of particle-laden flows.The algorithm is applicable to perform simulations of dilute suspensions of small inertial particles in ...This paper presents an Eulerian-Lagrangian algorithm for direct numerical simulation(DNS)of particle-laden flows.The algorithm is applicable to perform simulations of dilute suspensions of small inertial particles in turbulent carrier flow.The Eulerian framework numerically resolves turbulent carrier flow using a parallelized,finite-volume DNS solver on a staggered Cartesian grid.Particles are tracked using a point-particle method utilizing a Lagrangian particle tracking(LPT)algorithm.The proposed Eulerian-Lagrangian algorithm is validated using an inertial particle-laden turbulent channel flow for different Stokes number cases.The particle concentration profiles and higher-order statistics of the carrier and dispersed phases agree well with the benchmark results.We investigated the effect of fluid velocity interpolation and numerical integration schemes of particle tracking algorithms on particle dispersion statistics.The suitability of fluid velocity interpolation schemes for predicting the particle dispersion statistics is discussed in the framework of the particle tracking algorithm coupled to the finite-volume solver.In addition,we present parallelization strategies implemented in the algorithm and evaluate their parallel performance.展开更多
Accurate prediction of flood events is important for flood control and risk management.Machine learning techniques contributed greatly to advances in flood predictions,and existing studies mainly focused on predicting...Accurate prediction of flood events is important for flood control and risk management.Machine learning techniques contributed greatly to advances in flood predictions,and existing studies mainly focused on predicting flood resource variables using single or hybrid machine learning techniques.However,class-based flood predictions have rarely been investigated,which can aid in quickly diagnosing comprehensive flood characteristics and proposing targeted management strategies.This study proposed a prediction approach of flood regime metrics and event classes coupling machine learning algorithms with clustering-deduced membership degrees.Five algorithms were adopted for this exploration.Results showed that the class membership degrees accurately determined event classes with class hit rates up to 100%,compared with the four classes clustered from nine regime metrics.The nonlinear algorithms(Multiple Linear Regression,Random Forest,and least squares-Support Vector Machine)outperformed the linear techniques(Multiple Linear Regression and Stepwise Regression)in predicting flood regime metrics.The proposed approach well predicted flood event classes with average class hit rates of 66.0%-85.4%and 47.2%-76.0%in calibration and validation periods,respectively,particularly for the slow and late flood events.The predictive capability of the proposed prediction approach for flood regime metrics and classes was considerably stronger than that of hydrological modeling approach.展开更多
Optimization problems are prevalent in various fields of science and engineering,with several real-world applications characterized by high dimensionality and complex search landscapes.Starfish optimization algorithm(...Optimization problems are prevalent in various fields of science and engineering,with several real-world applications characterized by high dimensionality and complex search landscapes.Starfish optimization algorithm(SFOA)is a recently optimizer inspired by swarm intelligence,which is effective for numerical optimization,but it may encounter premature and local convergence for complex optimization problems.To address these challenges,this paper proposes the multi-strategy enhanced crested porcupine-starfish optimization algorithm(MCPSFOA).The core innovation of MCPSFOA lies in employing a hybrid strategy to improve SFOA,which integrates the exploratory mechanisms of SFOA with the diverse search capacity of the Crested Porcupine Optimizer(CPO).This synergy enhances MCPSFOA’s ability to navigate complex and multimodal search spaces.To further prevent premature convergence,MCPSFOA incorporates Lévy flight,leveraging its characteristic long and short jump patterns to enable large-scale exploration and escape from local optima.Subsequently,Gaussian mutation is applied for precise solution tuning,introducing controlled perturbations that enhance accuracy and mitigate the risk of insufficient exploitation.Notably,the population diversity enhancement mechanism periodically identifies and resets stagnant individuals,thereby consistently revitalizing population variety throughout the optimization process.MCPSFOA is rigorously evaluated on 24 classical benchmark functions(including high-dimensional cases),the CEC2017 suite,and the CEC2022 suite.MCPSFOA achieves superior overall performance with Friedman mean ranks of 2.208,2.310 and 2.417 on these benchmark functions,outperforming 11 state-of-the-art algorithms.Furthermore,the practical applicability of MCPSFOA is confirmed through its successful application to five engineering optimization cases,where it also yields excellent results.In conclusion,MCPSFOA is not only a highly effective and reliable optimizer for benchmark functions,but also a practical tool for solving real-world optimization problems.展开更多
Existing feature selection methods for intrusion detection systems in the Industrial Internet of Things often suffer from local optimality and high computational complexity.These challenges hinder traditional IDS from...Existing feature selection methods for intrusion detection systems in the Industrial Internet of Things often suffer from local optimality and high computational complexity.These challenges hinder traditional IDS from effectively extracting features while maintaining detection accuracy.This paper proposes an industrial Internet ofThings intrusion detection feature selection algorithm based on an improved whale optimization algorithm(GSLDWOA).The aim is to address the problems that feature selection algorithms under high-dimensional data are prone to,such as local optimality,long detection time,and reduced accuracy.First,the initial population’s diversity is increased using the Gaussian Mutation mechanism.Then,Non-linear Shrinking Factor balances global exploration and local development,avoiding premature convergence.Lastly,Variable-step Levy Flight operator and Dynamic Differential Evolution strategy are introduced to improve the algorithm’s search efficiency and convergence accuracy in highdimensional feature space.Experiments on the NSL-KDD and WUSTL-IIoT-2021 datasets demonstrate that the feature subset selected by GSLDWOA significantly improves detection performance.Compared to the traditional WOA algorithm,the detection rate and F1-score increased by 3.68%and 4.12%.On the WUSTL-IIoT-2021 dataset,accuracy,recall,and F1-score all exceed 99.9%.展开更多
Concrete-filled steel tubes(CFST)are widely utilized in civil engineering due to their superior load-bearing capacity,ductility,and seismic resistance.However,existing design codes,such as AISC and Eurocode 4,tend to ...Concrete-filled steel tubes(CFST)are widely utilized in civil engineering due to their superior load-bearing capacity,ductility,and seismic resistance.However,existing design codes,such as AISC and Eurocode 4,tend to be excessively conservative as they fail to account for the composite action between the steel tube and the concrete core.To address this limitation,this study proposes a hybrid model that integrates XGBoost with the Pied Kingfisher Optimizer(PKO),a nature-inspired algorithm,to enhance the accuracy of shear strength prediction for CFST columns.Additionally,quantile regression is employed to construct prediction intervals for the ultimate shear force,while the Asymmetric Squared Error Loss(ASEL)function is incorporated to mitigate overestimation errors.The computational results demonstrate that the PKO-XGBoost model delivers superior predictive accuracy,achieving a Mean Absolute Percentage Error(MAPE)of 4.431%and R2 of 0.9925 on the test set.Furthermore,the ASEL-PKO-XGBoost model substantially reduces overestimation errors to 28.26%,with negligible impact on predictive performance.Additionally,based on the Genetic Algorithm(GA)and existing equation models,a strength equation model is developed,achieving markedly higher accuracy than existing models(R^(2)=0.934).Lastly,web-based Graphical User Interfaces(GUIs)were developed to enable real-time prediction.展开更多
Impact craters are important for understanding the evolution of lunar geologic and surface erosion rates,among other functions.However,the morphological characteristics of these micro impact craters are not obvious an...Impact craters are important for understanding the evolution of lunar geologic and surface erosion rates,among other functions.However,the morphological characteristics of these micro impact craters are not obvious and they are numerous,resulting in low detection accuracy by deep learning models.Therefore,we proposed a new multi-scale fusion crater detection algorithm(MSF-CDA)based on the YOLO11 to improve the accuracy of lunar impact crater detection,especially for small craters with a diameter of<1 km.Using the images taken by the LROC(Lunar Reconnaissance Orbiter Camera)at the Chang’e-4(CE-4)landing area,we constructed three separate datasets for craters with diameters of 0-70 m,70-140 m,and>140 m.We then trained three submodels separately with these three datasets.Additionally,we designed a slicing-amplifying-slicing strategy to enhance the ability to extract features from small craters.To handle redundant predictions,we proposed a new Non-Maximum Suppression with Area Filtering method to fuse the results in overlapping targets within the multi-scale submodels.Finally,our new MSF-CDA method achieved high detection performance,with the Precision,Recall,and F1 score having values of 0.991,0.987,and 0.989,respectively,perfectly addressing the problems induced by the lesser features and sample imbalance of small craters.Our MSF-CDA can provide strong data support for more in-depth study of the geological evolution of the lunar surface and finer geological age estimations.This strategy can also be used to detect other small objects with lesser features and sample imbalance problems.We detected approximately 500,000 impact craters in an area of approximately 214 km2 around the CE-4 landing area.By statistically analyzing the new data,we updated the distribution function of the number and diameter of impact craters.Finally,we identified the most suitable lighting conditions for detecting impact crater targets by analyzing the effect of different lighting conditions on the detection accuracy.展开更多
The traditional A^(*)algorithm exhibits a low efficiency in the path planning of unmanned surface vehicles(USVs).In addition,the path planned presents numerous redundant inflection waypoints,and the security is low,wh...The traditional A^(*)algorithm exhibits a low efficiency in the path planning of unmanned surface vehicles(USVs).In addition,the path planned presents numerous redundant inflection waypoints,and the security is low,which is not conducive to the control of USV and also affects navigation safety.In this paper,these problems were addressed through the following improvements.First,the path search angle and security were comprehensively considered,and a security expansion strategy of nodes based on the 5×5 neighborhood was proposed.The A^(*)algorithm search neighborhood was expanded from 3×3 to 5×5,and safe nodes were screened out for extension via the node security expansion strategy.This algorithm can also optimize path search angles while improving path security.Second,the distance from the current node to the target node was introduced into the heuristic function.The efficiency of the A^(*)algorithm was improved,and the path was smoothed using the Floyd algorithm.For the dynamic adjustment of the weight to improve the efficiency of DWA,the distance from the USV to the target point was introduced into the evaluation function of the dynamic-window approach(DWA)algorithm.Finally,combined with the local target point selection strategy,the optimized DWA algorithm was performed for local path planning.The experimental results show the smooth and safe path planned by the fusion algorithm,which can successfully avoid dynamic obstacles and is effective and feasible in path planning for USVs.展开更多
In the deployment of wireless networks in two-dimensional outdoor campus spaces,aiming at the problem of efficient coverage of the monitoring area by limited number of access points(APs),this paper proposes a deployme...In the deployment of wireless networks in two-dimensional outdoor campus spaces,aiming at the problem of efficient coverage of the monitoring area by limited number of access points(APs),this paper proposes a deployment method of multi-objective optimization with virtual force fusion bat algorithm(VFBA)using the classical four-node regular distribution as an entry point.The introduction of Lévy flight strategy for bat position updating helps to maintain the population diversity,reduce the premature maturity problem caused by population convergence,avoid the over aggregation of individuals in the local optimal region,and enhance the superiority in global search;the virtual force algorithm simulates the attraction and repulsion between individuals,which enables individual bats to precisely locate the optimal solution within the search space.At the same time,the fusion effect of virtual force prompts the bat individuals to move faster to the potential optimal solution.To validate the effectiveness of the fusion algorithm,the benchmark test function is selected for simulation testing.Finally,the simulation result verifies that the VFBA achieves superior coverage and effectively reduces node redundancy compared to the other three regular layout methods.The VFBA also shows better coverage results when compared to other optimization algorithms.展开更多
In disaster relief operations,multiple UAVs can be used to search for trapped people.In recent years,many researchers have proposed machine le arning-based algorithms,sampling-based algorithms,and heuristic algorithms...In disaster relief operations,multiple UAVs can be used to search for trapped people.In recent years,many researchers have proposed machine le arning-based algorithms,sampling-based algorithms,and heuristic algorithms to solve the problem of multi-UAV path planning.The Dung Beetle Optimization(DBO)algorithm has been widely applied due to its diverse search patterns in the above algorithms.However,the update strategies for the rolling and thieving dung beetles of the DBO algorithm are overly simplistic,potentially leading to an inability to fully explore the search space and a tendency to converge to local optima,thereby not guaranteeing the discovery of the optimal path.To address these issues,we propose an improved DBO algorithm guided by the Landmark Operator(LODBO).Specifically,we first use tent mapping to update the population strategy,which enables the algorithm to generate initial solutions with enhanced diversity within the search space.Second,we expand the search range of the rolling ball dung beetle by using the landmark factor.Finally,by using the adaptive factor that changes with the number of iterations.,we improve the global search ability of the stealing dung beetle,making it more likely to escape from local optima.To verify the effectiveness of the proposed method,extensive simulation experiments are conducted,and the result shows that the LODBO algorithm can obtain the optimal path using the shortest time compared with the Genetic Algorithm(GA),the Gray Wolf Optimizer(GWO),the Whale Optimization Algorithm(WOA)and the original DBO algorithm in the disaster search and rescue task set.展开更多
Accurate short-term wind power forecast technique plays a crucial role in maintaining the safety and economic efficiency of smart grids.Although numerous studies have employed various methods to forecast wind power,th...Accurate short-term wind power forecast technique plays a crucial role in maintaining the safety and economic efficiency of smart grids.Although numerous studies have employed various methods to forecast wind power,there remains a research gap in leveraging swarm intelligence algorithms to optimize the hyperparameters of the Transformer model for wind power prediction.To improve the accuracy of short-term wind power forecast,this paper proposes a hybrid short-term wind power forecast approach named STL-IAOA-iTransformer,which is based on seasonal and trend decomposition using LOESS(STL)and iTransformer model optimized by improved arithmetic optimization algorithm(IAOA).First,to fully extract the power data features,STL is used to decompose the original data into components with less redundant information.The extracted components as well as the weather data are then input into iTransformer for short-term wind power forecast.The final predicted short-term wind power curve is obtained by combining the predicted components.To improve the model accuracy,IAOA is employed to optimize the hyperparameters of iTransformer.The proposed approach is validated using real-generation data from different seasons and different power stations inNorthwest China,and ablation experiments have been conducted.Furthermore,to validate the superiority of the proposed approach under different wind characteristics,real power generation data fromsouthwestChina are utilized for experiments.Thecomparative results with the other six state-of-the-art prediction models in experiments show that the proposed model well fits the true value of generation series and achieves high prediction accuracy.展开更多
Previous studies have shown that deep learning is very effective in detecting known attacks.However,when facing unknown attacks,models such as Deep Neural Networks(DNN)combined with Long Short-Term Memory(LSTM),Convol...Previous studies have shown that deep learning is very effective in detecting known attacks.However,when facing unknown attacks,models such as Deep Neural Networks(DNN)combined with Long Short-Term Memory(LSTM),Convolutional Neural Networks(CNN)combined with LSTM,and so on are built by simple stacking,which has the problems of feature loss,low efficiency,and low accuracy.Therefore,this paper proposes an autonomous detectionmodel for Distributed Denial of Service attacks,Multi-Scale Convolutional Neural Network-Bidirectional Gated Recurrent Units-Single Headed Attention(MSCNN-BiGRU-SHA),which is based on a Multistrategy Integrated Zebra Optimization Algorithm(MI-ZOA).The model undergoes training and testing with the CICDDoS2019 dataset,and its performance is evaluated on a new GINKS2023 dataset.The hyperparameters for Conv_filter and GRU_unit are optimized using the Multi-strategy Integrated Zebra Optimization Algorithm(MIZOA).The experimental results show that the test accuracy of the MSCNN-BiGRU-SHA model based on the MIZOA proposed in this paper is as high as 0.9971 in the CICDDoS 2019 dataset.The evaluation accuracy of the new dataset GINKS2023 created in this paper is 0.9386.Compared to the MSCNN-BiGRU-SHA model based on the Zebra Optimization Algorithm(ZOA),the detection accuracy on the GINKS2023 dataset has improved by 5.81%,precisionhas increasedby 1.35%,the recallhas improvedby 9%,and theF1scorehas increasedby 5.55%.Compared to the MSCNN-BiGRU-SHA models developed using Grid Search,Random Search,and Bayesian Optimization,the MSCNN-BiGRU-SHA model optimized with the MI-ZOA exhibits better performance in terms of accuracy,precision,recall,and F1 score.展开更多
In this paper,we prove that Euclid's algorithm,Bezout's equation and Divi-sion algorithm are equivalent to each other.Our result shows that Euclid has preliminarily established the theory of divisibility and t...In this paper,we prove that Euclid's algorithm,Bezout's equation and Divi-sion algorithm are equivalent to each other.Our result shows that Euclid has preliminarily established the theory of divisibility and the greatest common divisor.We further provided several suggestions for teaching.展开更多
In order to improve the efficiency of cloud-based web services,an improved plant growth simulation algorithm scheduling model.This model first used mathematical methods to describe the relationships between cloud-base...In order to improve the efficiency of cloud-based web services,an improved plant growth simulation algorithm scheduling model.This model first used mathematical methods to describe the relationships between cloud-based web services and the constraints of system resources.Then,a light-induced plant growth simulation algorithm was established.The performance of the algorithm was compared through several plant types,and the best plant model was selected as the setting for the system.Experimental results show that when the number of test cloud-based web services reaches 2048,the model being 2.14 times faster than PSO,2.8 times faster than the ant colony algorithm,2.9 times faster than the bee colony algorithm,and a remarkable 8.38 times faster than the genetic algorithm.展开更多
The advent of microgrids in modern energy systems heralds a promising era of resilience,sustainability,and efficiency.Within the realm of grid-tied microgrids,the selection of an optimal optimization algorithm is crit...The advent of microgrids in modern energy systems heralds a promising era of resilience,sustainability,and efficiency.Within the realm of grid-tied microgrids,the selection of an optimal optimization algorithm is critical for effective energy management,particularly in economic dispatching.This study compares the performance of Particle Swarm Optimization(PSO)and Genetic Algorithms(GA)in microgrid energy management systems,implemented using MATLAB tools.Through a comprehensive review of the literature and sim-ulations conducted in MATLAB,the study analyzes performance metrics,convergence speed,and the overall efficacy of GA and PSO,with a focus on economic dispatching tasks.Notably,a significant distinction emerges between the cost curves generated by the two algo-rithms for microgrid operation,with the PSO algorithm consistently resulting in lower costs due to its effective economic dispatching capabilities.Specifically,the utilization of the PSO approach could potentially lead to substantial savings on the power bill,amounting to approximately$15.30 in this evaluation.Thefindings provide insights into the strengths and limitations of each algorithm within the complex dynamics of grid-tied microgrids,thereby assisting stakeholders and researchers in arriving at informed decisions.This study contributes to the discourse on sustainable energy management by offering actionable guidance for the advancement of grid-tied micro-grid technologies through MATLAB-implemented optimization algorithms.展开更多
Parameter extraction of photovoltaic(PV)models is crucial for the planning,optimization,and control of PV systems.Although some methods using meta-heuristic algorithms have been proposed to determine these parameters,...Parameter extraction of photovoltaic(PV)models is crucial for the planning,optimization,and control of PV systems.Although some methods using meta-heuristic algorithms have been proposed to determine these parameters,the robustness of solutions obtained by these methods faces great challenges when the complexity of the PV model increases.The unstable results will affect the reliable operation and maintenance strategies of PV systems.In response to this challenge,an improved rime optimization algorithm with enhanced exploration and exploitation,termed TERIME,is proposed for robust and accurate parameter identification for various PV models.Specifically,the differential evolution mutation operator is integrated in the exploration phase to enhance the population diversity.Meanwhile,a new exploitation strategy incorporating randomization and neighborhood strategies simultaneously is developed to maintain the balance of exploitation width and depth.The TERIME algorithm is applied to estimate the optimal parameters of the single diode model,double diode model,and triple diode model combined with the Lambert-W function for three PV cell and module types including RTC France,Photo Watt-PWP 201 and S75.According to the statistical analysis in 100 runs,the proposed algorithm achieves more accurate and robust parameter estimations than other techniques to various PV models in varying environmental conditions.All of our source codes are publicly available at https://github.com/dirge1/TERIME.展开更多
The Internet of Things(IoT)has emerged as an important future technology.IoT-Fog is a new computing paradigm that processes IoT data on servers close to the source of the data.In IoT-Fog computing,resource allocation ...The Internet of Things(IoT)has emerged as an important future technology.IoT-Fog is a new computing paradigm that processes IoT data on servers close to the source of the data.In IoT-Fog computing,resource allocation and independent task scheduling aim to deliver short response time services demanded by the IoT devices and performed by fog servers.The heterogeneity of the IoT-Fog resources and the huge amount of data that needs to be processed by the IoT-Fog tasks make scheduling fog computing tasks a challenging problem.This study proposes an Adaptive Firefly Algorithm(AFA)for dependent task scheduling in IoT-Fog computing.The proposed AFA is a modified version of the standard Firefly Algorithm(FA),considering the execution times of the submitted tasks,the impact of synchronization requirements,and the communication time between dependent tasks.As IoT-Fog computing depends mainly on distributed fog node servers that receive tasks in a dynamic manner,tackling the communications and synchronization issues between dependent tasks is becoming a challenging problem.The proposed AFA aims to address the dynamic nature of IoT-Fog computing environments.The proposed AFA mechanism considers a dynamic light absorption coefficient to control the decrease in attractiveness over iterations.The proposed AFA mechanism performance was benchmarked against the standard Firefly Algorithm(FA),Puma Optimizer(PO),Genetic Algorithm(GA),and Ant Colony Optimization(ACO)through simulations under light,typical,and heavy workload scenarios.In heavy workloads,the proposed AFA mechanism obtained the shortest average execution time,968.98 ms compared to 970.96,1352.87,1247.28,and 1773.62 of FA,PO,GA,and ACO,respectively.The simulation results demonstrate the proposed AFA’s ability to rapidly converge to optimal solutions,emphasizing its adaptability and efficiency in typical and heavy workloads.展开更多
This study presents a real-time tracking algorithm derived from the retina algorithm,designed for the rapid,real-time tracking of straight-line particle trajectories.These trajectories are detected by pixel detectors ...This study presents a real-time tracking algorithm derived from the retina algorithm,designed for the rapid,real-time tracking of straight-line particle trajectories.These trajectories are detected by pixel detectors to localize single-event effects in two-dimensional space.Initially,we developed a retina algorithm to track the trajectory of a single heavy ion and achieved a positional accuracy of 40μm.This was accomplished by analyzing trajectory samples from the simulations using a pixel sensor with a 72×72 pixel array and an 83μm pixel pitch.Subsequently,we refined this approach to create an iterative retina algorithm for tracking multiple heavy-ion trajectories in single events.This iterative version demonstrated a tracking efficiency of over 97%,with a positional resolution comparable to that of single-track events.Furthermore,it exhibits significant parallelism,requires fewer resources,and is ideally suited for implementation in field-programmable gate arrays on board-level systems,facilitating real-time online trajectory tracking.展开更多
Nowadays,abnormal traffic detection for Software-Defined Networking(SDN)faces the challenges of large data volume and high dimensionality.Since traditional machine learning-based detection methods have the problem of ...Nowadays,abnormal traffic detection for Software-Defined Networking(SDN)faces the challenges of large data volume and high dimensionality.Since traditional machine learning-based detection methods have the problem of data redundancy,the Metaheuristic Algorithm(MA)is introduced to select features beforemachine learning to reduce the dimensionality of data.Since a Tyrannosaurus Optimization Algorithm(TROA)has the advantages of few parameters,simple implementation,and fast convergence,and it shows better results in feature selection,TROA can be applied to abnormal traffic detection for SDN.However,TROA suffers frominsufficient global search capability,is easily trapped in local optimums,and has poor search accuracy.Then,this paper tries to improve TROA,namely the Improved Tyrannosaurus Optimization Algorithm(ITROA).It proposes a metaheuristic-driven abnormal traffic detection model for SDN based on ITROA.Finally,the validity of the ITROA is verified by the benchmark function and the UCI dataset,and the feature selection optimization operation is performed on the InSDN dataset by ITROA and other MAs to obtain the optimized feature subset for SDN abnormal traffic detection.The experiment shows that the performance of the proposed ITROA outperforms compared MAs in terms of the metaheuristic-driven model for SDN,achieving an accuracy of 99.37%on binary classification and 96.73%on multiclassification.展开更多
文摘Optimization is the key to obtaining efficient utilization of resources in structural design.Due to the complex nature of truss systems,this study presents a method based on metaheuristic modelling that minimises structural weight under stress and frequency constraints.Two new algorithms,the Red Kite Optimization Algorithm(ROA)and Secretary Bird Optimization Algorithm(SBOA),are utilized on five benchmark trusses with 10,18,37,72,and 200-bar trusses.Both algorithms are evaluated against benchmarks in the literature.The results indicate that SBOA always reaches a lighter optimal.Designs with reducing structural weight ranging from 0.02%to 0.15%compared to ROA,and up to 6%–8%as compared to conventional algorithms.In addition,SBOA can achieve 15%–20%faster convergence speed and 10%–18%reduction in computational time with a smaller standard deviation over independent runs,which demonstrates its robustness and reliability.It is indicated that the adaptive exploration mechanism of SBOA,especially its Levy flight–based search strategy,can obviously improve optimization performance for low-and high-dimensional trusses.The research has implications in the context of promoting bio-inspired optimization techniques by demonstrating the viability of SBOA,a reliable model for large-scale structural design that provides significant enhancements in performance and convergence behavior.
文摘Aiming to solve the steering instability and hysteresis of agricultural robots in the process of movement,a fusion PID control method of particle swarm optimization(PSO)and genetic algorithm(GA)was proposed.The fusion algorithm took advantage of the fast optimization ability of PSO to optimize the population screening link of GA.The Simulink simulation results showed that the convergence of the fitness function of the fusion algorithm was accelerated,the system response adjustment time was reduced,and the overshoot was almost zero.Then the algorithm was applied to the steering test of agricultural robot in various scenes.After modeling the steering system of agricultural robot,the steering test results in the unloaded suspended state showed that the PID control based on fusion algorithm reduced the rise time,response adjustment time and overshoot of the system,and improved the response speed and stability of the system,compared with the artificial trial and error PID control and the PID control based on GA.The actual road steering test results showed that the PID control response rise time based on the fusion algorithm was the shortest,about 4.43 s.When the target pulse number was set to 100,the actual mean value in the steady-state regulation stage was about 102.9,which was the closest to the target value among the three control methods,and the overshoot was reduced at the same time.The steering test results under various scene states showed that the PID control based on the proposed fusion algorithm had good anti-interference ability,it can adapt to the changes of environment and load and improve the performance of the control system.It was effective in the steering control of agricultural robot.This method can provide a reference for the precise steering control of other robots.
基金supported by the P.G.Senapathy Center for Computing Resources at IIT Madrasfunding provided by the Ministry of Education,Government of Indiasupported by the National Natural Science Foundation of China(Grant Nos.12388101,12472224 and 92252104).
文摘This paper presents an Eulerian-Lagrangian algorithm for direct numerical simulation(DNS)of particle-laden flows.The algorithm is applicable to perform simulations of dilute suspensions of small inertial particles in turbulent carrier flow.The Eulerian framework numerically resolves turbulent carrier flow using a parallelized,finite-volume DNS solver on a staggered Cartesian grid.Particles are tracked using a point-particle method utilizing a Lagrangian particle tracking(LPT)algorithm.The proposed Eulerian-Lagrangian algorithm is validated using an inertial particle-laden turbulent channel flow for different Stokes number cases.The particle concentration profiles and higher-order statistics of the carrier and dispersed phases agree well with the benchmark results.We investigated the effect of fluid velocity interpolation and numerical integration schemes of particle tracking algorithms on particle dispersion statistics.The suitability of fluid velocity interpolation schemes for predicting the particle dispersion statistics is discussed in the framework of the particle tracking algorithm coupled to the finite-volume solver.In addition,we present parallelization strategies implemented in the algorithm and evaluate their parallel performance.
基金National Key Research and Development Program of China,No.2023YFC3006704National Natural Science Foundation of China,No.42171047CAS-CSIRO Partnership Joint Project of 2024,No.177GJHZ2023097MI。
文摘Accurate prediction of flood events is important for flood control and risk management.Machine learning techniques contributed greatly to advances in flood predictions,and existing studies mainly focused on predicting flood resource variables using single or hybrid machine learning techniques.However,class-based flood predictions have rarely been investigated,which can aid in quickly diagnosing comprehensive flood characteristics and proposing targeted management strategies.This study proposed a prediction approach of flood regime metrics and event classes coupling machine learning algorithms with clustering-deduced membership degrees.Five algorithms were adopted for this exploration.Results showed that the class membership degrees accurately determined event classes with class hit rates up to 100%,compared with the four classes clustered from nine regime metrics.The nonlinear algorithms(Multiple Linear Regression,Random Forest,and least squares-Support Vector Machine)outperformed the linear techniques(Multiple Linear Regression and Stepwise Regression)in predicting flood regime metrics.The proposed approach well predicted flood event classes with average class hit rates of 66.0%-85.4%and 47.2%-76.0%in calibration and validation periods,respectively,particularly for the slow and late flood events.The predictive capability of the proposed prediction approach for flood regime metrics and classes was considerably stronger than that of hydrological modeling approach.
基金supported by the National Natural Science Foundation of China(Grant No.12402139,No.52368070)supported by Hainan Provincial Natural Science Foundation of China(Grant No.524QN223)+3 种基金Scientific Research Startup Foundation of Hainan University(Grant No.RZ2300002710)State Key Laboratory of Structural Analysis,Optimization and CAE Software for Industrial Equipment,Dalian University of Technology(Grant No.GZ24107)the Horizontal Research Project(Grant No.HD-KYH-2024022)Innovative Research Projects for Postgraduate Students in Hainan Province(Grant No.Hys2025-217).
文摘Optimization problems are prevalent in various fields of science and engineering,with several real-world applications characterized by high dimensionality and complex search landscapes.Starfish optimization algorithm(SFOA)is a recently optimizer inspired by swarm intelligence,which is effective for numerical optimization,but it may encounter premature and local convergence for complex optimization problems.To address these challenges,this paper proposes the multi-strategy enhanced crested porcupine-starfish optimization algorithm(MCPSFOA).The core innovation of MCPSFOA lies in employing a hybrid strategy to improve SFOA,which integrates the exploratory mechanisms of SFOA with the diverse search capacity of the Crested Porcupine Optimizer(CPO).This synergy enhances MCPSFOA’s ability to navigate complex and multimodal search spaces.To further prevent premature convergence,MCPSFOA incorporates Lévy flight,leveraging its characteristic long and short jump patterns to enable large-scale exploration and escape from local optima.Subsequently,Gaussian mutation is applied for precise solution tuning,introducing controlled perturbations that enhance accuracy and mitigate the risk of insufficient exploitation.Notably,the population diversity enhancement mechanism periodically identifies and resets stagnant individuals,thereby consistently revitalizing population variety throughout the optimization process.MCPSFOA is rigorously evaluated on 24 classical benchmark functions(including high-dimensional cases),the CEC2017 suite,and the CEC2022 suite.MCPSFOA achieves superior overall performance with Friedman mean ranks of 2.208,2.310 and 2.417 on these benchmark functions,outperforming 11 state-of-the-art algorithms.Furthermore,the practical applicability of MCPSFOA is confirmed through its successful application to five engineering optimization cases,where it also yields excellent results.In conclusion,MCPSFOA is not only a highly effective and reliable optimizer for benchmark functions,but also a practical tool for solving real-world optimization problems.
基金supported by the Major Science and Technology Programs in Henan Province(No.241100210100)Henan Provincial Science and Technology Research Project(No.252102211085,No.252102211105)+3 种基金Endogenous Security Cloud Network Convergence R&D Center(No.602431011PQ1)The Special Project for Research and Development in Key Areas of Guangdong Province(No.2021ZDZX1098)The Stabilization Support Program of Science,Technology and Innovation Commission of Shenzhen Municipality(No.20231128083944001)The Key scientific research projects of Henan higher education institutions(No.24A520042).
文摘Existing feature selection methods for intrusion detection systems in the Industrial Internet of Things often suffer from local optimality and high computational complexity.These challenges hinder traditional IDS from effectively extracting features while maintaining detection accuracy.This paper proposes an industrial Internet ofThings intrusion detection feature selection algorithm based on an improved whale optimization algorithm(GSLDWOA).The aim is to address the problems that feature selection algorithms under high-dimensional data are prone to,such as local optimality,long detection time,and reduced accuracy.First,the initial population’s diversity is increased using the Gaussian Mutation mechanism.Then,Non-linear Shrinking Factor balances global exploration and local development,avoiding premature convergence.Lastly,Variable-step Levy Flight operator and Dynamic Differential Evolution strategy are introduced to improve the algorithm’s search efficiency and convergence accuracy in highdimensional feature space.Experiments on the NSL-KDD and WUSTL-IIoT-2021 datasets demonstrate that the feature subset selected by GSLDWOA significantly improves detection performance.Compared to the traditional WOA algorithm,the detection rate and F1-score increased by 3.68%and 4.12%.On the WUSTL-IIoT-2021 dataset,accuracy,recall,and F1-score all exceed 99.9%.
基金funded by United Arab Emirates University(UAEU)under the UAEU-AUA grant number G00004577(12N145)with the corresponding grant at Universiti Malaya(UM)under grant number IF019-2024.
文摘Concrete-filled steel tubes(CFST)are widely utilized in civil engineering due to their superior load-bearing capacity,ductility,and seismic resistance.However,existing design codes,such as AISC and Eurocode 4,tend to be excessively conservative as they fail to account for the composite action between the steel tube and the concrete core.To address this limitation,this study proposes a hybrid model that integrates XGBoost with the Pied Kingfisher Optimizer(PKO),a nature-inspired algorithm,to enhance the accuracy of shear strength prediction for CFST columns.Additionally,quantile regression is employed to construct prediction intervals for the ultimate shear force,while the Asymmetric Squared Error Loss(ASEL)function is incorporated to mitigate overestimation errors.The computational results demonstrate that the PKO-XGBoost model delivers superior predictive accuracy,achieving a Mean Absolute Percentage Error(MAPE)of 4.431%and R2 of 0.9925 on the test set.Furthermore,the ASEL-PKO-XGBoost model substantially reduces overestimation errors to 28.26%,with negligible impact on predictive performance.Additionally,based on the Genetic Algorithm(GA)and existing equation models,a strength equation model is developed,achieving markedly higher accuracy than existing models(R^(2)=0.934).Lastly,web-based Graphical User Interfaces(GUIs)were developed to enable real-time prediction.
基金the National Key Research and Development Program of China (Grant No.2022YFF0711400)the National Space Science Data Center Youth Open Project (Grant No. NSSDC2302001)
文摘Impact craters are important for understanding the evolution of lunar geologic and surface erosion rates,among other functions.However,the morphological characteristics of these micro impact craters are not obvious and they are numerous,resulting in low detection accuracy by deep learning models.Therefore,we proposed a new multi-scale fusion crater detection algorithm(MSF-CDA)based on the YOLO11 to improve the accuracy of lunar impact crater detection,especially for small craters with a diameter of<1 km.Using the images taken by the LROC(Lunar Reconnaissance Orbiter Camera)at the Chang’e-4(CE-4)landing area,we constructed three separate datasets for craters with diameters of 0-70 m,70-140 m,and>140 m.We then trained three submodels separately with these three datasets.Additionally,we designed a slicing-amplifying-slicing strategy to enhance the ability to extract features from small craters.To handle redundant predictions,we proposed a new Non-Maximum Suppression with Area Filtering method to fuse the results in overlapping targets within the multi-scale submodels.Finally,our new MSF-CDA method achieved high detection performance,with the Precision,Recall,and F1 score having values of 0.991,0.987,and 0.989,respectively,perfectly addressing the problems induced by the lesser features and sample imbalance of small craters.Our MSF-CDA can provide strong data support for more in-depth study of the geological evolution of the lunar surface and finer geological age estimations.This strategy can also be used to detect other small objects with lesser features and sample imbalance problems.We detected approximately 500,000 impact craters in an area of approximately 214 km2 around the CE-4 landing area.By statistically analyzing the new data,we updated the distribution function of the number and diameter of impact craters.Finally,we identified the most suitable lighting conditions for detecting impact crater targets by analyzing the effect of different lighting conditions on the detection accuracy.
基金Supported by the EDD of China(No.80912020104)the Science and Technology Commission of Shanghai Municipality(No.22ZR1427700 and No.23692106900).
文摘The traditional A^(*)algorithm exhibits a low efficiency in the path planning of unmanned surface vehicles(USVs).In addition,the path planned presents numerous redundant inflection waypoints,and the security is low,which is not conducive to the control of USV and also affects navigation safety.In this paper,these problems were addressed through the following improvements.First,the path search angle and security were comprehensively considered,and a security expansion strategy of nodes based on the 5×5 neighborhood was proposed.The A^(*)algorithm search neighborhood was expanded from 3×3 to 5×5,and safe nodes were screened out for extension via the node security expansion strategy.This algorithm can also optimize path search angles while improving path security.Second,the distance from the current node to the target node was introduced into the heuristic function.The efficiency of the A^(*)algorithm was improved,and the path was smoothed using the Floyd algorithm.For the dynamic adjustment of the weight to improve the efficiency of DWA,the distance from the USV to the target point was introduced into the evaluation function of the dynamic-window approach(DWA)algorithm.Finally,combined with the local target point selection strategy,the optimized DWA algorithm was performed for local path planning.The experimental results show the smooth and safe path planned by the fusion algorithm,which can successfully avoid dynamic obstacles and is effective and feasible in path planning for USVs.
基金supported in part by the National Natural Science Foundation of China under Grant No.62271453in part by the National Natural Science Foundation of China No.62101512+2 种基金in part by the Central Support for Local Projects under Grant No.YDZJSX2024D031in part by Project supported by the Shanxi Provincial Foundation for Leaders of Disciplines in Science,China under Grant No.2024Q022in part by Shanxi Province Patent Conversion Special Plan Funding Projects under Grant No.202405004。
文摘In the deployment of wireless networks in two-dimensional outdoor campus spaces,aiming at the problem of efficient coverage of the monitoring area by limited number of access points(APs),this paper proposes a deployment method of multi-objective optimization with virtual force fusion bat algorithm(VFBA)using the classical four-node regular distribution as an entry point.The introduction of Lévy flight strategy for bat position updating helps to maintain the population diversity,reduce the premature maturity problem caused by population convergence,avoid the over aggregation of individuals in the local optimal region,and enhance the superiority in global search;the virtual force algorithm simulates the attraction and repulsion between individuals,which enables individual bats to precisely locate the optimal solution within the search space.At the same time,the fusion effect of virtual force prompts the bat individuals to move faster to the potential optimal solution.To validate the effectiveness of the fusion algorithm,the benchmark test function is selected for simulation testing.Finally,the simulation result verifies that the VFBA achieves superior coverage and effectively reduces node redundancy compared to the other three regular layout methods.The VFBA also shows better coverage results when compared to other optimization algorithms.
基金supported by the National Natural Science Foundation of China(No.62373027).
文摘In disaster relief operations,multiple UAVs can be used to search for trapped people.In recent years,many researchers have proposed machine le arning-based algorithms,sampling-based algorithms,and heuristic algorithms to solve the problem of multi-UAV path planning.The Dung Beetle Optimization(DBO)algorithm has been widely applied due to its diverse search patterns in the above algorithms.However,the update strategies for the rolling and thieving dung beetles of the DBO algorithm are overly simplistic,potentially leading to an inability to fully explore the search space and a tendency to converge to local optima,thereby not guaranteeing the discovery of the optimal path.To address these issues,we propose an improved DBO algorithm guided by the Landmark Operator(LODBO).Specifically,we first use tent mapping to update the population strategy,which enables the algorithm to generate initial solutions with enhanced diversity within the search space.Second,we expand the search range of the rolling ball dung beetle by using the landmark factor.Finally,by using the adaptive factor that changes with the number of iterations.,we improve the global search ability of the stealing dung beetle,making it more likely to escape from local optima.To verify the effectiveness of the proposed method,extensive simulation experiments are conducted,and the result shows that the LODBO algorithm can obtain the optimal path using the shortest time compared with the Genetic Algorithm(GA),the Gray Wolf Optimizer(GWO),the Whale Optimization Algorithm(WOA)and the original DBO algorithm in the disaster search and rescue task set.
基金supported by Yunnan Provincial Basic Research Project(202401AT070344,202301AT070443)National Natural Science Foundation of China(62263014,52207105)+1 种基金Yunnan Lancang-Mekong International Electric Power Technology Joint Laboratory(202203AP140001)Major Science and Technology Projects in Yunnan Province(202402AG050006).
文摘Accurate short-term wind power forecast technique plays a crucial role in maintaining the safety and economic efficiency of smart grids.Although numerous studies have employed various methods to forecast wind power,there remains a research gap in leveraging swarm intelligence algorithms to optimize the hyperparameters of the Transformer model for wind power prediction.To improve the accuracy of short-term wind power forecast,this paper proposes a hybrid short-term wind power forecast approach named STL-IAOA-iTransformer,which is based on seasonal and trend decomposition using LOESS(STL)and iTransformer model optimized by improved arithmetic optimization algorithm(IAOA).First,to fully extract the power data features,STL is used to decompose the original data into components with less redundant information.The extracted components as well as the weather data are then input into iTransformer for short-term wind power forecast.The final predicted short-term wind power curve is obtained by combining the predicted components.To improve the model accuracy,IAOA is employed to optimize the hyperparameters of iTransformer.The proposed approach is validated using real-generation data from different seasons and different power stations inNorthwest China,and ablation experiments have been conducted.Furthermore,to validate the superiority of the proposed approach under different wind characteristics,real power generation data fromsouthwestChina are utilized for experiments.Thecomparative results with the other six state-of-the-art prediction models in experiments show that the proposed model well fits the true value of generation series and achieves high prediction accuracy.
基金supported by Science and Technology Innovation Programfor Postgraduate Students in IDP Subsidized by Fundamental Research Funds for the Central Universities(Project No.ZY20240335)support of the Research Project of the Key Technology of Malicious Code Detection Based on Data Mining in APT Attack(Project No.2022IT173)the Research Project of the Big Data Sensitive Information Supervision Technology Based on Convolutional Neural Network(Project No.2022011033).
文摘Previous studies have shown that deep learning is very effective in detecting known attacks.However,when facing unknown attacks,models such as Deep Neural Networks(DNN)combined with Long Short-Term Memory(LSTM),Convolutional Neural Networks(CNN)combined with LSTM,and so on are built by simple stacking,which has the problems of feature loss,low efficiency,and low accuracy.Therefore,this paper proposes an autonomous detectionmodel for Distributed Denial of Service attacks,Multi-Scale Convolutional Neural Network-Bidirectional Gated Recurrent Units-Single Headed Attention(MSCNN-BiGRU-SHA),which is based on a Multistrategy Integrated Zebra Optimization Algorithm(MI-ZOA).The model undergoes training and testing with the CICDDoS2019 dataset,and its performance is evaluated on a new GINKS2023 dataset.The hyperparameters for Conv_filter and GRU_unit are optimized using the Multi-strategy Integrated Zebra Optimization Algorithm(MIZOA).The experimental results show that the test accuracy of the MSCNN-BiGRU-SHA model based on the MIZOA proposed in this paper is as high as 0.9971 in the CICDDoS 2019 dataset.The evaluation accuracy of the new dataset GINKS2023 created in this paper is 0.9386.Compared to the MSCNN-BiGRU-SHA model based on the Zebra Optimization Algorithm(ZOA),the detection accuracy on the GINKS2023 dataset has improved by 5.81%,precisionhas increasedby 1.35%,the recallhas improvedby 9%,and theF1scorehas increasedby 5.55%.Compared to the MSCNN-BiGRU-SHA models developed using Grid Search,Random Search,and Bayesian Optimization,the MSCNN-BiGRU-SHA model optimized with the MI-ZOA exhibits better performance in terms of accuracy,precision,recall,and F1 score.
基金Supported by the Natural Science Foundation of Chongqing(General Program,NO.CSTB2022NSCQ-MSX0884)Discipline Teaching Special Project of Yangtze Normal University(csxkjx14)。
文摘In this paper,we prove that Euclid's algorithm,Bezout's equation and Divi-sion algorithm are equivalent to each other.Our result shows that Euclid has preliminarily established the theory of divisibility and the greatest common divisor.We further provided several suggestions for teaching.
基金Shanxi Province Higher Education Science and Technology Innovation Fund Project(2022-676)Shanxi Soft Science Program Research Fund Project(2016041008-6)。
文摘In order to improve the efficiency of cloud-based web services,an improved plant growth simulation algorithm scheduling model.This model first used mathematical methods to describe the relationships between cloud-based web services and the constraints of system resources.Then,a light-induced plant growth simulation algorithm was established.The performance of the algorithm was compared through several plant types,and the best plant model was selected as the setting for the system.Experimental results show that when the number of test cloud-based web services reaches 2048,the model being 2.14 times faster than PSO,2.8 times faster than the ant colony algorithm,2.9 times faster than the bee colony algorithm,and a remarkable 8.38 times faster than the genetic algorithm.
文摘The advent of microgrids in modern energy systems heralds a promising era of resilience,sustainability,and efficiency.Within the realm of grid-tied microgrids,the selection of an optimal optimization algorithm is critical for effective energy management,particularly in economic dispatching.This study compares the performance of Particle Swarm Optimization(PSO)and Genetic Algorithms(GA)in microgrid energy management systems,implemented using MATLAB tools.Through a comprehensive review of the literature and sim-ulations conducted in MATLAB,the study analyzes performance metrics,convergence speed,and the overall efficacy of GA and PSO,with a focus on economic dispatching tasks.Notably,a significant distinction emerges between the cost curves generated by the two algo-rithms for microgrid operation,with the PSO algorithm consistently resulting in lower costs due to its effective economic dispatching capabilities.Specifically,the utilization of the PSO approach could potentially lead to substantial savings on the power bill,amounting to approximately$15.30 in this evaluation.Thefindings provide insights into the strengths and limitations of each algorithm within the complex dynamics of grid-tied microgrids,thereby assisting stakeholders and researchers in arriving at informed decisions.This study contributes to the discourse on sustainable energy management by offering actionable guidance for the advancement of grid-tied micro-grid technologies through MATLAB-implemented optimization algorithms.
基金supported by the National Natural Science Foundation of China[grant number 51775020]the Science Challenge Project[grant number.TZ2018007]+2 种基金the National Natural Science Foundation of China[grant number 62073009]the Postdoctoral Fellowship Program of CPSF[grant number GZC20233365]the Fundamental Research Funds for Central Universities[grant number JKF-20240559].
文摘Parameter extraction of photovoltaic(PV)models is crucial for the planning,optimization,and control of PV systems.Although some methods using meta-heuristic algorithms have been proposed to determine these parameters,the robustness of solutions obtained by these methods faces great challenges when the complexity of the PV model increases.The unstable results will affect the reliable operation and maintenance strategies of PV systems.In response to this challenge,an improved rime optimization algorithm with enhanced exploration and exploitation,termed TERIME,is proposed for robust and accurate parameter identification for various PV models.Specifically,the differential evolution mutation operator is integrated in the exploration phase to enhance the population diversity.Meanwhile,a new exploitation strategy incorporating randomization and neighborhood strategies simultaneously is developed to maintain the balance of exploitation width and depth.The TERIME algorithm is applied to estimate the optimal parameters of the single diode model,double diode model,and triple diode model combined with the Lambert-W function for three PV cell and module types including RTC France,Photo Watt-PWP 201 and S75.According to the statistical analysis in 100 runs,the proposed algorithm achieves more accurate and robust parameter estimations than other techniques to various PV models in varying environmental conditions.All of our source codes are publicly available at https://github.com/dirge1/TERIME.
基金the Deanship of Graduate Studies and Scientific Research at Najran University for funding this work under the Easy Funding Program grant code(NU/EFP/SERC/13/166).
文摘The Internet of Things(IoT)has emerged as an important future technology.IoT-Fog is a new computing paradigm that processes IoT data on servers close to the source of the data.In IoT-Fog computing,resource allocation and independent task scheduling aim to deliver short response time services demanded by the IoT devices and performed by fog servers.The heterogeneity of the IoT-Fog resources and the huge amount of data that needs to be processed by the IoT-Fog tasks make scheduling fog computing tasks a challenging problem.This study proposes an Adaptive Firefly Algorithm(AFA)for dependent task scheduling in IoT-Fog computing.The proposed AFA is a modified version of the standard Firefly Algorithm(FA),considering the execution times of the submitted tasks,the impact of synchronization requirements,and the communication time between dependent tasks.As IoT-Fog computing depends mainly on distributed fog node servers that receive tasks in a dynamic manner,tackling the communications and synchronization issues between dependent tasks is becoming a challenging problem.The proposed AFA aims to address the dynamic nature of IoT-Fog computing environments.The proposed AFA mechanism considers a dynamic light absorption coefficient to control the decrease in attractiveness over iterations.The proposed AFA mechanism performance was benchmarked against the standard Firefly Algorithm(FA),Puma Optimizer(PO),Genetic Algorithm(GA),and Ant Colony Optimization(ACO)through simulations under light,typical,and heavy workload scenarios.In heavy workloads,the proposed AFA mechanism obtained the shortest average execution time,968.98 ms compared to 970.96,1352.87,1247.28,and 1773.62 of FA,PO,GA,and ACO,respectively.The simulation results demonstrate the proposed AFA’s ability to rapidly converge to optimal solutions,emphasizing its adaptability and efficiency in typical and heavy workloads.
基金supported by the National Natural Science Foundation of China(No.12205224)the Research Foundation of Education Bureau of Hubei Province China(No.Q20221703)+1 种基金the National Natural Science Foundation of China(Nos.12035006,U2032140)the National Key Research and Development Program of China(No.2020YFE0202000)。
文摘This study presents a real-time tracking algorithm derived from the retina algorithm,designed for the rapid,real-time tracking of straight-line particle trajectories.These trajectories are detected by pixel detectors to localize single-event effects in two-dimensional space.Initially,we developed a retina algorithm to track the trajectory of a single heavy ion and achieved a positional accuracy of 40μm.This was accomplished by analyzing trajectory samples from the simulations using a pixel sensor with a 72×72 pixel array and an 83μm pixel pitch.Subsequently,we refined this approach to create an iterative retina algorithm for tracking multiple heavy-ion trajectories in single events.This iterative version demonstrated a tracking efficiency of over 97%,with a positional resolution comparable to that of single-track events.Furthermore,it exhibits significant parallelism,requires fewer resources,and is ideally suited for implementation in field-programmable gate arrays on board-level systems,facilitating real-time online trajectory tracking.
基金supported by the National Natural Science Foundation of China under Grant 61602162the Hubei Provincial Science and Technology Plan Project under Grant 2023BCB041.
文摘Nowadays,abnormal traffic detection for Software-Defined Networking(SDN)faces the challenges of large data volume and high dimensionality.Since traditional machine learning-based detection methods have the problem of data redundancy,the Metaheuristic Algorithm(MA)is introduced to select features beforemachine learning to reduce the dimensionality of data.Since a Tyrannosaurus Optimization Algorithm(TROA)has the advantages of few parameters,simple implementation,and fast convergence,and it shows better results in feature selection,TROA can be applied to abnormal traffic detection for SDN.However,TROA suffers frominsufficient global search capability,is easily trapped in local optimums,and has poor search accuracy.Then,this paper tries to improve TROA,namely the Improved Tyrannosaurus Optimization Algorithm(ITROA).It proposes a metaheuristic-driven abnormal traffic detection model for SDN based on ITROA.Finally,the validity of the ITROA is verified by the benchmark function and the UCI dataset,and the feature selection optimization operation is performed on the InSDN dataset by ITROA and other MAs to obtain the optimized feature subset for SDN abnormal traffic detection.The experiment shows that the performance of the proposed ITROA outperforms compared MAs in terms of the metaheuristic-driven model for SDN,achieving an accuracy of 99.37%on binary classification and 96.73%on multiclassification.