Complex network models are frequently employed for simulating and studyingdiverse real-world complex systems.Among these models,scale-free networks typically exhibit greater fragility to malicious attacks.Consequently...Complex network models are frequently employed for simulating and studyingdiverse real-world complex systems.Among these models,scale-free networks typically exhibit greater fragility to malicious attacks.Consequently,enhancing the robustness of scale-free networks has become a pressing issue.To address this problem,this paper proposes a Multi-Granularity Integration Algorithm(MGIA),which aims to improve the robustness of scale-free networks while keeping the initial degree of each node unchanged,ensuring network connectivity and avoiding the generation of multiple edges.The algorithm generates a multi-granularity structure from the initial network to be optimized,then uses different optimization strategies to optimize the networks at various granular layers in this structure,and finally realizes the information exchange between different granular layers,thereby further enhancing the optimization effect.We propose new network refresh,crossover,and mutation operators to ensure that the optimized network satisfies the given constraints.Meanwhile,we propose new network similarity and network dissimilarity evaluation metrics to improve the effectiveness of the optimization operators in the algorithm.In the experiments,the MGIA enhances the robustness of the scale-free network by 67.6%.This improvement is approximately 17.2%higher than the optimization effects achieved by eight currently existing complex network robustness optimization algorithms.展开更多
The advent of microgrids in modern energy systems heralds a promising era of resilience,sustainability,and efficiency.Within the realm of grid-tied microgrids,the selection of an optimal optimization algorithm is crit...The advent of microgrids in modern energy systems heralds a promising era of resilience,sustainability,and efficiency.Within the realm of grid-tied microgrids,the selection of an optimal optimization algorithm is critical for effective energy management,particularly in economic dispatching.This study compares the performance of Particle Swarm Optimization(PSO)and Genetic Algorithms(GA)in microgrid energy management systems,implemented using MATLAB tools.Through a comprehensive review of the literature and sim-ulations conducted in MATLAB,the study analyzes performance metrics,convergence speed,and the overall efficacy of GA and PSO,with a focus on economic dispatching tasks.Notably,a significant distinction emerges between the cost curves generated by the two algo-rithms for microgrid operation,with the PSO algorithm consistently resulting in lower costs due to its effective economic dispatching capabilities.Specifically,the utilization of the PSO approach could potentially lead to substantial savings on the power bill,amounting to approximately$15.30 in this evaluation.Thefindings provide insights into the strengths and limitations of each algorithm within the complex dynamics of grid-tied microgrids,thereby assisting stakeholders and researchers in arriving at informed decisions.This study contributes to the discourse on sustainable energy management by offering actionable guidance for the advancement of grid-tied micro-grid technologies through MATLAB-implemented optimization algorithms.展开更多
Software defect prediction(SDP)aims to find a reliable method to predict defects in specific software projects and help software engineers allocate limited resources to release high-quality software products.Software ...Software defect prediction(SDP)aims to find a reliable method to predict defects in specific software projects and help software engineers allocate limited resources to release high-quality software products.Software defect prediction can be effectively performed using traditional features,but there are some redundant or irrelevant features in them(the presence or absence of this feature has little effect on the prediction results).These problems can be solved using feature selection.However,existing feature selection methods have shortcomings such as insignificant dimensionality reduction effect and low classification accuracy of the selected optimal feature subset.In order to reduce the impact of these shortcomings,this paper proposes a new feature selection method Cubic TraverseMa Beluga whale optimization algorithm(CTMBWO)based on the improved Beluga whale optimization algorithm(BWO).The goal of this study is to determine how well the CTMBWO can extract the features that are most important for correctly predicting software defects,improve the accuracy of fault prediction,reduce the number of the selected feature and mitigate the risk of overfitting,thereby achieving more efficient resource utilization and better distribution of test workload.The CTMBWO comprises three main stages:preprocessing the dataset,selecting relevant features,and evaluating the classification performance of the model.The novel feature selection method can effectively improve the performance of SDP.This study performs experiments on two software defect datasets(PROMISE,NASA)and shows the method’s classification performance using four detailed evaluation metrics,Accuracy,F1-score,MCC,AUC and Recall.The results indicate that the approach presented in this paper achieves outstanding classification performance on both datasets and has significant improvement over the baseline models.展开更多
The liquid cooling system(LCS)of fuel cells is challenged by significant time delays,model uncertainties,pump and fan coupling,and frequent disturbances,leading to overshoot and control oscillations that degrade tempe...The liquid cooling system(LCS)of fuel cells is challenged by significant time delays,model uncertainties,pump and fan coupling,and frequent disturbances,leading to overshoot and control oscillations that degrade temperature regulation performance.To address these challenges,we propose a composite control scheme combining fuzzy logic and a variable-gain generalized supertwisting algorithm(VG-GSTA).Firstly,a one-dimensional(1D)fuzzy logic controler(FLC)for the pump ensures stable coolant flow,while a two-dimensional(2D)FLC for the fan regulates the stack temperature near the reference value.The VG-GSTA is then introduced to eliminate steady-state errors,offering resistance to disturbances and minimizing control oscillations.The equilibrium optimizer is used to fine-tune VG-GSTA parameters.Co-simulation verifies the effectiveness of our method,demonstrating its advantages in terms of disturbance immunity,overshoot suppression,tracking accuracy and response speed.展开更多
This paper presents an Eulerian-Lagrangian algorithm for direct numerical simulation(DNS)of particle-laden flows.The algorithm is applicable to perform simulations of dilute suspensions of small inertial particles in ...This paper presents an Eulerian-Lagrangian algorithm for direct numerical simulation(DNS)of particle-laden flows.The algorithm is applicable to perform simulations of dilute suspensions of small inertial particles in turbulent carrier flow.The Eulerian framework numerically resolves turbulent carrier flow using a parallelized,finite-volume DNS solver on a staggered Cartesian grid.Particles are tracked using a point-particle method utilizing a Lagrangian particle tracking(LPT)algorithm.The proposed Eulerian-Lagrangian algorithm is validated using an inertial particle-laden turbulent channel flow for different Stokes number cases.The particle concentration profiles and higher-order statistics of the carrier and dispersed phases agree well with the benchmark results.We investigated the effect of fluid velocity interpolation and numerical integration schemes of particle tracking algorithms on particle dispersion statistics.The suitability of fluid velocity interpolation schemes for predicting the particle dispersion statistics is discussed in the framework of the particle tracking algorithm coupled to the finite-volume solver.In addition,we present parallelization strategies implemented in the algorithm and evaluate their parallel performance.展开更多
Internet of things networks often suffer from early node failures and short lifespan due to energy limits.Traditional routing methods are not enough.This work proposes a new hybrid algorithm called ACOGA.It combines A...Internet of things networks often suffer from early node failures and short lifespan due to energy limits.Traditional routing methods are not enough.This work proposes a new hybrid algorithm called ACOGA.It combines Ant Colony Optimization(ACO)and the Greedy Algorithm(GA).ACO finds smart paths while Greedy makes quick decisions.This improves energy use and performance.ACOGA outperforms Hybrid Energy-Efficient(HEE)and Adaptive Lossless Data Compression(ALDC)algorithms.After 500 rounds,only 5%of ACOGA’s nodes are dead,compared to 15%for HEE and 20%for ALDC.The network using ACOGA runs for 1200 rounds before the first nodes fail.HEE lasts 900 rounds and ALDC only 850.ACOGA saves at least 15%more energy by better distributing the load.It also achieves a 98%packet delivery rate.The method works well in mixed IoT networks like Smart Water Management Systems(SWMS).These systems have different power levels and communication ranges.The simulation of proposed model has been done in MATLAB simulator.The results show that that the proposed model outperform then the existing models.展开更多
Cloud computing has become an essential technology for the management and processing of large datasets,offering scalability,high availability,and fault tolerance.However,optimizing data replication across multiple dat...Cloud computing has become an essential technology for the management and processing of large datasets,offering scalability,high availability,and fault tolerance.However,optimizing data replication across multiple data centers poses a significant challenge,especially when balancing opposing goals such as latency,storage costs,energy consumption,and network efficiency.This study introduces a novel Dynamic Optimization Algorithm called Dynamic Multi-Objective Gannet Optimization(DMGO),designed to enhance data replication efficiency in cloud environments.Unlike traditional static replication systems,DMGO adapts dynamically to variations in network conditions,system demand,and resource availability.The approach utilizes multi-objective optimization approaches to efficiently balance data access latency,storage efficiency,and operational costs.DMGO consistently evaluates data center performance and adjusts replication algorithms in real time to guarantee optimal system efficiency.Experimental evaluations conducted in a simulated cloud environment demonstrate that DMGO significantly outperforms conventional static algorithms,achieving faster data access,lower storage overhead,reduced energy consumption,and improved scalability.The proposed methodology offers a robust and adaptable solution for modern cloud systems,ensuring efficient resource consumption while maintaining high performance.展开更多
针对在现场可编程门阵列(Field Programmable Gate Array,FPGA)上实现基于极化敏感阵列的多重信号分类(Multiple Signal Classification,MUSIC)算法进行二维波达方向(Direction of Arrival,DOA)和二维极化参数联合估计时,硬件资源占用...针对在现场可编程门阵列(Field Programmable Gate Array,FPGA)上实现基于极化敏感阵列的多重信号分类(Multiple Signal Classification,MUSIC)算法进行二维波达方向(Direction of Arrival,DOA)和二维极化参数联合估计时,硬件资源占用大、运行时间长的问题,提出了一种基于极化MUSIC算法的四维参数联合估计FPGA实现架构。该架构包括信号协方差矩阵计算模块、Jacobi旋转模块、噪声子空间提取模块、两级空间谱搜索模块和极化参数计算模块。Jacobi旋转模块被拆分为多个可复用模块,并采用查找表模块生成旋转矩阵。一级空间谱搜索模块通过二维DOA搜索初步确定信源的角度信息。二级空间谱搜索模块根据一级搜索的角度结果确定二级搜索区域各点的极化信息,并计算该区域的四维空间谱,区域内最小值对应的四维参数信息即为最终估计的信源方向角、俯仰角、极化辅助角和极化相位角。仿真结果表明,与传统极化MUSIC算法的四维搜索算法相比,该架构避免了大量四维空间谱计算,同时保证了四维参数估计的精度,显著减少了运行时间和硬件资源消耗。展开更多
Aiming to solve the steering instability and hysteresis of agricultural robots in the process of movement,a fusion PID control method of particle swarm optimization(PSO)and genetic algorithm(GA)was proposed.The fusion...Aiming to solve the steering instability and hysteresis of agricultural robots in the process of movement,a fusion PID control method of particle swarm optimization(PSO)and genetic algorithm(GA)was proposed.The fusion algorithm took advantage of the fast optimization ability of PSO to optimize the population screening link of GA.The Simulink simulation results showed that the convergence of the fitness function of the fusion algorithm was accelerated,the system response adjustment time was reduced,and the overshoot was almost zero.Then the algorithm was applied to the steering test of agricultural robot in various scenes.After modeling the steering system of agricultural robot,the steering test results in the unloaded suspended state showed that the PID control based on fusion algorithm reduced the rise time,response adjustment time and overshoot of the system,and improved the response speed and stability of the system,compared with the artificial trial and error PID control and the PID control based on GA.The actual road steering test results showed that the PID control response rise time based on the fusion algorithm was the shortest,about 4.43 s.When the target pulse number was set to 100,the actual mean value in the steady-state regulation stage was about 102.9,which was the closest to the target value among the three control methods,and the overshoot was reduced at the same time.The steering test results under various scene states showed that the PID control based on the proposed fusion algorithm had good anti-interference ability,it can adapt to the changes of environment and load and improve the performance of the control system.It was effective in the steering control of agricultural robot.This method can provide a reference for the precise steering control of other robots.展开更多
基金National Natural Science Foundation of China(11971211,12171388).
文摘Complex network models are frequently employed for simulating and studyingdiverse real-world complex systems.Among these models,scale-free networks typically exhibit greater fragility to malicious attacks.Consequently,enhancing the robustness of scale-free networks has become a pressing issue.To address this problem,this paper proposes a Multi-Granularity Integration Algorithm(MGIA),which aims to improve the robustness of scale-free networks while keeping the initial degree of each node unchanged,ensuring network connectivity and avoiding the generation of multiple edges.The algorithm generates a multi-granularity structure from the initial network to be optimized,then uses different optimization strategies to optimize the networks at various granular layers in this structure,and finally realizes the information exchange between different granular layers,thereby further enhancing the optimization effect.We propose new network refresh,crossover,and mutation operators to ensure that the optimized network satisfies the given constraints.Meanwhile,we propose new network similarity and network dissimilarity evaluation metrics to improve the effectiveness of the optimization operators in the algorithm.In the experiments,the MGIA enhances the robustness of the scale-free network by 67.6%.This improvement is approximately 17.2%higher than the optimization effects achieved by eight currently existing complex network robustness optimization algorithms.
文摘The advent of microgrids in modern energy systems heralds a promising era of resilience,sustainability,and efficiency.Within the realm of grid-tied microgrids,the selection of an optimal optimization algorithm is critical for effective energy management,particularly in economic dispatching.This study compares the performance of Particle Swarm Optimization(PSO)and Genetic Algorithms(GA)in microgrid energy management systems,implemented using MATLAB tools.Through a comprehensive review of the literature and sim-ulations conducted in MATLAB,the study analyzes performance metrics,convergence speed,and the overall efficacy of GA and PSO,with a focus on economic dispatching tasks.Notably,a significant distinction emerges between the cost curves generated by the two algo-rithms for microgrid operation,with the PSO algorithm consistently resulting in lower costs due to its effective economic dispatching capabilities.Specifically,the utilization of the PSO approach could potentially lead to substantial savings on the power bill,amounting to approximately$15.30 in this evaluation.Thefindings provide insights into the strengths and limitations of each algorithm within the complex dynamics of grid-tied microgrids,thereby assisting stakeholders and researchers in arriving at informed decisions.This study contributes to the discourse on sustainable energy management by offering actionable guidance for the advancement of grid-tied micro-grid technologies through MATLAB-implemented optimization algorithms.
文摘Software defect prediction(SDP)aims to find a reliable method to predict defects in specific software projects and help software engineers allocate limited resources to release high-quality software products.Software defect prediction can be effectively performed using traditional features,but there are some redundant or irrelevant features in them(the presence or absence of this feature has little effect on the prediction results).These problems can be solved using feature selection.However,existing feature selection methods have shortcomings such as insignificant dimensionality reduction effect and low classification accuracy of the selected optimal feature subset.In order to reduce the impact of these shortcomings,this paper proposes a new feature selection method Cubic TraverseMa Beluga whale optimization algorithm(CTMBWO)based on the improved Beluga whale optimization algorithm(BWO).The goal of this study is to determine how well the CTMBWO can extract the features that are most important for correctly predicting software defects,improve the accuracy of fault prediction,reduce the number of the selected feature and mitigate the risk of overfitting,thereby achieving more efficient resource utilization and better distribution of test workload.The CTMBWO comprises three main stages:preprocessing the dataset,selecting relevant features,and evaluating the classification performance of the model.The novel feature selection method can effectively improve the performance of SDP.This study performs experiments on two software defect datasets(PROMISE,NASA)and shows the method’s classification performance using four detailed evaluation metrics,Accuracy,F1-score,MCC,AUC and Recall.The results indicate that the approach presented in this paper achieves outstanding classification performance on both datasets and has significant improvement over the baseline models.
基金Supported by the Major Science and Technology Project of Jilin Province(20220301010GX)the International Scientific and Technological Cooperation(20240402071GH).
文摘The liquid cooling system(LCS)of fuel cells is challenged by significant time delays,model uncertainties,pump and fan coupling,and frequent disturbances,leading to overshoot and control oscillations that degrade temperature regulation performance.To address these challenges,we propose a composite control scheme combining fuzzy logic and a variable-gain generalized supertwisting algorithm(VG-GSTA).Firstly,a one-dimensional(1D)fuzzy logic controler(FLC)for the pump ensures stable coolant flow,while a two-dimensional(2D)FLC for the fan regulates the stack temperature near the reference value.The VG-GSTA is then introduced to eliminate steady-state errors,offering resistance to disturbances and minimizing control oscillations.The equilibrium optimizer is used to fine-tune VG-GSTA parameters.Co-simulation verifies the effectiveness of our method,demonstrating its advantages in terms of disturbance immunity,overshoot suppression,tracking accuracy and response speed.
基金supported by the P.G.Senapathy Center for Computing Resources at IIT Madrasfunding provided by the Ministry of Education,Government of Indiasupported by the National Natural Science Foundation of China(Grant Nos.12388101,12472224 and 92252104).
文摘This paper presents an Eulerian-Lagrangian algorithm for direct numerical simulation(DNS)of particle-laden flows.The algorithm is applicable to perform simulations of dilute suspensions of small inertial particles in turbulent carrier flow.The Eulerian framework numerically resolves turbulent carrier flow using a parallelized,finite-volume DNS solver on a staggered Cartesian grid.Particles are tracked using a point-particle method utilizing a Lagrangian particle tracking(LPT)algorithm.The proposed Eulerian-Lagrangian algorithm is validated using an inertial particle-laden turbulent channel flow for different Stokes number cases.The particle concentration profiles and higher-order statistics of the carrier and dispersed phases agree well with the benchmark results.We investigated the effect of fluid velocity interpolation and numerical integration schemes of particle tracking algorithms on particle dispersion statistics.The suitability of fluid velocity interpolation schemes for predicting the particle dispersion statistics is discussed in the framework of the particle tracking algorithm coupled to the finite-volume solver.In addition,we present parallelization strategies implemented in the algorithm and evaluate their parallel performance.
文摘Internet of things networks often suffer from early node failures and short lifespan due to energy limits.Traditional routing methods are not enough.This work proposes a new hybrid algorithm called ACOGA.It combines Ant Colony Optimization(ACO)and the Greedy Algorithm(GA).ACO finds smart paths while Greedy makes quick decisions.This improves energy use and performance.ACOGA outperforms Hybrid Energy-Efficient(HEE)and Adaptive Lossless Data Compression(ALDC)algorithms.After 500 rounds,only 5%of ACOGA’s nodes are dead,compared to 15%for HEE and 20%for ALDC.The network using ACOGA runs for 1200 rounds before the first nodes fail.HEE lasts 900 rounds and ALDC only 850.ACOGA saves at least 15%more energy by better distributing the load.It also achieves a 98%packet delivery rate.The method works well in mixed IoT networks like Smart Water Management Systems(SWMS).These systems have different power levels and communication ranges.The simulation of proposed model has been done in MATLAB simulator.The results show that that the proposed model outperform then the existing models.
文摘Cloud computing has become an essential technology for the management and processing of large datasets,offering scalability,high availability,and fault tolerance.However,optimizing data replication across multiple data centers poses a significant challenge,especially when balancing opposing goals such as latency,storage costs,energy consumption,and network efficiency.This study introduces a novel Dynamic Optimization Algorithm called Dynamic Multi-Objective Gannet Optimization(DMGO),designed to enhance data replication efficiency in cloud environments.Unlike traditional static replication systems,DMGO adapts dynamically to variations in network conditions,system demand,and resource availability.The approach utilizes multi-objective optimization approaches to efficiently balance data access latency,storage efficiency,and operational costs.DMGO consistently evaluates data center performance and adjusts replication algorithms in real time to guarantee optimal system efficiency.Experimental evaluations conducted in a simulated cloud environment demonstrate that DMGO significantly outperforms conventional static algorithms,achieving faster data access,lower storage overhead,reduced energy consumption,and improved scalability.The proposed methodology offers a robust and adaptable solution for modern cloud systems,ensuring efficient resource consumption while maintaining high performance.
文摘针对在现场可编程门阵列(Field Programmable Gate Array,FPGA)上实现基于极化敏感阵列的多重信号分类(Multiple Signal Classification,MUSIC)算法进行二维波达方向(Direction of Arrival,DOA)和二维极化参数联合估计时,硬件资源占用大、运行时间长的问题,提出了一种基于极化MUSIC算法的四维参数联合估计FPGA实现架构。该架构包括信号协方差矩阵计算模块、Jacobi旋转模块、噪声子空间提取模块、两级空间谱搜索模块和极化参数计算模块。Jacobi旋转模块被拆分为多个可复用模块,并采用查找表模块生成旋转矩阵。一级空间谱搜索模块通过二维DOA搜索初步确定信源的角度信息。二级空间谱搜索模块根据一级搜索的角度结果确定二级搜索区域各点的极化信息,并计算该区域的四维空间谱,区域内最小值对应的四维参数信息即为最终估计的信源方向角、俯仰角、极化辅助角和极化相位角。仿真结果表明,与传统极化MUSIC算法的四维搜索算法相比,该架构避免了大量四维空间谱计算,同时保证了四维参数估计的精度,显著减少了运行时间和硬件资源消耗。
文摘Aiming to solve the steering instability and hysteresis of agricultural robots in the process of movement,a fusion PID control method of particle swarm optimization(PSO)and genetic algorithm(GA)was proposed.The fusion algorithm took advantage of the fast optimization ability of PSO to optimize the population screening link of GA.The Simulink simulation results showed that the convergence of the fitness function of the fusion algorithm was accelerated,the system response adjustment time was reduced,and the overshoot was almost zero.Then the algorithm was applied to the steering test of agricultural robot in various scenes.After modeling the steering system of agricultural robot,the steering test results in the unloaded suspended state showed that the PID control based on fusion algorithm reduced the rise time,response adjustment time and overshoot of the system,and improved the response speed and stability of the system,compared with the artificial trial and error PID control and the PID control based on GA.The actual road steering test results showed that the PID control response rise time based on the fusion algorithm was the shortest,about 4.43 s.When the target pulse number was set to 100,the actual mean value in the steady-state regulation stage was about 102.9,which was the closest to the target value among the three control methods,and the overshoot was reduced at the same time.The steering test results under various scene states showed that the PID control based on the proposed fusion algorithm had good anti-interference ability,it can adapt to the changes of environment and load and improve the performance of the control system.It was effective in the steering control of agricultural robot.This method can provide a reference for the precise steering control of other robots.