In order to overcome the limitations of a unitary reference station in mobile communication positioning network differential barometric altimetry(DBA)and broaden the action scope of the reference station and improve p...In order to overcome the limitations of a unitary reference station in mobile communication positioning network differential barometric altimetry(DBA)and broaden the action scope of the reference station and improve positioning accuracy of elevation,an integrated interpolation algorithm model based on generalized extended approximation(GEA)algorithm and Kriging interpolation in time-space domain of reference station is proposed.In the time domain,barometric measured data is considered the maximum value estimated by bilateral extension to avoid wrong direction of estimation,which is approaching true value.In the spatial domain,barometric relevance among multiple reference stations is utilized,the weighted coefficients of multiple reference stations is calculated by the integrated algorithm model based on the GEA algorithm and Kriging interpolation.The impact of each reference station to the measured station is quantified,so that a virtual reference station is constructed,which can overcome the limitation of barometric correction by a unitary reference station.In addition,the measurement error due to irregular change in atmospheric pressure will be eliminated.展开更多
In order to estimate intensity and direction of pressure acted on oil well in horizontal, this paper proposes an algorithm of ellipse fitting based on least square method for sections of oil well with ultrasonic imagi...In order to estimate intensity and direction of pressure acted on oil well in horizontal, this paper proposes an algorithm of ellipse fitting based on least square method for sections of oil well with ultrasonic imaging logging data. Five parameters about ellipse corresponding to the sections of oil well are extracted and their geometry deformations and direction are estimated. The tested results show that the fitting algorithm is robust and the estimating error of parameter is small, which have very important practical significance on processing the well logging.展开更多
Heuristic optimization algorithms have been widely used in solving complex optimization problems in various fields such as engineering,economics,and computer science.These algorithms are designed to find high-quality ...Heuristic optimization algorithms have been widely used in solving complex optimization problems in various fields such as engineering,economics,and computer science.These algorithms are designed to find high-quality solutions efficiently by balancing exploration of the search space and exploitation of promising solutions.While heuristic optimization algorithms vary in their specific details,they often exhibit common patterns that are essential to their effectiveness.This paper aims to analyze and explore common patterns in heuristic optimization algorithms.Through a comprehensive review of the literature,we identify the patterns that are commonly observed in these algorithms,including initialization,local search,diversity maintenance,adaptation,and stochasticity.For each pattern,we describe the motivation behind it,its implementation,and its impact on the search process.To demonstrate the utility of our analysis,we identify these patterns in multiple heuristic optimization algorithms.For each case study,we analyze how the patterns are implemented in the algorithm and how they contribute to its performance.Through these case studies,we show how our analysis can be used to understand the behavior of heuristic optimization algorithms and guide the design of new algorithms.Our analysis reveals that patterns in heuristic optimization algorithms are essential to their effectiveness.By understanding and incorporating these patterns into the design of new algorithms,researchers can develop more efficient and effective optimization algorithms.展开更多
Coal consumption curve of the thermal power plant can reflect the function relationship between the coal consumption of unit and load, which plays a key role for research on unit economic operation and load optimal di...Coal consumption curve of the thermal power plant can reflect the function relationship between the coal consumption of unit and load, which plays a key role for research on unit economic operation and load optimal dispatch. Now get coal consumption curve is generally obtained by least square method, but which are static curve and these curves remain unchanged for a long time, and make them are incompatible with the actual operation situation of the unit. Furthermore, coal consumption has the characteristics of typical nonlinear and time varying, sometimes the least square method does not work for nonlinear complex problems. For these problems, a method of coal consumption curve fitting of the thermal power plant units based on genetic algorithm is proposed. The residual analysis method is used for data detection;quadratic function is employed to the objective function;appropriate parameters such as initial population size, crossover rate and mutation rate are set;the unit’s actual coal consumption curves are fitted, and comparing the proposed method with least squares method, the results indicate that fitting effect of the former is better than the latter, and further indicate that the proposed method to do curve fitting can best approximate known data in a certain significance, and they can real-timely reflect the interdependence between power output and coal consumption.展开更多
Cloud computing has become an essential technology for the management and processing of large datasets,offering scalability,high availability,and fault tolerance.However,optimizing data replication across multiple dat...Cloud computing has become an essential technology for the management and processing of large datasets,offering scalability,high availability,and fault tolerance.However,optimizing data replication across multiple data centers poses a significant challenge,especially when balancing opposing goals such as latency,storage costs,energy consumption,and network efficiency.This study introduces a novel Dynamic Optimization Algorithm called Dynamic Multi-Objective Gannet Optimization(DMGO),designed to enhance data replication efficiency in cloud environments.Unlike traditional static replication systems,DMGO adapts dynamically to variations in network conditions,system demand,and resource availability.The approach utilizes multi-objective optimization approaches to efficiently balance data access latency,storage efficiency,and operational costs.DMGO consistently evaluates data center performance and adjusts replication algorithms in real time to guarantee optimal system efficiency.Experimental evaluations conducted in a simulated cloud environment demonstrate that DMGO significantly outperforms conventional static algorithms,achieving faster data access,lower storage overhead,reduced energy consumption,and improved scalability.The proposed methodology offers a robust and adaptable solution for modern cloud systems,ensuring efficient resource consumption while maintaining high performance.展开更多
Let p be a prime. For any finite p-group G, the deep transfers T H,G ' : H / H ' → G ' / G " from the maximal subgroups H of index (G:H) = p in G to the derived subgroup G ' are introduced as an ...Let p be a prime. For any finite p-group G, the deep transfers T H,G ' : H / H ' → G ' / G " from the maximal subgroups H of index (G:H) = p in G to the derived subgroup G ' are introduced as an innovative tool for identifying G uniquely by means of the family of kernels ùd(G) =(ker(T H,G ')) (G: H) = p. For all finite 3-groups G of coclass cc(G) = 1, the family ùd(G) is determined explicitly. The results are applied to the Galois groups G =Gal(F3 (∞)/ F) of the Hilbert 3-class towers of all real quadratic fields F = Q(√d) with fundamental discriminants d > 1, 3-class group Cl3(F) □ C3 × C3, and total 3-principalization in each of their four unramified cyclic cubic extensions E/F. A systematic statistical evaluation is given for the complete range 1 d 7, and a few exceptional cases are pointed out for 1 d 8.展开更多
3D ground-penetrating radar has been widely used in urban road underground disease detection due to its nondestructive,efficient,and intuitive results.However,the 3D imaging of the underground target body presents the...3D ground-penetrating radar has been widely used in urban road underground disease detection due to its nondestructive,efficient,and intuitive results.However,the 3D imaging of the underground target body presents the edge plate phenomenon due to the space between the 3D radar array antennas.Consequently,direct 3D imaging using detection results cannot reflect underground spatial distribution characteristics.Due to the wide-beam polarization of the ground-penetrating radar antenna,the emission of electromagnetic waves with a specific width decreases the strong middle energy on both sides gradually.Therefore,a bicubic high-precision 3D target body slice-imaging fitting algorithm with changing trend characteristics is constructed by combining the subsurface target characteristics with the changing spatial morphology trends.Using the wide-angle polarization antenna’s characteristics in the algorithm to build the trend factor between the measurement lines,the target body change trend and the edge detail portrayal achieve a 3D ground-penetrating radar-detection target high-precision fitting.Compared with other traditional fitting techniques,the fitting error is small.This paper conducts experiments and analyses on GpaMax 3D forward modeling and 3D ground-penetrating measured radar data.The experiments show that the improved bicubic fitting algorithm can eff ectively improve the accuracy of underground target slice imaging and the 3D ground-penetrating radar’s anomaly interpretation.展开更多
Non-technical losses(NTL)of electric power are a serious problem for electric distribution companies.The solution determines the cost,stability,reliability,and quality of the supplied electricity.The widespread use of...Non-technical losses(NTL)of electric power are a serious problem for electric distribution companies.The solution determines the cost,stability,reliability,and quality of the supplied electricity.The widespread use of advanced metering infrastructure(AMI)and Smart Grid allows all participants in the distribution grid to store and track electricity consumption.During the research,a machine learning model is developed that allows analyzing and predicting the probability of NTL for each consumer of the distribution grid based on daily electricity consumption readings.This model is an ensemble meta-algorithm(stacking)that generalizes the algorithms of random forest,LightGBM,and a homogeneous ensemble of artificial neural networks.The best accuracy of the proposed meta-algorithm in comparison to basic classifiers is experimentally confirmed on the test sample.Such a model,due to good accuracy indicators(ROC-AUC-0.88),can be used as a methodological basis for a decision support system,the purpose of which is to form a sample of suspected NTL sources.The use of such a sample will allow the top management of electric distribution companies to increase the efficiency of raids by performers,making them targeted and accurate,which should contribute to the fight against NTL and the sustainable development of the electric power industry.展开更多
Two new regularization algorithms for solving the first-kind Volterra integral equation, which describes the pressure-rate deconvolution problem in well test data interpretation, are developed in this paper. The main ...Two new regularization algorithms for solving the first-kind Volterra integral equation, which describes the pressure-rate deconvolution problem in well test data interpretation, are developed in this paper. The main features of the problem are the strong nonuniform scale of the solution and large errors (up to 15%) in the input data. In both algorithms, the solution is represented as decomposition on special basic functions, which satisfy given a priori information on solution, and this idea allow us significantly to improve the quality of approximate solution and simplify solving the minimization problem. The theoretical details of the algorithms, as well as the results of numerical experiments for proving robustness of the algorithms, are presented.展开更多
This paper presents a binary gravitational search algorithm (BGSA) is applied to solve the problem of optimal allotment of DG sets and Shunt capacitors in radial distribution systems. The problem is formulated as a no...This paper presents a binary gravitational search algorithm (BGSA) is applied to solve the problem of optimal allotment of DG sets and Shunt capacitors in radial distribution systems. The problem is formulated as a nonlinear constrained single-objective optimization problem where the total line loss (TLL) and the total voltage deviations (TVD) are to be minimized separately by incorporating optimal placement of DG units and shunt capacitors with constraints which include limits on voltage, sizes of installed capacitors and DG. This BGSA is applied on the balanced IEEE 10 Bus distribution network and the results are compared with conventional binary particle swarm optimization.展开更多
提出了一种基于最小二乘支持向量机的织物剪切性能预测模型,并且采用遗传算法进行最小二乘支持向量机的参数优化,将获得的样本进行归一化处理后,将其输入预测模型以得到预测结果.仿真结果表明,基于最小二乘支持向量机的预测模型比BP神...提出了一种基于最小二乘支持向量机的织物剪切性能预测模型,并且采用遗传算法进行最小二乘支持向量机的参数优化,将获得的样本进行归一化处理后,将其输入预测模型以得到预测结果.仿真结果表明,基于最小二乘支持向量机的预测模型比BP神经网络和线性回归方法具有更高的精度和范化能力.
Abstract:
A new method is proposed to predict the fabric shearing property with least square support vector machines ( LS-SVM ). The genetic algorithm is investigated to select the parameters of LS-SVM models as a means of improving the LS- SVM prediction. After normalizing the sampling data, the sampling data are inputted into the model to gain the prediction result. The simulation results show the prediction model gives better forecasting accuracy and generalization ability than BP neural network and linear regression method.展开更多
This study presents a novel hybrid topology optimization and mold design framework that integrates process fitting,runner system optimization,and structural analysis to significantly enhance the performance of injecti...This study presents a novel hybrid topology optimization and mold design framework that integrates process fitting,runner system optimization,and structural analysis to significantly enhance the performance of injection-molded parts.At its core,the framework employs a greedy algorithm that generates runner systems based on adjacency and shortest path principles,leading to improvements in both mechanical strength and material efficiency.The design optimization is validated through a series of rigorous experimental tests,including three-point bending and torsion tests performed on key-socket frames,ensuring that the optimized designs meet practical performance requirements.A critical innovation of the framework is the development of the Adjacent Element Temperature-Driven Prestress Algorithm(AETDPA),which refines the prediction of mechanical failure and strength fitting.This algorithm has been shown to deliver mesh-independent accuracy,thereby enhancing the reliability of simulation results across various design iterations.The framework’s adaptability is further demonstrated by its ability to adjust optimization methods based on the unique geometry of each part,thus accelerating the overall design process while ensuring struc-tural integrity.In addition to its immediate applications in injection molding,the study explores the potential extension of this framework to metal additive manufacturing,opening new avenues for its use in advanced manufacturing technologies.Numerical simulations,including finite element analysis,support the experimental findings and confirm that the optimized designs provide a balanced combination of strength,durability,and efficiency.Furthermore,the integration challenges with existing injection molding practices are addressed,underscoring the framework’s scalability and industrial relevance.Overall,this hybrid topology optimization framework offers a computationally efficient and robust solution for advanced manufacturing applications,promising significant improvements in design efficiency,cost-effectiveness,and product performance.Future work will focus on further enhancing algorithm robustness and exploring additional applications across diverse manufacturing processes.展开更多
The contradiction of variable step size least mean square(LMS)algorithm between fast convergence speed and small steady-state error has always existed.So,a new algorithm based on the combination of logarithmic and sym...The contradiction of variable step size least mean square(LMS)algorithm between fast convergence speed and small steady-state error has always existed.So,a new algorithm based on the combination of logarithmic and symbolic function and step size factor is proposed.It establishes a new updating method of step factor that is related to step factor and error signal.This work makes an analysis from 3 aspects:theoretical analysis,theoretical verification and specific experiments.The experimental results show that the proposed algorithm is superior to other variable step size algorithms in convergence speed and steady-state error.展开更多
Numerous cryptographic algorithms (ElGamal, Rabin, RSA, NTRU etc) require multiple computations of modulo multiplicative inverses. This paper describes and validates a new algorithm, called the Enhanced Euclid Algorit...Numerous cryptographic algorithms (ElGamal, Rabin, RSA, NTRU etc) require multiple computations of modulo multiplicative inverses. This paper describes and validates a new algorithm, called the Enhanced Euclid Algorithm, for modular multiplicative inverse (MMI). Analysis of the proposed algorithm shows that it is more efficient than the Extended Euclid algorithm (XEA). In addition, if a MMI does not exist, then it is not necessary to use the Backtracking procedure in the proposed algorithm;this case requires fewer operations on every step (divisions, multiplications, additions, assignments and push operations on stack), than the XEA. Overall, XEA uses more multiplications, additions, assignments and twice as many variables than the proposed algorithm.展开更多
The word“spatial”fundamentally relates to human existence,evolution,and activity in terrestrial and even celestial spaces.After reviewing the spatial features of many areas,the paper describes basics of high level m...The word“spatial”fundamentally relates to human existence,evolution,and activity in terrestrial and even celestial spaces.After reviewing the spatial features of many areas,the paper describes basics of high level model and technology called Spatial Grasp for dealing with large distributed systems,which can provide spatial vision,awareness,management,control,and even consciousness.The technology description includes its key Spatial Grasp Language(SGL),self-evolution of recursive SGL scenarios,and implementation of SGL interpreter converting distributed networked systems into powerful spatial engines.Examples of typical spatial scenarios in SGL include finding shortest path tree and shortest path between network nodes,collecting proper information throughout the whole world,elimination of multiple targets by intelligent teams of chasers,and withstanding cyber attacks in distributed networked systems.Also this paper compares Spatial Grasp model with traditional algorithms,confirming universality of the former for any spatial systems,while the latter just tools for concrete applications.展开更多
In this paper we consider a parallel algorithm that detects the maximizer of unimodal function f(x) computable at every point on unbounded interval (0, ∞). The algorithm consists of two modes: scanning and detecting....In this paper we consider a parallel algorithm that detects the maximizer of unimodal function f(x) computable at every point on unbounded interval (0, ∞). The algorithm consists of two modes: scanning and detecting. Search diagrams are introduced as a way to describe parallel searching algorithms on unbounded intervals. Dynamic programming equations, combined with a series of liner programming problems, describe relations between results for every pair of successive evaluations of function f in parallel. Properties of optimal search strategies are derived from these equations. The worst-case complexity analysis shows that, if the maximizer is located on a priori unknown interval (n-1], then it can be detected after cp(n)=「2log「p/2」+1(n+1)」-1 parallel evaluations of f(x), where p is the number of processors.展开更多
This work proposes a novel approach for multi-type optimal placement of flexible AC transmission system(FACTS) devices so as to optimize multi-objective voltage stability problem. The current study discusses a way for...This work proposes a novel approach for multi-type optimal placement of flexible AC transmission system(FACTS) devices so as to optimize multi-objective voltage stability problem. The current study discusses a way for locating and setting of thyristor controlled series capacitor(TCSC) and static var compensator(SVC) using the multi-objective optimization approach named strength pareto multi-objective evolutionary algorithm(SPMOEA). Maximization of the static voltage stability margin(SVSM) and minimizations of real power losses(RPL) and load voltage deviation(LVD) are taken as the goals or three objective functions, when optimally locating multi-type FACTS devices. The performance and effectiveness of the proposed approach has been validated by the simulation results of the IEEE 30-bus and IEEE 118-bus test systems. The proposed approach is compared with non-dominated sorting particle swarm optimization(NSPSO) algorithm. This comparison confirms the usefulness of the multi-objective proposed technique that makes it promising for determination of combinatorial problems of FACTS devices location and setting in large scale power systems.展开更多
A non-orthogonal multiple access(NOMA) power allocation scheme on the basis of the sparrow search algorithm(SSA) is proposed in this work. Specifically, the logarithmic utility function is utilized to address the pote...A non-orthogonal multiple access(NOMA) power allocation scheme on the basis of the sparrow search algorithm(SSA) is proposed in this work. Specifically, the logarithmic utility function is utilized to address the potential fairness issue that may arise from the maximum sum-rate based objective function and the optical power constraints are set considering the non-negativity of the transmit signal, the requirement of the human eyes safety and all users' quality of service(Qo S). Then, the SSA is utilized to solve this optimization problem. Moreover, to demonstrate the superiority of the proposed strategy, it is compared with the fixed power allocation(FPA) and the gain ratio power allocation(GRPA) schemes. Results show that regardless of the number of users considered, the sum-rate achieved by SSA consistently outperforms that of FPA and GRPA schemes. Specifically, compared to FPA and GRPA schemes, the sum-rate obtained by SSA is increased by 40.45% and 53.44% when the number of users is 7, respectively. The proposed SSA also has better performance in terms of user fairness. This work will benefit the design and development of the NOMA-visible light communication(VLC) systems.展开更多
The Cross-domain Heuristic Search Challenge(CHeSC)is a competition focused on creating efficient search algorithms adaptable to diverse problem domains.Selection hyper-heuristics are a class of algorithms that dynamic...The Cross-domain Heuristic Search Challenge(CHeSC)is a competition focused on creating efficient search algorithms adaptable to diverse problem domains.Selection hyper-heuristics are a class of algorithms that dynamically choose heuristics during the search process.Numerous selection hyper-heuristics have different imple-mentation strategies.However,comparisons between them are lacking in the literature,and previous works have not highlighted the beneficial and detrimental implementation methods of different components.The question is how to effectively employ them to produce an efficient search heuristic.Furthermore,the algorithms that competed in the inaugural CHeSC have not been collectively reviewed.This work conducts a review analysis of the top twenty competitors from this competition to identify effective and ineffective strategies influencing algorithmic performance.A summary of the main characteristics and classification of the algorithms is presented.The analysis underlines efficient and inefficient methods in eight key components,including search points,search phases,heuristic selection,move acceptance,feedback,Tabu mechanism,restart mechanism,and low-level heuristic parameter control.This review analyzes the components referencing the competition’s final leaderboard and discusses future research directions for these components.The effective approaches,identified as having the highest quality index,are mixed search point,iterated search phases,relay hybridization selection,threshold acceptance,mixed learning,Tabu heuristics,stochastic restart,and dynamic parameters.Findings are also compared with recent trends in hyper-heuristics.This work enhances the understanding of selection hyper-heuristics,offering valuable insights for researchers and practitioners aiming to develop effective search algorithms for diverse problem domains.展开更多
基金Supported by the National Key Research Program of China"Collaborative Precision Positioning Project"(2016YFB 0501900)the National Natural Science Foundation of China(11603041)+1 种基金Key Laboratory of Cognitive Radio and Information Processing,Ministry of Education(Guilin University of Electronic Technology)Guangxi Key Laboratory of Precision Navigation Technology and Application
文摘In order to overcome the limitations of a unitary reference station in mobile communication positioning network differential barometric altimetry(DBA)and broaden the action scope of the reference station and improve positioning accuracy of elevation,an integrated interpolation algorithm model based on generalized extended approximation(GEA)algorithm and Kriging interpolation in time-space domain of reference station is proposed.In the time domain,barometric measured data is considered the maximum value estimated by bilateral extension to avoid wrong direction of estimation,which is approaching true value.In the spatial domain,barometric relevance among multiple reference stations is utilized,the weighted coefficients of multiple reference stations is calculated by the integrated algorithm model based on the GEA algorithm and Kriging interpolation.The impact of each reference station to the measured station is quantified,so that a virtual reference station is constructed,which can overcome the limitation of barometric correction by a unitary reference station.In addition,the measurement error due to irregular change in atmospheric pressure will be eliminated.
文摘In order to estimate intensity and direction of pressure acted on oil well in horizontal, this paper proposes an algorithm of ellipse fitting based on least square method for sections of oil well with ultrasonic imaging logging data. Five parameters about ellipse corresponding to the sections of oil well are extracted and their geometry deformations and direction are estimated. The tested results show that the fitting algorithm is robust and the estimating error of parameter is small, which have very important practical significance on processing the well logging.
文摘Heuristic optimization algorithms have been widely used in solving complex optimization problems in various fields such as engineering,economics,and computer science.These algorithms are designed to find high-quality solutions efficiently by balancing exploration of the search space and exploitation of promising solutions.While heuristic optimization algorithms vary in their specific details,they often exhibit common patterns that are essential to their effectiveness.This paper aims to analyze and explore common patterns in heuristic optimization algorithms.Through a comprehensive review of the literature,we identify the patterns that are commonly observed in these algorithms,including initialization,local search,diversity maintenance,adaptation,and stochasticity.For each pattern,we describe the motivation behind it,its implementation,and its impact on the search process.To demonstrate the utility of our analysis,we identify these patterns in multiple heuristic optimization algorithms.For each case study,we analyze how the patterns are implemented in the algorithm and how they contribute to its performance.Through these case studies,we show how our analysis can be used to understand the behavior of heuristic optimization algorithms and guide the design of new algorithms.Our analysis reveals that patterns in heuristic optimization algorithms are essential to their effectiveness.By understanding and incorporating these patterns into the design of new algorithms,researchers can develop more efficient and effective optimization algorithms.
文摘Coal consumption curve of the thermal power plant can reflect the function relationship between the coal consumption of unit and load, which plays a key role for research on unit economic operation and load optimal dispatch. Now get coal consumption curve is generally obtained by least square method, but which are static curve and these curves remain unchanged for a long time, and make them are incompatible with the actual operation situation of the unit. Furthermore, coal consumption has the characteristics of typical nonlinear and time varying, sometimes the least square method does not work for nonlinear complex problems. For these problems, a method of coal consumption curve fitting of the thermal power plant units based on genetic algorithm is proposed. The residual analysis method is used for data detection;quadratic function is employed to the objective function;appropriate parameters such as initial population size, crossover rate and mutation rate are set;the unit’s actual coal consumption curves are fitted, and comparing the proposed method with least squares method, the results indicate that fitting effect of the former is better than the latter, and further indicate that the proposed method to do curve fitting can best approximate known data in a certain significance, and they can real-timely reflect the interdependence between power output and coal consumption.
文摘Cloud computing has become an essential technology for the management and processing of large datasets,offering scalability,high availability,and fault tolerance.However,optimizing data replication across multiple data centers poses a significant challenge,especially when balancing opposing goals such as latency,storage costs,energy consumption,and network efficiency.This study introduces a novel Dynamic Optimization Algorithm called Dynamic Multi-Objective Gannet Optimization(DMGO),designed to enhance data replication efficiency in cloud environments.Unlike traditional static replication systems,DMGO adapts dynamically to variations in network conditions,system demand,and resource availability.The approach utilizes multi-objective optimization approaches to efficiently balance data access latency,storage efficiency,and operational costs.DMGO consistently evaluates data center performance and adjusts replication algorithms in real time to guarantee optimal system efficiency.Experimental evaluations conducted in a simulated cloud environment demonstrate that DMGO significantly outperforms conventional static algorithms,achieving faster data access,lower storage overhead,reduced energy consumption,and improved scalability.The proposed methodology offers a robust and adaptable solution for modern cloud systems,ensuring efficient resource consumption while maintaining high performance.
文摘Let p be a prime. For any finite p-group G, the deep transfers T H,G ' : H / H ' → G ' / G " from the maximal subgroups H of index (G:H) = p in G to the derived subgroup G ' are introduced as an innovative tool for identifying G uniquely by means of the family of kernels ùd(G) =(ker(T H,G ')) (G: H) = p. For all finite 3-groups G of coclass cc(G) = 1, the family ùd(G) is determined explicitly. The results are applied to the Galois groups G =Gal(F3 (∞)/ F) of the Hilbert 3-class towers of all real quadratic fields F = Q(√d) with fundamental discriminants d > 1, 3-class group Cl3(F) □ C3 × C3, and total 3-principalization in each of their four unramified cyclic cubic extensions E/F. A systematic statistical evaluation is given for the complete range 1 d 7, and a few exceptional cases are pointed out for 1 d 8.
基金supported by The National Key Research and Development Program of China (2021YFC3090304)The Fundamental Research Funds for the Central Universities,China University of Mining and Technology-Beijing (8000150A073).
文摘3D ground-penetrating radar has been widely used in urban road underground disease detection due to its nondestructive,efficient,and intuitive results.However,the 3D imaging of the underground target body presents the edge plate phenomenon due to the space between the 3D radar array antennas.Consequently,direct 3D imaging using detection results cannot reflect underground spatial distribution characteristics.Due to the wide-beam polarization of the ground-penetrating radar antenna,the emission of electromagnetic waves with a specific width decreases the strong middle energy on both sides gradually.Therefore,a bicubic high-precision 3D target body slice-imaging fitting algorithm with changing trend characteristics is constructed by combining the subsurface target characteristics with the changing spatial morphology trends.Using the wide-angle polarization antenna’s characteristics in the algorithm to build the trend factor between the measurement lines,the target body change trend and the edge detail portrayal achieve a 3D ground-penetrating radar-detection target high-precision fitting.Compared with other traditional fitting techniques,the fitting error is small.This paper conducts experiments and analyses on GpaMax 3D forward modeling and 3D ground-penetrating measured radar data.The experiments show that the improved bicubic fitting algorithm can eff ectively improve the accuracy of underground target slice imaging and the 3D ground-penetrating radar’s anomaly interpretation.
文摘Non-technical losses(NTL)of electric power are a serious problem for electric distribution companies.The solution determines the cost,stability,reliability,and quality of the supplied electricity.The widespread use of advanced metering infrastructure(AMI)and Smart Grid allows all participants in the distribution grid to store and track electricity consumption.During the research,a machine learning model is developed that allows analyzing and predicting the probability of NTL for each consumer of the distribution grid based on daily electricity consumption readings.This model is an ensemble meta-algorithm(stacking)that generalizes the algorithms of random forest,LightGBM,and a homogeneous ensemble of artificial neural networks.The best accuracy of the proposed meta-algorithm in comparison to basic classifiers is experimentally confirmed on the test sample.Such a model,due to good accuracy indicators(ROC-AUC-0.88),can be used as a methodological basis for a decision support system,the purpose of which is to form a sample of suspected NTL sources.The use of such a sample will allow the top management of electric distribution companies to increase the efficiency of raids by performers,making them targeted and accurate,which should contribute to the fight against NTL and the sustainable development of the electric power industry.
文摘Two new regularization algorithms for solving the first-kind Volterra integral equation, which describes the pressure-rate deconvolution problem in well test data interpretation, are developed in this paper. The main features of the problem are the strong nonuniform scale of the solution and large errors (up to 15%) in the input data. In both algorithms, the solution is represented as decomposition on special basic functions, which satisfy given a priori information on solution, and this idea allow us significantly to improve the quality of approximate solution and simplify solving the minimization problem. The theoretical details of the algorithms, as well as the results of numerical experiments for proving robustness of the algorithms, are presented.
文摘This paper presents a binary gravitational search algorithm (BGSA) is applied to solve the problem of optimal allotment of DG sets and Shunt capacitors in radial distribution systems. The problem is formulated as a nonlinear constrained single-objective optimization problem where the total line loss (TLL) and the total voltage deviations (TVD) are to be minimized separately by incorporating optimal placement of DG units and shunt capacitors with constraints which include limits on voltage, sizes of installed capacitors and DG. This BGSA is applied on the balanced IEEE 10 Bus distribution network and the results are compared with conventional binary particle swarm optimization.
文摘提出了一种基于最小二乘支持向量机的织物剪切性能预测模型,并且采用遗传算法进行最小二乘支持向量机的参数优化,将获得的样本进行归一化处理后,将其输入预测模型以得到预测结果.仿真结果表明,基于最小二乘支持向量机的预测模型比BP神经网络和线性回归方法具有更高的精度和范化能力.
Abstract:
A new method is proposed to predict the fabric shearing property with least square support vector machines ( LS-SVM ). The genetic algorithm is investigated to select the parameters of LS-SVM models as a means of improving the LS- SVM prediction. After normalizing the sampling data, the sampling data are inputted into the model to gain the prediction result. The simulation results show the prediction model gives better forecasting accuracy and generalization ability than BP neural network and linear regression method.
文摘This study presents a novel hybrid topology optimization and mold design framework that integrates process fitting,runner system optimization,and structural analysis to significantly enhance the performance of injection-molded parts.At its core,the framework employs a greedy algorithm that generates runner systems based on adjacency and shortest path principles,leading to improvements in both mechanical strength and material efficiency.The design optimization is validated through a series of rigorous experimental tests,including three-point bending and torsion tests performed on key-socket frames,ensuring that the optimized designs meet practical performance requirements.A critical innovation of the framework is the development of the Adjacent Element Temperature-Driven Prestress Algorithm(AETDPA),which refines the prediction of mechanical failure and strength fitting.This algorithm has been shown to deliver mesh-independent accuracy,thereby enhancing the reliability of simulation results across various design iterations.The framework’s adaptability is further demonstrated by its ability to adjust optimization methods based on the unique geometry of each part,thus accelerating the overall design process while ensuring struc-tural integrity.In addition to its immediate applications in injection molding,the study explores the potential extension of this framework to metal additive manufacturing,opening new avenues for its use in advanced manufacturing technologies.Numerical simulations,including finite element analysis,support the experimental findings and confirm that the optimized designs provide a balanced combination of strength,durability,and efficiency.Furthermore,the integration challenges with existing injection molding practices are addressed,underscoring the framework’s scalability and industrial relevance.Overall,this hybrid topology optimization framework offers a computationally efficient and robust solution for advanced manufacturing applications,promising significant improvements in design efficiency,cost-effectiveness,and product performance.Future work will focus on further enhancing algorithm robustness and exploring additional applications across diverse manufacturing processes.
基金the National Natural Science Foundation of China(No.51575328,61503232).
文摘The contradiction of variable step size least mean square(LMS)algorithm between fast convergence speed and small steady-state error has always existed.So,a new algorithm based on the combination of logarithmic and symbolic function and step size factor is proposed.It establishes a new updating method of step factor that is related to step factor and error signal.This work makes an analysis from 3 aspects:theoretical analysis,theoretical verification and specific experiments.The experimental results show that the proposed algorithm is superior to other variable step size algorithms in convergence speed and steady-state error.
文摘Numerous cryptographic algorithms (ElGamal, Rabin, RSA, NTRU etc) require multiple computations of modulo multiplicative inverses. This paper describes and validates a new algorithm, called the Enhanced Euclid Algorithm, for modular multiplicative inverse (MMI). Analysis of the proposed algorithm shows that it is more efficient than the Extended Euclid algorithm (XEA). In addition, if a MMI does not exist, then it is not necessary to use the Backtracking procedure in the proposed algorithm;this case requires fewer operations on every step (divisions, multiplications, additions, assignments and push operations on stack), than the XEA. Overall, XEA uses more multiplications, additions, assignments and twice as many variables than the proposed algorithm.
文摘The word“spatial”fundamentally relates to human existence,evolution,and activity in terrestrial and even celestial spaces.After reviewing the spatial features of many areas,the paper describes basics of high level model and technology called Spatial Grasp for dealing with large distributed systems,which can provide spatial vision,awareness,management,control,and even consciousness.The technology description includes its key Spatial Grasp Language(SGL),self-evolution of recursive SGL scenarios,and implementation of SGL interpreter converting distributed networked systems into powerful spatial engines.Examples of typical spatial scenarios in SGL include finding shortest path tree and shortest path between network nodes,collecting proper information throughout the whole world,elimination of multiple targets by intelligent teams of chasers,and withstanding cyber attacks in distributed networked systems.Also this paper compares Spatial Grasp model with traditional algorithms,confirming universality of the former for any spatial systems,while the latter just tools for concrete applications.
文摘In this paper we consider a parallel algorithm that detects the maximizer of unimodal function f(x) computable at every point on unbounded interval (0, ∞). The algorithm consists of two modes: scanning and detecting. Search diagrams are introduced as a way to describe parallel searching algorithms on unbounded intervals. Dynamic programming equations, combined with a series of liner programming problems, describe relations between results for every pair of successive evaluations of function f in parallel. Properties of optimal search strategies are derived from these equations. The worst-case complexity analysis shows that, if the maximizer is located on a priori unknown interval (n-1], then it can be detected after cp(n)=「2log「p/2」+1(n+1)」-1 parallel evaluations of f(x), where p is the number of processors.
文摘This work proposes a novel approach for multi-type optimal placement of flexible AC transmission system(FACTS) devices so as to optimize multi-objective voltage stability problem. The current study discusses a way for locating and setting of thyristor controlled series capacitor(TCSC) and static var compensator(SVC) using the multi-objective optimization approach named strength pareto multi-objective evolutionary algorithm(SPMOEA). Maximization of the static voltage stability margin(SVSM) and minimizations of real power losses(RPL) and load voltage deviation(LVD) are taken as the goals or three objective functions, when optimally locating multi-type FACTS devices. The performance and effectiveness of the proposed approach has been validated by the simulation results of the IEEE 30-bus and IEEE 118-bus test systems. The proposed approach is compared with non-dominated sorting particle swarm optimization(NSPSO) algorithm. This comparison confirms the usefulness of the multi-objective proposed technique that makes it promising for determination of combinatorial problems of FACTS devices location and setting in large scale power systems.
基金supported by the Cooperative Research Project between China Coal Energy Research Institute Co.,Ltd. and Xidian University (No.N-KY-HX-1101-202302-00725)the Key Research and Development Program of Shaanxi Province (No.2017ZDCXL-GY-06-02)。
文摘A non-orthogonal multiple access(NOMA) power allocation scheme on the basis of the sparrow search algorithm(SSA) is proposed in this work. Specifically, the logarithmic utility function is utilized to address the potential fairness issue that may arise from the maximum sum-rate based objective function and the optical power constraints are set considering the non-negativity of the transmit signal, the requirement of the human eyes safety and all users' quality of service(Qo S). Then, the SSA is utilized to solve this optimization problem. Moreover, to demonstrate the superiority of the proposed strategy, it is compared with the fixed power allocation(FPA) and the gain ratio power allocation(GRPA) schemes. Results show that regardless of the number of users considered, the sum-rate achieved by SSA consistently outperforms that of FPA and GRPA schemes. Specifically, compared to FPA and GRPA schemes, the sum-rate obtained by SSA is increased by 40.45% and 53.44% when the number of users is 7, respectively. The proposed SSA also has better performance in terms of user fairness. This work will benefit the design and development of the NOMA-visible light communication(VLC) systems.
基金funded by Ministry of Higher Education(MoHE)Malaysia,under Transdisciplinary Research Grant Scheme(TRGS/1/2019/UKM/01/4/2).
文摘The Cross-domain Heuristic Search Challenge(CHeSC)is a competition focused on creating efficient search algorithms adaptable to diverse problem domains.Selection hyper-heuristics are a class of algorithms that dynamically choose heuristics during the search process.Numerous selection hyper-heuristics have different imple-mentation strategies.However,comparisons between them are lacking in the literature,and previous works have not highlighted the beneficial and detrimental implementation methods of different components.The question is how to effectively employ them to produce an efficient search heuristic.Furthermore,the algorithms that competed in the inaugural CHeSC have not been collectively reviewed.This work conducts a review analysis of the top twenty competitors from this competition to identify effective and ineffective strategies influencing algorithmic performance.A summary of the main characteristics and classification of the algorithms is presented.The analysis underlines efficient and inefficient methods in eight key components,including search points,search phases,heuristic selection,move acceptance,feedback,Tabu mechanism,restart mechanism,and low-level heuristic parameter control.This review analyzes the components referencing the competition’s final leaderboard and discusses future research directions for these components.The effective approaches,identified as having the highest quality index,are mixed search point,iterated search phases,relay hybridization selection,threshold acceptance,mixed learning,Tabu heuristics,stochastic restart,and dynamic parameters.Findings are also compared with recent trends in hyper-heuristics.This work enhances the understanding of selection hyper-heuristics,offering valuable insights for researchers and practitioners aiming to develop effective search algorithms for diverse problem domains.