Algorithms are the primary component of Artificial Intelligence(AI).The algorithm is the process in AI that imitates the human mind to solve problems.Currently evaluating the performance of AI is achieved by evaluatin...Algorithms are the primary component of Artificial Intelligence(AI).The algorithm is the process in AI that imitates the human mind to solve problems.Currently evaluating the performance of AI is achieved by evaluating AI algorithms by metric scores on data sets.However the evaluation of algorithms in AI is challenging because the evaluation of the same type of algorithm has many data sets and evaluation metrics.Different algorithms may have individual strengths and weaknesses in evaluation metric scores on separate data sets,lacking the credibility and validity of the evaluation.Moreover,evaluation of algorithms requires repeated experiments on different data sets,reducing the attention of researchers to the research of the algorithms itself.Crucially,this approach to evaluating comparative metric scores does not take into account the algorithm’s ability to solve problems.And the classical algorithm evaluation of time and space complexity is not suitable for evaluating AI algorithms.Because classical algorithms input is infinite numbers,whereas AI algorithms input is a data set,which is limited and multifarious.According to the AI algorithm evaluation without response to the problem solving capability,this paper summarizes the features of AI algorithm evaluation and proposes an AI evaluation method that incorporates the problem-solving capabilities of algorithms.展开更多
In recent years, particle swarm optimization (PSO) has received widespread attention in feature selection due to its simplicity and potential for global search. However, in traditional PSO, particles primarily update ...In recent years, particle swarm optimization (PSO) has received widespread attention in feature selection due to its simplicity and potential for global search. However, in traditional PSO, particles primarily update based on two extreme values: personal best and global best, which limits the diversity of information. Ideally, particles should learn from multiple advantageous particles to enhance interactivity and optimization efficiency. Accordingly, this paper proposes a PSO that simulates the evolutionary dynamics of species survival in mountain peak ecology (PEPSO) for feature selection. Based on the pyramid topology, the algorithm simulates the features of mountain peak ecology in nature and the competitive-cooperative strategies among species. According to the principles of the algorithm, the population is first adaptively divided into many subgroups based on the fitness level of particles. Then, particles within each subgroup are divided into three different types based on their evolutionary levels, employing different adaptive inertia weight rules and dynamic learning mechanisms to define distinct learning modes. Consequently, all particles play their respective roles in promoting the global optimization performance of the algorithm, similar to different species in the ecological pattern of mountain peaks. Experimental validation of the PEPSO performance was conducted on 18 public datasets. The experimental results demonstrate that the PEPSO outperforms other PSO variant-based feature selection methods and mainstream feature selection methods based on intelligent optimization algorithms in terms of overall performance in global search capability, classification accuracy, and reduction of feature space dimensions. Wilcoxon signed-rank test also confirms the excellent performance of the PEPSO.展开更多
Casing damage resulting from sand production in unconsolidated sandstone reservoirs can significantly impact the average production of oil wells.However,the prediction task remains challenging due to the complex damag...Casing damage resulting from sand production in unconsolidated sandstone reservoirs can significantly impact the average production of oil wells.However,the prediction task remains challenging due to the complex damage mechanism caused by sand production.This paper presents an innovative approach that combines feature selection(FS)with boosting algorithms to accurately predict casing damage in unconsolidated sandstone reservoirs.A novel TriScore FS technique is developed,combining mRMR,Random Forest,and F-test.The approach integrates three distinct feature selection approaches—TriScore,wrapper,and hybrid TriScore-wrapper and four interpretable Boosting models(AdaBoost,XGBoost,LightGBM,CatBoost).Moreover,shapley additive explanations(SHAP)was used to identify the most significant features across engineering,geological,and production features.The CatBoost model,using the Hybrid TriScore-rapper G_(1)G_(2)FS method,showed exceptional performance in analyzing data from the Gangxi Oilfield.It achieved the highestaccuracy(95.5%)and recall rate(89.7%)compared to other tested models.Casing service time,casing wall thickness,and perforation density were selected as the top three most important features.This framework enhances predictive robustness and is an effective tool for policymakers and energy analysts,confirming its capability to deliver reliable casing damage forecasts.展开更多
In this paper,we propose a new full-Newton step feasible interior-point algorithm for the special weighted linear complementarity problems.The proposed algorithm employs the technique of algebraic equivalent transform...In this paper,we propose a new full-Newton step feasible interior-point algorithm for the special weighted linear complementarity problems.The proposed algorithm employs the technique of algebraic equivalent transformation to derive the search direction.It is shown that the proximity measure reduces quadratically at each iteration.Moreover,the iteration bound of the algorithm is as good as the best-known polynomial complexity for these types of problems.Furthermore,numerical results are presented to show the efficiency of the proposed algorithm.展开更多
Software defect prediction(SDP)aims to find a reliable method to predict defects in specific software projects and help software engineers allocate limited resources to release high-quality software products.Software ...Software defect prediction(SDP)aims to find a reliable method to predict defects in specific software projects and help software engineers allocate limited resources to release high-quality software products.Software defect prediction can be effectively performed using traditional features,but there are some redundant or irrelevant features in them(the presence or absence of this feature has little effect on the prediction results).These problems can be solved using feature selection.However,existing feature selection methods have shortcomings such as insignificant dimensionality reduction effect and low classification accuracy of the selected optimal feature subset.In order to reduce the impact of these shortcomings,this paper proposes a new feature selection method Cubic TraverseMa Beluga whale optimization algorithm(CTMBWO)based on the improved Beluga whale optimization algorithm(BWO).The goal of this study is to determine how well the CTMBWO can extract the features that are most important for correctly predicting software defects,improve the accuracy of fault prediction,reduce the number of the selected feature and mitigate the risk of overfitting,thereby achieving more efficient resource utilization and better distribution of test workload.The CTMBWO comprises three main stages:preprocessing the dataset,selecting relevant features,and evaluating the classification performance of the model.The novel feature selection method can effectively improve the performance of SDP.This study performs experiments on two software defect datasets(PROMISE,NASA)and shows the method’s classification performance using four detailed evaluation metrics,Accuracy,F1-score,MCC,AUC and Recall.The results indicate that the approach presented in this paper achieves outstanding classification performance on both datasets and has significant improvement over the baseline models.展开更多
In recent years,feature selection(FS)optimization of high-dimensional gene expression data has become one of the most promising approaches for cancer prediction and classification.This work reviews FS and classificati...In recent years,feature selection(FS)optimization of high-dimensional gene expression data has become one of the most promising approaches for cancer prediction and classification.This work reviews FS and classification methods that utilize evolutionary algorithms(EAs)for gene expression profiles in cancer or medical applications based on research motivations,challenges,and recommendations.Relevant studies were retrieved from four major academic databases-IEEE,Scopus,Springer,and ScienceDirect-using the keywords‘cancer classification’,‘optimization’,‘FS’,and‘gene expression profile’.A total of 67 papers were finally selected with key advancements identified as follows:(1)The majority of papers(44.8%)focused on developing algorithms and models for FS and classification.(2)The second category encompassed studies on biomarker identification by EAs,including 20 papers(30%).(3)The third category comprised works that applied FS to cancer data for decision support system purposes,addressing high-dimensional data and the formulation of chromosome length.These studies accounted for 12%of the total number of studies.(4)The remaining three papers(4.5%)were reviews and surveys focusing on models and developments in prediction and classification optimization for cancer classification under current technical conditions.This review highlights the importance of optimizing FS in EAs to manage high-dimensional data effectively.Despite recent advancements,significant limitations remain:the dynamic formulation of chromosome length remains an underexplored area.Thus,further research is needed on dynamic-length chromosome techniques for more sophisticated biomarker gene selection techniques.The findings suggest that further advancements in dynamic chromosome length formulations and adaptive algorithms could enhance cancer classification accuracy and efficiency.展开更多
With the birth of Software-Defined Networking(SDN),integration of both SDN and traditional architectures becomes the development trend of computer networks.Network intrusion detection faces challenges in dealing with ...With the birth of Software-Defined Networking(SDN),integration of both SDN and traditional architectures becomes the development trend of computer networks.Network intrusion detection faces challenges in dealing with complex attacks in SDN environments,thus to address the network security issues from the viewpoint of Artificial Intelligence(AI),this paper introduces the Crayfish Optimization Algorithm(COA)to the field of intrusion detection for both SDN and traditional network architectures,and based on the characteristics of the original COA,an Improved Crayfish Optimization Algorithm(ICOA)is proposed by integrating strategies of elite reverse learning,Levy flight,crowding factor and parameter modification.The ICOA is then utilized for AI-integrated feature selection of intrusion detection for both SDN and traditional network architectures,to reduce the dimensionality of the data and improve the performance of network intrusion detection.Finally,the performance evaluation is performed by testing not only the NSL-KDD dataset and the UNSW-NB 15 dataset for traditional networks but also the InSDN dataset for SDN-based networks.Experimental results show that ICOA improves the accuracy by 0.532%and 2.928%respectively compared with GWO and COA in traditional networks.In SDN networks,the accuracy of ICOA is 0.25%and 0.3%higher than COA and PSO.These findings collectively indicate that AI-integrated feature selection based on the proposed ICOA can promote network intrusion detection for both SDN and traditional architectures.展开更多
To diagnose the feasibility of the solution of a job-shop scheduling problem(JSSP),a test algorithm based on diagraph and heuristic search is developed and verified through a case study.Meanwhile,a new repair algori...To diagnose the feasibility of the solution of a job-shop scheduling problem(JSSP),a test algorithm based on diagraph and heuristic search is developed and verified through a case study.Meanwhile,a new repair algorithm for modifying an infeasible solution of the JSSP to become a feasible solution is proposed for the general JSSP.The computational complexity of the test algorithm and the repair algorithm is both O(n) under the worst-case scenario,and O(2J+M) for the repair algorithm under the best-case scenario.The repair algorithm is not limited to specific optimization methods,such as local tabu search,genetic algorithms and shifting bottleneck procedures for job shop scheduling,but applicable to generic infeasible solutions for the JSSP to achieve feasibility.展开更多
[Objective] The aim was to study the feature extraction of stored-grain insects based on ant colony optimization and support vector machine algorithm, and to explore the feasibility of the feature extraction of stored...[Objective] The aim was to study the feature extraction of stored-grain insects based on ant colony optimization and support vector machine algorithm, and to explore the feasibility of the feature extraction of stored-grain insects. [Method] Through the analysis of feature extraction in the image recognition of the stored-grain insects, the recognition accuracy of the cross-validation training model in support vector machine (SVM) algorithm was taken as an important factor of the evaluation principle of feature extraction of stored-grain insects. The ant colony optimization (ACO) algorithm was applied to the automatic feature extraction of stored-grain insects. [Result] The algorithm extracted the optimal feature subspace of seven features from the 17 morphological features, including area and perimeter. The ninety image samples of the stored-grain insects were automatically recognized by the optimized SVM classifier, and the recognition accuracy was over 95%. [Conclusion] The experiment shows that the application of ant colony optimization to the feature extraction of grain insects is practical and feasible.展开更多
Combining the adaptive shrinkage genetic algorithm in the feasible region with the imaging of apparent vertical conductance differential, we have inverted the TEM conductive thin layer. The result of the inversion dem...Combining the adaptive shrinkage genetic algorithm in the feasible region with the imaging of apparent vertical conductance differential, we have inverted the TEM conductive thin layer. The result of the inversion demonstrates that by adaptive shrinkage in the feasible region, the calculation speed accelerates and the calculation precision improves. To a certain extent, in this method we surmount the transient electromagnetic sounding equivalence and reduced equivalence scope. Comparison of the inverted result with the forward curve clearly shows that we can image the conductive thin layer.展开更多
A globally convergent infeasible-interior-point predictor-corrector algorithm is presented for the second-order cone programming (SOCP) by using the Alizadeh- Haeberly-Overton (AHO) search direction. This algorith...A globally convergent infeasible-interior-point predictor-corrector algorithm is presented for the second-order cone programming (SOCP) by using the Alizadeh- Haeberly-Overton (AHO) search direction. This algorithm does not require the feasibility of the initial points and iteration points. Under suitable assumptions, it is shown that the algorithm can find an -approximate solution of an SOCP in at most O(√n ln(ε0/ε)) iterations. The iteration-complexity bound of our algorithm is almost the same as the best known bound of feasible interior point algorithms for the SOCP.展开更多
With the rapid increase in the number of three-dimensional (3D) models each year, to quickly and easily find the part desired has become a big challenge of enterprises. Meanwhile, many methods and algorithms have be...With the rapid increase in the number of three-dimensional (3D) models each year, to quickly and easily find the part desired has become a big challenge of enterprises. Meanwhile, many methods and algorithms have been proposed for part retrieval. However, most of the existing methods are designed lbr mechanical parts, and can not be well worked for sheet metal part re- trieval. An approach to feature-based retrieval of sheet metal parts is presented. Firstly, the features frequently used in sheet metal part design are chosen as the "'key words" in retrieval. Based on those features, a relative position model is built to express the different relationships of the features in 3D space. Secondly, a description method of the model is studied. With the descrip- tion method the relative position of features in sheet metal parts can be expressed by four location description matrices. Thirdly, based on the relative position model and location description matrices, the equivalent definition of relationships of two feature groups is given which is the basis to calculate the similarity of two sheet metal parts. Next, the tbrmula of retrieval algorithm for sheet metal parts is given. Finally, a prototype system is developed to test and verify the effectiveness of the retrieval method suggested. Experiments verify that the new method is able to meet the requirements of searching sheet metal parts and possesses potentials in practical application.展开更多
This paper presents an all-parametric model of radar target in optic region, in which the localized scattering center's frequency and aspect angle dependent scattering level, distance and azimuth locations are mod...This paper presents an all-parametric model of radar target in optic region, in which the localized scattering center's frequency and aspect angle dependent scattering level, distance and azimuth locations are modeled as the feature vectors. And the traditional TLS-Prony algorithm is modified to extract these feature vectors. The analysis of Cramer-Rao bound shows that the modified algorithm not only improves the restriction of high signal-to-noise ratio(SNR)threshold of traditional TLS-Prony algorithm, but also is suitable to the extraction of big damped coefficients and high-resolution estimation of near separation poles. Finally, an illustrative example is presented to verify its practicability in the applications. The experimental results show that the method developed can not only recognize two airplane-like targets with similar shape at low SNR, but also compress the original radar data with high fidelity.展开更多
A new method, SVD-TLS extending Prony algorithm, is introduced for extracting UWB radar target features. The method is a modified classical Prony method based on singular value decomposition and total least squares th...A new method, SVD-TLS extending Prony algorithm, is introduced for extracting UWB radar target features. The method is a modified classical Prony method based on singular value decomposition and total least squares that can improve robust for spectrum estimation. Simulation results show that poles and residuum of target echo can be extracted effectively using this method, and at the same time, random noises can be restrained to some degree. It is applicable for target feature extraction such as UWB radar or other high resolution range radars.展开更多
Due to requirements and necessities in digital image research, image matching is considered as a key, essential and complicating point especially for machine learning. According to its convenience and facility, the mo...Due to requirements and necessities in digital image research, image matching is considered as a key, essential and complicating point especially for machine learning. According to its convenience and facility, the most applied algorithm for image feature point extraction and matching is Speeded-Up Robust Feature (SURF). The enhancement for scale invariant feature transform (SIFT) algorithm promotes the effectiveness of the algorithm as well as facilitates the possibility, while the application of the algorithm is being applied in a present time computer vision system. In this research work, the aim of SURF algorithm is to extract image features, and we have incorporated RANSAC algorithm to filter matching points. The images were juxtaposed and asserted experiments utilizing pertinent image improvement methods. The idea based on merging improvement technology through SURF algorithm is put forward to get better quality of feature points matching the efficiency and appropriate image improvement methods are adopted for different feature images which are compared and verified by experiments. Some results have been explained there which are the effects of lighting on the underexposed and overexposed images.展开更多
The advent of Big Data has rendered Machine Learning tasks more intricate as they frequently involve higher-dimensional data.Feature Selection(FS)methods can abate the complexity of the data and enhance the accuracy,g...The advent of Big Data has rendered Machine Learning tasks more intricate as they frequently involve higher-dimensional data.Feature Selection(FS)methods can abate the complexity of the data and enhance the accuracy,generalizability,and interpretability of models.Meta-heuristic algorithms are often utilized for FS tasks due to their low requirements and efficient performance.This paper introduces an augmented Forensic-Based Investigation algorithm(DCFBI)that incorporates a Dynamic Individual Selection(DIS)and crisscross(CC)mechanism to improve the pursuit phase of the FBI.Moreover,a binary version of DCFBI(BDCFBI)is applied to FS.Experiments conducted on IEEE CEC 2017 with other metaheuristics demonstrate that DCFBI surpasses them in search capability.The influence of different mechanisms on the original FBI is analyzed on benchmark functions,while its scalability is verified by comparing it with the original FBI on benchmarks with varied dimensions.BDCFBI is then applied to 18 real datasets from the UCI machine learning database and the Wieslaw dataset to select near-optimal features,which are then compared with six renowned binary metaheuristics.The results show that BDCFBI can be more competitive than similar methods and acquire a subset of features with superior classification accuracy.展开更多
A new spectral matching algorithm is proposed by us- ing nonsubsampled contourlet transform and scale-invariant fea- ture transform. The nonsubsampled contourlet transform is used to decompose an image into a low freq...A new spectral matching algorithm is proposed by us- ing nonsubsampled contourlet transform and scale-invariant fea- ture transform. The nonsubsampled contourlet transform is used to decompose an image into a low frequency image and several high frequency images, and the scale-invariant feature transform is employed to extract feature points from the low frequency im- age. A proximity matrix is constructed for the feature points of two related images. By singular value decomposition of the proximity matrix, a matching matrix (or matching result) reflecting the match- ing degree among feature points is obtained. Experimental results indicate that the proposed algorithm can reduce time complexity and possess a higher accuracy.展开更多
Prediction plays a vital role in decision making. Correct prediction leads to right decision making to save the life, energy,efforts, money and time. The right decision prevents physical and material losses and it is ...Prediction plays a vital role in decision making. Correct prediction leads to right decision making to save the life, energy,efforts, money and time. The right decision prevents physical and material losses and it is practiced in all the fields including medical,finance, environmental studies, engineering and emerging technologies. Prediction is carried out by a model called classifier. The predictive accuracy of the classifier highly depends on the training datasets utilized for training the classifier. The irrelevant and redundant features of the training dataset reduce the accuracy of the classifier. Hence, the irrelevant and redundant features must be removed from the training dataset through the process known as feature selection. This paper proposes a feature selection algorithm namely unsupervised learning with ranking based feature selection(FSULR). It removes redundant features by clustering and eliminates irrelevant features by statistical measures to select the most significant features from the training dataset. The performance of this proposed algorithm is compared with the other seven feature selection algorithms by well known classifiers namely naive Bayes(NB),instance based(IB1) and tree based J48. Experimental results show that the proposed algorithm yields better prediction accuracy for classifiers.展开更多
This paper deals with a bi-extrapolated subgradient projection algorithm by intro- ducing two extrapolated factors in the iterative step to solve the multiple-sets split feasibility problem. The strategy is intend to ...This paper deals with a bi-extrapolated subgradient projection algorithm by intro- ducing two extrapolated factors in the iterative step to solve the multiple-sets split feasibility problem. The strategy is intend to improve the convergence. And its convergence is proved un- der some suitable conditions. Numerical results illustrate that the bi-extrapolated subgradient projection algorithm converges more quickly than the existing algorithms.展开更多
: This paper proposes a new sequential similarity detection algorithm (SSDA), which can overcome matching error caused by grayscale distortion; meanwhile, time consumption is much less than that of regular algorith...: This paper proposes a new sequential similarity detection algorithm (SSDA), which can overcome matching error caused by grayscale distortion; meanwhile, time consumption is much less than that of regular algorithms based on image feature. The algorithm adopts Sobel operator to deal with subgraph and template image, and regards the region which has maximum relevance as final result. In order to solve time-consuming problem existing in original algorithm, a coarse-to-fine matching method is put forward. Besides, the location correlation keeps updating and remains the minimum value in the whole scanning process, which can significantly decrease time consumption. Experiments show that the algorithm proposed in this article can not only overcome gray distortion, but also ensure accuracy. Time consumption is at least one time orders of magnitude shorter than that of primal algorithm.展开更多
基金funded by the General Program of the National Natural Science Foundation of China grant number[62277022].
文摘Algorithms are the primary component of Artificial Intelligence(AI).The algorithm is the process in AI that imitates the human mind to solve problems.Currently evaluating the performance of AI is achieved by evaluating AI algorithms by metric scores on data sets.However the evaluation of algorithms in AI is challenging because the evaluation of the same type of algorithm has many data sets and evaluation metrics.Different algorithms may have individual strengths and weaknesses in evaluation metric scores on separate data sets,lacking the credibility and validity of the evaluation.Moreover,evaluation of algorithms requires repeated experiments on different data sets,reducing the attention of researchers to the research of the algorithms itself.Crucially,this approach to evaluating comparative metric scores does not take into account the algorithm’s ability to solve problems.And the classical algorithm evaluation of time and space complexity is not suitable for evaluating AI algorithms.Because classical algorithms input is infinite numbers,whereas AI algorithms input is a data set,which is limited and multifarious.According to the AI algorithm evaluation without response to the problem solving capability,this paper summarizes the features of AI algorithm evaluation and proposes an AI evaluation method that incorporates the problem-solving capabilities of algorithms.
文摘In recent years, particle swarm optimization (PSO) has received widespread attention in feature selection due to its simplicity and potential for global search. However, in traditional PSO, particles primarily update based on two extreme values: personal best and global best, which limits the diversity of information. Ideally, particles should learn from multiple advantageous particles to enhance interactivity and optimization efficiency. Accordingly, this paper proposes a PSO that simulates the evolutionary dynamics of species survival in mountain peak ecology (PEPSO) for feature selection. Based on the pyramid topology, the algorithm simulates the features of mountain peak ecology in nature and the competitive-cooperative strategies among species. According to the principles of the algorithm, the population is first adaptively divided into many subgroups based on the fitness level of particles. Then, particles within each subgroup are divided into three different types based on their evolutionary levels, employing different adaptive inertia weight rules and dynamic learning mechanisms to define distinct learning modes. Consequently, all particles play their respective roles in promoting the global optimization performance of the algorithm, similar to different species in the ecological pattern of mountain peaks. Experimental validation of the PEPSO performance was conducted on 18 public datasets. The experimental results demonstrate that the PEPSO outperforms other PSO variant-based feature selection methods and mainstream feature selection methods based on intelligent optimization algorithms in terms of overall performance in global search capability, classification accuracy, and reduction of feature space dimensions. Wilcoxon signed-rank test also confirms the excellent performance of the PEPSO.
基金funded by the National Natural Science Foundation Project(Grant No.52274015)the National Science and Technology Major Project(Grant No.2025ZD1402205)。
文摘Casing damage resulting from sand production in unconsolidated sandstone reservoirs can significantly impact the average production of oil wells.However,the prediction task remains challenging due to the complex damage mechanism caused by sand production.This paper presents an innovative approach that combines feature selection(FS)with boosting algorithms to accurately predict casing damage in unconsolidated sandstone reservoirs.A novel TriScore FS technique is developed,combining mRMR,Random Forest,and F-test.The approach integrates three distinct feature selection approaches—TriScore,wrapper,and hybrid TriScore-wrapper and four interpretable Boosting models(AdaBoost,XGBoost,LightGBM,CatBoost).Moreover,shapley additive explanations(SHAP)was used to identify the most significant features across engineering,geological,and production features.The CatBoost model,using the Hybrid TriScore-rapper G_(1)G_(2)FS method,showed exceptional performance in analyzing data from the Gangxi Oilfield.It achieved the highestaccuracy(95.5%)and recall rate(89.7%)compared to other tested models.Casing service time,casing wall thickness,and perforation density were selected as the top three most important features.This framework enhances predictive robustness and is an effective tool for policymakers and energy analysts,confirming its capability to deliver reliable casing damage forecasts.
基金Supported by the Optimisation Theory and Algorithm Research Team(Grant No.23kytdzd004)University Science Research Project of Anhui Province(Grant No.2024AH050631)the General Programs for Young Teacher Cultivation of Educational Commission of Anhui Province(Grant No.YQYB2023090).
文摘In this paper,we propose a new full-Newton step feasible interior-point algorithm for the special weighted linear complementarity problems.The proposed algorithm employs the technique of algebraic equivalent transformation to derive the search direction.It is shown that the proximity measure reduces quadratically at each iteration.Moreover,the iteration bound of the algorithm is as good as the best-known polynomial complexity for these types of problems.Furthermore,numerical results are presented to show the efficiency of the proposed algorithm.
文摘Software defect prediction(SDP)aims to find a reliable method to predict defects in specific software projects and help software engineers allocate limited resources to release high-quality software products.Software defect prediction can be effectively performed using traditional features,but there are some redundant or irrelevant features in them(the presence or absence of this feature has little effect on the prediction results).These problems can be solved using feature selection.However,existing feature selection methods have shortcomings such as insignificant dimensionality reduction effect and low classification accuracy of the selected optimal feature subset.In order to reduce the impact of these shortcomings,this paper proposes a new feature selection method Cubic TraverseMa Beluga whale optimization algorithm(CTMBWO)based on the improved Beluga whale optimization algorithm(BWO).The goal of this study is to determine how well the CTMBWO can extract the features that are most important for correctly predicting software defects,improve the accuracy of fault prediction,reduce the number of the selected feature and mitigate the risk of overfitting,thereby achieving more efficient resource utilization and better distribution of test workload.The CTMBWO comprises three main stages:preprocessing the dataset,selecting relevant features,and evaluating the classification performance of the model.The novel feature selection method can effectively improve the performance of SDP.This study performs experiments on two software defect datasets(PROMISE,NASA)and shows the method’s classification performance using four detailed evaluation metrics,Accuracy,F1-score,MCC,AUC and Recall.The results indicate that the approach presented in this paper achieves outstanding classification performance on both datasets and has significant improvement over the baseline models.
基金funded by the Ministry of Higher Education of Malaysia,grant number FRGS/1/2022/ICT02/UPSI/02/1.
文摘In recent years,feature selection(FS)optimization of high-dimensional gene expression data has become one of the most promising approaches for cancer prediction and classification.This work reviews FS and classification methods that utilize evolutionary algorithms(EAs)for gene expression profiles in cancer or medical applications based on research motivations,challenges,and recommendations.Relevant studies were retrieved from four major academic databases-IEEE,Scopus,Springer,and ScienceDirect-using the keywords‘cancer classification’,‘optimization’,‘FS’,and‘gene expression profile’.A total of 67 papers were finally selected with key advancements identified as follows:(1)The majority of papers(44.8%)focused on developing algorithms and models for FS and classification.(2)The second category encompassed studies on biomarker identification by EAs,including 20 papers(30%).(3)The third category comprised works that applied FS to cancer data for decision support system purposes,addressing high-dimensional data and the formulation of chromosome length.These studies accounted for 12%of the total number of studies.(4)The remaining three papers(4.5%)were reviews and surveys focusing on models and developments in prediction and classification optimization for cancer classification under current technical conditions.This review highlights the importance of optimizing FS in EAs to manage high-dimensional data effectively.Despite recent advancements,significant limitations remain:the dynamic formulation of chromosome length remains an underexplored area.Thus,further research is needed on dynamic-length chromosome techniques for more sophisticated biomarker gene selection techniques.The findings suggest that further advancements in dynamic chromosome length formulations and adaptive algorithms could enhance cancer classification accuracy and efficiency.
基金supported by the National Natural Science Foundation of China under Grant 61602162the Hubei Provincial Science and Technology Plan Project under Grant 2023BCB041.
文摘With the birth of Software-Defined Networking(SDN),integration of both SDN and traditional architectures becomes the development trend of computer networks.Network intrusion detection faces challenges in dealing with complex attacks in SDN environments,thus to address the network security issues from the viewpoint of Artificial Intelligence(AI),this paper introduces the Crayfish Optimization Algorithm(COA)to the field of intrusion detection for both SDN and traditional network architectures,and based on the characteristics of the original COA,an Improved Crayfish Optimization Algorithm(ICOA)is proposed by integrating strategies of elite reverse learning,Levy flight,crowding factor and parameter modification.The ICOA is then utilized for AI-integrated feature selection of intrusion detection for both SDN and traditional network architectures,to reduce the dimensionality of the data and improve the performance of network intrusion detection.Finally,the performance evaluation is performed by testing not only the NSL-KDD dataset and the UNSW-NB 15 dataset for traditional networks but also the InSDN dataset for SDN-based networks.Experimental results show that ICOA improves the accuracy by 0.532%and 2.928%respectively compared with GWO and COA in traditional networks.In SDN networks,the accuracy of ICOA is 0.25%and 0.3%higher than COA and PSO.These findings collectively indicate that AI-integrated feature selection based on the proposed ICOA can promote network intrusion detection for both SDN and traditional architectures.
基金The US National Science Foundation (No. CMMI-0408390, CMMI-0644552)the Research Fellowship for International Young Scientists (No. 51050110143)+2 种基金the Fok Ying-Tong Education Foundation(No. 114024)the Natural Science Foundation of Jiangsu Province (No.BK2009015)the Postdoctoral Science Foundation of Jiangsu Province (No.0901005C)
文摘To diagnose the feasibility of the solution of a job-shop scheduling problem(JSSP),a test algorithm based on diagraph and heuristic search is developed and verified through a case study.Meanwhile,a new repair algorithm for modifying an infeasible solution of the JSSP to become a feasible solution is proposed for the general JSSP.The computational complexity of the test algorithm and the repair algorithm is both O(n) under the worst-case scenario,and O(2J+M) for the repair algorithm under the best-case scenario.The repair algorithm is not limited to specific optimization methods,such as local tabu search,genetic algorithms and shifting bottleneck procedures for job shop scheduling,but applicable to generic infeasible solutions for the JSSP to achieve feasibility.
基金Supported by the National Natural Science Foundation of China(31101085)the Program for Young Core Teachers of Colleges in Henan(2011GGJS-094)the Scientific Research Project for the High Level Talents,North China University of Water Conservancy and Hydroelectric Power~~
文摘[Objective] The aim was to study the feature extraction of stored-grain insects based on ant colony optimization and support vector machine algorithm, and to explore the feasibility of the feature extraction of stored-grain insects. [Method] Through the analysis of feature extraction in the image recognition of the stored-grain insects, the recognition accuracy of the cross-validation training model in support vector machine (SVM) algorithm was taken as an important factor of the evaluation principle of feature extraction of stored-grain insects. The ant colony optimization (ACO) algorithm was applied to the automatic feature extraction of stored-grain insects. [Result] The algorithm extracted the optimal feature subspace of seven features from the 17 morphological features, including area and perimeter. The ninety image samples of the stored-grain insects were automatically recognized by the optimized SVM classifier, and the recognition accuracy was over 95%. [Conclusion] The experiment shows that the application of ant colony optimization to the feature extraction of grain insects is practical and feasible.
文摘Combining the adaptive shrinkage genetic algorithm in the feasible region with the imaging of apparent vertical conductance differential, we have inverted the TEM conductive thin layer. The result of the inversion demonstrates that by adaptive shrinkage in the feasible region, the calculation speed accelerates and the calculation precision improves. To a certain extent, in this method we surmount the transient electromagnetic sounding equivalence and reduced equivalence scope. Comparison of the inverted result with the forward curve clearly shows that we can image the conductive thin layer.
基金the National Science Foundation(60574075, 60674108)
文摘A globally convergent infeasible-interior-point predictor-corrector algorithm is presented for the second-order cone programming (SOCP) by using the Alizadeh- Haeberly-Overton (AHO) search direction. This algorithm does not require the feasibility of the initial points and iteration points. Under suitable assumptions, it is shown that the algorithm can find an -approximate solution of an SOCP in at most O(√n ln(ε0/ε)) iterations. The iteration-complexity bound of our algorithm is almost the same as the best known bound of feasible interior point algorithms for the SOCP.
基金National High-tech Research and Development Program of China (2009AA043302)
文摘With the rapid increase in the number of three-dimensional (3D) models each year, to quickly and easily find the part desired has become a big challenge of enterprises. Meanwhile, many methods and algorithms have been proposed for part retrieval. However, most of the existing methods are designed lbr mechanical parts, and can not be well worked for sheet metal part re- trieval. An approach to feature-based retrieval of sheet metal parts is presented. Firstly, the features frequently used in sheet metal part design are chosen as the "'key words" in retrieval. Based on those features, a relative position model is built to express the different relationships of the features in 3D space. Secondly, a description method of the model is studied. With the descrip- tion method the relative position of features in sheet metal parts can be expressed by four location description matrices. Thirdly, based on the relative position model and location description matrices, the equivalent definition of relationships of two feature groups is given which is the basis to calculate the similarity of two sheet metal parts. Next, the tbrmula of retrieval algorithm for sheet metal parts is given. Finally, a prototype system is developed to test and verify the effectiveness of the retrieval method suggested. Experiments verify that the new method is able to meet the requirements of searching sheet metal parts and possesses potentials in practical application.
文摘This paper presents an all-parametric model of radar target in optic region, in which the localized scattering center's frequency and aspect angle dependent scattering level, distance and azimuth locations are modeled as the feature vectors. And the traditional TLS-Prony algorithm is modified to extract these feature vectors. The analysis of Cramer-Rao bound shows that the modified algorithm not only improves the restriction of high signal-to-noise ratio(SNR)threshold of traditional TLS-Prony algorithm, but also is suitable to the extraction of big damped coefficients and high-resolution estimation of near separation poles. Finally, an illustrative example is presented to verify its practicability in the applications. The experimental results show that the method developed can not only recognize two airplane-like targets with similar shape at low SNR, but also compress the original radar data with high fidelity.
文摘A new method, SVD-TLS extending Prony algorithm, is introduced for extracting UWB radar target features. The method is a modified classical Prony method based on singular value decomposition and total least squares that can improve robust for spectrum estimation. Simulation results show that poles and residuum of target echo can be extracted effectively using this method, and at the same time, random noises can be restrained to some degree. It is applicable for target feature extraction such as UWB radar or other high resolution range radars.
文摘Due to requirements and necessities in digital image research, image matching is considered as a key, essential and complicating point especially for machine learning. According to its convenience and facility, the most applied algorithm for image feature point extraction and matching is Speeded-Up Robust Feature (SURF). The enhancement for scale invariant feature transform (SIFT) algorithm promotes the effectiveness of the algorithm as well as facilitates the possibility, while the application of the algorithm is being applied in a present time computer vision system. In this research work, the aim of SURF algorithm is to extract image features, and we have incorporated RANSAC algorithm to filter matching points. The images were juxtaposed and asserted experiments utilizing pertinent image improvement methods. The idea based on merging improvement technology through SURF algorithm is put forward to get better quality of feature points matching the efficiency and appropriate image improvement methods are adopted for different feature images which are compared and verified by experiments. Some results have been explained there which are the effects of lighting on the underexposed and overexposed images.
基金supported by Special Fund of Fundamental Scientific Research Business Expense for Higher School of Central Government(ZY20180119)the Natural Science Foundation of Zhejiang Province(LZ22F020005)+1 种基金the Natural Science Foundation of Hebei Province(D2022512001)National Natural Science Foundation of China(42164002,62076185).
文摘The advent of Big Data has rendered Machine Learning tasks more intricate as they frequently involve higher-dimensional data.Feature Selection(FS)methods can abate the complexity of the data and enhance the accuracy,generalizability,and interpretability of models.Meta-heuristic algorithms are often utilized for FS tasks due to their low requirements and efficient performance.This paper introduces an augmented Forensic-Based Investigation algorithm(DCFBI)that incorporates a Dynamic Individual Selection(DIS)and crisscross(CC)mechanism to improve the pursuit phase of the FBI.Moreover,a binary version of DCFBI(BDCFBI)is applied to FS.Experiments conducted on IEEE CEC 2017 with other metaheuristics demonstrate that DCFBI surpasses them in search capability.The influence of different mechanisms on the original FBI is analyzed on benchmark functions,while its scalability is verified by comparing it with the original FBI on benchmarks with varied dimensions.BDCFBI is then applied to 18 real datasets from the UCI machine learning database and the Wieslaw dataset to select near-optimal features,which are then compared with six renowned binary metaheuristics.The results show that BDCFBI can be more competitive than similar methods and acquire a subset of features with superior classification accuracy.
基金supported by the National Natural Science Foundation of China (6117212711071002)+1 种基金the Specialized Research Fund for the Doctoral Program of Higher Education (20113401110006)the Innovative Research Team of 211 Project in Anhui University (KJTD007A)
文摘A new spectral matching algorithm is proposed by us- ing nonsubsampled contourlet transform and scale-invariant fea- ture transform. The nonsubsampled contourlet transform is used to decompose an image into a low frequency image and several high frequency images, and the scale-invariant feature transform is employed to extract feature points from the low frequency im- age. A proximity matrix is constructed for the feature points of two related images. By singular value decomposition of the proximity matrix, a matching matrix (or matching result) reflecting the match- ing degree among feature points is obtained. Experimental results indicate that the proposed algorithm can reduce time complexity and possess a higher accuracy.
文摘Prediction plays a vital role in decision making. Correct prediction leads to right decision making to save the life, energy,efforts, money and time. The right decision prevents physical and material losses and it is practiced in all the fields including medical,finance, environmental studies, engineering and emerging technologies. Prediction is carried out by a model called classifier. The predictive accuracy of the classifier highly depends on the training datasets utilized for training the classifier. The irrelevant and redundant features of the training dataset reduce the accuracy of the classifier. Hence, the irrelevant and redundant features must be removed from the training dataset through the process known as feature selection. This paper proposes a feature selection algorithm namely unsupervised learning with ranking based feature selection(FSULR). It removes redundant features by clustering and eliminates irrelevant features by statistical measures to select the most significant features from the training dataset. The performance of this proposed algorithm is compared with the other seven feature selection algorithms by well known classifiers namely naive Bayes(NB),instance based(IB1) and tree based J48. Experimental results show that the proposed algorithm yields better prediction accuracy for classifiers.
基金Supported by Natural Science Foundation of Shanghai(14ZR1429200)National Science Foundation of China(11171221)+4 种基金Shanghai Leading Academic Discipline Project(XTKX2012)Innovation Program of Shanghai Municipal Education Commission(14YZ094)Doctoral Program Foundation of Institutions of Higher Educationof China(20123120110004)Doctoral Starting Projection of the University of Shanghai for Science and Technology(ID-10-303-002)Young Teacher Training Projection Program of Shanghai for Science and Technology
文摘This paper deals with a bi-extrapolated subgradient projection algorithm by intro- ducing two extrapolated factors in the iterative step to solve the multiple-sets split feasibility problem. The strategy is intend to improve the convergence. And its convergence is proved un- der some suitable conditions. Numerical results illustrate that the bi-extrapolated subgradient projection algorithm converges more quickly than the existing algorithms.
基金the National Natural Science Foundation of China(No.61165008)
文摘: This paper proposes a new sequential similarity detection algorithm (SSDA), which can overcome matching error caused by grayscale distortion; meanwhile, time consumption is much less than that of regular algorithms based on image feature. The algorithm adopts Sobel operator to deal with subgraph and template image, and regards the region which has maximum relevance as final result. In order to solve time-consuming problem existing in original algorithm, a coarse-to-fine matching method is put forward. Besides, the location correlation keeps updating and remains the minimum value in the whole scanning process, which can significantly decrease time consumption. Experiments show that the algorithm proposed in this article can not only overcome gray distortion, but also ensure accuracy. Time consumption is at least one time orders of magnitude shorter than that of primal algorithm.