Advanced intensity measures(IMs)based on an inelastic deformation spectrum improved the evaluation of the median engineering demand parameters(EDPs)and reduced dispersion.In this regard,an optimized two-degreefreedom(...Advanced intensity measures(IMs)based on an inelastic deformation spectrum improved the evaluation of the median engineering demand parameters(EDPs)and reduced dispersion.In this regard,an optimized two-degreefreedom(2DOF)modal pushover-based scaling procedure(2DMPS)has been developed for a nonlinear dynamic analysis of asymmetric in-plan buildings.The 2DMPS procedure scales ground motions to approach close enough to a target value of the inelastic displacement of the first-mode inelastic 2DOF modal stick,extended for structures with significant contributions of higher modes.Further,4-,6-and 13-story RC SMRF buildings were selected for analyses using ground motion records scaled by the 2DMPS procedure,the modal pushover-based scaling method(MPS),and ASCE/SEI 7-16 scaling procedures.The median values of EDPs on scaled records closely matched the benchmark results.The bias in the EDP values due to the scaled records in every group regarding their median value was lower than the dispersion of the 21 unscaled records.These results generally demonstrate the accuracy and efficiency of the 2DMPS method.Additionally,the 2DOF modal stick’s inelastic response spectra are better suited for calculating seismic demands for one-way asymmetric-plan structures than the SDOF inelastic response spectra.展开更多
The selection of a suitable navigation area is pivotal in aircraft scene matching guidance technology.This study addresses the challenge of identifying suitable reference image ranges for precise scene matching,which ...The selection of a suitable navigation area is pivotal in aircraft scene matching guidance technology.This study addresses the challenge of identifying suitable reference image ranges for precise scene matching,which is crucial for enhancing aircraft positioning accuracy.Traditional methods for image matchability analysis are often limited by their reliance on manual feature parameter design and threshold-based filtering,resulting in suboptimal accuracy and efficiency.This paper proposes a novel network architecture for selecting suitable navigation areas using image Matching Level Segmentation(MLSNet).The approach involves two key innovations:a method for generating segmentation labels that quantify matchability levels and an end-to-end network architecture for rapid and precise prediction of reference image matchability segmentation maps.The network includes two core modules:the saliency analysis module uses multi-layer convolutional networks to accurately detect image saliency features across various levels and scales;the multidimensional attention module utilizes attention mechanisms to focus on feature channels and spatial neighborhood scenes to assess the image’s matchability.Our method was rigorously tested on an extensive collection of remote sensing images,where it was benchmarked against a range of both traditional and cutting-edge deep learning methods.The findings indicate that MLSNet is significantly superior to traditional methods in accuracy and efficiency of matchability analysis,and is also relatively ahead of state-of-the-art deep learning models.展开更多
Landslide susceptibility prediction(LSP)is significantly affected by the uncertainty issue of landslide related conditioning factor selection.However,most of literature only performs comparative studies on a certain c...Landslide susceptibility prediction(LSP)is significantly affected by the uncertainty issue of landslide related conditioning factor selection.However,most of literature only performs comparative studies on a certain conditioning factor selection method rather than systematically study this uncertainty issue.Targeted,this study aims to systematically explore the influence rules of various commonly used conditioning factor selection methods on LSP,and on this basis to innovatively propose a principle with universal application for optimal selection of conditioning factors.An'yuan County in southern China is taken as example considering 431 landslides and 29 types of conditioning factors.Five commonly used factor selection methods,namely,the correlation analysis(CA),linear regression(LR),principal component analysis(PCA),rough set(RS)and artificial neural network(ANN),are applied to select the optimal factor combinations from the original 29 conditioning factors.The factor selection results are then used as inputs of four types of common machine learning models to construct 20 types of combined models,such as CA-multilayer perceptron,CA-random forest.Additionally,multifactor-based multilayer perceptron random forest models that selecting conditioning factors based on the proposed principle of“accurate data,rich types,clear significance,feasible operation and avoiding duplication”are constructed for comparisons.Finally,the LSP uncertainties are evaluated by the accuracy,susceptibility index distribution,etc.Results show that:(1)multifactor-based models have generally higher LSP performance and lower uncertainties than those of factors selection-based models;(2)Influence degree of different machine learning on LSP accuracy is greater than that of different factor selection methods.Conclusively,the above commonly used conditioning factor selection methods are not ideal for improving LSP performance and may complicate the LSP processes.In contrast,a satisfied combination of conditioning factors can be constructed according to the proposed principle.展开更多
Software defect prediction(SDP)aims to find a reliable method to predict defects in specific software projects and help software engineers allocate limited resources to release high-quality software products.Software ...Software defect prediction(SDP)aims to find a reliable method to predict defects in specific software projects and help software engineers allocate limited resources to release high-quality software products.Software defect prediction can be effectively performed using traditional features,but there are some redundant or irrelevant features in them(the presence or absence of this feature has little effect on the prediction results).These problems can be solved using feature selection.However,existing feature selection methods have shortcomings such as insignificant dimensionality reduction effect and low classification accuracy of the selected optimal feature subset.In order to reduce the impact of these shortcomings,this paper proposes a new feature selection method Cubic TraverseMa Beluga whale optimization algorithm(CTMBWO)based on the improved Beluga whale optimization algorithm(BWO).The goal of this study is to determine how well the CTMBWO can extract the features that are most important for correctly predicting software defects,improve the accuracy of fault prediction,reduce the number of the selected feature and mitigate the risk of overfitting,thereby achieving more efficient resource utilization and better distribution of test workload.The CTMBWO comprises three main stages:preprocessing the dataset,selecting relevant features,and evaluating the classification performance of the model.The novel feature selection method can effectively improve the performance of SDP.This study performs experiments on two software defect datasets(PROMISE,NASA)and shows the method’s classification performance using four detailed evaluation metrics,Accuracy,F1-score,MCC,AUC and Recall.The results indicate that the approach presented in this paper achieves outstanding classification performance on both datasets and has significant improvement over the baseline models.展开更多
Machine learning(ML)is increasingly applied for medical image processing with appropriate learning paradigms.These applications include analyzing images of various organs,such as the brain,lung,eye,etc.,to identify sp...Machine learning(ML)is increasingly applied for medical image processing with appropriate learning paradigms.These applications include analyzing images of various organs,such as the brain,lung,eye,etc.,to identify specific flaws/diseases for diagnosis.The primary concern of ML applications is the precise selection of flexible image features for pattern detection and region classification.Most of the extracted image features are irrelevant and lead to an increase in computation time.Therefore,this article uses an analytical learning paradigm to design a Congruent Feature Selection Method to select the most relevant image features.This process trains the learning paradigm using similarity and correlation-based features over different textural intensities and pixel distributions.The similarity between the pixels over the various distribution patterns with high indexes is recommended for disease diagnosis.Later,the correlation based on intensity and distribution is analyzed to improve the feature selection congruency.Therefore,the more congruent pixels are sorted in the descending order of the selection,which identifies better regions than the distribution.Now,the learning paradigm is trained using intensity and region-based similarity to maximize the chances of selection.Therefore,the probability of feature selection,regardless of the textures and medical image patterns,is improved.This process enhances the performance of ML applications for different medical image processing.The proposed method improves the accuracy,precision,and training rate by 13.19%,10.69%,and 11.06%,respectively,compared to other models for the selected dataset.The mean error and selection time is also reduced by 12.56%and 13.56%,respectively,compared to the same models and dataset.展开更多
Feature selection methods rooted in rough sets confront two notable limitations:their high computa-tional complexity and sensitivity to noise,rendering them impractical for managing large-scale and noisy datasets.The ...Feature selection methods rooted in rough sets confront two notable limitations:their high computa-tional complexity and sensitivity to noise,rendering them impractical for managing large-scale and noisy datasets.The primary issue stems from these methods’undue reliance on all samples.To overcome these challenges,we introduce the concept of cross-similarity grounded in a robust fuzzy relation and design a rapid and robust feature selection algorithm.Firstly,we construct a robust fuzzy relation by introducing a truncation parameter.Then,based on this fuzzy relation,we propose the concept of cross-similarity,which emphasizes the sample-to-sample similarity relations that uniquely determine feature importance,rather than considering all such relations equally.After studying the manifestations and properties of cross-similarity across different fuzzy granularities,we propose a forward greedy feature selection algorithm that leverages cross-similarity as the foundation for information measurement.This algorithm significantly reduces the time complexity from O(m2n2)to O(mn2).Experimental findings reveal that the average runtime of five state-of-the-art comparison algorithms is roughly 3.7 times longer than our algorithm,while our algorithm achieves an average accuracy that surpasses those of the five comparison algorithms by approximately 3.52%.This underscores the effectiveness of our approach.This paper paves the way for applying feature selection algorithms grounded in fuzzy rough sets to large-scale gene datasets.展开更多
The growing demand for privacy-preserving machine learning has positioned federated learning as a promising research paradigm,enabling the training of high-performance models across distributed data sources without co...The growing demand for privacy-preserving machine learning has positioned federated learning as a promising research paradigm,enabling the training of high-performance models across distributed data sources without compromising user privacy.However,despite its advantages,federated learning faces critical challenges arising from the heterogeneity and volatility of participating clients.In real-world scenarios,variations in client participation,data volume,computational capability,and communication reliability contribute to a highly dynamic training environment,which negatively impacts efficiency and convergence of the model.To address these challenges,this paper proposes a novel client selection method named CDE3.First,CDE3 employs a multidimensional model to comprehensively evaluate clients’contributions.Second,we enhance the classical Exp3 algorithm by incorporating a discount factor that exponentially decays historical contributions,thereby increasing the influence of recent client behavior in the selection process.Furthermore,we provide a theoretical analysis demonstrating a favorable regret bound for the proposed method.Extensive experiments conducted in volatile FL settings validate the effectiveness of CDE3,showing improved convergence speed and model accuracy compared with those of the baseline algorithms.These results confirm that CDE3 effectively mitigates volatility,enhancing the stability and efficiency of federated learning.展开更多
Astronomical site selection work is very hard.Unmanned technologies are important trends and solutions.We present a relatively easy method to plan a high reliability site selection which can extend the time from site ...Astronomical site selection work is very hard.Unmanned technologies are important trends and solutions.We present a relatively easy method to plan a high reliability site selection which can extend the time from site deployment to returning for maintaining by unmanned confirming the site.First,we redefine the reliability of a site selection deployment with the parameter of the trusty time,which means when we must return,and which can be relatively easy for estimating.The redefinition makes the reliability parameter as a Bayesian probability,and can be obtained by estimating besides testing,which makes the evaluation of each device's reliability much easier.Then we use block diagram tools in the Matlab Simulink software to construct structure diagram,and to link each component with relations of parallel,serial,protection,and so on.This makes the whole reliability value can be calculated at the time when we design or plan a site selection.We applied this concept and method in an actual site selection in Lenghu,Qinghai Province,China.The survey practice reveals its effectiveness and simpleness.展开更多
Selecting proper descriptors(also known feature selection,FS)is key in the process of establishing mechanical properties prediction model of hot-rolled microalloyed steels by using machine learning(ML)algorithm.FS met...Selecting proper descriptors(also known feature selection,FS)is key in the process of establishing mechanical properties prediction model of hot-rolled microalloyed steels by using machine learning(ML)algorithm.FS methods based on data-driving can reduce the redundancy of data features and improve the prediction accuracy of mechanical properties.Based on the collected data of hot-rolled microalloyed steels,the association rules are used to mine the correlation information between the data.High-quality feature subsets are selected by the proposed FS method(FS method based on genetic algorithm embedding,GAMIC).Compared with the common FS method,it is shown on dataset that GAMIC selects feature subsets more appropriately.Six different ML algorithms are trained and tested for mechanical properties prediction.The result shows that the root-mean-square error of yield strength,tensile strength and elongation based on limit gradient enhancement(XGBoost)algorithm is 21.95 MPa,20.85 MPa and 1.96%,the correlation coefficient(R^(2))is 0.969,0.968 and 0.830,and the mean absolute error is 16.84 MPa,15.83 MPa and 1.48%,respectively,showing the best prediction performance.Finally,SHapley Additive exPlanation is used to further explore the influence of feature variables on mechanical properties.GAMIC feature selection method proposed is universal,which provides a basis for the development of high-precision mechanical property prediction model.展开更多
Dynamic positioning systems(DPS)on marine vessels exhibit actuator redundancy,with more actuators than degrees of freedom.A control allocation unit is employed to address this redundancy.Practical systems often featur...Dynamic positioning systems(DPS)on marine vessels exhibit actuator redundancy,with more actuators than degrees of freedom.A control allocation unit is employed to address this redundancy.Practical systems often feature time-varying elements in the effectiveness matrix due to factors such as changing operating conditions,nonlinearity,and disturbances.Additionally,not all thrusters require engagement at each step to counteract disturbances and maintain position.Control efforts can be generated by selecting some thrusters based on their instant effectiveness,while others can remain on standby.Therefore,introducing a control allocation method that calculates the effectiveness matrix online and selects the most efficient thrusters could be effective.This paper introduces a fault-tolerant control allocation strategy for DPS with a varying effectiveness matrix.Specifically,the investigation focuses on a case study featuring eight azimuth thrusters used on a drilling rig.At each time step,the effective matrix is calculated online,followed by the selection of the four most effective thrusters based on the actuator effectiveness index,with the four serving as backups in case of a fault.The proposed strategy has been validated through simulation results,demonstrating advantages such as robustness against changes in the effectiveness matrix and reduced energy usage by the thrusters.展开更多
Employee selection is a critical process that directly impacts organizational performance,productivity,and strategic success.This comprehensive article explores the multifaceted landscape of contemporary employee sele...Employee selection is a critical process that directly impacts organizational performance,productivity,and strategic success.This comprehensive article explores the multifaceted landscape of contemporary employee selection tools,examining their theoretical foundations,practical applications,psychological underpinnings,and empirical effectiveness.By systematically analyzing various selection methodologies,this research provides insights into the complex decision-making processes that organizations employ to identify,evaluate,and recruit top talent.The article critically evaluates the reliability,validity,advantages,and limitations of each selection tool,offering a nuanced perspective on their implementation and potential organizational impact.展开更多
Advanced Persistent Threats(APTs)represent one of the most complex and dangerous categories of cyber-attacks characterised by their stealthy behaviour,long-term persistence,and ability to bypass traditional detection ...Advanced Persistent Threats(APTs)represent one of the most complex and dangerous categories of cyber-attacks characterised by their stealthy behaviour,long-term persistence,and ability to bypass traditional detection systems.The complexity of real-world network data poses significant challenges in detection.Machine learning models have shown promise in detecting APTs;however,their performance often suffers when trained on large datasets with redundant or irrelevant features.This study presents a novel,hybrid feature selection method designed to improve APT detection by reducing dimensionality while preserving the informative characteristics of the data.It combines Mutual Information(MI),Symmetric Uncertainty(SU)and Minimum Redundancy Maximum Relevance(mRMR)to enhance feature selection.MI and SU assess feature relevance,while mRMR maximises relevance and minimises redundancy,ensuring that the most impactful features are prioritised.This method addresses redundancy among selected features,improving the overall efficiency and effectiveness of the detection model.Experiments on a real-world APT datasets were conducted to evaluate the proposed method.Multiple classifiers including,Random Forest,Support Vector Machine(SVM),Gradient Boosting,and Neural Networks were used to assess classification performance.The results demonstrate that the proposed feature selection method significantly enhances detection accuracy compared to baseline models trained on the full feature set.The Random Forest algorithm achieved the highest performance,with near-perfect accuracy,precision,recall,and F1 scores(99.97%).The proposed adaptive thresholding algorithm within the selection method allows each classifier to benefit from a reduced and optimised feature space,resulting in improved training and predictive performance.This research offers a scalable and classifier-agnostic solution for dimensionality reduction in cybersecurity applications.展开更多
Data collected in fields such as cybersecurity and biomedicine often encounter high dimensionality and class imbalance.To address the problem of low classification accuracy for minority class samples arising from nume...Data collected in fields such as cybersecurity and biomedicine often encounter high dimensionality and class imbalance.To address the problem of low classification accuracy for minority class samples arising from numerous irrelevant and redundant features in high-dimensional imbalanced data,we proposed a novel feature selection method named AMF-SGSK based on adaptive multi-filter and subspace-based gaining sharing knowledge.Firstly,the balanced dataset was obtained by random under-sampling.Secondly,combining the feature importance score with the AUC score for each filter method,we proposed a concept called feature hardness to judge the importance of feature,which could adaptively select the essential features.Finally,the optimal feature subset was obtained by gaining sharing knowledge in multiple subspaces.This approach effectively achieved dimensionality reduction for high-dimensional imbalanced data.The experiment results on 30 benchmark imbalanced datasets showed that AMF-SGSK performed better than other eight commonly used algorithms including BGWO and IG-SSO in terms of F1-score,AUC,and G-mean.The mean values of F1-score,AUC,and Gmean for AMF-SGSK are 0.950,0.967,and 0.965,respectively,achieving the highest among all algorithms.And the mean value of Gmean is higher than those of IG-PSO,ReliefF-GWO,and BGOA by 3.72%,11.12%,and 20.06%,respectively.Furthermore,the selected feature ratio is below 0.01 across the selected ten datasets,further demonstrating the proposed method’s overall superiority over competing approaches.AMF-SGSK could adaptively remove irrelevant and redundant features and effectively improve the classification accuracy of high-dimensional imbalanced data,providing scientific and technological references for practical applications.展开更多
This study systematically conducted preparation optimization and performance investigations on Co-modified Ce/TiO_(2) catalysts,with a focus on examining how preparation methods and Co loading regulate the catalyst’s...This study systematically conducted preparation optimization and performance investigations on Co-modified Ce/TiO_(2) catalysts,with a focus on examining how preparation methods and Co loading regulate the catalyst’s low-temperature denitrification activity.After identifying optimal preparation parameters via condition screening,multiple characterization techniques-including BET,XRD,XPS,H_(2)-TPR and in situ DRIFTS-were employed to deeply analyze the catalyst’s physicochemical properties and reaction mechanism.Results demonstrated that compared to the impregnation and co-precipitation methods,the Ce-Co_(0.025)/TiO_(2)-SG catalyst(prepared by the sol-gel method with a Co/Ti mass ratio of 0.025)exhibited significantly superior denitrification activity:NO conversion remained stably above 95%in the 225−350℃ temperature range,and it displayed high N_(2) selectivity.Characterization analysis revealed that abundant surface oxygen vacancies,a high proportion of Ce^(3+) species,and prominent acidic sites collectively contributed to enhancing its low-temperature denitrification performance.This work provides reference value for the development of highly efficient low-temperature denitrification catalysts.展开更多
In this paper,we consider the maximal positive definite solution of the nonlinear matrix equation.By using the idea of Algorithm 2.1 in ZHANG(2013),a new inversion-free method with a stepsize parameter is proposed to ...In this paper,we consider the maximal positive definite solution of the nonlinear matrix equation.By using the idea of Algorithm 2.1 in ZHANG(2013),a new inversion-free method with a stepsize parameter is proposed to obtain the maximal positive definite solution of nonlinear matrix equation X+A^(*)X|^(-α)A=Q with the case 0<α≤1.Based on this method,a new iterative algorithm is developed,and its convergence proof is given.Finally,two numerical examples are provided to show the effectiveness of the proposed method.展开更多
The Financial Technology(FinTech)sector has witnessed rapid growth,resulting in increasingly complex and high-volume digital transactions.Although this expansion improves efficiency and accessibility,it also introduce...The Financial Technology(FinTech)sector has witnessed rapid growth,resulting in increasingly complex and high-volume digital transactions.Although this expansion improves efficiency and accessibility,it also introduces significant vulnerabilities,including fraud,money laundering,and market manipulation.Traditional anomaly detection techniques often fail to capture the relational and dynamic characteristics of financial data.Graph Neural Networks(GNNs),capable of modeling intricate interdependencies among entities,have emerged as a powerful framework for detecting subtle and sophisticated anomalies.However,the high-dimensionality and inherent noise of FinTech datasets demand robust feature selection strategies to improve model scalability,performance,and interpretability.This paper presents a comprehensive survey of GNN-based approaches for anomaly detection in FinTech,with an emphasis on the synergistic role of feature selection.We examine the theoretical foundations of GNNs,review state-of-the-art feature selection techniques,analyze their integration with GNNs,and categorize prevalent anomaly types in FinTech applications.In addition,we discuss practical implementation challenges,highlight representative case studies,and propose future research directions to advance the field of graph-based anomaly detection in financial systems.展开更多
This study presents an effective hybrid simulation approach for simulating broadband ground motion in complex near-fault locations.The approach utilizes a deterministic approach based on the spectral element method(SE...This study presents an effective hybrid simulation approach for simulating broadband ground motion in complex near-fault locations.The approach utilizes a deterministic approach based on the spectral element method(SEM),which is used to simulate low-frequency ground motion(f<1 Hz)by incorporating an innovative efficient discontinuous Galerkin(DG)method for grid division to accurately model basin sedimentary layers at reduced costs.It also introduces a comprehensive hybrid source model for high-frequency random scattering and a nonlinear analysis module for basin sedimentary layers.Deterministic outcomes are combined with modified three-dimensional stochastic finite fault method(3D-EXSIM)simulations of high-frequency ground motion(f>1 Hz).A fourth-order Butterworth filter with zero phase shift is employed for time-domain filtering of low-and high-frequency time series at a crossover frequency of 1 Hz,merging the low and high-frequency ground motions into a broadband time series.Taking an Ms 6.8 Luding earthquake,as an example,this hybrid method was used for a rapid and efficient simulation analysis of broadband ground motion in the region.The accuracy and efficiency of this hybrid method were verified through comparisons with actually observed station data and empirical attenuation curves.Deterministic method simulation results revealed the effects of mountainous topography,basin effects,nonlinear effects within the basin’s sedimentary layers,and a coupling interaction between the basin and the mountains.The findings are consistent with similar studies,showing that near-fault sedimentary basins significantly focus and amplify strong ground motion,and the soil’s nonlinear behavior in the basin influences ground motion to varying extents at different distances from the fault.The mountainous topography impacts the basin’s response to ground motion,leading to barrier effects.This research provides a scientific foundation for seismic zoning,urban planning,and seismic design in nearfault mountain basin regions.展开更多
With the increasing complexity of vehicular networks and the proliferation of connected vehicles,Federated Learning(FL)has emerged as a critical framework for decentralized model training while preserving data privacy...With the increasing complexity of vehicular networks and the proliferation of connected vehicles,Federated Learning(FL)has emerged as a critical framework for decentralized model training while preserving data privacy.However,efficient client selection and adaptive weight allocation in heterogeneous and non-IID environments remain challenging.To address these issues,we propose Federated Learning with Client Selection and Adaptive Weighting(FedCW),a novel algorithm that leverages adaptive client selection and dynamic weight allocation for optimizing model convergence in real-time vehicular networks.FedCW selects clients based on their Euclidean distance from the global model and dynamically adjusts aggregation weights to optimize both data diversity and model convergence.Experimental results show that FedCW significantly outperforms existing FL algorithms such as FedAvg,FedProx,and SCAFFOLD,particularly in non-IID settings,achieving faster convergence,higher accuracy,and reduced communication overhead.These findings demonstrate that FedCW provides an effective solution for enhancing the performance of FL in heterogeneous,edge-based computing environments.展开更多
High-dimensional data causes difficulties in machine learning due to high time consumption and large memory requirements.In particular,in amulti-label environment,higher complexity is required asmuch as the number of ...High-dimensional data causes difficulties in machine learning due to high time consumption and large memory requirements.In particular,in amulti-label environment,higher complexity is required asmuch as the number of labels.Moreover,an optimization problem that fully considers all dependencies between features and labels is difficult to solve.In this study,we propose a novel regression-basedmulti-label feature selectionmethod that integrates mutual information to better exploit the underlying data structure.By incorporating mutual information into the regression formulation,the model captures not only linear relationships but also complex non-linear dependencies.The proposed objective function simultaneously considers three types of relationships:(1)feature redundancy,(2)featurelabel relevance,and(3)inter-label dependency.These three quantities are computed usingmutual information,allowing the proposed formulation to capture nonlinear dependencies among variables.These three types of relationships are key factors in multi-label feature selection,and our method expresses them within a unified formulation,enabling efficient optimization while simultaneously accounting for all of them.To efficiently solve the proposed optimization problem under non-negativity constraints,we develop a gradient-based optimization algorithm with fast convergence.Theexperimental results on sevenmulti-label datasets show that the proposed method outperforms existingmulti-label feature selection techniques.展开更多
Populus species,important economic species combining rapid growth with broad ecological adaptability,play a critical role in sustainable forestry and bioenergy production.In this study,we performed whole-genome resequ...Populus species,important economic species combining rapid growth with broad ecological adaptability,play a critical role in sustainable forestry and bioenergy production.In this study,we performed whole-genome resequencing of 707 individuals from a full-sib family to develop comprehensive single nucleotide polymorphism(SNP)markers and constructed a high-density genetic linkage map of 19 linkage groups.The total genetic length of the map reached 3623.65 cM with an average marker interval of 0.34 cM.By integrating multidimensional phenotypic data,89 quantitative trait loci(QTL)associated with growth,wood physical and chemical properties,disease resistance,and leaf morphology traits were identified,with logarithm of odds(LOD)scores ranging from 3.13 to 21.72 Notably,pleiotropic analysis revealed significant colocaliza and phenotypic variance explained between 1.7% and 11.6%.-tion hotspots on chromosomes LG1,LG5,LG6,LG8,and LG14,with epistatic interaction network analysis confirming genetic basis of coordinated regulation across multiple traits.Functional annotation of 207 candidate genes showed that R2R3-MYB and bHLH transcription factors and pyruvate kinase-encoding genes were significantly enriched,suggesting crucial roles in lignin biosynthesis and carbon metabolic pathways.Allelic effect analysis indicated that the frequency of favorable alleles associated with target traits ranged from 0.20 to 0.55.Incorporation of QTL-derived favorable alleles as random effects into Bayesian-based genomic selection models led to an increase in prediction accuracy ranging from 1% to 21%,with Bayesian ridge regression as the best predictive model.This study provides valuable genomic resources and genetic insights for deciphering complex trait architecture and advancing molecular breeding in poplar.展开更多
文摘Advanced intensity measures(IMs)based on an inelastic deformation spectrum improved the evaluation of the median engineering demand parameters(EDPs)and reduced dispersion.In this regard,an optimized two-degreefreedom(2DOF)modal pushover-based scaling procedure(2DMPS)has been developed for a nonlinear dynamic analysis of asymmetric in-plan buildings.The 2DMPS procedure scales ground motions to approach close enough to a target value of the inelastic displacement of the first-mode inelastic 2DOF modal stick,extended for structures with significant contributions of higher modes.Further,4-,6-and 13-story RC SMRF buildings were selected for analyses using ground motion records scaled by the 2DMPS procedure,the modal pushover-based scaling method(MPS),and ASCE/SEI 7-16 scaling procedures.The median values of EDPs on scaled records closely matched the benchmark results.The bias in the EDP values due to the scaled records in every group regarding their median value was lower than the dispersion of the 21 unscaled records.These results generally demonstrate the accuracy and efficiency of the 2DMPS method.Additionally,the 2DOF modal stick’s inelastic response spectra are better suited for calculating seismic demands for one-way asymmetric-plan structures than the SDOF inelastic response spectra.
基金supported in part by the National Natural Science Foundation of China(No.42271446)in part by the Tianjin Key Laboratory of Rail Transit Navigation Positioning and Spatio-Temporary Big Data Technology,China(No.TKL2024B13)in part by the Science and Technology Program of Tianjin,China(No.24YFYSHZ00080)。
文摘The selection of a suitable navigation area is pivotal in aircraft scene matching guidance technology.This study addresses the challenge of identifying suitable reference image ranges for precise scene matching,which is crucial for enhancing aircraft positioning accuracy.Traditional methods for image matchability analysis are often limited by their reliance on manual feature parameter design and threshold-based filtering,resulting in suboptimal accuracy and efficiency.This paper proposes a novel network architecture for selecting suitable navigation areas using image Matching Level Segmentation(MLSNet).The approach involves two key innovations:a method for generating segmentation labels that quantify matchability levels and an end-to-end network architecture for rapid and precise prediction of reference image matchability segmentation maps.The network includes two core modules:the saliency analysis module uses multi-layer convolutional networks to accurately detect image saliency features across various levels and scales;the multidimensional attention module utilizes attention mechanisms to focus on feature channels and spatial neighborhood scenes to assess the image’s matchability.Our method was rigorously tested on an extensive collection of remote sensing images,where it was benchmarked against a range of both traditional and cutting-edge deep learning methods.The findings indicate that MLSNet is significantly superior to traditional methods in accuracy and efficiency of matchability analysis,and is also relatively ahead of state-of-the-art deep learning models.
基金funded by the Natural Science Foundation of China(Grant Nos.42377164 and 41972280)the Badong National Observation and Research Station of Geohazards(Grant No.BNORSG-202305).
文摘Landslide susceptibility prediction(LSP)is significantly affected by the uncertainty issue of landslide related conditioning factor selection.However,most of literature only performs comparative studies on a certain conditioning factor selection method rather than systematically study this uncertainty issue.Targeted,this study aims to systematically explore the influence rules of various commonly used conditioning factor selection methods on LSP,and on this basis to innovatively propose a principle with universal application for optimal selection of conditioning factors.An'yuan County in southern China is taken as example considering 431 landslides and 29 types of conditioning factors.Five commonly used factor selection methods,namely,the correlation analysis(CA),linear regression(LR),principal component analysis(PCA),rough set(RS)and artificial neural network(ANN),are applied to select the optimal factor combinations from the original 29 conditioning factors.The factor selection results are then used as inputs of four types of common machine learning models to construct 20 types of combined models,such as CA-multilayer perceptron,CA-random forest.Additionally,multifactor-based multilayer perceptron random forest models that selecting conditioning factors based on the proposed principle of“accurate data,rich types,clear significance,feasible operation and avoiding duplication”are constructed for comparisons.Finally,the LSP uncertainties are evaluated by the accuracy,susceptibility index distribution,etc.Results show that:(1)multifactor-based models have generally higher LSP performance and lower uncertainties than those of factors selection-based models;(2)Influence degree of different machine learning on LSP accuracy is greater than that of different factor selection methods.Conclusively,the above commonly used conditioning factor selection methods are not ideal for improving LSP performance and may complicate the LSP processes.In contrast,a satisfied combination of conditioning factors can be constructed according to the proposed principle.
文摘Software defect prediction(SDP)aims to find a reliable method to predict defects in specific software projects and help software engineers allocate limited resources to release high-quality software products.Software defect prediction can be effectively performed using traditional features,but there are some redundant or irrelevant features in them(the presence or absence of this feature has little effect on the prediction results).These problems can be solved using feature selection.However,existing feature selection methods have shortcomings such as insignificant dimensionality reduction effect and low classification accuracy of the selected optimal feature subset.In order to reduce the impact of these shortcomings,this paper proposes a new feature selection method Cubic TraverseMa Beluga whale optimization algorithm(CTMBWO)based on the improved Beluga whale optimization algorithm(BWO).The goal of this study is to determine how well the CTMBWO can extract the features that are most important for correctly predicting software defects,improve the accuracy of fault prediction,reduce the number of the selected feature and mitigate the risk of overfitting,thereby achieving more efficient resource utilization and better distribution of test workload.The CTMBWO comprises three main stages:preprocessing the dataset,selecting relevant features,and evaluating the classification performance of the model.The novel feature selection method can effectively improve the performance of SDP.This study performs experiments on two software defect datasets(PROMISE,NASA)and shows the method’s classification performance using four detailed evaluation metrics,Accuracy,F1-score,MCC,AUC and Recall.The results indicate that the approach presented in this paper achieves outstanding classification performance on both datasets and has significant improvement over the baseline models.
基金the Deanship of Scientifc Research at King Khalid University for funding this work through large group Research Project under grant number RGP2/421/45supported via funding from Prince Sattam bin Abdulaziz University project number(PSAU/2024/R/1446)+1 种基金supported by theResearchers Supporting Project Number(UM-DSR-IG-2023-07)Almaarefa University,Riyadh,Saudi Arabia.supported by the Basic Science Research Program through the National Research Foundation of Korea(NRF)funded by the Ministry of Education(No.2021R1F1A1055408).
文摘Machine learning(ML)is increasingly applied for medical image processing with appropriate learning paradigms.These applications include analyzing images of various organs,such as the brain,lung,eye,etc.,to identify specific flaws/diseases for diagnosis.The primary concern of ML applications is the precise selection of flexible image features for pattern detection and region classification.Most of the extracted image features are irrelevant and lead to an increase in computation time.Therefore,this article uses an analytical learning paradigm to design a Congruent Feature Selection Method to select the most relevant image features.This process trains the learning paradigm using similarity and correlation-based features over different textural intensities and pixel distributions.The similarity between the pixels over the various distribution patterns with high indexes is recommended for disease diagnosis.Later,the correlation based on intensity and distribution is analyzed to improve the feature selection congruency.Therefore,the more congruent pixels are sorted in the descending order of the selection,which identifies better regions than the distribution.Now,the learning paradigm is trained using intensity and region-based similarity to maximize the chances of selection.Therefore,the probability of feature selection,regardless of the textures and medical image patterns,is improved.This process enhances the performance of ML applications for different medical image processing.The proposed method improves the accuracy,precision,and training rate by 13.19%,10.69%,and 11.06%,respectively,compared to other models for the selected dataset.The mean error and selection time is also reduced by 12.56%and 13.56%,respectively,compared to the same models and dataset.
基金supported by the Anhui Provincial Department of Education University Research Project(2024AH051375)Research Project of Chizhou University(CZ2022ZRZ06)+1 种基金Anhui Province Natural Science Research Project of Colleges and Universities(2024AH051368)Excellent Scientific Research and Innovation Team of Anhui Colleges(2022AH010098).
文摘Feature selection methods rooted in rough sets confront two notable limitations:their high computa-tional complexity and sensitivity to noise,rendering them impractical for managing large-scale and noisy datasets.The primary issue stems from these methods’undue reliance on all samples.To overcome these challenges,we introduce the concept of cross-similarity grounded in a robust fuzzy relation and design a rapid and robust feature selection algorithm.Firstly,we construct a robust fuzzy relation by introducing a truncation parameter.Then,based on this fuzzy relation,we propose the concept of cross-similarity,which emphasizes the sample-to-sample similarity relations that uniquely determine feature importance,rather than considering all such relations equally.After studying the manifestations and properties of cross-similarity across different fuzzy granularities,we propose a forward greedy feature selection algorithm that leverages cross-similarity as the foundation for information measurement.This algorithm significantly reduces the time complexity from O(m2n2)to O(mn2).Experimental findings reveal that the average runtime of five state-of-the-art comparison algorithms is roughly 3.7 times longer than our algorithm,while our algorithm achieves an average accuracy that surpasses those of the five comparison algorithms by approximately 3.52%.This underscores the effectiveness of our approach.This paper paves the way for applying feature selection algorithms grounded in fuzzy rough sets to large-scale gene datasets.
基金funded by the Central University of Finance and Economics,Greater Bay Area Research Institute Project(No.YJY202303)the National Natural Science Foundation of China(No.61906220)the Ministry of Education of Humanities and Social Science Project(No.19YJCZH178).
文摘The growing demand for privacy-preserving machine learning has positioned federated learning as a promising research paradigm,enabling the training of high-performance models across distributed data sources without compromising user privacy.However,despite its advantages,federated learning faces critical challenges arising from the heterogeneity and volatility of participating clients.In real-world scenarios,variations in client participation,data volume,computational capability,and communication reliability contribute to a highly dynamic training environment,which negatively impacts efficiency and convergence of the model.To address these challenges,this paper proposes a novel client selection method named CDE3.First,CDE3 employs a multidimensional model to comprehensively evaluate clients’contributions.Second,we enhance the classical Exp3 algorithm by incorporating a discount factor that exponentially decays historical contributions,thereby increasing the influence of recent client behavior in the selection process.Furthermore,we provide a theoretical analysis demonstrating a favorable regret bound for the proposed method.Extensive experiments conducted in volatile FL settings validate the effectiveness of CDE3,showing improved convergence speed and model accuracy compared with those of the baseline algorithms.These results confirm that CDE3 effectively mitigates volatility,enhancing the stability and efficiency of federated learning.
基金supported by the Investigation of Technological Infrastructure Resources(No.2023FY101101)the National Natural Science Foundation of China(NSFC)(No.11073027 and No.12373104)。
文摘Astronomical site selection work is very hard.Unmanned technologies are important trends and solutions.We present a relatively easy method to plan a high reliability site selection which can extend the time from site deployment to returning for maintaining by unmanned confirming the site.First,we redefine the reliability of a site selection deployment with the parameter of the trusty time,which means when we must return,and which can be relatively easy for estimating.The redefinition makes the reliability parameter as a Bayesian probability,and can be obtained by estimating besides testing,which makes the evaluation of each device's reliability much easier.Then we use block diagram tools in the Matlab Simulink software to construct structure diagram,and to link each component with relations of parallel,serial,protection,and so on.This makes the whole reliability value can be calculated at the time when we design or plan a site selection.We applied this concept and method in an actual site selection in Lenghu,Qinghai Province,China.The survey practice reveals its effectiveness and simpleness.
基金supported by the National Key Research and Development Program of China(Grant No.2021YFB3702404)the National Natural Science Foundation of China(Grant No.52104370)+4 种基金the Reviving-Liaoning Excellence Plan(XLYC2203186)Science and Technology Special Projects of Liaoning Province(Grant No.2022JH25/10200001)the Postdoctoral Research Fund for Northeastern(Grant No.20210203)Independent Projects of Basic Scientific Research(ZZ2021005)CITIC Niobium Steel Development Award Fund(2022-M1824).
文摘Selecting proper descriptors(also known feature selection,FS)is key in the process of establishing mechanical properties prediction model of hot-rolled microalloyed steels by using machine learning(ML)algorithm.FS methods based on data-driving can reduce the redundancy of data features and improve the prediction accuracy of mechanical properties.Based on the collected data of hot-rolled microalloyed steels,the association rules are used to mine the correlation information between the data.High-quality feature subsets are selected by the proposed FS method(FS method based on genetic algorithm embedding,GAMIC).Compared with the common FS method,it is shown on dataset that GAMIC selects feature subsets more appropriately.Six different ML algorithms are trained and tested for mechanical properties prediction.The result shows that the root-mean-square error of yield strength,tensile strength and elongation based on limit gradient enhancement(XGBoost)algorithm is 21.95 MPa,20.85 MPa and 1.96%,the correlation coefficient(R^(2))is 0.969,0.968 and 0.830,and the mean absolute error is 16.84 MPa,15.83 MPa and 1.48%,respectively,showing the best prediction performance.Finally,SHapley Additive exPlanation is used to further explore the influence of feature variables on mechanical properties.GAMIC feature selection method proposed is universal,which provides a basis for the development of high-precision mechanical property prediction model.
文摘Dynamic positioning systems(DPS)on marine vessels exhibit actuator redundancy,with more actuators than degrees of freedom.A control allocation unit is employed to address this redundancy.Practical systems often feature time-varying elements in the effectiveness matrix due to factors such as changing operating conditions,nonlinearity,and disturbances.Additionally,not all thrusters require engagement at each step to counteract disturbances and maintain position.Control efforts can be generated by selecting some thrusters based on their instant effectiveness,while others can remain on standby.Therefore,introducing a control allocation method that calculates the effectiveness matrix online and selects the most efficient thrusters could be effective.This paper introduces a fault-tolerant control allocation strategy for DPS with a varying effectiveness matrix.Specifically,the investigation focuses on a case study featuring eight azimuth thrusters used on a drilling rig.At each time step,the effective matrix is calculated online,followed by the selection of the four most effective thrusters based on the actuator effectiveness index,with the four serving as backups in case of a fault.The proposed strategy has been validated through simulation results,demonstrating advantages such as robustness against changes in the effectiveness matrix and reduced energy usage by the thrusters.
文摘Employee selection is a critical process that directly impacts organizational performance,productivity,and strategic success.This comprehensive article explores the multifaceted landscape of contemporary employee selection tools,examining their theoretical foundations,practical applications,psychological underpinnings,and empirical effectiveness.By systematically analyzing various selection methodologies,this research provides insights into the complex decision-making processes that organizations employ to identify,evaluate,and recruit top talent.The article critically evaluates the reliability,validity,advantages,and limitations of each selection tool,offering a nuanced perspective on their implementation and potential organizational impact.
基金funded by Universiti Teknologi Malaysia under the UTM RA ICONIC Grant(Q.J130000.4351.09G61).
文摘Advanced Persistent Threats(APTs)represent one of the most complex and dangerous categories of cyber-attacks characterised by their stealthy behaviour,long-term persistence,and ability to bypass traditional detection systems.The complexity of real-world network data poses significant challenges in detection.Machine learning models have shown promise in detecting APTs;however,their performance often suffers when trained on large datasets with redundant or irrelevant features.This study presents a novel,hybrid feature selection method designed to improve APT detection by reducing dimensionality while preserving the informative characteristics of the data.It combines Mutual Information(MI),Symmetric Uncertainty(SU)and Minimum Redundancy Maximum Relevance(mRMR)to enhance feature selection.MI and SU assess feature relevance,while mRMR maximises relevance and minimises redundancy,ensuring that the most impactful features are prioritised.This method addresses redundancy among selected features,improving the overall efficiency and effectiveness of the detection model.Experiments on a real-world APT datasets were conducted to evaluate the proposed method.Multiple classifiers including,Random Forest,Support Vector Machine(SVM),Gradient Boosting,and Neural Networks were used to assess classification performance.The results demonstrate that the proposed feature selection method significantly enhances detection accuracy compared to baseline models trained on the full feature set.The Random Forest algorithm achieved the highest performance,with near-perfect accuracy,precision,recall,and F1 scores(99.97%).The proposed adaptive thresholding algorithm within the selection method allows each classifier to benefit from a reduced and optimised feature space,resulting in improved training and predictive performance.This research offers a scalable and classifier-agnostic solution for dimensionality reduction in cybersecurity applications.
基金supported by Fundamental Research Program of Shanxi Province(Nos.202203021211088,202403021212254,202403021221109)Graduate Research Innovation Project in Shanxi Province(No.2024KY616).
文摘Data collected in fields such as cybersecurity and biomedicine often encounter high dimensionality and class imbalance.To address the problem of low classification accuracy for minority class samples arising from numerous irrelevant and redundant features in high-dimensional imbalanced data,we proposed a novel feature selection method named AMF-SGSK based on adaptive multi-filter and subspace-based gaining sharing knowledge.Firstly,the balanced dataset was obtained by random under-sampling.Secondly,combining the feature importance score with the AUC score for each filter method,we proposed a concept called feature hardness to judge the importance of feature,which could adaptively select the essential features.Finally,the optimal feature subset was obtained by gaining sharing knowledge in multiple subspaces.This approach effectively achieved dimensionality reduction for high-dimensional imbalanced data.The experiment results on 30 benchmark imbalanced datasets showed that AMF-SGSK performed better than other eight commonly used algorithms including BGWO and IG-SSO in terms of F1-score,AUC,and G-mean.The mean values of F1-score,AUC,and Gmean for AMF-SGSK are 0.950,0.967,and 0.965,respectively,achieving the highest among all algorithms.And the mean value of Gmean is higher than those of IG-PSO,ReliefF-GWO,and BGOA by 3.72%,11.12%,and 20.06%,respectively.Furthermore,the selected feature ratio is below 0.01 across the selected ten datasets,further demonstrating the proposed method’s overall superiority over competing approaches.AMF-SGSK could adaptively remove irrelevant and redundant features and effectively improve the classification accuracy of high-dimensional imbalanced data,providing scientific and technological references for practical applications.
基金Supported by the National Key Research and Development Program of China (2023YFB4102903)。
文摘This study systematically conducted preparation optimization and performance investigations on Co-modified Ce/TiO_(2) catalysts,with a focus on examining how preparation methods and Co loading regulate the catalyst’s low-temperature denitrification activity.After identifying optimal preparation parameters via condition screening,multiple characterization techniques-including BET,XRD,XPS,H_(2)-TPR and in situ DRIFTS-were employed to deeply analyze the catalyst’s physicochemical properties and reaction mechanism.Results demonstrated that compared to the impregnation and co-precipitation methods,the Ce-Co_(0.025)/TiO_(2)-SG catalyst(prepared by the sol-gel method with a Co/Ti mass ratio of 0.025)exhibited significantly superior denitrification activity:NO conversion remained stably above 95%in the 225−350℃ temperature range,and it displayed high N_(2) selectivity.Characterization analysis revealed that abundant surface oxygen vacancies,a high proportion of Ce^(3+) species,and prominent acidic sites collectively contributed to enhancing its low-temperature denitrification performance.This work provides reference value for the development of highly efficient low-temperature denitrification catalysts.
基金Supported in part by Natural Science Foundation of Guangxi(2023GXNSFAA026246)in part by the Central Government's Guide to Local Science and Technology Development Fund(GuikeZY23055044)in part by the National Natural Science Foundation of China(62363003)。
文摘In this paper,we consider the maximal positive definite solution of the nonlinear matrix equation.By using the idea of Algorithm 2.1 in ZHANG(2013),a new inversion-free method with a stepsize parameter is proposed to obtain the maximal positive definite solution of nonlinear matrix equation X+A^(*)X|^(-α)A=Q with the case 0<α≤1.Based on this method,a new iterative algorithm is developed,and its convergence proof is given.Finally,two numerical examples are provided to show the effectiveness of the proposed method.
基金supported by Ho Chi Minh City Open University,Vietnam under grant number E2024.02.1CD and Suan Sunandha Rajabhat University,Thailand.
文摘The Financial Technology(FinTech)sector has witnessed rapid growth,resulting in increasingly complex and high-volume digital transactions.Although this expansion improves efficiency and accessibility,it also introduces significant vulnerabilities,including fraud,money laundering,and market manipulation.Traditional anomaly detection techniques often fail to capture the relational and dynamic characteristics of financial data.Graph Neural Networks(GNNs),capable of modeling intricate interdependencies among entities,have emerged as a powerful framework for detecting subtle and sophisticated anomalies.However,the high-dimensionality and inherent noise of FinTech datasets demand robust feature selection strategies to improve model scalability,performance,and interpretability.This paper presents a comprehensive survey of GNN-based approaches for anomaly detection in FinTech,with an emphasis on the synergistic role of feature selection.We examine the theoretical foundations of GNNs,review state-of-the-art feature selection techniques,analyze their integration with GNNs,and categorize prevalent anomaly types in FinTech applications.In addition,we discuss practical implementation challenges,highlight representative case studies,and propose future research directions to advance the field of graph-based anomaly detection in financial systems.
基金National Natural Science Foundation of China under Grant Nos.U2139208 and 52278516Key Laboratory of Earthquake Engineering and Engineering Vibration,China Earthquake Administration under Grant No.2024D15Key Laboratory of Soft Soil Characteristic and Engineering Environment,Tianjin Chengjian University under Grant No.2022SCEEKL003。
文摘This study presents an effective hybrid simulation approach for simulating broadband ground motion in complex near-fault locations.The approach utilizes a deterministic approach based on the spectral element method(SEM),which is used to simulate low-frequency ground motion(f<1 Hz)by incorporating an innovative efficient discontinuous Galerkin(DG)method for grid division to accurately model basin sedimentary layers at reduced costs.It also introduces a comprehensive hybrid source model for high-frequency random scattering and a nonlinear analysis module for basin sedimentary layers.Deterministic outcomes are combined with modified three-dimensional stochastic finite fault method(3D-EXSIM)simulations of high-frequency ground motion(f>1 Hz).A fourth-order Butterworth filter with zero phase shift is employed for time-domain filtering of low-and high-frequency time series at a crossover frequency of 1 Hz,merging the low and high-frequency ground motions into a broadband time series.Taking an Ms 6.8 Luding earthquake,as an example,this hybrid method was used for a rapid and efficient simulation analysis of broadband ground motion in the region.The accuracy and efficiency of this hybrid method were verified through comparisons with actually observed station data and empirical attenuation curves.Deterministic method simulation results revealed the effects of mountainous topography,basin effects,nonlinear effects within the basin’s sedimentary layers,and a coupling interaction between the basin and the mountains.The findings are consistent with similar studies,showing that near-fault sedimentary basins significantly focus and amplify strong ground motion,and the soil’s nonlinear behavior in the basin influences ground motion to varying extents at different distances from the fault.The mountainous topography impacts the basin’s response to ground motion,leading to barrier effects.This research provides a scientific foundation for seismic zoning,urban planning,and seismic design in nearfault mountain basin regions.
文摘With the increasing complexity of vehicular networks and the proliferation of connected vehicles,Federated Learning(FL)has emerged as a critical framework for decentralized model training while preserving data privacy.However,efficient client selection and adaptive weight allocation in heterogeneous and non-IID environments remain challenging.To address these issues,we propose Federated Learning with Client Selection and Adaptive Weighting(FedCW),a novel algorithm that leverages adaptive client selection and dynamic weight allocation for optimizing model convergence in real-time vehicular networks.FedCW selects clients based on their Euclidean distance from the global model and dynamically adjusts aggregation weights to optimize both data diversity and model convergence.Experimental results show that FedCW significantly outperforms existing FL algorithms such as FedAvg,FedProx,and SCAFFOLD,particularly in non-IID settings,achieving faster convergence,higher accuracy,and reduced communication overhead.These findings demonstrate that FedCW provides an effective solution for enhancing the performance of FL in heterogeneous,edge-based computing environments.
基金supported by Basic Science Research Program through the National Research Foundation of Korea(NRF)funded by the Ministry of Education(RS-2020-NR049579).
文摘High-dimensional data causes difficulties in machine learning due to high time consumption and large memory requirements.In particular,in amulti-label environment,higher complexity is required asmuch as the number of labels.Moreover,an optimization problem that fully considers all dependencies between features and labels is difficult to solve.In this study,we propose a novel regression-basedmulti-label feature selectionmethod that integrates mutual information to better exploit the underlying data structure.By incorporating mutual information into the regression formulation,the model captures not only linear relationships but also complex non-linear dependencies.The proposed objective function simultaneously considers three types of relationships:(1)feature redundancy,(2)featurelabel relevance,and(3)inter-label dependency.These three quantities are computed usingmutual information,allowing the proposed formulation to capture nonlinear dependencies among variables.These three types of relationships are key factors in multi-label feature selection,and our method expresses them within a unified formulation,enabling efficient optimization while simultaneously accounting for all of them.To efficiently solve the proposed optimization problem under non-negativity constraints,we develop a gradient-based optimization algorithm with fast convergence.Theexperimental results on sevenmulti-label datasets show that the proposed method outperforms existingmulti-label feature selection techniques.
基金supported by the National Key Research and Development Plan of China(2021YFD2200202)the Key Research and Development Project of Jiangsu Province,China(BE2021366).
文摘Populus species,important economic species combining rapid growth with broad ecological adaptability,play a critical role in sustainable forestry and bioenergy production.In this study,we performed whole-genome resequencing of 707 individuals from a full-sib family to develop comprehensive single nucleotide polymorphism(SNP)markers and constructed a high-density genetic linkage map of 19 linkage groups.The total genetic length of the map reached 3623.65 cM with an average marker interval of 0.34 cM.By integrating multidimensional phenotypic data,89 quantitative trait loci(QTL)associated with growth,wood physical and chemical properties,disease resistance,and leaf morphology traits were identified,with logarithm of odds(LOD)scores ranging from 3.13 to 21.72 Notably,pleiotropic analysis revealed significant colocaliza and phenotypic variance explained between 1.7% and 11.6%.-tion hotspots on chromosomes LG1,LG5,LG6,LG8,and LG14,with epistatic interaction network analysis confirming genetic basis of coordinated regulation across multiple traits.Functional annotation of 207 candidate genes showed that R2R3-MYB and bHLH transcription factors and pyruvate kinase-encoding genes were significantly enriched,suggesting crucial roles in lignin biosynthesis and carbon metabolic pathways.Allelic effect analysis indicated that the frequency of favorable alleles associated with target traits ranged from 0.20 to 0.55.Incorporation of QTL-derived favorable alleles as random effects into Bayesian-based genomic selection models led to an increase in prediction accuracy ranging from 1% to 21%,with Bayesian ridge regression as the best predictive model.This study provides valuable genomic resources and genetic insights for deciphering complex trait architecture and advancing molecular breeding in poplar.