Developing innovative capabilities in university students is essential for individual career success and broader societal advancement.This study introduces a predictive Feature Selection(FS)model named bWRBA-SVM-FS,wh...Developing innovative capabilities in university students is essential for individual career success and broader societal advancement.This study introduces a predictive Feature Selection(FS)model named bWRBA-SVM-FS,which combines an enhanced Bat Algorithm(BA)and Support Vector Machine(SVM).To enhance the optimization capability of BA,water follow search and random follow search are introduced to optimize the efficiency and accuracy of the feature subset search.Experimental validation conducted on the IEEE CEC 2017 benchmark functions and the talented innovative capacity dataset demonstrates the efficacy of the proposed method relative to peer and prominent machine learning models.The experimental results reveal that the predictive accuracy of the bWRBA-SVM-FS model is 97.503%,with a sensitivity of 98.391%.Our findings indicate significant predictors of innovation capacity,including project application goals,educational background,and interdisciplinary thinking abilities.The bWRBA-SVM-FS model offers effective strategies for talent selection in higher education,fostering the development of future research leaders.展开更多
Background:Despite the promise shown by large language models(LLMs)for standardized tasks,their multidimensional performance in real-world oncology decision-making remains unevaluated.This study aims to introduce a fr...Background:Despite the promise shown by large language models(LLMs)for standardized tasks,their multidimensional performance in real-world oncology decision-making remains unevaluated.This study aims to introduce a framework for evaluating LLMs and physician decisions in challenging lung cancer cases.Methods:We curated 50 challenging lung cancer cases(25 local and 25 published)classified as complex,rare,or refractory.Blinded three-dimensional,five-point Likert evaluations(1–5 for comprehensiveness,specificity,and readability)compared standalone LLMs(DeepSeek R1,Claude 3.5,Gemini 1.5,and GPT-4o),physicians by experience level(junior,intermediate,and senior),and AI-assisted juniors;intergroup differences and augmentation effects were analyzed statistically.Results:Of 50 challenging cases(18 complex,17 rare,and 15 refractory)rated by three experts,DeepSeek R1 achieved scores of 3.95±0.33,3.71±0.53,and 4.26±0.18 for comprehensiveness,specificity,and readability,respectively,positioning it between intermediate(3.68,3.68,3.75)and senior(4.50,4.64,4.53)physicians.GPT-4o and Claude 3.5 reached intermediate physician–level comprehensiveness(3.76±0.39,3.60±0.39)but junior-to-intermediate physician–level specificity(3.39±0.39,3.39±0.49).All LLMs scored higher on rare cases than intermediate physicians but fell below junior physicians in refractory-case specificity.AIassisted junior physicians showed marked gains in rare cases,with comprehensiveness rising from 2.32 to 4.29(84.8%),specificity from 2.24 to 4.26(90.8%),and readability from 2.76 to 4.59(66.0%),while specificity declined by 3.2%(3.17 to 3.07)in refractory cases.Error analysis showed complementary strengths,with physicians demonstrating reasoning stability and LLMs excelling in knowledge updating and risk management.Conclusions:LLMs performed variably in clinical decision-making tasks depending on case type,performing better in rare cases and worse in refractory cases requiring longitudinal reasoning.Complementary strengths between LLMs and physicians support case-and task-tailored human–AI collaboration.展开更多
Optimization algorithms are crucial for solving NP-hard problems in engineering and computational sciences.Metaheuristic algorithms,in particular,have proven highly effective in complex optimization scenarios characte...Optimization algorithms are crucial for solving NP-hard problems in engineering and computational sciences.Metaheuristic algorithms,in particular,have proven highly effective in complex optimization scenarios characterized by high dimensionality and intricate variable relationships.The Mountain Gazelle Optimizer(MGO)is notably effective but struggles to balance local search refinement and global space exploration,often leading to premature convergence and entrapment in local optima.This paper presents the Improved MGO(IMGO),which integrates three synergistic enhancements:dynamic chaos mapping using piecewise chaotic sequences to boost explo-ration diversity;Opposition-Based Learning(OBL)with adaptive,diversity-driven activation to speed up convergence;and structural refinements to the position update mechanisms to enhance exploitation.The IMGO underwent a comprehensive evaluation using 52 standardised benchmark functions and seven engineering optimization problems.Benchmark evaluations showed that IMGO achieved the highest rank in best solution quality for 31 functions,the highest rank in mean performance for 18 functions,and the highest rank in worst-case performance for 14 functions among 11 competing algorithms.Statistical validation using Wilcoxon signed-rank tests confirmed that IMGO outperformed individual competitors across 16 to 50 functions,depending on the algorithm.At the same time,Friedman ranking analysis placed IMGO with an average rank of 4.15,compared to the baseline MGO’s 4.38,establishing the best overall performance.The evaluation of engineering problems revealed consistent improvements,including an optimal cost of 1.6896 for the welded beam design vs.MGO’s 1.7249,a minimum cost of 5885.33 for the pressure vessel design vs.MGO’s 6300,and a minimum weight of 2964.52 kg for the speed reducer design vs.MGO’s 2990.00 kg.Ablation studies identified OBL as the strongest individual contributor,whereas complete integration achieved superior performance through synergistic interactions among components.Computational complexity analysis established an O(T×N×5×f(P))time complexity,representing a 1.25×increase in fitness evaluation relative to the baseline MGO,validating the favorable accuracy-efficiency trade-offs for practical optimization applications.展开更多
Environmental problems are intensifying due to the rapid growth of the population,industry,and urban infrastructure.This expansion has resulted in increased air and water pollution,intensified urban heat island effect...Environmental problems are intensifying due to the rapid growth of the population,industry,and urban infrastructure.This expansion has resulted in increased air and water pollution,intensified urban heat island effects,and greater runoff from parks and other green spaces.Addressing these challenges requires prioritizing green infrastructure and other sustainable urban development strategies.This study introduces a novel Integrated Decision Support System that combines Pythagorean Fuzzy Sets with the Advanced Alternative Ranking Order Method allowing for Two-Step Normalization(AAROM-TN),enhanced by a dual weighting strategy.The weighting approach integrates the Criteria Importance Through Intercriteria Correlation(CRITIC)method with the Criteria Importance through Means and Standard Deviation(CIMAS)technique.The originality of the proposed framework lies in its ability to objectively quantify criteria importance using CRITIC,incorporate decision-makers’preferences through CIMAS,and capture the uncertainty and hesitation inherent in human judgment via Pythagorean Fuzzy Sets.A case study evaluating green infrastructure alternatives in metropolitan regions demonstrates the applicability and effectiveness of the framework.A sensitivity analysis is conducted to examine how variations in criteria weights affect the rankings and to evaluate the robustness of the results.Furthermore,a comparative analysis highlights the practical and financial implications of each alternative by assessing their respective strengths and weaknesses.展开更多
Accurate prediction of concrete compressive strength is fundamental for optimizing mix designs,improving material utilization,and ensuring structural safety in modern construction.Traditional empirical methods often f...Accurate prediction of concrete compressive strength is fundamental for optimizing mix designs,improving material utilization,and ensuring structural safety in modern construction.Traditional empirical methods often fail to capture the non-linear relationships among concrete constituents,especially with the growing use of supple-mentary cementitious materials and recycled aggregates.This study presents an integrated machine learning framework for concrete strength prediction,combining advanced regression models—namely CatBoost—with metaheuristic optimization algorithms,with a particular focus on the Somersaulting Spider Optimizer(SSO).A comprehensive dataset encompassing diverse mix proportions and material types was used to evaluate baseline machine learning models,including CatBoost,XGBoost,ExtraTrees,and RandomForest.Among these,CatBoost demonstrated superior accuracy across multiple performance metrics.To further enhance predictive capability,several bio-inspired optimizers were employed for hyperparameter tuning.The SSO-CatBoost hybrid achieved the lowest mean squared error and highest correlation coefficients,outperforming other metaheuristic approaches such as Genetic Algorithm,Particle Swarm Optimization,and Grey Wolf Optimizer.Statistical significance was established through Analysis of Variance and Wilcoxon signed-rank testing,confirming the robustness of the optimized models.The proposed methodology not only delivers improved predictive performance but also offers a transparent framework for mix design optimization,supporting data-driven decision making in sustainable and resilient infrastructure development.展开更多
The cloud-fog computing paradigm has emerged as a novel hybrid computing model that integrates computational resources at both fog nodes and cloud servers to address the challenges posed by dynamic and heterogeneous c...The cloud-fog computing paradigm has emerged as a novel hybrid computing model that integrates computational resources at both fog nodes and cloud servers to address the challenges posed by dynamic and heterogeneous computing networks.Finding an optimal computational resource for task offloading and then executing efficiently is a critical issue to achieve a trade-off between energy consumption and transmission delay.In this network,the task processed at fog nodes reduces transmission delay.Still,it increases energy consumption,while routing tasks to the cloud server saves energy at the cost of higher communication delay.Moreover,the order in which offloaded tasks are executed affects the system’s efficiency.For instance,executing lower-priority tasks before higher-priority jobs can disturb the reliability and stability of the system.Therefore,an efficient strategy of optimal computation offloading and task scheduling is required for operational efficacy.In this paper,we introduced a multi-objective and enhanced version of Cheeta Optimizer(CO),namely(MoECO),to jointly optimize the computation offloading and task scheduling in cloud-fog networks to minimize two competing objectives,i.e.,energy consumption and communication delay.MoECO first assigns tasks to the optimal computational nodes and then the allocated tasks are scheduled for processing based on the task priority.The mathematical modelling of CO needs improvement in computation time and convergence speed.Therefore,MoECO is proposed to increase the search capability of agents by controlling the search strategy based on a leader’s location.The adaptive step length operator is adjusted to diversify the solution and thus improves the exploration phase,i.e.,global search strategy.Consequently,this prevents the algorithm from getting trapped in the local optimal solution.Moreover,the interaction factor during the exploitation phase is also adjusted based on the location of the prey instead of the adjacent Cheetah.This increases the exploitation capability of agents,i.e.,local search capability.Furthermore,MoECO employs a multi-objective Pareto-optimal front to simultaneously minimize designated objectives.Comprehensive simulations in MATLAB demonstrate that the proposed algorithm obtains multiple solutions via a Pareto-optimal front and achieves an efficient trade-off between optimization objectives compared to baseline methods.展开更多
Wireless Sensor Networks(WSNs)have become foundational in numerous real-world applications,ranging from environmental monitoring and industrial automation to healthcare systems and smart city development.As these netw...Wireless Sensor Networks(WSNs)have become foundational in numerous real-world applications,ranging from environmental monitoring and industrial automation to healthcare systems and smart city development.As these networks continue to grow in scale and complexity,the need for energy-efficient,scalable,and robust communication protocols becomes more critical than ever.Metaheuristic algorithms have shown significant promise in addressing these challenges,offering flexible and effective solutions for optimizing WSN performance.Among them,the Grey Wolf Optimizer(GWO)algorithm has attracted growing attention due to its simplicity,fast convergence,and strong global search capabilities.Accordingly,this survey provides an in-depth review of the applications of GWO and its variants for clustering,multi-hop routing,and hybrid cluster-based routing in WSNs.We categorize and analyze the existing GWO-based approaches across these key network optimization tasks,discussing the different problem formulations,decision variables,objective functions,and performance metrics used.In doing so,we examine standard GWO,multi-objective GWO,and hybrid GWO models that incorporate other computational intelligence techniques.Each method is evaluated based on how effectively it addresses the core constraints of WSNs,including energy consumption,communication overhead,and network lifetime.Finally,this survey outlines existing gaps in the literature and proposes potential future research directions aimed at enhancing the effectiveness and real-world applicability of GWO-based techniques for WSN clustering and routing.Our goal is to provide researchers and practitioners with a clear,structured understanding of the current state of GWO in WSNs and inspire further innovation in this evolving field.展开更多
With the rapid development of artificial intelligence,intelligent air combat maneuver decision-making(ACMD)has garnered global attention.Although deep reinforcement learning provides a promising approach to ACMD,exist...With the rapid development of artificial intelligence,intelligent air combat maneuver decision-making(ACMD)has garnered global attention.Although deep reinforcement learning provides a promising approach to ACMD,existing methods often suffer from rigid reward functions and limited adaptability to evolving adversarial strategies.Moreover,most research assumes open airspace,overlooking the influence of potential obstacles.In this paper,we address one-on-one within-visual-range ACMD in obstructed environments,and propose an improved Soft Actor-Critic(SAC)algorithm trained under a curriculum self-play framework.A maneuver strategy mirroring inference module is integrated to estimate each other's likely positions when visual obstruction occurs.By leveraging curriculum learning to guide progressive experience accumulation and self-play for adversarial evolution,our method enhances both training efficiency and tactical diversity.We further integrate an attention mechanism that dynamically adjusts the weights of sub-rewards,enabling the learned policy to adapt to rapidly changing air combat situations.Numerical simulations demonstrate that our enhanced SAC converges more quickly and achieves higher win rates than other baseline methods.An animation is available at bilibili.com/video/BV1BHVszHE98 for better illustration.展开更多
Reinforcement learning(RL)has been widely studied as an efficient class of machine learning methods for adaptive optimal control under uncertainties.In recent years,the applications of RL in optimised decision-making ...Reinforcement learning(RL)has been widely studied as an efficient class of machine learning methods for adaptive optimal control under uncertainties.In recent years,the applications of RL in optimised decision-making and motion control of intelligent vehicles have received increasing attention.Due to the complex and dynamic operating environments of intelligent vehicles,it is necessary to improve the learning efficiency and generalisation ability of RL-based decision and control algorithms under different conditions.This survey systematically examines the theoretical foundations,algorithmic advancements and practical challenges of applying RL to intelligent vehicle systems operating in complex and dynamic environments.The major algorithm frameworks of RL are first introduced,and the recent advances in RL-based decision-making and control of intelligent vehicles are overviewed.In addition to self-learning decision and control approaches using state measurements,the developments of DRL methods for end-to-end driving control of intelligent vehicles are summarised.The open problems and directions for further research works are also discussed.展开更多
With the increasing complexity of industrial automation,planetary gearboxes play a vital role in largescale equipment transmission systems,directly impacting operational efficiency and safety.Traditional maintenance s...With the increasing complexity of industrial automation,planetary gearboxes play a vital role in largescale equipment transmission systems,directly impacting operational efficiency and safety.Traditional maintenance strategies often struggle to accurately predict the degradation process of equipment,leading to excessive maintenance costs or potential failure risks.However,existing prediction methods based on statistical models are difficult to adapt to nonlinear degradation processes.To address these challenges,this study proposes a novel condition-based maintenance framework for planetary gearboxes.A comprehensive full-lifecycle degradation experiment was conducted to collect raw vibration signals,which were then processed using a temporal convolutional network autoencoder with multi-scale perception capability to extract deep temporal degradation features,enabling the collaborative extraction of longperiod meshing frequencies and short-term impact features from the vibration signals.Kernel principal component analysis was employed to fuse and normalize these features,enhancing the characterization of degradation progression.A nonlinear Wiener process was used to model the degradation trajectory,with a threshold decay function introduced to dynamically adjust maintenance strategies,and model parameters optimized through maximum likelihood estimation.Meanwhile,the maintenance strategy was optimized to minimize costs per unit time,determining the optimal maintenance timing and preventive maintenance threshold.The comprehensive indicator of degradation trends extracted by this method reaches 0.756,which is 41.2%higher than that of traditional time-domain features;the dynamic threshold strategy reduces the maintenance cost per unit time to 55.56,which is 8.9%better than that of the static threshold optimization.Experimental results demonstrate significant reductions in maintenance costs while enhancing system reliability and safety.This study realizes the organic integration of deep learning and reliability theory in the maintenance of planetary gearboxes,provides an interpretable solution for the predictive maintenance of complex mechanical systems,and promotes the development of condition-based maintenance strategies for planetary gearboxes.展开更多
Electrical parking lots(EPLs)play a vital role in the current energy system to achieve the decarbonization goal.This paper proposes a novel structure for integrating EPLs into a multi-carrier energy system(MCES)using ...Electrical parking lots(EPLs)play a vital role in the current energy system to achieve the decarbonization goal.This paper proposes a novel structure for integrating EPLs into a multi-carrier energy system(MCES)using a Stackelberg game theory approach.The bi-level optimization is used to model the Stackelberg game.Within this bi-level optimization model,the MCES operator minimizes the MCES cost by participating in the upstream energy market at the upper level,and the EPL operators maximize their profits by participating in the local energy market between the MCES operator and themselves at the lower level.At the upper level,the MCES operator faces uncertainties in the wind and PV systems.The bi-level multi-objective information gap decision theory(MO-IGDT)is employed to address uncertainties at the upper level of the Stackelberg game problem,resulting in a nested bi-level optimization model.The nested bi-level optimization problem is converted into a mixed-integer linear programming(MILP)optimization problem using Karush–Kuhn–Tucker(KKT)conditions.The main research assumptions pertain to EPLs’privacy and the KKT-based approach.The results demonstrate that increasing the incentive/penalty price for self-sufficiency programs from 0.0$/%to 0.2$/%,with a 50%self-sufficiency target,can reduce MCES operation costs by 10.19%.展开更多
Ground water is a crucial ecological resource and source of drinking water to a great percentage of theworld population.The quality of groundwater in an area with industrial emission and air pollution is an especially...Ground water is a crucial ecological resource and source of drinking water to a great percentage of theworld population.The quality of groundwater in an area with industrial emission and air pollution is an especiallyimportant issue that requires proper evaluation.This paper introduces a spatiotemporal deep learning model thatincorporates the use of metaheuristic optimization in predicting groundwater quality in various pollution contexts.Thegiven method is a combination of the Spatial-Temporal-Assisted Deep Belief Network(StaDBN)and a hybrid WhaleOptimization Algorithm and Tiki-Taka Algorithms(WOA-TTA)that would model intricate patterns of contamination.Historical ground water data sets with the hydrochemical data and time are preprocessed and pertinent and nonredundant features are determined with the Addax Optimization Algorithm(AOA).Spatial and temporal dependenciesare explicitly integrated in StaDBN architecture to facilitate representation learning,and network hyperparametersare optimized by the WOA-TTA module to increase the training efficiency and predictive performance.The modelwas coded in Python and tested based on common statistical measures,such as root mean square error(RMSE),Nash Sutcliffe efficiency(NSE),mean absolute error(MAE),and the correlation coefficient(R).The proposedGWQP-StaDBN-WOA-TTA framework demonstrates superior predictive performance and interpretability comparedto conventional machine learning and deep learning models,achieving higher correlation(R=0.963),improvedNash-Sutcliffe efficiency(NSE=0.84),and substantially lower prediction errors(MAE=0.29,RMSE=0.48),therebyvalidating its effectiveness for groundwater quality assessment under industrial and atmospheric pollution scenarios.展开更多
Magnetic Resonance Imaging(MRI)has a pivotal role in medical image analysis,for its ability in supporting disease detection and diagnosis.Fuzzy C-Means(FCM)clustering is widely used for MRI segmentation due to its abi...Magnetic Resonance Imaging(MRI)has a pivotal role in medical image analysis,for its ability in supporting disease detection and diagnosis.Fuzzy C-Means(FCM)clustering is widely used for MRI segmentation due to its ability to handle image uncertainty.However,the latter still has countless limitations,including sensitivity to initialization,susceptibility to local optima,and high computational cost.To address these limitations,this study integrates Grey Wolf Optimization(GWO)with FCM to enhance cluster center selection,improving segmentation accuracy and robustness.Moreover,to further refine optimization,Fuzzy Entropy Clustering was utilized for its distinctive features from other traditional objective functions.Fuzzy entropy effectively quantifies uncertainty,leading to more well-defined clusters,improved noise robustness,and better preservation of anatomical structures in MRI images.Despite these advantages,the iterative nature of GWO and FCM introduces significant computational overhead,which restricts their applicability to high-resolution medical images.To overcome this bottleneck,we propose a Parallelized-GWO-based FCM(P-GWO-FCM)approach using GPU acceleration,where both GWO optimization and FCM updates(centroid computation and membership matrix updates)are parallelized.By concurrently executing these processes,our approach efficiently distributes the computational workload,significantly reducing execution time while maintaining high segmentation accuracy.The proposed parallel method,P-GWO-FCM,was evaluated on both simulated and clinical brain MR images,focusing on segmenting white matter,gray matter,and cerebrospinal fluid regions.The results indicate significant improvements in segmentation accuracy,achieving a Jaccard Similarity(JS)of 0.92,a Partition Coefficient Index(PCI)of 0.91,a Partition Entropy Index(PEI)of 0.25,and a Davies-Bouldin Index(DBI)of 0.30.Experimental comparisons demonstrate that P-GWO-FCM outperforms existing methods in both segmentation accuracy and computational efficiency,making it a promising solution for real-time medical image segmentation.展开更多
In Wireless Sensor Networks(WSNs),survivability is a crucial issue that is greatly impacted by energy efficiency.Solutions that satisfy application objectives while extending network life are needed to address severe ...In Wireless Sensor Networks(WSNs),survivability is a crucial issue that is greatly impacted by energy efficiency.Solutions that satisfy application objectives while extending network life are needed to address severe energy constraints inWSNs.This paper presents an Adaptive Enhanced GreyWolf Optimizer(AEGWO)for energy-efficient cluster head(CH)selection that mitigates the exploration–exploitation imbalance,preserves population diversity,and avoids premature convergence inherent in baseline GWO.The AEGWO combines adaptive control of the parameter of the search pressure to accelerate convergence without stagnation,a hybrid velocity-momentum update based on the dynamics of PSO,and an intelligent mutation operator to maintain the diversity of the population.The search is guided by a multi-objective fitness,which aims at maximizing the residual energy,equal distribution of CH,minimizing the intra-cluster distance,desirable proximity to sinks,and enhancing the coverage.Simulations on 100 nodes homogeneousWSN Tested the proposed AEGWO under the same conditions with LEACH,GWO,IGWO,PSO,WOA,and GA,AEGWO significantly increases stability and lifetime compared to LEACHand other tested algorithms;it has the best first,half,and last node dead,and higher residual energy and smaller communication overhead.The findings prove that AEGWO provides sustainable energy management and better lifetime extension,which makes it a robust,flexible clustering protocol of large-scaleWSNs.展开更多
Under the influence of human activities,landscape fragmentation in the Wei River Basin(WRB)has become increasingly severe.Upstream development has intensified soil erosion,and industrial and agricultural pollution in ...Under the influence of human activities,landscape fragmentation in the Wei River Basin(WRB)has become increasingly severe.Upstream development has intensified soil erosion,and industrial and agricultural pollution in the middle reaches has degraded water quality.Rapid urbanization has further caused habitat fragmentation and biodiversity loss.Collectively,these challenges threaten human well-being and hinder sustainable development,making the construction and optimization of an ecological security pattern(ESP)urgently necessary.However,existing studies often fail to systematically integrate future landscape ecological risk(LER)assessment with ESP optimization.This study evaluated regional LER using the“ecological patches-ecological resistance surface(ERS)-ecological corridor”framework,combined with land-use predictions under three development scenarios,and optimized the ESP by adjusting the ERS and extracting ecological corridors.The results indicate that the LER in the WRB follows an“inverted N”distribution,with low-risk areas concentrated in forested mountain regions and high-risk areas mainly in cultivated land subject to intensive human activity.Across future scenarios,ESPs showed fewer ecological breakpoints and improved landscape connectivity than the 2020 baseline.Scenario-based differences emerged in the spatial configuration of ERS adjustments,with the ecological protection scenario yielding the lowest LER and most favorable ESP.This study demonstrates the deep integration of multi-scenario simulation with LER assessment,providing a new framework for ESP optimization.The findings have guiding significance for ecological protection and coordinated development in the WRB and offer a novel paradigm for sustainable development in ecologically fragile basins worldwide.展开更多
In the rapidly evolving technological landscape,state-owned enterprises(SOEs)encounter significant challenges in sustaining their competitiveness through efficient R&D management.Integrated Product Development(IPD...In the rapidly evolving technological landscape,state-owned enterprises(SOEs)encounter significant challenges in sustaining their competitiveness through efficient R&D management.Integrated Product Development(IPD),with its emphasis on cross-functional teamwork,concurrent engineering,and data-driven decision-making,has been widely recognized for enhancing R&D efficiency and product quality.However,the unique characteristics of SOEs pose challenges to the effective implementation of IPD.The advancement of big data and artificial intelligence technologies offers new opportunities for optimizing IPD R&D management through data-driven decision-making models.This paper constructs and validates a data-driven decision-making model tailored to the IPD R&D management of SOEs.By integrating data mining,machine learning,and other advanced analytical techniques,the model serves as a scientific and efficient decision-making tool.It aids SOEs in optimizing R&D resource allocation,shortening product development cycles,reducing R&D costs,and improving product quality and innovation.Moreover,this study contributes to a deeper theoretical understanding of the value of data-driven decision-making in the context of IPD.展开更多
Variable stiffness composites present a promising solution for mitigating impact loads via varying the fiber volume fraction layer-wise,thereby adjusting the panel's stiffness.Since each layer of the composite may...Variable stiffness composites present a promising solution for mitigating impact loads via varying the fiber volume fraction layer-wise,thereby adjusting the panel's stiffness.Since each layer of the composite may be affected by a different failure mode,the optimal fiber volume fraction to suppress damage initiation and evolution is different across the layers.This research examines how re-allocating the fibers layer-wise enhances the composites'impact resistance.In this study,constant stiffness panels with the same fiber volume fraction throughout the layers are compared to variable stiffness ones by varying volume fraction layer-wise.A method is established that utilizes numerical analysis coupled with optimization techniques to determine the optimal fiber volume fraction in both scenarios.Three different reinforcement fibers(Kevlar,carbon,and glass)embedded in epoxy resin were studied.Panels were manufactured and tested under various loading conditions to validate results.Kevlar reinforcement revealed the highest tensile toughness,followed by carbon and then glass fibers.Varying reinforcement volume fraction significantly influences failure modes.Higher fractions lead to matrix cracking and debonding,while lower fractions result in more fiber breakage.The optimal volume fraction for maximizing fiber breakage energy is around 45%,whereas it is about 90%for matrix cracking and debonding.A drop tower test was used to examine the composite structure's behavior under lowvelocity impact,confirming the superiority of Kevlar-reinforced composites with variable stiffness.Conversely,glass-reinforced composites with constant stiffness revealed the lowest performance with the highest deflection.Across all reinforcement materials,the variable stiffness structure consistently outperformed its constant stiffness counterpart.展开更多
The increasing integration of cyber-physical components in Industry 4.0 water infrastructures has heightened the risk of false data injection(FDI)attacks,posing critical threats to operational integrity,resource manag...The increasing integration of cyber-physical components in Industry 4.0 water infrastructures has heightened the risk of false data injection(FDI)attacks,posing critical threats to operational integrity,resource management,and public safety.Traditional detection mechanisms often struggle to generalize across heterogeneous environments or adapt to sophisticated,stealthy threats.To address these challenges,we propose a novel evolutionary optimized transformer-based deep reinforcement learning framework(Evo-Transformer-DRL)designed for robust and adaptive FDI detection in smart water infrastructures.The proposed architecture integrates three powerful paradigms:a transformer encoder for modeling complex temporal dependencies in multivariate time series,a DRL agent for learning optimal decision policies in dynamic environments,and an evolutionary optimizer to fine-tune model hyper-parameters.This synergy enhances detection performance while maintaining adaptability across varying data distributions.Specifically,hyper-parameters of both the transformer and DRL modules are optimized using an improved grey wolf optimizer(IGWO),ensuring a balanced trade-off between detection accuracy and computational efficiency.The model is trained and evaluated on three realistic Industry 4.0 water datasets:secure water treatment(SWaT),water distribution(WADI),and battle of the attack detection algorithms(BATADAL),which capture diverse attack scenarios in smart treatment and distribution systems.Comparative analysis against state-of-the-art baselines including Transformer,DRL,bidirectional encoder representations from transformers(BERT),convolutional neural network(CNN),long short-term memory(LSTM),and support vector machines(SVM)demonstrates that our proposed Evo-Transformer-DRL framework consistently outperforms others in key metrics such as accuracy,recall,area under the curve(AUC),and execution time.Notably,it achieves a maximum detection accuracy of 99.19%,highlighting its strong generalization capability across different testbeds.These results confirm the suitability of our hybrid framework for real-world Industry 4.0 deployment,where rapid adaptation,scalability,and reliability are paramount for securing critical infrastructure systems.展开更多
基金supported by the Zhejiang Province 14th Five Year Plan Teaching Reform Project(jg20220514).
文摘Developing innovative capabilities in university students is essential for individual career success and broader societal advancement.This study introduces a predictive Feature Selection(FS)model named bWRBA-SVM-FS,which combines an enhanced Bat Algorithm(BA)and Support Vector Machine(SVM).To enhance the optimization capability of BA,water follow search and random follow search are introduced to optimize the efficiency and accuracy of the feature subset search.Experimental validation conducted on the IEEE CEC 2017 benchmark functions and the talented innovative capacity dataset demonstrates the efficacy of the proposed method relative to peer and prominent machine learning models.The experimental results reveal that the predictive accuracy of the bWRBA-SVM-FS model is 97.503%,with a sensitivity of 98.391%.Our findings indicate significant predictors of innovation capacity,including project application goals,educational background,and interdisciplinary thinking abilities.The bWRBA-SVM-FS model offers effective strategies for talent selection in higher education,fostering the development of future research leaders.
文摘Background:Despite the promise shown by large language models(LLMs)for standardized tasks,their multidimensional performance in real-world oncology decision-making remains unevaluated.This study aims to introduce a framework for evaluating LLMs and physician decisions in challenging lung cancer cases.Methods:We curated 50 challenging lung cancer cases(25 local and 25 published)classified as complex,rare,or refractory.Blinded three-dimensional,five-point Likert evaluations(1–5 for comprehensiveness,specificity,and readability)compared standalone LLMs(DeepSeek R1,Claude 3.5,Gemini 1.5,and GPT-4o),physicians by experience level(junior,intermediate,and senior),and AI-assisted juniors;intergroup differences and augmentation effects were analyzed statistically.Results:Of 50 challenging cases(18 complex,17 rare,and 15 refractory)rated by three experts,DeepSeek R1 achieved scores of 3.95±0.33,3.71±0.53,and 4.26±0.18 for comprehensiveness,specificity,and readability,respectively,positioning it between intermediate(3.68,3.68,3.75)and senior(4.50,4.64,4.53)physicians.GPT-4o and Claude 3.5 reached intermediate physician–level comprehensiveness(3.76±0.39,3.60±0.39)but junior-to-intermediate physician–level specificity(3.39±0.39,3.39±0.49).All LLMs scored higher on rare cases than intermediate physicians but fell below junior physicians in refractory-case specificity.AIassisted junior physicians showed marked gains in rare cases,with comprehensiveness rising from 2.32 to 4.29(84.8%),specificity from 2.24 to 4.26(90.8%),and readability from 2.76 to 4.59(66.0%),while specificity declined by 3.2%(3.17 to 3.07)in refractory cases.Error analysis showed complementary strengths,with physicians demonstrating reasoning stability and LLMs excelling in knowledge updating and risk management.Conclusions:LLMs performed variably in clinical decision-making tasks depending on case type,performing better in rare cases and worse in refractory cases requiring longitudinal reasoning.Complementary strengths between LLMs and physicians support case-and task-tailored human–AI collaboration.
文摘Optimization algorithms are crucial for solving NP-hard problems in engineering and computational sciences.Metaheuristic algorithms,in particular,have proven highly effective in complex optimization scenarios characterized by high dimensionality and intricate variable relationships.The Mountain Gazelle Optimizer(MGO)is notably effective but struggles to balance local search refinement and global space exploration,often leading to premature convergence and entrapment in local optima.This paper presents the Improved MGO(IMGO),which integrates three synergistic enhancements:dynamic chaos mapping using piecewise chaotic sequences to boost explo-ration diversity;Opposition-Based Learning(OBL)with adaptive,diversity-driven activation to speed up convergence;and structural refinements to the position update mechanisms to enhance exploitation.The IMGO underwent a comprehensive evaluation using 52 standardised benchmark functions and seven engineering optimization problems.Benchmark evaluations showed that IMGO achieved the highest rank in best solution quality for 31 functions,the highest rank in mean performance for 18 functions,and the highest rank in worst-case performance for 14 functions among 11 competing algorithms.Statistical validation using Wilcoxon signed-rank tests confirmed that IMGO outperformed individual competitors across 16 to 50 functions,depending on the algorithm.At the same time,Friedman ranking analysis placed IMGO with an average rank of 4.15,compared to the baseline MGO’s 4.38,establishing the best overall performance.The evaluation of engineering problems revealed consistent improvements,including an optimal cost of 1.6896 for the welded beam design vs.MGO’s 1.7249,a minimum cost of 5885.33 for the pressure vessel design vs.MGO’s 6300,and a minimum weight of 2964.52 kg for the speed reducer design vs.MGO’s 2990.00 kg.Ablation studies identified OBL as the strongest individual contributor,whereas complete integration achieved superior performance through synergistic interactions among components.Computational complexity analysis established an O(T×N×5×f(P))time complexity,representing a 1.25×increase in fitness evaluation relative to the baseline MGO,validating the favorable accuracy-efficiency trade-offs for practical optimization applications.
基金supported by the Princess Nourah bint Abdulrahman University Researchers Supporting Project number(PNURSP2026R259)Princess Nourah bint Abdulrahman University,Riyadh,Saudi Arabia.Ashit Kumar Dutta would like to thank AlMaarefa University for supporting this research under project number MHIRSP2025017.
文摘Environmental problems are intensifying due to the rapid growth of the population,industry,and urban infrastructure.This expansion has resulted in increased air and water pollution,intensified urban heat island effects,and greater runoff from parks and other green spaces.Addressing these challenges requires prioritizing green infrastructure and other sustainable urban development strategies.This study introduces a novel Integrated Decision Support System that combines Pythagorean Fuzzy Sets with the Advanced Alternative Ranking Order Method allowing for Two-Step Normalization(AAROM-TN),enhanced by a dual weighting strategy.The weighting approach integrates the Criteria Importance Through Intercriteria Correlation(CRITIC)method with the Criteria Importance through Means and Standard Deviation(CIMAS)technique.The originality of the proposed framework lies in its ability to objectively quantify criteria importance using CRITIC,incorporate decision-makers’preferences through CIMAS,and capture the uncertainty and hesitation inherent in human judgment via Pythagorean Fuzzy Sets.A case study evaluating green infrastructure alternatives in metropolitan regions demonstrates the applicability and effectiveness of the framework.A sensitivity analysis is conducted to examine how variations in criteria weights affect the rankings and to evaluate the robustness of the results.Furthermore,a comparative analysis highlights the practical and financial implications of each alternative by assessing their respective strengths and weaknesses.
文摘Accurate prediction of concrete compressive strength is fundamental for optimizing mix designs,improving material utilization,and ensuring structural safety in modern construction.Traditional empirical methods often fail to capture the non-linear relationships among concrete constituents,especially with the growing use of supple-mentary cementitious materials and recycled aggregates.This study presents an integrated machine learning framework for concrete strength prediction,combining advanced regression models—namely CatBoost—with metaheuristic optimization algorithms,with a particular focus on the Somersaulting Spider Optimizer(SSO).A comprehensive dataset encompassing diverse mix proportions and material types was used to evaluate baseline machine learning models,including CatBoost,XGBoost,ExtraTrees,and RandomForest.Among these,CatBoost demonstrated superior accuracy across multiple performance metrics.To further enhance predictive capability,several bio-inspired optimizers were employed for hyperparameter tuning.The SSO-CatBoost hybrid achieved the lowest mean squared error and highest correlation coefficients,outperforming other metaheuristic approaches such as Genetic Algorithm,Particle Swarm Optimization,and Grey Wolf Optimizer.Statistical significance was established through Analysis of Variance and Wilcoxon signed-rank testing,confirming the robustness of the optimized models.The proposed methodology not only delivers improved predictive performance but also offers a transparent framework for mix design optimization,supporting data-driven decision making in sustainable and resilient infrastructure development.
基金appreciation to the Princess Nourah bint Abdulrahman University Researchers Supporting Project number(PNURSP2025R384)Princess Nourah bint Abdulrahman University,Riyadh,Saudi Arabia.
文摘The cloud-fog computing paradigm has emerged as a novel hybrid computing model that integrates computational resources at both fog nodes and cloud servers to address the challenges posed by dynamic and heterogeneous computing networks.Finding an optimal computational resource for task offloading and then executing efficiently is a critical issue to achieve a trade-off between energy consumption and transmission delay.In this network,the task processed at fog nodes reduces transmission delay.Still,it increases energy consumption,while routing tasks to the cloud server saves energy at the cost of higher communication delay.Moreover,the order in which offloaded tasks are executed affects the system’s efficiency.For instance,executing lower-priority tasks before higher-priority jobs can disturb the reliability and stability of the system.Therefore,an efficient strategy of optimal computation offloading and task scheduling is required for operational efficacy.In this paper,we introduced a multi-objective and enhanced version of Cheeta Optimizer(CO),namely(MoECO),to jointly optimize the computation offloading and task scheduling in cloud-fog networks to minimize two competing objectives,i.e.,energy consumption and communication delay.MoECO first assigns tasks to the optimal computational nodes and then the allocated tasks are scheduled for processing based on the task priority.The mathematical modelling of CO needs improvement in computation time and convergence speed.Therefore,MoECO is proposed to increase the search capability of agents by controlling the search strategy based on a leader’s location.The adaptive step length operator is adjusted to diversify the solution and thus improves the exploration phase,i.e.,global search strategy.Consequently,this prevents the algorithm from getting trapped in the local optimal solution.Moreover,the interaction factor during the exploitation phase is also adjusted based on the location of the prey instead of the adjacent Cheetah.This increases the exploitation capability of agents,i.e.,local search capability.Furthermore,MoECO employs a multi-objective Pareto-optimal front to simultaneously minimize designated objectives.Comprehensive simulations in MATLAB demonstrate that the proposed algorithm obtains multiple solutions via a Pareto-optimal front and achieves an efficient trade-off between optimization objectives compared to baseline methods.
文摘Wireless Sensor Networks(WSNs)have become foundational in numerous real-world applications,ranging from environmental monitoring and industrial automation to healthcare systems and smart city development.As these networks continue to grow in scale and complexity,the need for energy-efficient,scalable,and robust communication protocols becomes more critical than ever.Metaheuristic algorithms have shown significant promise in addressing these challenges,offering flexible and effective solutions for optimizing WSN performance.Among them,the Grey Wolf Optimizer(GWO)algorithm has attracted growing attention due to its simplicity,fast convergence,and strong global search capabilities.Accordingly,this survey provides an in-depth review of the applications of GWO and its variants for clustering,multi-hop routing,and hybrid cluster-based routing in WSNs.We categorize and analyze the existing GWO-based approaches across these key network optimization tasks,discussing the different problem formulations,decision variables,objective functions,and performance metrics used.In doing so,we examine standard GWO,multi-objective GWO,and hybrid GWO models that incorporate other computational intelligence techniques.Each method is evaluated based on how effectively it addresses the core constraints of WSNs,including energy consumption,communication overhead,and network lifetime.Finally,this survey outlines existing gaps in the literature and proposes potential future research directions aimed at enhancing the effectiveness and real-world applicability of GWO-based techniques for WSN clustering and routing.Our goal is to provide researchers and practitioners with a clear,structured understanding of the current state of GWO in WSNs and inspire further innovation in this evolving field.
基金support of the National Key Research and Development Plan(No.2021YFB3302501)the financial support of the National Science Foundation of China(No.12161076)the financial support of the Fundamental Research Funds for the Central Universities(No.DUT25GF207).
文摘With the rapid development of artificial intelligence,intelligent air combat maneuver decision-making(ACMD)has garnered global attention.Although deep reinforcement learning provides a promising approach to ACMD,existing methods often suffer from rigid reward functions and limited adaptability to evolving adversarial strategies.Moreover,most research assumes open airspace,overlooking the influence of potential obstacles.In this paper,we address one-on-one within-visual-range ACMD in obstructed environments,and propose an improved Soft Actor-Critic(SAC)algorithm trained under a curriculum self-play framework.A maneuver strategy mirroring inference module is integrated to estimate each other's likely positions when visual obstruction occurs.By leveraging curriculum learning to guide progressive experience accumulation and self-play for adversarial evolution,our method enhances both training efficiency and tactical diversity.We further integrate an attention mechanism that dynamically adjusts the weights of sub-rewards,enabling the learned policy to adapt to rapidly changing air combat situations.Numerical simulations demonstrate that our enhanced SAC converges more quickly and achieves higher win rates than other baseline methods.An animation is available at bilibili.com/video/BV1BHVszHE98 for better illustration.
基金supported by the National Natural Science Foundation of China under Grant T2521006,Grant 62403483,Grant 62533021 and Grant U24A20279.
文摘Reinforcement learning(RL)has been widely studied as an efficient class of machine learning methods for adaptive optimal control under uncertainties.In recent years,the applications of RL in optimised decision-making and motion control of intelligent vehicles have received increasing attention.Due to the complex and dynamic operating environments of intelligent vehicles,it is necessary to improve the learning efficiency and generalisation ability of RL-based decision and control algorithms under different conditions.This survey systematically examines the theoretical foundations,algorithmic advancements and practical challenges of applying RL to intelligent vehicle systems operating in complex and dynamic environments.The major algorithm frameworks of RL are first introduced,and the recent advances in RL-based decision-making and control of intelligent vehicles are overviewed.In addition to self-learning decision and control approaches using state measurements,the developments of DRL methods for end-to-end driving control of intelligent vehicles are summarised.The open problems and directions for further research works are also discussed.
基金funded by scientific research projects under Grant JY2024B011.
文摘With the increasing complexity of industrial automation,planetary gearboxes play a vital role in largescale equipment transmission systems,directly impacting operational efficiency and safety.Traditional maintenance strategies often struggle to accurately predict the degradation process of equipment,leading to excessive maintenance costs or potential failure risks.However,existing prediction methods based on statistical models are difficult to adapt to nonlinear degradation processes.To address these challenges,this study proposes a novel condition-based maintenance framework for planetary gearboxes.A comprehensive full-lifecycle degradation experiment was conducted to collect raw vibration signals,which were then processed using a temporal convolutional network autoencoder with multi-scale perception capability to extract deep temporal degradation features,enabling the collaborative extraction of longperiod meshing frequencies and short-term impact features from the vibration signals.Kernel principal component analysis was employed to fuse and normalize these features,enhancing the characterization of degradation progression.A nonlinear Wiener process was used to model the degradation trajectory,with a threshold decay function introduced to dynamically adjust maintenance strategies,and model parameters optimized through maximum likelihood estimation.Meanwhile,the maintenance strategy was optimized to minimize costs per unit time,determining the optimal maintenance timing and preventive maintenance threshold.The comprehensive indicator of degradation trends extracted by this method reaches 0.756,which is 41.2%higher than that of traditional time-domain features;the dynamic threshold strategy reduces the maintenance cost per unit time to 55.56,which is 8.9%better than that of the static threshold optimization.Experimental results demonstrate significant reductions in maintenance costs while enhancing system reliability and safety.This study realizes the organic integration of deep learning and reliability theory in the maintenance of planetary gearboxes,provides an interpretable solution for the predictive maintenance of complex mechanical systems,and promotes the development of condition-based maintenance strategies for planetary gearboxes.
基金supported by the first Cycle of ARG Grant No.ARG01-0504-230073,from the Qatar Research,Development and Innovation(QRDI)Council,Qatar.The findings herein reflect the work,and are solely the responsibility,of the authors.The authors also gratefully acknowledge support from Qatar University.
文摘Electrical parking lots(EPLs)play a vital role in the current energy system to achieve the decarbonization goal.This paper proposes a novel structure for integrating EPLs into a multi-carrier energy system(MCES)using a Stackelberg game theory approach.The bi-level optimization is used to model the Stackelberg game.Within this bi-level optimization model,the MCES operator minimizes the MCES cost by participating in the upstream energy market at the upper level,and the EPL operators maximize their profits by participating in the local energy market between the MCES operator and themselves at the lower level.At the upper level,the MCES operator faces uncertainties in the wind and PV systems.The bi-level multi-objective information gap decision theory(MO-IGDT)is employed to address uncertainties at the upper level of the Stackelberg game problem,resulting in a nested bi-level optimization model.The nested bi-level optimization problem is converted into a mixed-integer linear programming(MILP)optimization problem using Karush–Kuhn–Tucker(KKT)conditions.The main research assumptions pertain to EPLs’privacy and the KKT-based approach.The results demonstrate that increasing the incentive/penalty price for self-sufficiency programs from 0.0$/%to 0.2$/%,with a 50%self-sufficiency target,can reduce MCES operation costs by 10.19%.
基金Fund for funding this research work under Research Support Program for Central labs at King Khalid University through the project number CL/CO/B/6.
文摘Ground water is a crucial ecological resource and source of drinking water to a great percentage of theworld population.The quality of groundwater in an area with industrial emission and air pollution is an especiallyimportant issue that requires proper evaluation.This paper introduces a spatiotemporal deep learning model thatincorporates the use of metaheuristic optimization in predicting groundwater quality in various pollution contexts.Thegiven method is a combination of the Spatial-Temporal-Assisted Deep Belief Network(StaDBN)and a hybrid WhaleOptimization Algorithm and Tiki-Taka Algorithms(WOA-TTA)that would model intricate patterns of contamination.Historical ground water data sets with the hydrochemical data and time are preprocessed and pertinent and nonredundant features are determined with the Addax Optimization Algorithm(AOA).Spatial and temporal dependenciesare explicitly integrated in StaDBN architecture to facilitate representation learning,and network hyperparametersare optimized by the WOA-TTA module to increase the training efficiency and predictive performance.The modelwas coded in Python and tested based on common statistical measures,such as root mean square error(RMSE),Nash Sutcliffe efficiency(NSE),mean absolute error(MAE),and the correlation coefficient(R).The proposedGWQP-StaDBN-WOA-TTA framework demonstrates superior predictive performance and interpretability comparedto conventional machine learning and deep learning models,achieving higher correlation(R=0.963),improvedNash-Sutcliffe efficiency(NSE=0.84),and substantially lower prediction errors(MAE=0.29,RMSE=0.48),therebyvalidating its effectiveness for groundwater quality assessment under industrial and atmospheric pollution scenarios.
文摘Magnetic Resonance Imaging(MRI)has a pivotal role in medical image analysis,for its ability in supporting disease detection and diagnosis.Fuzzy C-Means(FCM)clustering is widely used for MRI segmentation due to its ability to handle image uncertainty.However,the latter still has countless limitations,including sensitivity to initialization,susceptibility to local optima,and high computational cost.To address these limitations,this study integrates Grey Wolf Optimization(GWO)with FCM to enhance cluster center selection,improving segmentation accuracy and robustness.Moreover,to further refine optimization,Fuzzy Entropy Clustering was utilized for its distinctive features from other traditional objective functions.Fuzzy entropy effectively quantifies uncertainty,leading to more well-defined clusters,improved noise robustness,and better preservation of anatomical structures in MRI images.Despite these advantages,the iterative nature of GWO and FCM introduces significant computational overhead,which restricts their applicability to high-resolution medical images.To overcome this bottleneck,we propose a Parallelized-GWO-based FCM(P-GWO-FCM)approach using GPU acceleration,where both GWO optimization and FCM updates(centroid computation and membership matrix updates)are parallelized.By concurrently executing these processes,our approach efficiently distributes the computational workload,significantly reducing execution time while maintaining high segmentation accuracy.The proposed parallel method,P-GWO-FCM,was evaluated on both simulated and clinical brain MR images,focusing on segmenting white matter,gray matter,and cerebrospinal fluid regions.The results indicate significant improvements in segmentation accuracy,achieving a Jaccard Similarity(JS)of 0.92,a Partition Coefficient Index(PCI)of 0.91,a Partition Entropy Index(PEI)of 0.25,and a Davies-Bouldin Index(DBI)of 0.30.Experimental comparisons demonstrate that P-GWO-FCM outperforms existing methods in both segmentation accuracy and computational efficiency,making it a promising solution for real-time medical image segmentation.
基金The Open Access publication fee for this article was fully covered by Abu Dhabi University.
文摘In Wireless Sensor Networks(WSNs),survivability is a crucial issue that is greatly impacted by energy efficiency.Solutions that satisfy application objectives while extending network life are needed to address severe energy constraints inWSNs.This paper presents an Adaptive Enhanced GreyWolf Optimizer(AEGWO)for energy-efficient cluster head(CH)selection that mitigates the exploration–exploitation imbalance,preserves population diversity,and avoids premature convergence inherent in baseline GWO.The AEGWO combines adaptive control of the parameter of the search pressure to accelerate convergence without stagnation,a hybrid velocity-momentum update based on the dynamics of PSO,and an intelligent mutation operator to maintain the diversity of the population.The search is guided by a multi-objective fitness,which aims at maximizing the residual energy,equal distribution of CH,minimizing the intra-cluster distance,desirable proximity to sinks,and enhancing the coverage.Simulations on 100 nodes homogeneousWSN Tested the proposed AEGWO under the same conditions with LEACH,GWO,IGWO,PSO,WOA,and GA,AEGWO significantly increases stability and lifetime compared to LEACHand other tested algorithms;it has the best first,half,and last node dead,and higher residual energy and smaller communication overhead.The findings prove that AEGWO provides sustainable energy management and better lifetime extension,which makes it a robust,flexible clustering protocol of large-scaleWSNs.
基金supported by the National Natural Science Foundation of China[Grant No.42361040].
文摘Under the influence of human activities,landscape fragmentation in the Wei River Basin(WRB)has become increasingly severe.Upstream development has intensified soil erosion,and industrial and agricultural pollution in the middle reaches has degraded water quality.Rapid urbanization has further caused habitat fragmentation and biodiversity loss.Collectively,these challenges threaten human well-being and hinder sustainable development,making the construction and optimization of an ecological security pattern(ESP)urgently necessary.However,existing studies often fail to systematically integrate future landscape ecological risk(LER)assessment with ESP optimization.This study evaluated regional LER using the“ecological patches-ecological resistance surface(ERS)-ecological corridor”framework,combined with land-use predictions under three development scenarios,and optimized the ESP by adjusting the ERS and extracting ecological corridors.The results indicate that the LER in the WRB follows an“inverted N”distribution,with low-risk areas concentrated in forested mountain regions and high-risk areas mainly in cultivated land subject to intensive human activity.Across future scenarios,ESPs showed fewer ecological breakpoints and improved landscape connectivity than the 2020 baseline.Scenario-based differences emerged in the spatial configuration of ERS adjustments,with the ecological protection scenario yielding the lowest LER and most favorable ESP.This study demonstrates the deep integration of multi-scenario simulation with LER assessment,providing a new framework for ESP optimization.The findings have guiding significance for ecological protection and coordinated development in the WRB and offer a novel paradigm for sustainable development in ecologically fragile basins worldwide.
文摘In the rapidly evolving technological landscape,state-owned enterprises(SOEs)encounter significant challenges in sustaining their competitiveness through efficient R&D management.Integrated Product Development(IPD),with its emphasis on cross-functional teamwork,concurrent engineering,and data-driven decision-making,has been widely recognized for enhancing R&D efficiency and product quality.However,the unique characteristics of SOEs pose challenges to the effective implementation of IPD.The advancement of big data and artificial intelligence technologies offers new opportunities for optimizing IPD R&D management through data-driven decision-making models.This paper constructs and validates a data-driven decision-making model tailored to the IPD R&D management of SOEs.By integrating data mining,machine learning,and other advanced analytical techniques,the model serves as a scientific and efficient decision-making tool.It aids SOEs in optimizing R&D resource allocation,shortening product development cycles,reducing R&D costs,and improving product quality and innovation.Moreover,this study contributes to a deeper theoretical understanding of the value of data-driven decision-making in the context of IPD.
基金funded by the American University of Sharjah.United Arab Emirates award number EN 9502-FRG19-M-E75。
文摘Variable stiffness composites present a promising solution for mitigating impact loads via varying the fiber volume fraction layer-wise,thereby adjusting the panel's stiffness.Since each layer of the composite may be affected by a different failure mode,the optimal fiber volume fraction to suppress damage initiation and evolution is different across the layers.This research examines how re-allocating the fibers layer-wise enhances the composites'impact resistance.In this study,constant stiffness panels with the same fiber volume fraction throughout the layers are compared to variable stiffness ones by varying volume fraction layer-wise.A method is established that utilizes numerical analysis coupled with optimization techniques to determine the optimal fiber volume fraction in both scenarios.Three different reinforcement fibers(Kevlar,carbon,and glass)embedded in epoxy resin were studied.Panels were manufactured and tested under various loading conditions to validate results.Kevlar reinforcement revealed the highest tensile toughness,followed by carbon and then glass fibers.Varying reinforcement volume fraction significantly influences failure modes.Higher fractions lead to matrix cracking and debonding,while lower fractions result in more fiber breakage.The optimal volume fraction for maximizing fiber breakage energy is around 45%,whereas it is about 90%for matrix cracking and debonding.A drop tower test was used to examine the composite structure's behavior under lowvelocity impact,confirming the superiority of Kevlar-reinforced composites with variable stiffness.Conversely,glass-reinforced composites with constant stiffness revealed the lowest performance with the highest deflection.Across all reinforcement materials,the variable stiffness structure consistently outperformed its constant stiffness counterpart.
文摘The increasing integration of cyber-physical components in Industry 4.0 water infrastructures has heightened the risk of false data injection(FDI)attacks,posing critical threats to operational integrity,resource management,and public safety.Traditional detection mechanisms often struggle to generalize across heterogeneous environments or adapt to sophisticated,stealthy threats.To address these challenges,we propose a novel evolutionary optimized transformer-based deep reinforcement learning framework(Evo-Transformer-DRL)designed for robust and adaptive FDI detection in smart water infrastructures.The proposed architecture integrates three powerful paradigms:a transformer encoder for modeling complex temporal dependencies in multivariate time series,a DRL agent for learning optimal decision policies in dynamic environments,and an evolutionary optimizer to fine-tune model hyper-parameters.This synergy enhances detection performance while maintaining adaptability across varying data distributions.Specifically,hyper-parameters of both the transformer and DRL modules are optimized using an improved grey wolf optimizer(IGWO),ensuring a balanced trade-off between detection accuracy and computational efficiency.The model is trained and evaluated on three realistic Industry 4.0 water datasets:secure water treatment(SWaT),water distribution(WADI),and battle of the attack detection algorithms(BATADAL),which capture diverse attack scenarios in smart treatment and distribution systems.Comparative analysis against state-of-the-art baselines including Transformer,DRL,bidirectional encoder representations from transformers(BERT),convolutional neural network(CNN),long short-term memory(LSTM),and support vector machines(SVM)demonstrates that our proposed Evo-Transformer-DRL framework consistently outperforms others in key metrics such as accuracy,recall,area under the curve(AUC),and execution time.Notably,it achieves a maximum detection accuracy of 99.19%,highlighting its strong generalization capability across different testbeds.These results confirm the suitability of our hybrid framework for real-world Industry 4.0 deployment,where rapid adaptation,scalability,and reliability are paramount for securing critical infrastructure systems.