Vehicle Edge Computing(VEC)and Cloud Computing(CC)significantly enhance the processing efficiency of delay-sensitive and computation-intensive applications by offloading compute-intensive tasks from resource-constrain...Vehicle Edge Computing(VEC)and Cloud Computing(CC)significantly enhance the processing efficiency of delay-sensitive and computation-intensive applications by offloading compute-intensive tasks from resource-constrained onboard devices to nearby Roadside Unit(RSU),thereby achieving lower delay and energy consumption.However,due to the limited storage capacity and energy budget of RSUs,it is challenging to meet the demands of the highly dynamic Internet of Vehicles(IoV)environment.Therefore,determining reasonable service caching and computation offloading strategies is crucial.To address this,this paper proposes a joint service caching scheme for cloud-edge collaborative IoV computation offloading.By modeling the dynamic optimization problem using Markov Decision Processes(MDP),the scheme jointly optimizes task delay,energy consumption,load balancing,and privacy entropy to achieve better quality of service.Additionally,a dynamic adaptive multi-objective deep reinforcement learning algorithm is proposed.Each Double Deep Q-Network(DDQN)agent obtains rewards for different objectives based on distinct reward functions and dynamically updates the objective weights by learning the value changes between objectives using Radial Basis Function Networks(RBFN),thereby efficiently approximating the Pareto-optimal decisions for multiple objectives.Extensive experiments demonstrate that the proposed algorithm can better coordinate the three-tier computing resources of cloud,edge,and vehicles.Compared to existing algorithms,the proposed method reduces task delay and energy consumption by 10.64%and 5.1%,respectively.展开更多
Designing compositions and processing of biodegradable magnesium(Mg)alloys to synergistically en-hance mechanical properties and corrosion resistance using conventional trial-and-error method is a challenging task.Thi...Designing compositions and processing of biodegradable magnesium(Mg)alloys to synergistically en-hance mechanical properties and corrosion resistance using conventional trial-and-error method is a challenging task.This study presents a Bayesian optimization(BO)-based multi-objective framework inte-grated with explainable machine learning(ML)to efficiently explore and optimize the high-dimensional design space of biodegradable Mg alloys.Using ultimate tensile strength(UTS),elongation(EL)and cor-rosion potential(E_(corr))as objective properties,the framework balances these conflicting objectives and identifies optimal solutions.A novel biodegradable Mg alloy(Mg-4.6Zn-0.3Y-0.2Mn-0.1Nd-0.1Gd,wt.%)was successfully designed,demonstrating a UTS of 320 MPa,EL of 22%and E_(corr) of−1.60 V(tested in 37℃ simulated body fluid).Compared to JDBM,the UTS has increased by 13 MPa,the EL has improved by 6.1%,and the E_(corr) has risen by 0.02 V.The experimental results presented close agreement with predicted values,validating the proposed framework.The Shapley Additive Explanation method was em-ployed to interpret the ML models,revealing extrusion temperature and Zn content as key parameters driving the optimization design.The strategy provided in this study is universal and offers a potential approach for addressing high-dimensional multi-objective optimization challenges in material develop-ment.展开更多
Achieving the simultaneous enhancement of strength and ductility in laser powder bed fused (LPBF-ed) titanium (Ti) is challenging due to the complex, high-dimensional parameter space and interactions between parameter...Achieving the simultaneous enhancement of strength and ductility in laser powder bed fused (LPBF-ed) titanium (Ti) is challenging due to the complex, high-dimensional parameter space and interactions between parameters and powders. Herein, a hybrid intelligent framework for process parameter optimization of LPBF-ed Ti with improved ultimate tensile strength (UTS) and elongation (EL) was proposed. It combines the data augmentation method (AVG ± EC × SD), the multi-model fusion stacking ensemble learning model (GBDT-BPNN-XGBoost), the interpretable machine learning method and the non-dominated ranking genetic algorithm (NSGA-Ⅱ). The GBDT-BPNN-XGBoost outperforms single models in predicting UTS and EL across the accuracy, generalization ability and stability. The SHAP analysis reveals that laser power (P) is the most important feature affecting both UTS and EL, and it has a positive impact on them when P < 220 W. The UTS and EL of samples fabricated by the optimal process parameters were 718 ± 5 MPa and 27.9 % ± 0.1 %, respectively. The outstanding strength-ductility balance is attributable to the forward stresses in hard α'-martensite and back stresses in soft αm'-martensite induced by the strain gradients of hetero-microstructure. The back stresses strengthen the soft αm'-martensite, improving the overall UTS. The forward stresses stimulate the activation of dislocations in hard α'-martensite and the generation of 〈c + a〉 dislocations, allowing the plastic strain to occur in hard regions and enhancing the overall ductility. This work provides a feasible strategy for multi-objective optimization and valuable insights into tailoring the microstructure for improving mechanical properties.展开更多
Accurate determination of rock mass parameters is essential for ensuring the accuracy of numericalsimulations. Displacement back-analysis is the most widely used method;however, the reliability of thecurrent approache...Accurate determination of rock mass parameters is essential for ensuring the accuracy of numericalsimulations. Displacement back-analysis is the most widely used method;however, the reliability of thecurrent approaches remains unsatisfactory. Therefore, in this paper, a multistage rock mass parameterback-analysis method, that considers the construction process and displacement losses is proposed andimplemented through the coupling of numerical simulation, auto-machine learning (AutoML), andmulti-objective optimization algorithms (MOOAs). First, a parametric modeling platform for mechanizedtwin tunnels is developed, generating a dataset through extensive numerical simulations. Next, theAutoML method is utilized to establish a surrogate model linking rock parameters and displacements.The tunnel construction process is divided into multiple stages, transforming the rock mass parameterback-analysis into a multi-objective optimization problem, for which multi-objective optimization algorithmsare introduced to obtain the rock mass parameters. The newly proposed rock mass parameterback-analysis method is validated in a mechanized twin tunnel project, and its accuracy and effectivenessare demonstrated. Compared with traditional single-stage back-analysis methods, the proposedmodel decreases the average absolute percentage error from 12.73% to 4.34%, significantly improving theaccuracy of the back-analysis. Moreover, although the accuracy of back analysis significantly increaseswith the number of construction stages considered, the back analysis time is acceptable. This studyprovides a new method for displacement back analysis that is efficient and accurate, thereby paving theway for precise parameter determination in numerical simulations.展开更多
This paper investigates a distributed heterogeneous hybrid blocking flow-shop scheduling problem(DHHBFSP)designed to minimize the total tardiness and total energy consumption simultaneously,and proposes an improved pr...This paper investigates a distributed heterogeneous hybrid blocking flow-shop scheduling problem(DHHBFSP)designed to minimize the total tardiness and total energy consumption simultaneously,and proposes an improved proximal policy optimization(IPPO)method to make real-time decisions for the DHHBFSP.A multi-objective Markov decision process is modeled for the DHHBFSP,where the reward function is represented by a vector with dynamic weights instead of the common objectiverelated scalar value.A factory agent(FA)is formulated for each factory to select unscheduled jobs and is trained by the proposed IPPO to improve the decision quality.Multiple FAs work asynchronously to allocate jobs that arrive randomly at the shop.A two-stage training strategy is introduced in the IPPO,which learns from both single-and dual-policy data for better data utilization.The proposed IPPO is tested on randomly generated instances and compared with variants of the basic proximal policy optimization(PPO),dispatch rules,multi-objective metaheuristics,and multi-agent reinforcement learning methods.Extensive experimental results suggest that the proposed strategies offer significant improvements to the basic PPO,and the proposed IPPO outperforms the state-of-the-art scheduling methods in both convergence and solution quality.展开更多
The biomass and coal co-pyrolysis (BCP) technology combines the advantages of both resources, achieving efficient resource complementarity, reducing reliance on coal, and minimizing pollutant emissions. However, this ...The biomass and coal co-pyrolysis (BCP) technology combines the advantages of both resources, achieving efficient resource complementarity, reducing reliance on coal, and minimizing pollutant emissions. However, this process still encounters numerous challenges in attaining optimal economic and environmental performance. Therefore, an ensemble learning (EL) framework is proposed for the BCP process in this study to optimize the synergistic benefits while minimizing negative environmental impacts. Six different ensemble learning models are developed to investigate the impact of input features, such as biomass characteristics, coal characteristics, and pyrolysis conditions on the product profit and CO_(2) emissions of the BCP processes. The Optuna method is further employed to automatically optimize the hyperparameters of BCP process models for enhancing their predictive accuracy and robustness. The results indicate that the categorical boosting (CAB) model of the BCP process has demonstrated exceptional performance in accurately predicting its product profit and CO_(2) emission (R2>0.92) after undergoing five-fold cross-validation. To enhance the interpretability of this preferred model, the Shapley additive explanations and partial dependence plot analyses are conducted to evaluate the impact and importance of biomass characteristics, coal characteristics, and pyrolysis conditions on the product profitability and CO_(2) emissions of the BCP processes. Finally, the preferred model coupled with a reference vector guided evolutionary algorithm is carried to identify the optimal conditions for maximizing the product profit of BCP process products while minimizing CO_(2) emissions. It indicates the optimal BCP process can achieve high product profits (5290.85 CNY·t−1) and low CO_(2) emissions (7.45 kg·t^(−1)).展开更多
In the parallel steering coordination control strategy for path tracking,it is difficult to match the current driver steering model using the fixed parameters with the actual driver,and the designed steering coordinat...In the parallel steering coordination control strategy for path tracking,it is difficult to match the current driver steering model using the fixed parameters with the actual driver,and the designed steering coordination control strategy under a single objective and simple conditions is difficult to adapt to the multi-dimensional state variables’input.In this paper,we propose a deep reinforcement learning algorithm-based multi-objective parallel human-machine steering coordination strategy for path tracking considering driver misoperation and external disturbance.Firstly,the driver steering mathematical model is constructed based on the driver preview characteristics and steering delay response,and the driver characteristic parameters are fitted after collecting the actual driver driving data.Secondly,considering that the vehicle is susceptible to the influence of external disturbances during the driving process,the Tube MPC(Tube Model Predictive Control)based path tracking steering controller is designed based on the vehicle system dynamics error model.After verifying that the driver steering model meets the driver steering operation characteristics,DQN(Deep Q-network),DDPG(Deep Deterministic Policy Gradient)and TD3(Twin Delayed Deep Deterministic Policy Gradient)deep reinforcement learning algorithms are utilized to design a multi-objective parallel steering coordination strategy which satisfies the multi-dimensional state variables’input of the vehicle.Finally,the tracking accuracy,lateral safety,human-machine conflict and driver steering load evaluation index are designed in different driver operation states and different road environments,and the performance of the parallel steering coordination control strategies with different deep reinforcement learning algorithms and fuzzy algorithms are compared by simulations and hardware in the loop experiments.The results show that the parallel steering collaborative strategy based on a deep reinforcement learning algorithm can more effectively assist the driver in tracking the target path under lateral wind interference and driver misoperation,and the TD3-based coordination control strategy has better overall performance.展开更多
Cyclohexene is an important raw material in the production of nylon.Selective hydrogenation of benzene is a key method for preparing cyclohexene.However,the Ru catalysts used in current industrial processes still face...Cyclohexene is an important raw material in the production of nylon.Selective hydrogenation of benzene is a key method for preparing cyclohexene.However,the Ru catalysts used in current industrial processes still face challenges,including high metal usage,high process costs,and low cyclohexene yield.This study utilizes existing literature data combined with machine learning methods to analyze the factors influencing benzene conversion,cyclohexene selectivity,and yield in the benzene hydrogenation to cyclohexene reaction.It constructs predictive models based on XGBoost and Random Forest algorithms.After analysis,it was found that reaction time,Ru content,and space velocity are key factors influencing cyclohexene yield,selectivity,and benzene conversion.Shapley Additive Explanations(SHAP)analysis and feature importance analysis further revealed the contribution of each variable to the reaction outcomes.Additionally,we randomly generated one million variable combinations using the Dirichlet distribution to attempt to predict high-yield catalyst formulations.This paper provides new insights into the application of machine learning in heterogeneous catalysis and offers some reference for further research.展开更多
The uplift resistance of the soil overlying shield tunnels significantly impacts their anti-floating stability.However,research on uplift resistance concerning special-shaped shield tunnels is limited.This study combi...The uplift resistance of the soil overlying shield tunnels significantly impacts their anti-floating stability.However,research on uplift resistance concerning special-shaped shield tunnels is limited.This study combines numerical simulation with machine learning techniques to explore this issue.It presents a summary of special-shaped tunnel geometries and introduces a shape coefficient.Through the finite element software,Plaxis3D,the study simulates six key parameters—shape coefficient,burial depth ratio,tunnel’s longest horizontal length,internal friction angle,cohesion,and soil submerged bulk density—that impact uplift resistance across different conditions.Employing XGBoost and ANN methods,the feature importance of each parameter was analyzed based on the numerical simulation results.The findings demonstrate that a tunnel shape more closely resembling a circle leads to reduced uplift resistance in the overlying soil,whereas other parameters exhibit the contrary effects.Furthermore,the study reveals a diminishing trend in the feature importance of buried depth ratio,internal friction angle,tunnel longest horizontal length,cohesion,soil submerged bulk density,and shape coefficient in influencing uplift resistance.展开更多
Federated Learning(FL)has become a leading decentralized solution that enables multiple clients to train a model in a collaborative environment without directly sharing raw data,making it suitable for privacy-sensitiv...Federated Learning(FL)has become a leading decentralized solution that enables multiple clients to train a model in a collaborative environment without directly sharing raw data,making it suitable for privacy-sensitive applications such as healthcare,finance,and smart systems.As the field continues to evolve,the research field has become more complex and scattered,covering different system designs,training methods,and privacy techniques.This survey is organized around the three core challenges:how the data is distributed,how models are synchronized,and how to defend against attacks.It provides a structured and up-to-date review of FL research from 2023 to 2025,offering a unified taxonomy that categorizes works by data distribution(Horizontal FL,Vertical FL,Federated Transfer Learning,and Personalized FL),training synchronization(synchronous and asynchronous FL),optimization strategies,and threat models(data leakage and poisoning attacks).In particular,we summarize the latest contributions in Vertical FL frameworks for secure multi-party learning,communication-efficient Horizontal FL,and domain-adaptive Federated Transfer Learning.Furthermore,we examine synchronization techniques addressing system heterogeneity,including straggler mitigation in synchronous FL and staleness management in asynchronous FL.The survey covers security threats in FL,such as gradient inversion,membership inference,and poisoning attacks,as well as their defense strategies that include privacy-preserving aggregation and anomaly detection.The paper concludes by outlining unresolved issues and highlighting challenges in handling personalized models,scalability,and real-world adoption.展开更多
Split Learning(SL)has been promoted as a promising collaborative machine learning technique designed to address data privacy and resource efficiency.Specifically,neural networks are divided into client and server subn...Split Learning(SL)has been promoted as a promising collaborative machine learning technique designed to address data privacy and resource efficiency.Specifically,neural networks are divided into client and server subnetworks in order to mitigate the exposure of sensitive data and reduce the overhead on client devices,thereby making SL particularly suitable for resource-constrained devices.Although SL prevents the direct transmission of raw data,it does not alleviate entirely the risk of privacy breaches.In fact,the data intermediately transmitted to the server sub-model may include patterns or information that could reveal sensitive data.Moreover,achieving a balance between model utility and data privacy has emerged as a challenging problem.In this article,we propose a novel defense approach that combines:(i)Adversarial learning,and(ii)Network channel pruning.In particular,the proposed adversarial learning approach is specifically designed to reduce the risk of private data exposure while maintaining high performance for the utility task.On the other hand,the suggested channel pruning enables the model to adaptively adjust and reactivate pruned channels while conducting adversarial training.The integration of these two techniques reduces the informativeness of the intermediate data transmitted by the client sub-model,thereby enhancing its robustness against attribute inference attacks without adding significant computational overhead,making it wellsuited for IoT devices,mobile platforms,and Internet of Vehicles(IoV)scenarios.The proposed defense approach was evaluated using EfficientNet-B0,a widely adopted compact model,along with three benchmark datasets.The obtained results showcased its superior defense capability against attribute inference attacks compared to existing state-of-the-art methods.This research’s findings demonstrated the effectiveness of the proposed channel pruning-based adversarial training approach in achieving the intended compromise between utility and privacy within SL frameworks.In fact,the classification accuracy attained by the attackers witnessed a drastic decrease of 70%.展开更多
Knowledge distillation has become a standard technique for compressing large language models into efficient student models,but existing methods often struggle to balance prediction accuracy with explanation quality.Re...Knowledge distillation has become a standard technique for compressing large language models into efficient student models,but existing methods often struggle to balance prediction accuracy with explanation quality.Recent approaches such as Distilling Step-by-Step(DSbS)introduce explanation supervision,yet they apply it in a uniform manner that may not fully exploit the different learning dynamics of prediction and explanation.In this work,we propose a task-structured curriculum learning(TSCL)framework that structures training into three sequential phases:(i)prediction-only,to establish stable feature representations;(ii)joint prediction-explanation,to align task outputs with rationale generation;and(iii)explanation-only,to refine the quality of rationales.This design provides a simple but effective modification to DSbS,requiring no architectural changes and adding negligible training cost.We justify the phase scheduling with ablation studies and convergence analysis,showing that an initial prediction-heavy stage followed by a balanced joint phase improves both stability and explanation alignment.Extensive experiments on five datasets(e-SNLI,ANLI,CommonsenseQA,SVAMP,and MedNLI)demonstrate that TSCL consistently outperforms strong baselines,achieving gains of+1.7-2.6 points in accuracy and 0.8-1.2 in ROUGE-L,corresponding to relative error reductions of up to 21%.Beyond lexical metrics,human evaluation and ERASERstyle faithfulness diagnostics confirm that TSCL produces more faithful and informative explanations.Comparative training curves further reveal faster convergence and lower variance across seeds.Efficiency analysis shows less than 3%overhead in wall-clock training time and no additional inference cost,making the approach practical for realworld deployment.This study demonstrates that a simple task-structured curriculum can significantly improve the effectiveness of knowledge distillation.By separating and sequencing objectives,TSCL achieves a better balance between accuracy,stability,and explanation quality.The framework generalizes across domains,including medical NLI,and offers a principled recipe for future applications in multimodal reasoning and reinforcement learning.展开更多
Latest digital advancements have intensified the necessity for adaptive,data-driven and socially-centered learning ecosystems.This paper presents the formulation of a cross-platform,innovative,gamified and personalize...Latest digital advancements have intensified the necessity for adaptive,data-driven and socially-centered learning ecosystems.This paper presents the formulation of a cross-platform,innovative,gamified and personalized Learning Ecosystem,which integrates 3D/VR environments,as well as machine learning algorithms,and business intelligence frameworks to enhance learner-centered education and inferenced decision-making.This Learning System makes use of immersive,analytically assessed virtual learning spaces,therefore facilitating real-time monitoring of not just learning performance,but also overall engagement and behavioral patterns,via a comprehensive set of sustainability-oriented ESG-aligned Key Performance Indicators(KPIs).Machine learning models support predictive analysis,personalized feedback,and hybrid recommendation mechanisms,whilst dedicated dashboards translate complex educational data into actionable insights for all Use Cases of the System(Educational Institutions,Educators and Learners).Additionally,the presented Learning System introduces a structured Mentoring and Consulting Subsystem,thence reinforcing human-centered guidance alongside automated intelligence.The Platform’s modular architecture and simulation-centered evaluation approach actively support personalized,and continuously optimized learning pathways.Thence,it exemplifies a mature,adaptive Learning Ecosystem,supporting immersive technologies,analytics,and pedagogical support,hence,contributing to contemporary digital learning innovation and sociotechnical transformation in education.展开更多
Automated grading of dandruff severity is a clinically significant but challenging task due to the inherent ordinal nature of severity levels and the high prevalence of label noise from subjective expert annotations.S...Automated grading of dandruff severity is a clinically significant but challenging task due to the inherent ordinal nature of severity levels and the high prevalence of label noise from subjective expert annotations.Standard classification methods fail to address these dual challenges,limiting their real-world performance.In this paper,a novel,three-phase training framework is proposed that learns a robust ordinal classifier directly from noisy labels.The approach synergistically combines a rank-based ordinal regression backbone with a cooperative,semi-supervised learning strategy to dynamically partition the data into clean and noisy subsets.A hybrid training objective is then employed,applying a supervised ordinal loss to the clean set.The noisy set is simultaneously trained using a dualobjective that combines a semi-supervised ordinal loss with a parallel,label-agnostic contrastive loss.This design allows themodel to learn fromthe entire noisy subset while using contrastive learning to mitigate the risk of error propagation frompotentially corrupt supervision.Extensive experiments on a new,large-scale,multi-site clinical dataset validate our approach.Themethod achieves state-of-the-art performance with 80.71%accuracy and a 76.86%F1-score,significantly outperforming existing approaches,including a 2.26%improvement over the strongest baseline method.This work provides not only a robust solution for a practical medical imaging problem but also a generalizable framework for other tasks plagued by noisy ordinal labels.展开更多
Adversarial Reinforcement Learning(ARL)models for intelligent devices and Network Intrusion Detection Systems(NIDS)improve systemresilience against sophisticated cyber-attacks.As a core component of ARL,Adversarial Tr...Adversarial Reinforcement Learning(ARL)models for intelligent devices and Network Intrusion Detection Systems(NIDS)improve systemresilience against sophisticated cyber-attacks.As a core component of ARL,Adversarial Training(AT)enables NIDS agents to discover and prevent newattack paths by exposing them to competing examples,thereby increasing detection accuracy,reducing False Positives(FPs),and enhancing network security.To develop robust decision-making capabilities for real-world network disruptions and hostile activity,NIDS agents are trained in adversarial scenarios to monitor the current state and notify management of any abnormal or malicious activity.The accuracy and timeliness of the IDS were crucial to the network’s availability and reliability at this time.This paper analyzes ARL applications in NIDS,revealing State-of-The-Art(SoTA)methodology,issues,and future research prospects.This includes Reinforcement Machine Learning(RML)-based NIDS,which enables an agent to interact with the environment to achieve a goal,andDeep Reinforcement Learning(DRL)-based NIDS,which can solve complex decision-making problems.Additionally,this survey study addresses cybersecurity adversarial circumstances and their importance for ARL and NIDS.Architectural design,RL algorithms,feature representation,and training methodologies are examined in the ARL-NIDS study.This comprehensive study evaluates ARL for intelligent NIDS research,benefiting cybersecurity researchers,practitioners,and policymakers.The report promotes cybersecurity defense research and innovation.展开更多
Task scheduling in cloud computing is a multi-objective optimization problem,often involving conflicting objectives such as minimizing execution time,reducing operational cost,and maximizing resource utilization.Howev...Task scheduling in cloud computing is a multi-objective optimization problem,often involving conflicting objectives such as minimizing execution time,reducing operational cost,and maximizing resource utilization.However,traditional approaches frequently rely on single-objective optimization methods which are insufficient for capturing the complexity of such problems.To address this limitation,we introduce MDMOSA(Multi-objective Dwarf Mongoose Optimization with Simulated Annealing),a hybrid that integrates multi-objective optimization for efficient task scheduling in Infrastructure-as-a-Service(IaaS)cloud environments.MDMOSA harmonizes the exploration capabilities of the biologically inspired Dwarf Mongoose Optimization(DMO)with the exploitation strengths of Simulated Annealing(SA),achieving a balanced search process.The algorithm aims to optimize task allocation by reducing makespan and financial cost while improving system resource utilization.We evaluate MDMOSA through extensive simulations using the real-world Google Cloud Jobs(GoCJ)dataset within the CloudSim environment.Comparative analysis against benchmarked algorithms such as SMOACO,MOTSGWO,and MFPAGWO reveals that MDMOSA consistently achieves superior performance in terms of scheduling efficiency,cost-effectiveness,and scalability.These results confirm the potential of MDMOSA as a robust and adaptable solution for resource scheduling in dynamic and heterogeneous cloud computing infrastructures.展开更多
As urbanization continues to accelerate,the challenges associated with managing transportation in metropolitan areas become increasingly complex.The surge in population density contributes to traffic congestion,impact...As urbanization continues to accelerate,the challenges associated with managing transportation in metropolitan areas become increasingly complex.The surge in population density contributes to traffic congestion,impacting travel experiences and posing safety risks.Smart urban transportation management emerges as a strategic solution,conceptualized here as a multidimensional big data problem.The success of this strategy hinges on the effective collection of information from diverse,extensive,and heterogeneous data sources,necessitating the implementation of full⁃stack Information and Communication Technology(ICT)solutions.The main idea of the work is to investigate the current technologies of Intelligent Transportation Systems(ITS)and enhance the safety of urban transportation systems.Machine learning models,trained on historical data,can predict traffic congestion,allowing for the implementation of preventive measures.Deep learning architectures,with their ability to handle complex data representations,further refine traffic predictions,contributing to more accurate and dynamic transportation management.The background of this research underscores the challenges posed by traffic congestion in metropolitan areas and emphasizes the need for advanced technological solutions.By integrating GPS and GIS technologies with machine learning algorithms,this work aims to pay attention to the development of intelligent transportation systems that not only address current challenges but also pave the way for future advancements in urban transportation management.展开更多
In a wide range of engineering applications,complex constrained multi-objective optimization problems(CMOPs)present significant challenges,as the complexity of constraints often hampers algorithmic convergence and red...In a wide range of engineering applications,complex constrained multi-objective optimization problems(CMOPs)present significant challenges,as the complexity of constraints often hampers algorithmic convergence and reduces population diversity.To address these challenges,we propose a novel algorithm named Constraint IntensityDriven Evolutionary Multitasking(CIDEMT),which employs a two-stage,tri-task framework to dynamically integrates problem structure and knowledge transfer.In the first stage,three cooperative tasks are designed to explore the Constrained Pareto Front(CPF),the Unconstrained Pareto Front(UPF),and theε-relaxed constraint boundary,respectively.A CPF-UPF relationship classifier is employed to construct a problem-type-aware evolutionary strategy pool.At the end of the first stage,each task selects strategies from this strategy pool based on the specific type of problem,thereby guiding the subsequent evolutionary process.In the second stage,while each task continues to evolve,aτ-driven knowledge transfer mechanism is introduced to selectively incorporate effective solutions across tasks.enhancing the convergence and feasibility of the main task.Extensive experiments conducted on 32 benchmark problems from three test suites(LIRCMOP,DASCMOP,and DOC)demonstrate that CIDEMT achieves the best Inverted Generational Distance(IGD)values on 24 problems and the best Hypervolume values(HV)on 22 problems.Furthermore,CIDEMT significantly outperforms six state-of-the-art constrained multi-objective evolutionary algorithms(CMOEAs).These results confirm CIDEMT’s superiority in promoting convergence,diversity,and robustness in solving complex CMOPs.展开更多
Deployable Composite Thin-Walled Structures(DCTWS)are widely used in space applications due to their ability to compactly fold and self-deploy in orbit,enabled by cutouts.Cutout design is crucial for balancing structu...Deployable Composite Thin-Walled Structures(DCTWS)are widely used in space applications due to their ability to compactly fold and self-deploy in orbit,enabled by cutouts.Cutout design is crucial for balancing structural rigidity and flexibility,ensuring material integrity during large deformations,and providing adequate load-bearing capacity and stability once deployed.Most research has focused on optimizing cutout size and shape,while topology optimization offers a broader design space.However,the anisotropic properties of woven composite laminates,complex failure criteria,and multi-performance optimization needs have limited the exploration of topology optimization in this field.This work derives the sensitivities of bending stiffness,critical buckling load,and the failure index of woven composite materials with respect to element density,and formulates both single-objective and multi-objective topology optimization models using a linear weighted aggregation approach.The developed method was integrated with the commercial finite element software ABAQUS via a Python script,allowing efficient application to cutout design in various DCTWS configurations to maximize bending stiffness and critical buckling load under material failure constraints.Optimization of a classical tubular hinge resulted in improvements of 107.7%in bending stiffness and 420.5%in critical buckling load compared to level-set topology optimization results reported in the literature,validating the effectiveness of the approach.To facilitate future research and encourage the broader adoption of topology optimization techniques in DCTWS design,the source code for this work is made publicly available via a Git Hub link:https://github.com/jinhao-ok1/Topo-for-DCTWS.git.展开更多
基金supported by Key Science and Technology Program of Henan Province,China(Grant Nos.242102210147,242102210027)Fujian Province Young and Middle aged Teacher Education Research Project(Science and Technology Category)(No.JZ240101)(Corresponding author:Dong Yuan).
文摘Vehicle Edge Computing(VEC)and Cloud Computing(CC)significantly enhance the processing efficiency of delay-sensitive and computation-intensive applications by offloading compute-intensive tasks from resource-constrained onboard devices to nearby Roadside Unit(RSU),thereby achieving lower delay and energy consumption.However,due to the limited storage capacity and energy budget of RSUs,it is challenging to meet the demands of the highly dynamic Internet of Vehicles(IoV)environment.Therefore,determining reasonable service caching and computation offloading strategies is crucial.To address this,this paper proposes a joint service caching scheme for cloud-edge collaborative IoV computation offloading.By modeling the dynamic optimization problem using Markov Decision Processes(MDP),the scheme jointly optimizes task delay,energy consumption,load balancing,and privacy entropy to achieve better quality of service.Additionally,a dynamic adaptive multi-objective deep reinforcement learning algorithm is proposed.Each Double Deep Q-Network(DDQN)agent obtains rewards for different objectives based on distinct reward functions and dynamically updates the objective weights by learning the value changes between objectives using Radial Basis Function Networks(RBFN),thereby efficiently approximating the Pareto-optimal decisions for multiple objectives.Extensive experiments demonstrate that the proposed algorithm can better coordinate the three-tier computing resources of cloud,edge,and vehicles.Compared to existing algorithms,the proposed method reduces task delay and energy consumption by 10.64%and 5.1%,respectively.
基金financially supported by the National Natu-ral Science Foundation of China(No.52301133)the China Post-doctoral Science Foundation(No.2023M730276)+1 种基金the Young Elite Scientists Sponsorship Program by China Association for Science and Technology(No.YESS20210415)the Graduate Innovation Pro-gram of Chongqing University of Science and Technology(No.YKJCX2320218).
文摘Designing compositions and processing of biodegradable magnesium(Mg)alloys to synergistically en-hance mechanical properties and corrosion resistance using conventional trial-and-error method is a challenging task.This study presents a Bayesian optimization(BO)-based multi-objective framework inte-grated with explainable machine learning(ML)to efficiently explore and optimize the high-dimensional design space of biodegradable Mg alloys.Using ultimate tensile strength(UTS),elongation(EL)and cor-rosion potential(E_(corr))as objective properties,the framework balances these conflicting objectives and identifies optimal solutions.A novel biodegradable Mg alloy(Mg-4.6Zn-0.3Y-0.2Mn-0.1Nd-0.1Gd,wt.%)was successfully designed,demonstrating a UTS of 320 MPa,EL of 22%and E_(corr) of−1.60 V(tested in 37℃ simulated body fluid).Compared to JDBM,the UTS has increased by 13 MPa,the EL has improved by 6.1%,and the E_(corr) has risen by 0.02 V.The experimental results presented close agreement with predicted values,validating the proposed framework.The Shapley Additive Explanation method was em-ployed to interpret the ML models,revealing extrusion temperature and Zn content as key parameters driving the optimization design.The strategy provided in this study is universal and offers a potential approach for addressing high-dimensional multi-objective optimization challenges in material develop-ment.
基金supported by the National Natural Sci-ence Foundation of China(Nos.52274359 and 52304379)the China National Postdoctoral Program for Innovative Talents(No.BX20220034)+2 种基金the China Postdoctoral Science Foundation(No.2022M720403)the AECC University Research Cooperation Project(No.HFZL2021CXY021)the Interdisciplinary Research Project for Young Teachers of USTB(Fundamental Research Funds for the Central Universities)(No.FRF-IDRY-23-025).
文摘Achieving the simultaneous enhancement of strength and ductility in laser powder bed fused (LPBF-ed) titanium (Ti) is challenging due to the complex, high-dimensional parameter space and interactions between parameters and powders. Herein, a hybrid intelligent framework for process parameter optimization of LPBF-ed Ti with improved ultimate tensile strength (UTS) and elongation (EL) was proposed. It combines the data augmentation method (AVG ± EC × SD), the multi-model fusion stacking ensemble learning model (GBDT-BPNN-XGBoost), the interpretable machine learning method and the non-dominated ranking genetic algorithm (NSGA-Ⅱ). The GBDT-BPNN-XGBoost outperforms single models in predicting UTS and EL across the accuracy, generalization ability and stability. The SHAP analysis reveals that laser power (P) is the most important feature affecting both UTS and EL, and it has a positive impact on them when P < 220 W. The UTS and EL of samples fabricated by the optimal process parameters were 718 ± 5 MPa and 27.9 % ± 0.1 %, respectively. The outstanding strength-ductility balance is attributable to the forward stresses in hard α'-martensite and back stresses in soft αm'-martensite induced by the strain gradients of hetero-microstructure. The back stresses strengthen the soft αm'-martensite, improving the overall UTS. The forward stresses stimulate the activation of dislocations in hard α'-martensite and the generation of 〈c + a〉 dislocations, allowing the plastic strain to occur in hard regions and enhancing the overall ductility. This work provides a feasible strategy for multi-objective optimization and valuable insights into tailoring the microstructure for improving mechanical properties.
基金supported by the National Natural Science Foundation of China(Grant Nos.52090081,52079068)the State Key Laboratory of Hydroscience and Hydraulic Engineering(Grant No.2021-KY-04).
文摘Accurate determination of rock mass parameters is essential for ensuring the accuracy of numericalsimulations. Displacement back-analysis is the most widely used method;however, the reliability of thecurrent approaches remains unsatisfactory. Therefore, in this paper, a multistage rock mass parameterback-analysis method, that considers the construction process and displacement losses is proposed andimplemented through the coupling of numerical simulation, auto-machine learning (AutoML), andmulti-objective optimization algorithms (MOOAs). First, a parametric modeling platform for mechanizedtwin tunnels is developed, generating a dataset through extensive numerical simulations. Next, theAutoML method is utilized to establish a surrogate model linking rock parameters and displacements.The tunnel construction process is divided into multiple stages, transforming the rock mass parameterback-analysis into a multi-objective optimization problem, for which multi-objective optimization algorithmsare introduced to obtain the rock mass parameters. The newly proposed rock mass parameterback-analysis method is validated in a mechanized twin tunnel project, and its accuracy and effectivenessare demonstrated. Compared with traditional single-stage back-analysis methods, the proposedmodel decreases the average absolute percentage error from 12.73% to 4.34%, significantly improving theaccuracy of the back-analysis. Moreover, although the accuracy of back analysis significantly increaseswith the number of construction stages considered, the back analysis time is acceptable. This studyprovides a new method for displacement back analysis that is efficient and accurate, thereby paving theway for precise parameter determination in numerical simulations.
基金partially supported by the National Key Research and Development Program of the Ministry of Science and Technology of China(2022YFE0114200)the National Natural Science Foundation of China(U20A6004).
文摘This paper investigates a distributed heterogeneous hybrid blocking flow-shop scheduling problem(DHHBFSP)designed to minimize the total tardiness and total energy consumption simultaneously,and proposes an improved proximal policy optimization(IPPO)method to make real-time decisions for the DHHBFSP.A multi-objective Markov decision process is modeled for the DHHBFSP,where the reward function is represented by a vector with dynamic weights instead of the common objectiverelated scalar value.A factory agent(FA)is formulated for each factory to select unscheduled jobs and is trained by the proposed IPPO to improve the decision quality.Multiple FAs work asynchronously to allocate jobs that arrive randomly at the shop.A two-stage training strategy is introduced in the IPPO,which learns from both single-and dual-policy data for better data utilization.The proposed IPPO is tested on randomly generated instances and compared with variants of the basic proximal policy optimization(PPO),dispatch rules,multi-objective metaheuristics,and multi-agent reinforcement learning methods.Extensive experimental results suggest that the proposed strategies offer significant improvements to the basic PPO,and the proposed IPPO outperforms the state-of-the-art scheduling methods in both convergence and solution quality.
基金support from the National Natural Science Foundation of China(22108052).
文摘The biomass and coal co-pyrolysis (BCP) technology combines the advantages of both resources, achieving efficient resource complementarity, reducing reliance on coal, and minimizing pollutant emissions. However, this process still encounters numerous challenges in attaining optimal economic and environmental performance. Therefore, an ensemble learning (EL) framework is proposed for the BCP process in this study to optimize the synergistic benefits while minimizing negative environmental impacts. Six different ensemble learning models are developed to investigate the impact of input features, such as biomass characteristics, coal characteristics, and pyrolysis conditions on the product profit and CO_(2) emissions of the BCP processes. The Optuna method is further employed to automatically optimize the hyperparameters of BCP process models for enhancing their predictive accuracy and robustness. The results indicate that the categorical boosting (CAB) model of the BCP process has demonstrated exceptional performance in accurately predicting its product profit and CO_(2) emission (R2>0.92) after undergoing five-fold cross-validation. To enhance the interpretability of this preferred model, the Shapley additive explanations and partial dependence plot analyses are conducted to evaluate the impact and importance of biomass characteristics, coal characteristics, and pyrolysis conditions on the product profitability and CO_(2) emissions of the BCP processes. Finally, the preferred model coupled with a reference vector guided evolutionary algorithm is carried to identify the optimal conditions for maximizing the product profit of BCP process products while minimizing CO_(2) emissions. It indicates the optimal BCP process can achieve high product profits (5290.85 CNY·t−1) and low CO_(2) emissions (7.45 kg·t^(−1)).
基金Supported by National Natural Science Foundation of China(Grant Nos.U22A20246,52372382)Hefei Municipal Natural Science Foundation(Grant No.2022008)+1 种基金the Open Fund of State Key Laboratory of Mechanical Behavior and System Safety of Traffic Engineering Structures(Grant No.KF2023-06)S&T Program of Hebei(Grant No.225676162GH).
文摘In the parallel steering coordination control strategy for path tracking,it is difficult to match the current driver steering model using the fixed parameters with the actual driver,and the designed steering coordination control strategy under a single objective and simple conditions is difficult to adapt to the multi-dimensional state variables’input.In this paper,we propose a deep reinforcement learning algorithm-based multi-objective parallel human-machine steering coordination strategy for path tracking considering driver misoperation and external disturbance.Firstly,the driver steering mathematical model is constructed based on the driver preview characteristics and steering delay response,and the driver characteristic parameters are fitted after collecting the actual driver driving data.Secondly,considering that the vehicle is susceptible to the influence of external disturbances during the driving process,the Tube MPC(Tube Model Predictive Control)based path tracking steering controller is designed based on the vehicle system dynamics error model.After verifying that the driver steering model meets the driver steering operation characteristics,DQN(Deep Q-network),DDPG(Deep Deterministic Policy Gradient)and TD3(Twin Delayed Deep Deterministic Policy Gradient)deep reinforcement learning algorithms are utilized to design a multi-objective parallel steering coordination strategy which satisfies the multi-dimensional state variables’input of the vehicle.Finally,the tracking accuracy,lateral safety,human-machine conflict and driver steering load evaluation index are designed in different driver operation states and different road environments,and the performance of the parallel steering coordination control strategies with different deep reinforcement learning algorithms and fuzzy algorithms are compared by simulations and hardware in the loop experiments.The results show that the parallel steering collaborative strategy based on a deep reinforcement learning algorithm can more effectively assist the driver in tracking the target path under lateral wind interference and driver misoperation,and the TD3-based coordination control strategy has better overall performance.
基金Supported by CAS Basic and Interdisciplinary Frontier Scientific Research Pilot Project(XDB1190300,XDB1190302)Youth Innovation Promotion Association CAS(Y2021056)+1 种基金Joint Fund of the Yulin University and the Dalian National Laboratory for Clean Energy(YLU-DNL Fund 2022007)The special fund for Science and Technology Innovation Teams of Shanxi Province(202304051001007)。
文摘Cyclohexene is an important raw material in the production of nylon.Selective hydrogenation of benzene is a key method for preparing cyclohexene.However,the Ru catalysts used in current industrial processes still face challenges,including high metal usage,high process costs,and low cyclohexene yield.This study utilizes existing literature data combined with machine learning methods to analyze the factors influencing benzene conversion,cyclohexene selectivity,and yield in the benzene hydrogenation to cyclohexene reaction.It constructs predictive models based on XGBoost and Random Forest algorithms.After analysis,it was found that reaction time,Ru content,and space velocity are key factors influencing cyclohexene yield,selectivity,and benzene conversion.Shapley Additive Explanations(SHAP)analysis and feature importance analysis further revealed the contribution of each variable to the reaction outcomes.Additionally,we randomly generated one million variable combinations using the Dirichlet distribution to attempt to predict high-yield catalyst formulations.This paper provides new insights into the application of machine learning in heterogeneous catalysis and offers some reference for further research.
基金Guangzhou Metro Scientific Research Project(No.JT204-100111-23001)Chongqing Municipal Special Project for Technological Innovation and Application Development(No.CSTB2022TIAD-KPX0101)Science and Technology Research and Development Program of China State Railway Group Co.,Ltd.(No.N2023G045)。
文摘The uplift resistance of the soil overlying shield tunnels significantly impacts their anti-floating stability.However,research on uplift resistance concerning special-shaped shield tunnels is limited.This study combines numerical simulation with machine learning techniques to explore this issue.It presents a summary of special-shaped tunnel geometries and introduces a shape coefficient.Through the finite element software,Plaxis3D,the study simulates six key parameters—shape coefficient,burial depth ratio,tunnel’s longest horizontal length,internal friction angle,cohesion,and soil submerged bulk density—that impact uplift resistance across different conditions.Employing XGBoost and ANN methods,the feature importance of each parameter was analyzed based on the numerical simulation results.The findings demonstrate that a tunnel shape more closely resembling a circle leads to reduced uplift resistance in the overlying soil,whereas other parameters exhibit the contrary effects.Furthermore,the study reveals a diminishing trend in the feature importance of buried depth ratio,internal friction angle,tunnel longest horizontal length,cohesion,soil submerged bulk density,and shape coefficient in influencing uplift resistance.
文摘Federated Learning(FL)has become a leading decentralized solution that enables multiple clients to train a model in a collaborative environment without directly sharing raw data,making it suitable for privacy-sensitive applications such as healthcare,finance,and smart systems.As the field continues to evolve,the research field has become more complex and scattered,covering different system designs,training methods,and privacy techniques.This survey is organized around the three core challenges:how the data is distributed,how models are synchronized,and how to defend against attacks.It provides a structured and up-to-date review of FL research from 2023 to 2025,offering a unified taxonomy that categorizes works by data distribution(Horizontal FL,Vertical FL,Federated Transfer Learning,and Personalized FL),training synchronization(synchronous and asynchronous FL),optimization strategies,and threat models(data leakage and poisoning attacks).In particular,we summarize the latest contributions in Vertical FL frameworks for secure multi-party learning,communication-efficient Horizontal FL,and domain-adaptive Federated Transfer Learning.Furthermore,we examine synchronization techniques addressing system heterogeneity,including straggler mitigation in synchronous FL and staleness management in asynchronous FL.The survey covers security threats in FL,such as gradient inversion,membership inference,and poisoning attacks,as well as their defense strategies that include privacy-preserving aggregation and anomaly detection.The paper concludes by outlining unresolved issues and highlighting challenges in handling personalized models,scalability,and real-world adoption.
基金supported by a grant(No.CRPG-25-2054)under the Cybersecurity Research and Innovation Pioneers Initiative,provided by the National Cybersecurity Authority(NCA)in the Kingdom of Saudi Arabia.
文摘Split Learning(SL)has been promoted as a promising collaborative machine learning technique designed to address data privacy and resource efficiency.Specifically,neural networks are divided into client and server subnetworks in order to mitigate the exposure of sensitive data and reduce the overhead on client devices,thereby making SL particularly suitable for resource-constrained devices.Although SL prevents the direct transmission of raw data,it does not alleviate entirely the risk of privacy breaches.In fact,the data intermediately transmitted to the server sub-model may include patterns or information that could reveal sensitive data.Moreover,achieving a balance between model utility and data privacy has emerged as a challenging problem.In this article,we propose a novel defense approach that combines:(i)Adversarial learning,and(ii)Network channel pruning.In particular,the proposed adversarial learning approach is specifically designed to reduce the risk of private data exposure while maintaining high performance for the utility task.On the other hand,the suggested channel pruning enables the model to adaptively adjust and reactivate pruned channels while conducting adversarial training.The integration of these two techniques reduces the informativeness of the intermediate data transmitted by the client sub-model,thereby enhancing its robustness against attribute inference attacks without adding significant computational overhead,making it wellsuited for IoT devices,mobile platforms,and Internet of Vehicles(IoV)scenarios.The proposed defense approach was evaluated using EfficientNet-B0,a widely adopted compact model,along with three benchmark datasets.The obtained results showcased its superior defense capability against attribute inference attacks compared to existing state-of-the-art methods.This research’s findings demonstrated the effectiveness of the proposed channel pruning-based adversarial training approach in achieving the intended compromise between utility and privacy within SL frameworks.In fact,the classification accuracy attained by the attackers witnessed a drastic decrease of 70%.
文摘Knowledge distillation has become a standard technique for compressing large language models into efficient student models,but existing methods often struggle to balance prediction accuracy with explanation quality.Recent approaches such as Distilling Step-by-Step(DSbS)introduce explanation supervision,yet they apply it in a uniform manner that may not fully exploit the different learning dynamics of prediction and explanation.In this work,we propose a task-structured curriculum learning(TSCL)framework that structures training into three sequential phases:(i)prediction-only,to establish stable feature representations;(ii)joint prediction-explanation,to align task outputs with rationale generation;and(iii)explanation-only,to refine the quality of rationales.This design provides a simple but effective modification to DSbS,requiring no architectural changes and adding negligible training cost.We justify the phase scheduling with ablation studies and convergence analysis,showing that an initial prediction-heavy stage followed by a balanced joint phase improves both stability and explanation alignment.Extensive experiments on five datasets(e-SNLI,ANLI,CommonsenseQA,SVAMP,and MedNLI)demonstrate that TSCL consistently outperforms strong baselines,achieving gains of+1.7-2.6 points in accuracy and 0.8-1.2 in ROUGE-L,corresponding to relative error reductions of up to 21%.Beyond lexical metrics,human evaluation and ERASERstyle faithfulness diagnostics confirm that TSCL produces more faithful and informative explanations.Comparative training curves further reveal faster convergence and lower variance across seeds.Efficiency analysis shows less than 3%overhead in wall-clock training time and no additional inference cost,making the approach practical for realworld deployment.This study demonstrates that a simple task-structured curriculum can significantly improve the effectiveness of knowledge distillation.By separating and sequencing objectives,TSCL achieves a better balance between accuracy,stability,and explanation quality.The framework generalizes across domains,including medical NLI,and offers a principled recipe for future applications in multimodal reasoning and reinforcement learning.
文摘Latest digital advancements have intensified the necessity for adaptive,data-driven and socially-centered learning ecosystems.This paper presents the formulation of a cross-platform,innovative,gamified and personalized Learning Ecosystem,which integrates 3D/VR environments,as well as machine learning algorithms,and business intelligence frameworks to enhance learner-centered education and inferenced decision-making.This Learning System makes use of immersive,analytically assessed virtual learning spaces,therefore facilitating real-time monitoring of not just learning performance,but also overall engagement and behavioral patterns,via a comprehensive set of sustainability-oriented ESG-aligned Key Performance Indicators(KPIs).Machine learning models support predictive analysis,personalized feedback,and hybrid recommendation mechanisms,whilst dedicated dashboards translate complex educational data into actionable insights for all Use Cases of the System(Educational Institutions,Educators and Learners).Additionally,the presented Learning System introduces a structured Mentoring and Consulting Subsystem,thence reinforcing human-centered guidance alongside automated intelligence.The Platform’s modular architecture and simulation-centered evaluation approach actively support personalized,and continuously optimized learning pathways.Thence,it exemplifies a mature,adaptive Learning Ecosystem,supporting immersive technologies,analytics,and pedagogical support,hence,contributing to contemporary digital learning innovation and sociotechnical transformation in education.
文摘Automated grading of dandruff severity is a clinically significant but challenging task due to the inherent ordinal nature of severity levels and the high prevalence of label noise from subjective expert annotations.Standard classification methods fail to address these dual challenges,limiting their real-world performance.In this paper,a novel,three-phase training framework is proposed that learns a robust ordinal classifier directly from noisy labels.The approach synergistically combines a rank-based ordinal regression backbone with a cooperative,semi-supervised learning strategy to dynamically partition the data into clean and noisy subsets.A hybrid training objective is then employed,applying a supervised ordinal loss to the clean set.The noisy set is simultaneously trained using a dualobjective that combines a semi-supervised ordinal loss with a parallel,label-agnostic contrastive loss.This design allows themodel to learn fromthe entire noisy subset while using contrastive learning to mitigate the risk of error propagation frompotentially corrupt supervision.Extensive experiments on a new,large-scale,multi-site clinical dataset validate our approach.Themethod achieves state-of-the-art performance with 80.71%accuracy and a 76.86%F1-score,significantly outperforming existing approaches,including a 2.26%improvement over the strongest baseline method.This work provides not only a robust solution for a practical medical imaging problem but also a generalizable framework for other tasks plagued by noisy ordinal labels.
文摘Adversarial Reinforcement Learning(ARL)models for intelligent devices and Network Intrusion Detection Systems(NIDS)improve systemresilience against sophisticated cyber-attacks.As a core component of ARL,Adversarial Training(AT)enables NIDS agents to discover and prevent newattack paths by exposing them to competing examples,thereby increasing detection accuracy,reducing False Positives(FPs),and enhancing network security.To develop robust decision-making capabilities for real-world network disruptions and hostile activity,NIDS agents are trained in adversarial scenarios to monitor the current state and notify management of any abnormal or malicious activity.The accuracy and timeliness of the IDS were crucial to the network’s availability and reliability at this time.This paper analyzes ARL applications in NIDS,revealing State-of-The-Art(SoTA)methodology,issues,and future research prospects.This includes Reinforcement Machine Learning(RML)-based NIDS,which enables an agent to interact with the environment to achieve a goal,andDeep Reinforcement Learning(DRL)-based NIDS,which can solve complex decision-making problems.Additionally,this survey study addresses cybersecurity adversarial circumstances and their importance for ARL and NIDS.Architectural design,RL algorithms,feature representation,and training methodologies are examined in the ARL-NIDS study.This comprehensive study evaluates ARL for intelligent NIDS research,benefiting cybersecurity researchers,practitioners,and policymakers.The report promotes cybersecurity defense research and innovation.
文摘Task scheduling in cloud computing is a multi-objective optimization problem,often involving conflicting objectives such as minimizing execution time,reducing operational cost,and maximizing resource utilization.However,traditional approaches frequently rely on single-objective optimization methods which are insufficient for capturing the complexity of such problems.To address this limitation,we introduce MDMOSA(Multi-objective Dwarf Mongoose Optimization with Simulated Annealing),a hybrid that integrates multi-objective optimization for efficient task scheduling in Infrastructure-as-a-Service(IaaS)cloud environments.MDMOSA harmonizes the exploration capabilities of the biologically inspired Dwarf Mongoose Optimization(DMO)with the exploitation strengths of Simulated Annealing(SA),achieving a balanced search process.The algorithm aims to optimize task allocation by reducing makespan and financial cost while improving system resource utilization.We evaluate MDMOSA through extensive simulations using the real-world Google Cloud Jobs(GoCJ)dataset within the CloudSim environment.Comparative analysis against benchmarked algorithms such as SMOACO,MOTSGWO,and MFPAGWO reveals that MDMOSA consistently achieves superior performance in terms of scheduling efficiency,cost-effectiveness,and scalability.These results confirm the potential of MDMOSA as a robust and adaptable solution for resource scheduling in dynamic and heterogeneous cloud computing infrastructures.
文摘As urbanization continues to accelerate,the challenges associated with managing transportation in metropolitan areas become increasingly complex.The surge in population density contributes to traffic congestion,impacting travel experiences and posing safety risks.Smart urban transportation management emerges as a strategic solution,conceptualized here as a multidimensional big data problem.The success of this strategy hinges on the effective collection of information from diverse,extensive,and heterogeneous data sources,necessitating the implementation of full⁃stack Information and Communication Technology(ICT)solutions.The main idea of the work is to investigate the current technologies of Intelligent Transportation Systems(ITS)and enhance the safety of urban transportation systems.Machine learning models,trained on historical data,can predict traffic congestion,allowing for the implementation of preventive measures.Deep learning architectures,with their ability to handle complex data representations,further refine traffic predictions,contributing to more accurate and dynamic transportation management.The background of this research underscores the challenges posed by traffic congestion in metropolitan areas and emphasizes the need for advanced technological solutions.By integrating GPS and GIS technologies with machine learning algorithms,this work aims to pay attention to the development of intelligent transportation systems that not only address current challenges but also pave the way for future advancements in urban transportation management.
基金supported by the National Natural Science Foundation of China under Grant No.61972040the Science and Technology Research and Development Project funded by China Railway Material Trade Group Luban Company.
文摘In a wide range of engineering applications,complex constrained multi-objective optimization problems(CMOPs)present significant challenges,as the complexity of constraints often hampers algorithmic convergence and reduces population diversity.To address these challenges,we propose a novel algorithm named Constraint IntensityDriven Evolutionary Multitasking(CIDEMT),which employs a two-stage,tri-task framework to dynamically integrates problem structure and knowledge transfer.In the first stage,three cooperative tasks are designed to explore the Constrained Pareto Front(CPF),the Unconstrained Pareto Front(UPF),and theε-relaxed constraint boundary,respectively.A CPF-UPF relationship classifier is employed to construct a problem-type-aware evolutionary strategy pool.At the end of the first stage,each task selects strategies from this strategy pool based on the specific type of problem,thereby guiding the subsequent evolutionary process.In the second stage,while each task continues to evolve,aτ-driven knowledge transfer mechanism is introduced to selectively incorporate effective solutions across tasks.enhancing the convergence and feasibility of the main task.Extensive experiments conducted on 32 benchmark problems from three test suites(LIRCMOP,DASCMOP,and DOC)demonstrate that CIDEMT achieves the best Inverted Generational Distance(IGD)values on 24 problems and the best Hypervolume values(HV)on 22 problems.Furthermore,CIDEMT significantly outperforms six state-of-the-art constrained multi-objective evolutionary algorithms(CMOEAs).These results confirm CIDEMT’s superiority in promoting convergence,diversity,and robustness in solving complex CMOPs.
基金supported by the National Natural Science Foundation of China(No.12202295)the International(Regional)Cooperation and Exchange Projects of the National Natural Science Foundation of China(No.W2421002)+2 种基金the Sichuan Science and Technology Program(No.2025ZNSFSC0845)Zhejiang Provincial Natural Science Foundation of China(No.ZCLZ24A0201)the Fundamental Research Funds for the Provincial Universities of Zhejiang(No.GK249909299001-004)。
文摘Deployable Composite Thin-Walled Structures(DCTWS)are widely used in space applications due to their ability to compactly fold and self-deploy in orbit,enabled by cutouts.Cutout design is crucial for balancing structural rigidity and flexibility,ensuring material integrity during large deformations,and providing adequate load-bearing capacity and stability once deployed.Most research has focused on optimizing cutout size and shape,while topology optimization offers a broader design space.However,the anisotropic properties of woven composite laminates,complex failure criteria,and multi-performance optimization needs have limited the exploration of topology optimization in this field.This work derives the sensitivities of bending stiffness,critical buckling load,and the failure index of woven composite materials with respect to element density,and formulates both single-objective and multi-objective topology optimization models using a linear weighted aggregation approach.The developed method was integrated with the commercial finite element software ABAQUS via a Python script,allowing efficient application to cutout design in various DCTWS configurations to maximize bending stiffness and critical buckling load under material failure constraints.Optimization of a classical tubular hinge resulted in improvements of 107.7%in bending stiffness and 420.5%in critical buckling load compared to level-set topology optimization results reported in the literature,validating the effectiveness of the approach.To facilitate future research and encourage the broader adoption of topology optimization techniques in DCTWS design,the source code for this work is made publicly available via a Git Hub link:https://github.com/jinhao-ok1/Topo-for-DCTWS.git.