Tropospheric zenith wet delay(ZWD)plays a vital role in the analysis of space geodetic observations.In recent years,machine learning methods have been increasingly applied to improve the accuracy of ZWD calculations.H...Tropospheric zenith wet delay(ZWD)plays a vital role in the analysis of space geodetic observations.In recent years,machine learning methods have been increasingly applied to improve the accuracy of ZWD calculations.However,a single machine learning model has limited generalization capabilities.To address these limitations,this study introduces a novel machine learning fusion(MLF)algorithm with stronger generalization capabilities to enhance ZWD modeling and prediction accuracy.The MLF algorithm utilizes a two-layer structure integrating extra trees(ET),backpropagation neural network(BPNN),and linear regression models.By comparing the root mean square error(RMSE)of these models,we found that both ET-based and MLF-based models outperform RF-based and BPNN-based models in terms of internal and external accuracy,across both surface meteorological data-based and blind models.The improvement in exte rnal accuracy is particularly significant in the blind models.Our re sults show that the MLF(with an RMSE of 3.93 cm)and ET(3.99 cm)models outperform the traditional GPT3model(4.07 cm),while the RF(4.21 cm)and BPNN(4.14 cm)have worse external accuracies than the GPT3 model.It is worth noting that the BPNN suffered from overfitting during external accuracy tests,which was avoided by the MLF.In summary,regardless of the availability of surface meteorological data,the MLF-based empirical models demonstrate superior internal and external accuracy compared to the other tested models in this study.展开更多
BACKGROUND:This study aims to develop and validate a machine learning-based in-hospital mortality predictive model for acute aortic syndrome(AAS)in the emergency department(ED)and to derive a simplifi ed version suita...BACKGROUND:This study aims to develop and validate a machine learning-based in-hospital mortality predictive model for acute aortic syndrome(AAS)in the emergency department(ED)and to derive a simplifi ed version suitable for rapid clinical application.METHODS:In this multi-center retrospective cohort study,AAS patient data from three hospitals were analyzed.The modeling cohort included data from the First Affiliated Hospital of Zhengzhou University and the People’s Hospital of Xinjiang Uygur Autonomous Region,with Peking University Third Hospital data serving as the external test set.Four machine learning algorithms—logistic regression(LR),multilayer perceptron(MLP),Gaussian naive Bayes(GNB),and random forest(RF)—were used to develop predictive models based on 34 early-accessible clinical variables.A simplifi ed model was then derived based on fi ve key variables(Stanford type,pericardial eff usion,asymmetric peripheral arterial pulsation,decreased bowel sounds,and dyspnea)via Least Absolute Shrinkage and Selection Operator(LASSO)regression to improve ED applicability.RESULTS:A total of 929 patients were included in the modeling cohort,and 210 were included in the external test set.Four machine learning models based on 34 clinical variables were developed,achieving internal and external validation AUCs of 0.85-0.90 and 0.73-0.85,respectively.The simplifi ed model incorporating fi ve key variables demonstrated internal and external validation AUCs of 0.71-0.86 and 0.75-0.78,respectively.Both models showed robust calibration and predictive stability across datasets.CONCLUSION:Both kinds of models were built based on machine learning tools,and proved to have certain prediction performance and extrapolation.展开更多
The rapid advancement of machine learning based tight-binding Hamiltonian(MLTB)methods has opened new avenues for efficient and accurate electronic structure simulations,particularly in large-scale systems and long-ti...The rapid advancement of machine learning based tight-binding Hamiltonian(MLTB)methods has opened new avenues for efficient and accurate electronic structure simulations,particularly in large-scale systems and long-time scenarios.This review begins with a concise overview of traditional tight-binding(TB)models,including both(semi-)empirical and first-principles approaches,establishing the foundation for understanding MLTB developments.We then present a systematic classification of existing MLTB methodologies,grouped into two major categories:direct prediction of TB Hamiltonian elements and inference of empirical parameters.A comparative analysis with other ML-based electronic structure models is also provided,highlighting the advancement of MLTB approaches.Finally,we explore the emerging MLTB application ecosystem,highlighting how the integration of MLTB models with a diverse suite of post-processing tools from linear-scaling solvers to quantum transport frameworks and molecular dynamics interfaces is essential for tackling complex scientific problems across different domains.The continued advancement of this integrated paradigm promises to accelerate materials discovery and open new frontiers in the predictive simulation of complex quantum phenomena.展开更多
This study aims to develop an accurate and robust machine learning model to predict the carbonation depth of fly ash concrete,overcoming the limitations of traditional predictive methods.Five ensemble-based models,suc...This study aims to develop an accurate and robust machine learning model to predict the carbonation depth of fly ash concrete,overcoming the limitations of traditional predictive methods.Five ensemble-based models,such as adaptive boosting(AdaBoost),categorical boosting(CatBoost),gradient boosting regressor(GBR),hist gradient boosting regressor(HistGBR),and extreme gradient boosting(XGBoost),were developed and optimized using 729 high-quality dataset points incorporating seven input parameters,including cement,CO_(2),exposure time,water-binder ratio,fly ash,curing time,and compressive strength.Several performance evaluation metrics were used to compare the models.The GBR model emerged as the best-performing model,based on high coefficient of determination(R^(2))values and balanced error metrics across both validation and testing datasets.While all models performed exceptionally well on the training data,GBR demonstrated superior generalization capability,with R^(2) values of 0.9438 on the validation set and 0.9310 on the testing set.Furthermore,its low mean squared error(MSE),root mean square error(RMSE),mean absolute error(MAE),and median absolute error(MdAE)confirmed its robustness and accuracy.Moreover,shapley additive explanations(SHAP)analysis enhanced the interpretability of predictions,highlighting the curing time and exposure time as the most critical drivers of carbonation depth.展开更多
Oxide dispersion strengthened(ODS)alloys are extensively used owing to high thermostability and creep strength contributed from uniformly dispersed fine oxides particles.However,the existence of these strengthening pa...Oxide dispersion strengthened(ODS)alloys are extensively used owing to high thermostability and creep strength contributed from uniformly dispersed fine oxides particles.However,the existence of these strengthening particles also deteriorates the processability and it is of great importance to establish accurate processing maps to guide the thermomechanical processes to enhance the formability.In this study,we performed particle swarm optimization-based back propagation artificial neural network model to predict the high temperature flow behavior of 0.25wt%Al2O3 particle-reinforced Cu alloys,and compared the accuracy with that of derived by Arrhenius-type constitutive model and back propagation artificial neural network model.To train these models,we obtained the raw data by fabricating ODS Cu alloys using the internal oxidation and reduction method,and conducting systematic hot compression tests between 400 and800℃with strain rates of 10^(-2)-10 S^(-1).At last,processing maps for ODS Cu alloys were proposed by combining processing parameters,mechanical behavior,microstructure characterization,and the modeling results achieved a coefficient of determination higher than>99%.展开更多
Background Cotton is one of the most important commercial crops after food crops,especially in countries like India,where it’s grown extensively under rainfed conditions.Because of its usage in multiple industries,su...Background Cotton is one of the most important commercial crops after food crops,especially in countries like India,where it’s grown extensively under rainfed conditions.Because of its usage in multiple industries,such as textile,medicine,and automobile industries,it has greater commercial importance.The crop’s performance is greatly influenced by prevailing weather dynamics.As climate changes,assessing how weather changes affect crop performance is essential.Among various techniques that are available,crop models are the most effective and widely used tools for predicting yields.Results This study compares statistical and machine learning models to assess their ability to predict cotton yield across major producing districts of Karnataka,India,utilizing a long-term dataset spanning from 1990 to 2023 that includes yield and weather factors.The artificial neural networks(ANNs)performed superiorly with acceptable yield deviations ranging within±10%during both vegetative stage(F1)and mid stage(F2)for cotton.The model evaluation metrics such as root mean square error(RMSE),normalized root mean square error(nRMSE),and modelling efficiency(EF)were also within the acceptance limits in most districts.Furthermore,the tested ANN model was used to assess the importance of the dominant weather factors influencing crop yield in each district.Specifically,the use of morning relative humidity as an individual parameter and its interaction with maximum and minimum tempera-ture had a major influence on cotton yield in most of the yield predicted districts.These differences highlighted the differential interactions of weather factors in each district for cotton yield formation,highlighting individual response of each weather factor under different soils and management conditions over the major cotton growing districts of Karnataka.Conclusions Compared with statistical models,machine learning models such as ANNs proved higher efficiency in forecasting the cotton yield due to their ability to consider the interactive effects of weather factors on yield forma-tion at different growth stages.This highlights the best suitability of ANNs for yield forecasting in rainfed conditions and for the study on relative impacts of weather factors on yield.Thus,the study aims to provide valuable insights to support stakeholders in planning effective crop management strategies and formulating relevant policies.展开更多
Assessing the stability of slopes is one of the crucial tasks of geotechnical engineering for assessing and managing risks related to natural hazards,directly affecting safety and sustainable development.This study pr...Assessing the stability of slopes is one of the crucial tasks of geotechnical engineering for assessing and managing risks related to natural hazards,directly affecting safety and sustainable development.This study primarily focuses on developing robust and practical hybrid models to predict the slope stability status of circular failure mode.For this purpose,three robust models were developed using a database including 627 case histories of slope stability status.The models were developed using the random forest(RF),support vector machine(SVM),and extreme gradient boosting(XGB)techniques,employing 5-fold cross validation approach.To enhance the performance of models,this study employs Bayesian optimizer(BO)to fine-tuning their hyperparameters.The results indicate that the performance order of the three developed models is RF-BO>SVM-BO>XGB-BO.Furthermore,comparing the developed models with previous models,it was found that the RF-BO model can effectively determine the slope stability status with outstanding performance.This implies that the RF-BO model could serve as a dependable tool for project managers,assisting in the evaluation of slope stability during both the design and operational phases of projects,despite the inherent challenges in this domain.The results regarding the importance of influencing parameters indicate that cohesion,friction angle,and slope height exert the most significant impact on slope stability status.This suggests that concentrating on these parameters and employing the RF-BO model can effectively mitigate the severity of geohazards in the short-term and contribute to the attainment of long-term sustainable development objectives.展开更多
The nonlinearity of hedonic datasets demands flexible automated valuation models to appraise housing prices accurately,and artificial intelligence models have been employed in mass appraisal to this end.However,they h...The nonlinearity of hedonic datasets demands flexible automated valuation models to appraise housing prices accurately,and artificial intelligence models have been employed in mass appraisal to this end.However,they have been referred to as“blackbox”models owing to difficulties associated with interpretation.In this study,we compared the results of traditional hedonic pricing models with those of machine learning algorithms,e.g.,random forest and deep neural network models.Commonly implemented measures,e.g.,Gini importance and permutation importance,provide only the magnitude of each explanatory variable’s importance,which results in ambiguous interpretability.To address this issue,we employed the SHapley Additive exPlanation(SHAP)method and explored its effectiveness through comparisons with traditionally explainable measures in hedonic pricing models.The results demonstrated that(1)the random forest model with the SHAP method could be a reliable instrument for appraising housing prices with high accuracy and sufficient interpretability,(2)the interpretable results retrieved from the SHAP method can be consolidated by the support of statistical evidence,and(3)housing characteristics and local amenities are primary contributors in property valuation,which is consistent with the findings of previous studies.Thus,our novel methodological framework and robust findings provide informative insights into the use of machine learning methods in property valuation based on the comparative analysis.展开更多
Understanding spatial heterogeneity in groundwater responses to multiple factors is critical for water resource management in coastal cities.Daily groundwater depth(GWD)data from 43 wells(2018-2022)were collected in t...Understanding spatial heterogeneity in groundwater responses to multiple factors is critical for water resource management in coastal cities.Daily groundwater depth(GWD)data from 43 wells(2018-2022)were collected in three coastal cities in Jiangsu Province,China.Seasonal and Trend decomposition using Loess(STL)together with wavelet analysis and empirical mode decomposition were applied to identify tide-influenced wells while remaining wells were grouped by hierarchical clustering analysis(HCA).Machine learning models were developed to predict GWD,then their response to natural conditions and human activities was assessed by the Shapley Additive exPlanations(SHAP)method.Results showed that eXtreme Gradient Boosting(XGB)was superior to other models in terms of prediction performance and computational efficiency(R^(2)>0.95).GWD in Yancheng and southern Lianyungang were greater than those in Nantong,exhibiting larger fluctuations.Groundwater within 5 km of the coastline was affected by tides,with more pronounced effects in agricultural areas compared to urban areas.Shallow groundwater(3-7 m depth)responded immediately(0-1 day)to rainfall,primarily influenced by farmland and topography(slope and distance from rivers).Rainfall recharge to groundwater peaked at 50%farmland coverage,but this effect was suppressed by high temperatures(>30℃)which intensified as distance from rivers increased,especially in forest and grassland.Deep groundwater(>10 m)showed delayed responses to rainfall(1-4 days)and temperature(10-15 days),with GDP as the primary influence,followed by agricultural irrigation and population density.Farmland helped to maintain stable GWD in low population density regions,while excessive farmland coverage(>90%)led to overexploitation.In the early stages of GDP development,increased industrial and agricultural water demand led to GWD decline,but as GDP levels significantly improved,groundwater consumption pressure gradually eased.This methodological framework is applicable not only to coastal cities in China but also could be extended to coastal regions worldwide.展开更多
BACKGROUND Severe dengue children with critical complications have been attributed to high mortality rates,varying from approximately 1%to over 20%.To date,there is a lack of data on machine-learning-based algorithms ...BACKGROUND Severe dengue children with critical complications have been attributed to high mortality rates,varying from approximately 1%to over 20%.To date,there is a lack of data on machine-learning-based algorithms for predicting the risk of inhospital mortality in children with dengue shock syndrome(DSS).AIM To develop machine-learning models to estimate the risk of death in hospitalized children with DSS.METHODS This single-center retrospective study was conducted at tertiary Children’s Hospital No.2 in Viet Nam,between 2013 and 2022.The primary outcome was the in-hospital mortality rate in children with DSS admitted to the pediatric intensive care unit(PICU).Nine significant features were predetermined for further analysis using machine learning models.An oversampling method was used to enhance the model performance.Supervised models,including logistic regression,Naïve Bayes,Random Forest(RF),K-nearest neighbors,Decision Tree and Extreme Gradient Boosting(XGBoost),were employed to develop predictive models.The Shapley Additive Explanation was used to determine the degree of contribution of the features.RESULTS In total,1278 PICU-admitted children with complete data were included in the analysis.The median patient age was 8.1 years(interquartile range:5.4-10.7).Thirty-nine patients(3%)died.The RF and XGboost models demonstrated the highest performance.The Shapley Addictive Explanations model revealed that the most important predictive features included younger age,female patients,presence of underlying diseases,severe transaminitis,severe bleeding,low platelet counts requiring platelet transfusion,elevated levels of international normalized ratio,blood lactate and serum creatinine,large volume of resuscitation fluid and a high vasoactive inotropic score(>30).CONCLUSION We developed robust machine learning-based models to estimate the risk of death in hospitalized children with DSS.The study findings are applicable to the design of management schemes to enhance survival outcomes of patients with DSS.展开更多
Floods and storm surges pose significant threats to coastal regions worldwide,demanding timely and accurate early warning systems(EWS)for disaster preparedness.Traditional numerical and statistical methods often fall ...Floods and storm surges pose significant threats to coastal regions worldwide,demanding timely and accurate early warning systems(EWS)for disaster preparedness.Traditional numerical and statistical methods often fall short in capturing complex,nonlinear,and real-time environmental dynamics.In recent years,machine learning(ML)and deep learning(DL)techniques have emerged as promising alternatives for enhancing the accuracy,speed,and scalability of EWS.This review critically evaluates the evolution of ML models—such as Artificial Neural Networks(ANN),Convolutional Neural Networks(CNN),and Long Short-Term Memory(LSTM)—in coastal flood prediction,highlighting their architectures,data requirements,performance metrics,and implementation challenges.A unique contribution of this work is the synthesis of real-time deployment challenges including latency,edge-cloud tradeoffs,and policy-level integration,areas often overlooked in prior literature.Furthermore,the review presents a comparative framework of model performance across different geographic and hydrologic settings,offering actionable insights for researchers and practitioners.Limitations of current AI-driven models,such as interpretability,data scarcity,and generalization across regions,are discussed in detail.Finally,the paper outlines future research directions including hybrid modelling,transfer learning,explainable AI,and policy-aware alert systems.By bridging technical performance and operational feasibility,this review aims to guide the development of next-generation intelligent EWS for resilient and adaptive coastal management.展开更多
In the event of nuclear accidents and incidents,when emergency resources are scarce,rapid and high-throughput biodosimeters for massive population triage and estimation are essential to guide medical treatment.Lymphoc...In the event of nuclear accidents and incidents,when emergency resources are scarce,rapid and high-throughput biodosimeters for massive population triage and estimation are essential to guide medical treatment.Lymphocyte dynamics,chromosome aberration analysis,and micronucleus assays are mainly used to estimate the biological dose of radiation[1].However,these technologies require highly trained personnel to perform and interpret and have the limitations of time consumption and low throughput,underscoring the urgent need for the development of radiation biomarkers and early classification.Dose and temporal responses as well as efficient triage models are important facets of radiation biodosimeters.展开更多
This paper aims to conduct a systematic literature review(SLR)using an artificial intelligence(AI)approach to predict and diagnose diabetes mellitus.After reviewing the literature published from 2015–2025,the paper a...This paper aims to conduct a systematic literature review(SLR)using an artificial intelligence(AI)approach to predict and diagnose diabetes mellitus.After reviewing the literature published from 2015–2025,the paper aims to identify the most effective AI techniques,the most used datasets,the most widely used data preprocessing techniques,and the most common issues.After analyzing the literature,it has been found that convolutional neural networks(CNNs)and long short-term memory(LSTM)networks are deep learning models that have shown high accuracy in diabetes prediction.Recursive feature elimination(RFE)and SMOTE are feature selection techniques that have significantly improved model accuracy,training time,and interpretability.Amidst this technological advancement,some existing issues persist:data imbalance,the inapplicability of techniques,computational limitations,and a lack of real-time application in a healthcare environment.The literature review has also identified the need for robust,interpretable,and scalable AI systems capable of handling large volumes of data,including real-world data,in the healthcare industry.Furthermore,it has been identified that the benefits should be integrated with wearable health monitoring systems and the development of privacy-preserving models to ensure continuous,secure,and proactive diabetes management.展开更多
Objectives:Decisions regarding CT after nCCRT for locally advanced rectal cancer(LARC)are challenging due to limited evidence guiding treatment.This study aimed to(i)evaluate the predictive performance of machine lear...Objectives:Decisions regarding CT after nCCRT for locally advanced rectal cancer(LARC)are challenging due to limited evidence guiding treatment.This study aimed to(i)evaluate the predictive performance of machine learning(ML)models in patients treated with neoadjuvant concurrent chemoradiotherapy(nCCRT)alone vs.those receiving nCCRT plus chemotherapy(CT),(ii)identify features associated with treatment improvement,and(iii)derive ML-based thresholds for treatment response.Methods:This retrospective study included 409 patients with LARC treated at three affiliated hospitals of Taipei Medical University.Patients were categorised into two groups:nCCRT alone followed by surgery(n=182)and nCCRT plus additional CT(n=227).Thirty-four baseline demographic,tumor,and laboratory variables were analysed.Four ML algorithms(K-Star,Random Forest,Multilayer Perceptron,and Random Committee)were evaluated,while five feature-ranking algorithms identified influential attributes among improved patients across both treatments.Decision Stump and AdaBoostM1 were applied to derive threshold-based patterns.Results:K-Star achieved the highest accuracy for nCCRT alone(80.8%;AUC=0.89),while Random Committee performed best for nCCRT plus CT(77.3%;AUC=0.84).Clinical N stage(cN)ranked highest,followed by Sodium(Na),Glutamic pyruvic transaminase,estimated glomerular filtration rate,body weight,red blood cell count,mean corpuscular hemoglobin concentration,and blood urea nitrogen.Threshold patterns suggested that CT-related improvement aligned with higher lymphocyte percentage and lower platelet distribution width,whereas nCCRT-only improvement aligned with elevated eGFR,GPT,and cN=2.Conclusions:ML-based analysis identified key predictors and demonstrated good model performance,supporting individualised post-nCCRT chemotherapy decisions.展开更多
Lithium-ion batteries(LIBs)are widely deployed,from grid-scale storage to electric vehicles.LIBs remain stationary most of their service life,where calendar aging degrades capacity.Understanding the mechanisms of LIB ...Lithium-ion batteries(LIBs)are widely deployed,from grid-scale storage to electric vehicles.LIBs remain stationary most of their service life,where calendar aging degrades capacity.Understanding the mechanisms of LIB calendar aging is crucial for extending battery lifespan.However,LIB calendar aging is influenced by multiple factors,including battery material,its state,and storage environment.Calendar aging experiments are also time-consuming,costly,and lack standardized testing conditions.This study employs a data-driven approach to establish a cross-scale database linking materials,side-reaction mechanisms,and calendar aging of LIBs.MELODI(Mechanism-informed,Explainable,Learning-based Optimization for Degradation Identification)is proposed to identify calendar aging mechanisms and quantify the effects of multi-scale factors.Results reveal that cathode material loss drives up to 91.42%of calendar aging degradation in high-nickel(Ni)batteries,while solid electrolyte interphase growth dominates in lithium iron phosphate(LFP)and low-Ni batteries,contributing up to 82.43%of degradation in LFP batteries and 99.10%of decay in low-Ni batteries,respectively.This study systematically quantifies calendar aging in commercial LIBs under varying materials,states of charge,and temperatures.These findings offer quantitative guidance for experimental design or battery use,and implications for emerging applications like aerial robotics,vehicle-to-grid,and embodied intelligence systems.展开更多
The efficient and safe delivery of messenger RNA(m RNA)therapeutics remains a critical challenge for clinical translation,driving the need for advanced carrier design.Ionizable amphiphilic Janus dendrimers(IAJDs)repre...The efficient and safe delivery of messenger RNA(m RNA)therapeutics remains a critical challenge for clinical translation,driving the need for advanced carrier design.Ionizable amphiphilic Janus dendrimers(IAJDs)represent a promising class of carriers;however,their structural complexity and limited available datasets hinder systematic exploration and optimization.In this study,we established a tailored machinelearning framework to investigate the structure-function relationships of IAJDs under a constrained data regime(n=231).Conventional molecular fingerprints were found to be suboptimal for representing these macromolecules,motivating the adoption of count-based descriptors and systematic ablation analyses to disentangle the contributions of the substructural features.These experiments identified key functional motifs underlying transfection performance and provided interpretable insights into the IAJD design principles.Complementing these handcrafted descriptors,we further applied deep learning-based molecular embeddings,which captured higher-order chemical semantics and significantly improved predictive accuracy.Collectively,these advances demonstrate that both refined fingerprinting and representation learning approaches can overcome data limitations,enabling the reliable prediction of IAJD activity while offering mechanistic interpretability.This study illustrates the potential of data-driven strategies as hypothesis-generation and prioritization tools for the design of next-generation m RNA delivery systems.展开更多
Landslides pose a significant threat in the mountainous regions of Nepal.Landslide susceptibility maps are commonly used to identify potential landslide zones by statistically analyzing geological,topographical,and hy...Landslides pose a significant threat in the mountainous regions of Nepal.Landslide susceptibility maps are commonly used to identify potential landslide zones by statistically analyzing geological,topographical,and hydrological factors,assuming that similar conditions may trigger future failures.While such maps provide valuable insights into landslide-triggering conditions,they are limited in assessing risk to settlements and infrastructure located downslope or in valley bottoms.This study integrates machine learning based landslide susceptibility with numerical runout modeling to provide a comprehensive landslide hazard assessment in the Bhotekoshi watershed,overcoming the limitations of traditional models that focus solely on statistical susceptibility.To conduct the susceptibility analysis,a total of 439 landslides were mapped from 2012 to 2021 using satellite images.Of these,70%were used for training two machine learning(ML)models:random forest and Xtreme Gradient Boosting(XGBoost),and the remaining 30%were used for validation.Among the two ML models,Random Forest model demonstrated slightly superior performance,achieving higher predictive accuracy.After the machine learning susceptibility analysis,the study transitions into a regional-scale landslide runout analysis.First,a back analysis of the past landslide event was conducted to fine-tune the model parameters(internal angle of friction and basal friction angle)and validate performance of the runout model.Following the back analysis,the regional-scale numerical modeling of landslide runout was conducted by designating areas classified as the highest susceptibility class in the Random Forest susceptibility map as potential release zones.This approach allows for a detailed examination of landslide propagation and potential impacts along the downslope settlements and infrastructures.The analysis clearly demonstrates that integrating both machine learning and numerical runout methods significantly increases the estimated exposure of population,buildings,and roads within the very high hazard class compared to relying solely on susceptibility methods.Specifically,population exposure rises from 360 to 7743,buildings increase from 97 to 2771,and road exposure expands from 41 to 251 km.This result highlights the significant risk of underestimating exposure in the analyses that solely rely on landslide susceptibility models.Integration of susceptibility and runout analysis improves landslide risk assessment,aiding in land-use planning and disaster mitigation strategies.展开更多
This study investigates the uncertain dynamic characterization of hybrid composite plates by employing advanced machine-assisted finite element methodologies.Hybrid composites,widely used in aerospace,automotive,and s...This study investigates the uncertain dynamic characterization of hybrid composite plates by employing advanced machine-assisted finite element methodologies.Hybrid composites,widely used in aerospace,automotive,and structural applications,often face variability in material properties,geometric configurations,and manufacturing processes,leading to uncertainty in their dynamic response.To address this,three surrogate-based machine learning approaches like radial basis function(RBF),multivariate adaptive regression splines(MARS),and polynomial neural networks(PNN)are integrated with a finite element framework to efficiently capture the stochastic behavior of these plates.The research focuses on predicting the first three natural frequencies under material uncertainties,which are critical to ensuring structural reliability.Monte Carlo simulation(MCS)is used as a benchmark for generating probabilistic datasets,including mean values,standard deviations,and probability density functions.The surrogate models are then trained and validated against these datasets,enabling accurate representation of uncertainty with substantially fewer samples compared to conventionalMCS.Among the methods studied,the RBFmodel demonstrates superior performance,closely approximating MCS results with a reduced sample size,thereby achieving significant computational savings.The proposed framework not only reduces computational time and costs but also maintains high predictive accuracy,making it well-suited for complex engineering systems.Beyond free vibration analysis,the methodology can be extended to more sophisticated scenarios,such as forced vibration,damping effects,and nonlinear structural responses.Overall,this work presents a computationally efficient and robust approach for surrogate-based uncertainty quantification,advancing the analysis and design of hybrid composite structures under uncertainty.展开更多
Accurate prediction of concrete compressive strength is fundamental for optimizing mix designs,improving material utilization,and ensuring structural safety in modern construction.Traditional empirical methods often f...Accurate prediction of concrete compressive strength is fundamental for optimizing mix designs,improving material utilization,and ensuring structural safety in modern construction.Traditional empirical methods often fail to capture the non-linear relationships among concrete constituents,especially with the growing use of supple-mentary cementitious materials and recycled aggregates.This study presents an integrated machine learning framework for concrete strength prediction,combining advanced regression models—namely CatBoost—with metaheuristic optimization algorithms,with a particular focus on the Somersaulting Spider Optimizer(SSO).A comprehensive dataset encompassing diverse mix proportions and material types was used to evaluate baseline machine learning models,including CatBoost,XGBoost,ExtraTrees,and RandomForest.Among these,CatBoost demonstrated superior accuracy across multiple performance metrics.To further enhance predictive capability,several bio-inspired optimizers were employed for hyperparameter tuning.The SSO-CatBoost hybrid achieved the lowest mean squared error and highest correlation coefficients,outperforming other metaheuristic approaches such as Genetic Algorithm,Particle Swarm Optimization,and Grey Wolf Optimizer.Statistical significance was established through Analysis of Variance and Wilcoxon signed-rank testing,confirming the robustness of the optimized models.The proposed methodology not only delivers improved predictive performance but also offers a transparent framework for mix design optimization,supporting data-driven decision making in sustainable and resilient infrastructure development.展开更多
Accurate prediction of flood events is important for flood control and risk management.Machine learning techniques contributed greatly to advances in flood predictions,and existing studies mainly focused on predicting...Accurate prediction of flood events is important for flood control and risk management.Machine learning techniques contributed greatly to advances in flood predictions,and existing studies mainly focused on predicting flood resource variables using single or hybrid machine learning techniques.However,class-based flood predictions have rarely been investigated,which can aid in quickly diagnosing comprehensive flood characteristics and proposing targeted management strategies.This study proposed a prediction approach of flood regime metrics and event classes coupling machine learning algorithms with clustering-deduced membership degrees.Five algorithms were adopted for this exploration.Results showed that the class membership degrees accurately determined event classes with class hit rates up to 100%,compared with the four classes clustered from nine regime metrics.The nonlinear algorithms(Multiple Linear Regression,Random Forest,and least squares-Support Vector Machine)outperformed the linear techniques(Multiple Linear Regression and Stepwise Regression)in predicting flood regime metrics.The proposed approach well predicted flood event classes with average class hit rates of 66.0%-85.4%and 47.2%-76.0%in calibration and validation periods,respectively,particularly for the slow and late flood events.The predictive capability of the proposed prediction approach for flood regime metrics and classes was considerably stronger than that of hydrological modeling approach.展开更多
基金funded by National Natural Science Foundation of China Key Program(12431014)Key Project of Hunan Education Department(22A0126)+1 种基金Natural Science Foundation of Hunan Province(2022JJ30555)Postgraduate Scientific Research Innovation Project of Xiangtan University(XDCX2024Y172)。
文摘Tropospheric zenith wet delay(ZWD)plays a vital role in the analysis of space geodetic observations.In recent years,machine learning methods have been increasingly applied to improve the accuracy of ZWD calculations.However,a single machine learning model has limited generalization capabilities.To address these limitations,this study introduces a novel machine learning fusion(MLF)algorithm with stronger generalization capabilities to enhance ZWD modeling and prediction accuracy.The MLF algorithm utilizes a two-layer structure integrating extra trees(ET),backpropagation neural network(BPNN),and linear regression models.By comparing the root mean square error(RMSE)of these models,we found that both ET-based and MLF-based models outperform RF-based and BPNN-based models in terms of internal and external accuracy,across both surface meteorological data-based and blind models.The improvement in exte rnal accuracy is particularly significant in the blind models.Our re sults show that the MLF(with an RMSE of 3.93 cm)and ET(3.99 cm)models outperform the traditional GPT3model(4.07 cm),while the RF(4.21 cm)and BPNN(4.14 cm)have worse external accuracies than the GPT3 model.It is worth noting that the BPNN suffered from overfitting during external accuracy tests,which was avoided by the MLF.In summary,regardless of the availability of surface meteorological data,the MLF-based empirical models demonstrate superior internal and external accuracy compared to the other tested models in this study.
基金supported by the special fund of the National Clinical Key Specialty Construction Program[(2022)301-2305].
文摘BACKGROUND:This study aims to develop and validate a machine learning-based in-hospital mortality predictive model for acute aortic syndrome(AAS)in the emergency department(ED)and to derive a simplifi ed version suitable for rapid clinical application.METHODS:In this multi-center retrospective cohort study,AAS patient data from three hospitals were analyzed.The modeling cohort included data from the First Affiliated Hospital of Zhengzhou University and the People’s Hospital of Xinjiang Uygur Autonomous Region,with Peking University Third Hospital data serving as the external test set.Four machine learning algorithms—logistic regression(LR),multilayer perceptron(MLP),Gaussian naive Bayes(GNB),and random forest(RF)—were used to develop predictive models based on 34 early-accessible clinical variables.A simplifi ed model was then derived based on fi ve key variables(Stanford type,pericardial eff usion,asymmetric peripheral arterial pulsation,decreased bowel sounds,and dyspnea)via Least Absolute Shrinkage and Selection Operator(LASSO)regression to improve ED applicability.RESULTS:A total of 929 patients were included in the modeling cohort,and 210 were included in the external test set.Four machine learning models based on 34 clinical variables were developed,achieving internal and external validation AUCs of 0.85-0.90 and 0.73-0.85,respectively.The simplifi ed model incorporating fi ve key variables demonstrated internal and external validation AUCs of 0.71-0.86 and 0.75-0.78,respectively.Both models showed robust calibration and predictive stability across datasets.CONCLUSION:Both kinds of models were built based on machine learning tools,and proved to have certain prediction performance and extrapolation.
基金supported by the Advanced Materials-National Science and Technology Major Project(Grant No.2025ZD0618401)the National Natural Science Foundation of China(Grant No.12504285)+1 种基金the Natural Science Foundation of Jiangsu Province(Grant No.BK20250472)NFSG grant from BITS-Pilani,Dubai campus。
文摘The rapid advancement of machine learning based tight-binding Hamiltonian(MLTB)methods has opened new avenues for efficient and accurate electronic structure simulations,particularly in large-scale systems and long-time scenarios.This review begins with a concise overview of traditional tight-binding(TB)models,including both(semi-)empirical and first-principles approaches,establishing the foundation for understanding MLTB developments.We then present a systematic classification of existing MLTB methodologies,grouped into two major categories:direct prediction of TB Hamiltonian elements and inference of empirical parameters.A comparative analysis with other ML-based electronic structure models is also provided,highlighting the advancement of MLTB approaches.Finally,we explore the emerging MLTB application ecosystem,highlighting how the integration of MLTB models with a diverse suite of post-processing tools from linear-scaling solvers to quantum transport frameworks and molecular dynamics interfaces is essential for tackling complex scientific problems across different domains.The continued advancement of this integrated paradigm promises to accelerate materials discovery and open new frontiers in the predictive simulation of complex quantum phenomena.
文摘This study aims to develop an accurate and robust machine learning model to predict the carbonation depth of fly ash concrete,overcoming the limitations of traditional predictive methods.Five ensemble-based models,such as adaptive boosting(AdaBoost),categorical boosting(CatBoost),gradient boosting regressor(GBR),hist gradient boosting regressor(HistGBR),and extreme gradient boosting(XGBoost),were developed and optimized using 729 high-quality dataset points incorporating seven input parameters,including cement,CO_(2),exposure time,water-binder ratio,fly ash,curing time,and compressive strength.Several performance evaluation metrics were used to compare the models.The GBR model emerged as the best-performing model,based on high coefficient of determination(R^(2))values and balanced error metrics across both validation and testing datasets.While all models performed exceptionally well on the training data,GBR demonstrated superior generalization capability,with R^(2) values of 0.9438 on the validation set and 0.9310 on the testing set.Furthermore,its low mean squared error(MSE),root mean square error(RMSE),mean absolute error(MAE),and median absolute error(MdAE)confirmed its robustness and accuracy.Moreover,shapley additive explanations(SHAP)analysis enhanced the interpretability of predictions,highlighting the curing time and exposure time as the most critical drivers of carbonation depth.
基金financial support of the National Natural Science Foundation of China(No.52371103)the Fundamental Research Funds for the Central Universities,China(No.2242023K40028)+1 种基金the Open Research Fund of Jiangsu Key Laboratory for Advanced Metallic Materials,China(No.AMM2023B01).financial support of the Research Fund of Shihezi Key Laboratory of AluminumBased Advanced Materials,China(No.2023PT02)financial support of Guangdong Province Science and Technology Major Project,China(No.2021B0301030005)。
文摘Oxide dispersion strengthened(ODS)alloys are extensively used owing to high thermostability and creep strength contributed from uniformly dispersed fine oxides particles.However,the existence of these strengthening particles also deteriorates the processability and it is of great importance to establish accurate processing maps to guide the thermomechanical processes to enhance the formability.In this study,we performed particle swarm optimization-based back propagation artificial neural network model to predict the high temperature flow behavior of 0.25wt%Al2O3 particle-reinforced Cu alloys,and compared the accuracy with that of derived by Arrhenius-type constitutive model and back propagation artificial neural network model.To train these models,we obtained the raw data by fabricating ODS Cu alloys using the internal oxidation and reduction method,and conducting systematic hot compression tests between 400 and800℃with strain rates of 10^(-2)-10 S^(-1).At last,processing maps for ODS Cu alloys were proposed by combining processing parameters,mechanical behavior,microstructure characterization,and the modeling results achieved a coefficient of determination higher than>99%.
基金funded through India Meteorological Department,New Delhi,India under the Forecasting Agricultural output using Space,Agrometeorol ogy and Land based observations(FASAL)project and fund number:No.ASC/FASAL/KT-11/01/HQ-2010.
文摘Background Cotton is one of the most important commercial crops after food crops,especially in countries like India,where it’s grown extensively under rainfed conditions.Because of its usage in multiple industries,such as textile,medicine,and automobile industries,it has greater commercial importance.The crop’s performance is greatly influenced by prevailing weather dynamics.As climate changes,assessing how weather changes affect crop performance is essential.Among various techniques that are available,crop models are the most effective and widely used tools for predicting yields.Results This study compares statistical and machine learning models to assess their ability to predict cotton yield across major producing districts of Karnataka,India,utilizing a long-term dataset spanning from 1990 to 2023 that includes yield and weather factors.The artificial neural networks(ANNs)performed superiorly with acceptable yield deviations ranging within±10%during both vegetative stage(F1)and mid stage(F2)for cotton.The model evaluation metrics such as root mean square error(RMSE),normalized root mean square error(nRMSE),and modelling efficiency(EF)were also within the acceptance limits in most districts.Furthermore,the tested ANN model was used to assess the importance of the dominant weather factors influencing crop yield in each district.Specifically,the use of morning relative humidity as an individual parameter and its interaction with maximum and minimum tempera-ture had a major influence on cotton yield in most of the yield predicted districts.These differences highlighted the differential interactions of weather factors in each district for cotton yield formation,highlighting individual response of each weather factor under different soils and management conditions over the major cotton growing districts of Karnataka.Conclusions Compared with statistical models,machine learning models such as ANNs proved higher efficiency in forecasting the cotton yield due to their ability to consider the interactive effects of weather factors on yield forma-tion at different growth stages.This highlights the best suitability of ANNs for yield forecasting in rainfed conditions and for the study on relative impacts of weather factors on yield.Thus,the study aims to provide valuable insights to support stakeholders in planning effective crop management strategies and formulating relevant policies.
文摘Assessing the stability of slopes is one of the crucial tasks of geotechnical engineering for assessing and managing risks related to natural hazards,directly affecting safety and sustainable development.This study primarily focuses on developing robust and practical hybrid models to predict the slope stability status of circular failure mode.For this purpose,three robust models were developed using a database including 627 case histories of slope stability status.The models were developed using the random forest(RF),support vector machine(SVM),and extreme gradient boosting(XGB)techniques,employing 5-fold cross validation approach.To enhance the performance of models,this study employs Bayesian optimizer(BO)to fine-tuning their hyperparameters.The results indicate that the performance order of the three developed models is RF-BO>SVM-BO>XGB-BO.Furthermore,comparing the developed models with previous models,it was found that the RF-BO model can effectively determine the slope stability status with outstanding performance.This implies that the RF-BO model could serve as a dependable tool for project managers,assisting in the evaluation of slope stability during both the design and operational phases of projects,despite the inherent challenges in this domain.The results regarding the importance of influencing parameters indicate that cohesion,friction angle,and slope height exert the most significant impact on slope stability status.This suggests that concentrating on these parameters and employing the RF-BO model can effectively mitigate the severity of geohazards in the short-term and contribute to the attainment of long-term sustainable development objectives.
基金supported by the National Research Foundation of Korea grant funded by the Korea government(MSIT)(RS-2025-16067531:Kwangwon Ahn)Hankuk University of Foreign Studies Research Fund(0f 2025:Sihyun An).
文摘The nonlinearity of hedonic datasets demands flexible automated valuation models to appraise housing prices accurately,and artificial intelligence models have been employed in mass appraisal to this end.However,they have been referred to as“blackbox”models owing to difficulties associated with interpretation.In this study,we compared the results of traditional hedonic pricing models with those of machine learning algorithms,e.g.,random forest and deep neural network models.Commonly implemented measures,e.g.,Gini importance and permutation importance,provide only the magnitude of each explanatory variable’s importance,which results in ambiguous interpretability.To address this issue,we employed the SHapley Additive exPlanation(SHAP)method and explored its effectiveness through comparisons with traditionally explainable measures in hedonic pricing models.The results demonstrated that(1)the random forest model with the SHAP method could be a reliable instrument for appraising housing prices with high accuracy and sufficient interpretability,(2)the interpretable results retrieved from the SHAP method can be consolidated by the support of statistical evidence,and(3)housing characteristics and local amenities are primary contributors in property valuation,which is consistent with the findings of previous studies.Thus,our novel methodological framework and robust findings provide informative insights into the use of machine learning methods in property valuation based on the comparative analysis.
基金supported by the Natural Science Foundation of Jiangsu province,China(BK20240937)the Belt and Road Special Foundation of the National Key Laboratory of Water Disaster Prevention(2022491411,2021491811)the Basal Research Fund of Central Public Welfare Scientific Institution of Nanjing Hydraulic Research Institute(Y223006).
文摘Understanding spatial heterogeneity in groundwater responses to multiple factors is critical for water resource management in coastal cities.Daily groundwater depth(GWD)data from 43 wells(2018-2022)were collected in three coastal cities in Jiangsu Province,China.Seasonal and Trend decomposition using Loess(STL)together with wavelet analysis and empirical mode decomposition were applied to identify tide-influenced wells while remaining wells were grouped by hierarchical clustering analysis(HCA).Machine learning models were developed to predict GWD,then their response to natural conditions and human activities was assessed by the Shapley Additive exPlanations(SHAP)method.Results showed that eXtreme Gradient Boosting(XGB)was superior to other models in terms of prediction performance and computational efficiency(R^(2)>0.95).GWD in Yancheng and southern Lianyungang were greater than those in Nantong,exhibiting larger fluctuations.Groundwater within 5 km of the coastline was affected by tides,with more pronounced effects in agricultural areas compared to urban areas.Shallow groundwater(3-7 m depth)responded immediately(0-1 day)to rainfall,primarily influenced by farmland and topography(slope and distance from rivers).Rainfall recharge to groundwater peaked at 50%farmland coverage,but this effect was suppressed by high temperatures(>30℃)which intensified as distance from rivers increased,especially in forest and grassland.Deep groundwater(>10 m)showed delayed responses to rainfall(1-4 days)and temperature(10-15 days),with GDP as the primary influence,followed by agricultural irrigation and population density.Farmland helped to maintain stable GWD in low population density regions,while excessive farmland coverage(>90%)led to overexploitation.In the early stages of GDP development,increased industrial and agricultural water demand led to GWD decline,but as GDP levels significantly improved,groundwater consumption pressure gradually eased.This methodological framework is applicable not only to coastal cities in China but also could be extended to coastal regions worldwide.
文摘BACKGROUND Severe dengue children with critical complications have been attributed to high mortality rates,varying from approximately 1%to over 20%.To date,there is a lack of data on machine-learning-based algorithms for predicting the risk of inhospital mortality in children with dengue shock syndrome(DSS).AIM To develop machine-learning models to estimate the risk of death in hospitalized children with DSS.METHODS This single-center retrospective study was conducted at tertiary Children’s Hospital No.2 in Viet Nam,between 2013 and 2022.The primary outcome was the in-hospital mortality rate in children with DSS admitted to the pediatric intensive care unit(PICU).Nine significant features were predetermined for further analysis using machine learning models.An oversampling method was used to enhance the model performance.Supervised models,including logistic regression,Naïve Bayes,Random Forest(RF),K-nearest neighbors,Decision Tree and Extreme Gradient Boosting(XGBoost),were employed to develop predictive models.The Shapley Additive Explanation was used to determine the degree of contribution of the features.RESULTS In total,1278 PICU-admitted children with complete data were included in the analysis.The median patient age was 8.1 years(interquartile range:5.4-10.7).Thirty-nine patients(3%)died.The RF and XGboost models demonstrated the highest performance.The Shapley Addictive Explanations model revealed that the most important predictive features included younger age,female patients,presence of underlying diseases,severe transaminitis,severe bleeding,low platelet counts requiring platelet transfusion,elevated levels of international normalized ratio,blood lactate and serum creatinine,large volume of resuscitation fluid and a high vasoactive inotropic score(>30).CONCLUSION We developed robust machine learning-based models to estimate the risk of death in hospitalized children with DSS.The study findings are applicable to the design of management schemes to enhance survival outcomes of patients with DSS.
文摘Floods and storm surges pose significant threats to coastal regions worldwide,demanding timely and accurate early warning systems(EWS)for disaster preparedness.Traditional numerical and statistical methods often fall short in capturing complex,nonlinear,and real-time environmental dynamics.In recent years,machine learning(ML)and deep learning(DL)techniques have emerged as promising alternatives for enhancing the accuracy,speed,and scalability of EWS.This review critically evaluates the evolution of ML models—such as Artificial Neural Networks(ANN),Convolutional Neural Networks(CNN),and Long Short-Term Memory(LSTM)—in coastal flood prediction,highlighting their architectures,data requirements,performance metrics,and implementation challenges.A unique contribution of this work is the synthesis of real-time deployment challenges including latency,edge-cloud tradeoffs,and policy-level integration,areas often overlooked in prior literature.Furthermore,the review presents a comparative framework of model performance across different geographic and hydrologic settings,offering actionable insights for researchers and practitioners.Limitations of current AI-driven models,such as interpretability,data scarcity,and generalization across regions,are discussed in detail.Finally,the paper outlines future research directions including hybrid modelling,transfer learning,explainable AI,and policy-aware alert systems.By bridging technical performance and operational feasibility,this review aims to guide the development of next-generation intelligent EWS for resilient and adaptive coastal management.
基金supported by the National Natural Science Foundation of China(No.82173463)the Young Scholar Science Foundation of China CDC(No.2023A204)the NIRP Research Fund for Young Scholars(No.2024NB01).
文摘In the event of nuclear accidents and incidents,when emergency resources are scarce,rapid and high-throughput biodosimeters for massive population triage and estimation are essential to guide medical treatment.Lymphocyte dynamics,chromosome aberration analysis,and micronucleus assays are mainly used to estimate the biological dose of radiation[1].However,these technologies require highly trained personnel to perform and interpret and have the limitations of time consumption and low throughput,underscoring the urgent need for the development of radiation biomarkers and early classification.Dose and temporal responses as well as efficient triage models are important facets of radiation biodosimeters.
文摘This paper aims to conduct a systematic literature review(SLR)using an artificial intelligence(AI)approach to predict and diagnose diabetes mellitus.After reviewing the literature published from 2015–2025,the paper aims to identify the most effective AI techniques,the most used datasets,the most widely used data preprocessing techniques,and the most common issues.After analyzing the literature,it has been found that convolutional neural networks(CNNs)and long short-term memory(LSTM)networks are deep learning models that have shown high accuracy in diabetes prediction.Recursive feature elimination(RFE)and SMOTE are feature selection techniques that have significantly improved model accuracy,training time,and interpretability.Amidst this technological advancement,some existing issues persist:data imbalance,the inapplicability of techniques,computational limitations,and a lack of real-time application in a healthcare environment.The literature review has also identified the need for robust,interpretable,and scalable AI systems capable of handling large volumes of data,including real-world data,in the healthcare industry.Furthermore,it has been identified that the benefits should be integrated with wearable health monitoring systems and the development of privacy-preserving models to ensure continuous,secure,and proactive diabetes management.
基金funded by the Australian National Health and Medical Research Council(Grant No.GNT1192469)supported by the Research Technology Services at the University of New South Wales Sydney,Google Cloud Research(Award No.GCP19980904)。
文摘Objectives:Decisions regarding CT after nCCRT for locally advanced rectal cancer(LARC)are challenging due to limited evidence guiding treatment.This study aimed to(i)evaluate the predictive performance of machine learning(ML)models in patients treated with neoadjuvant concurrent chemoradiotherapy(nCCRT)alone vs.those receiving nCCRT plus chemotherapy(CT),(ii)identify features associated with treatment improvement,and(iii)derive ML-based thresholds for treatment response.Methods:This retrospective study included 409 patients with LARC treated at three affiliated hospitals of Taipei Medical University.Patients were categorised into two groups:nCCRT alone followed by surgery(n=182)and nCCRT plus additional CT(n=227).Thirty-four baseline demographic,tumor,and laboratory variables were analysed.Four ML algorithms(K-Star,Random Forest,Multilayer Perceptron,and Random Committee)were evaluated,while five feature-ranking algorithms identified influential attributes among improved patients across both treatments.Decision Stump and AdaBoostM1 were applied to derive threshold-based patterns.Results:K-Star achieved the highest accuracy for nCCRT alone(80.8%;AUC=0.89),while Random Committee performed best for nCCRT plus CT(77.3%;AUC=0.84).Clinical N stage(cN)ranked highest,followed by Sodium(Na),Glutamic pyruvic transaminase,estimated glomerular filtration rate,body weight,red blood cell count,mean corpuscular hemoglobin concentration,and blood urea nitrogen.Threshold patterns suggested that CT-related improvement aligned with higher lymphocyte percentage and lower platelet distribution width,whereas nCCRT-only improvement aligned with elevated eGFR,GPT,and cN=2.Conclusions:ML-based analysis identified key predictors and demonstrated good model performance,supporting individualised post-nCCRT chemotherapy decisions.
基金supported by the National Key Research and Development Program of China(2024YFE0213000)the Postdoctoral Innovative Talents Support Program(BX20240232)+1 种基金the Natural Science Foundation of China for Young Scholars(72304031)the Fundamental Research Funds for the Central Universities(FRF-TP-22-024A1).
文摘Lithium-ion batteries(LIBs)are widely deployed,from grid-scale storage to electric vehicles.LIBs remain stationary most of their service life,where calendar aging degrades capacity.Understanding the mechanisms of LIB calendar aging is crucial for extending battery lifespan.However,LIB calendar aging is influenced by multiple factors,including battery material,its state,and storage environment.Calendar aging experiments are also time-consuming,costly,and lack standardized testing conditions.This study employs a data-driven approach to establish a cross-scale database linking materials,side-reaction mechanisms,and calendar aging of LIBs.MELODI(Mechanism-informed,Explainable,Learning-based Optimization for Degradation Identification)is proposed to identify calendar aging mechanisms and quantify the effects of multi-scale factors.Results reveal that cathode material loss drives up to 91.42%of calendar aging degradation in high-nickel(Ni)batteries,while solid electrolyte interphase growth dominates in lithium iron phosphate(LFP)and low-Ni batteries,contributing up to 82.43%of degradation in LFP batteries and 99.10%of decay in low-Ni batteries,respectively.This study systematically quantifies calendar aging in commercial LIBs under varying materials,states of charge,and temperatures.These findings offer quantitative guidance for experimental design or battery use,and implications for emerging applications like aerial robotics,vehicle-to-grid,and embodied intelligence systems.
基金supported by the National Key Research and Development Program of China(No.2024YFF0508600)the National Natural Science Foundation of China(Nos.22305081,32571562 and 52572303)+2 种基金Leading Talents in Shanghai in 2018,Shanghai Sailing Program(No.23YF1408600)111 project(No.B14018)Fundamental Research Funds for the Central Universities of China。
文摘The efficient and safe delivery of messenger RNA(m RNA)therapeutics remains a critical challenge for clinical translation,driving the need for advanced carrier design.Ionizable amphiphilic Janus dendrimers(IAJDs)represent a promising class of carriers;however,their structural complexity and limited available datasets hinder systematic exploration and optimization.In this study,we established a tailored machinelearning framework to investigate the structure-function relationships of IAJDs under a constrained data regime(n=231).Conventional molecular fingerprints were found to be suboptimal for representing these macromolecules,motivating the adoption of count-based descriptors and systematic ablation analyses to disentangle the contributions of the substructural features.These experiments identified key functional motifs underlying transfection performance and provided interpretable insights into the IAJD design principles.Complementing these handcrafted descriptors,we further applied deep learning-based molecular embeddings,which captured higher-order chemical semantics and significantly improved predictive accuracy.Collectively,these advances demonstrate that both refined fingerprinting and representation learning approaches can overcome data limitations,enabling the reliable prediction of IAJD activity while offering mechanistic interpretability.This study illustrates the potential of data-driven strategies as hypothesis-generation and prioritization tools for the design of next-generation m RNA delivery systems.
基金China Scholarship Council(CSC)for providing a fully funded post-graduate study in institute of mountain hazards and environment UCASsupported by the National Natural Science Foundation of China(Grant Nos.42361144880)+3 种基金the Science and Technology Program of Xizang(Grant No.XZ202402ZD0001)the Basic Research Program of Qinghai Province(2024-ZJ-904)the Postdoctoral Fellowship Programs of CPSF(Grant Nos.GZC20232571,2024M753153)the International Cooperation Overseas Platform Project,CAS(Grant No.131C11KYSB20200033).
文摘Landslides pose a significant threat in the mountainous regions of Nepal.Landslide susceptibility maps are commonly used to identify potential landslide zones by statistically analyzing geological,topographical,and hydrological factors,assuming that similar conditions may trigger future failures.While such maps provide valuable insights into landslide-triggering conditions,they are limited in assessing risk to settlements and infrastructure located downslope or in valley bottoms.This study integrates machine learning based landslide susceptibility with numerical runout modeling to provide a comprehensive landslide hazard assessment in the Bhotekoshi watershed,overcoming the limitations of traditional models that focus solely on statistical susceptibility.To conduct the susceptibility analysis,a total of 439 landslides were mapped from 2012 to 2021 using satellite images.Of these,70%were used for training two machine learning(ML)models:random forest and Xtreme Gradient Boosting(XGBoost),and the remaining 30%were used for validation.Among the two ML models,Random Forest model demonstrated slightly superior performance,achieving higher predictive accuracy.After the machine learning susceptibility analysis,the study transitions into a regional-scale landslide runout analysis.First,a back analysis of the past landslide event was conducted to fine-tune the model parameters(internal angle of friction and basal friction angle)and validate performance of the runout model.Following the back analysis,the regional-scale numerical modeling of landslide runout was conducted by designating areas classified as the highest susceptibility class in the Random Forest susceptibility map as potential release zones.This approach allows for a detailed examination of landslide propagation and potential impacts along the downslope settlements and infrastructures.The analysis clearly demonstrates that integrating both machine learning and numerical runout methods significantly increases the estimated exposure of population,buildings,and roads within the very high hazard class compared to relying solely on susceptibility methods.Specifically,population exposure rises from 360 to 7743,buildings increase from 97 to 2771,and road exposure expands from 41 to 251 km.This result highlights the significant risk of underestimating exposure in the analyses that solely rely on landslide susceptibility models.Integration of susceptibility and runout analysis improves landslide risk assessment,aiding in land-use planning and disaster mitigation strategies.
文摘This study investigates the uncertain dynamic characterization of hybrid composite plates by employing advanced machine-assisted finite element methodologies.Hybrid composites,widely used in aerospace,automotive,and structural applications,often face variability in material properties,geometric configurations,and manufacturing processes,leading to uncertainty in their dynamic response.To address this,three surrogate-based machine learning approaches like radial basis function(RBF),multivariate adaptive regression splines(MARS),and polynomial neural networks(PNN)are integrated with a finite element framework to efficiently capture the stochastic behavior of these plates.The research focuses on predicting the first three natural frequencies under material uncertainties,which are critical to ensuring structural reliability.Monte Carlo simulation(MCS)is used as a benchmark for generating probabilistic datasets,including mean values,standard deviations,and probability density functions.The surrogate models are then trained and validated against these datasets,enabling accurate representation of uncertainty with substantially fewer samples compared to conventionalMCS.Among the methods studied,the RBFmodel demonstrates superior performance,closely approximating MCS results with a reduced sample size,thereby achieving significant computational savings.The proposed framework not only reduces computational time and costs but also maintains high predictive accuracy,making it well-suited for complex engineering systems.Beyond free vibration analysis,the methodology can be extended to more sophisticated scenarios,such as forced vibration,damping effects,and nonlinear structural responses.Overall,this work presents a computationally efficient and robust approach for surrogate-based uncertainty quantification,advancing the analysis and design of hybrid composite structures under uncertainty.
文摘Accurate prediction of concrete compressive strength is fundamental for optimizing mix designs,improving material utilization,and ensuring structural safety in modern construction.Traditional empirical methods often fail to capture the non-linear relationships among concrete constituents,especially with the growing use of supple-mentary cementitious materials and recycled aggregates.This study presents an integrated machine learning framework for concrete strength prediction,combining advanced regression models—namely CatBoost—with metaheuristic optimization algorithms,with a particular focus on the Somersaulting Spider Optimizer(SSO).A comprehensive dataset encompassing diverse mix proportions and material types was used to evaluate baseline machine learning models,including CatBoost,XGBoost,ExtraTrees,and RandomForest.Among these,CatBoost demonstrated superior accuracy across multiple performance metrics.To further enhance predictive capability,several bio-inspired optimizers were employed for hyperparameter tuning.The SSO-CatBoost hybrid achieved the lowest mean squared error and highest correlation coefficients,outperforming other metaheuristic approaches such as Genetic Algorithm,Particle Swarm Optimization,and Grey Wolf Optimizer.Statistical significance was established through Analysis of Variance and Wilcoxon signed-rank testing,confirming the robustness of the optimized models.The proposed methodology not only delivers improved predictive performance but also offers a transparent framework for mix design optimization,supporting data-driven decision making in sustainable and resilient infrastructure development.
基金National Key Research and Development Program of China,No.2023YFC3006704National Natural Science Foundation of China,No.42171047CAS-CSIRO Partnership Joint Project of 2024,No.177GJHZ2023097MI。
文摘Accurate prediction of flood events is important for flood control and risk management.Machine learning techniques contributed greatly to advances in flood predictions,and existing studies mainly focused on predicting flood resource variables using single or hybrid machine learning techniques.However,class-based flood predictions have rarely been investigated,which can aid in quickly diagnosing comprehensive flood characteristics and proposing targeted management strategies.This study proposed a prediction approach of flood regime metrics and event classes coupling machine learning algorithms with clustering-deduced membership degrees.Five algorithms were adopted for this exploration.Results showed that the class membership degrees accurately determined event classes with class hit rates up to 100%,compared with the four classes clustered from nine regime metrics.The nonlinear algorithms(Multiple Linear Regression,Random Forest,and least squares-Support Vector Machine)outperformed the linear techniques(Multiple Linear Regression and Stepwise Regression)in predicting flood regime metrics.The proposed approach well predicted flood event classes with average class hit rates of 66.0%-85.4%and 47.2%-76.0%in calibration and validation periods,respectively,particularly for the slow and late flood events.The predictive capability of the proposed prediction approach for flood regime metrics and classes was considerably stronger than that of hydrological modeling approach.