The uplift resistance of the soil overlying shield tunnels significantly impacts their anti-floating stability.However,research on uplift resistance concerning special-shaped shield tunnels is limited.This study combi...The uplift resistance of the soil overlying shield tunnels significantly impacts their anti-floating stability.However,research on uplift resistance concerning special-shaped shield tunnels is limited.This study combines numerical simulation with machine learning techniques to explore this issue.It presents a summary of special-shaped tunnel geometries and introduces a shape coefficient.Through the finite element software,Plaxis3D,the study simulates six key parameters—shape coefficient,burial depth ratio,tunnel’s longest horizontal length,internal friction angle,cohesion,and soil submerged bulk density—that impact uplift resistance across different conditions.Employing XGBoost and ANN methods,the feature importance of each parameter was analyzed based on the numerical simulation results.The findings demonstrate that a tunnel shape more closely resembling a circle leads to reduced uplift resistance in the overlying soil,whereas other parameters exhibit the contrary effects.Furthermore,the study reveals a diminishing trend in the feature importance of buried depth ratio,internal friction angle,tunnel longest horizontal length,cohesion,soil submerged bulk density,and shape coefficient in influencing uplift resistance.展开更多
Cyclohexene is an important raw material in the production of nylon.Selective hydrogenation of benzene is a key method for preparing cyclohexene.However,the Ru catalysts used in current industrial processes still face...Cyclohexene is an important raw material in the production of nylon.Selective hydrogenation of benzene is a key method for preparing cyclohexene.However,the Ru catalysts used in current industrial processes still face challenges,including high metal usage,high process costs,and low cyclohexene yield.This study utilizes existing literature data combined with machine learning methods to analyze the factors influencing benzene conversion,cyclohexene selectivity,and yield in the benzene hydrogenation to cyclohexene reaction.It constructs predictive models based on XGBoost and Random Forest algorithms.After analysis,it was found that reaction time,Ru content,and space velocity are key factors influencing cyclohexene yield,selectivity,and benzene conversion.Shapley Additive Explanations(SHAP)analysis and feature importance analysis further revealed the contribution of each variable to the reaction outcomes.Additionally,we randomly generated one million variable combinations using the Dirichlet distribution to attempt to predict high-yield catalyst formulations.This paper provides new insights into the application of machine learning in heterogeneous catalysis and offers some reference for further research.展开更多
In the pharmaceutical field,machine learning can play an important role in drug development,production and treatment.Co-crystallization techniques have shown promising potential to enhance the properties of active pha...In the pharmaceutical field,machine learning can play an important role in drug development,production and treatment.Co-crystallization techniques have shown promising potential to enhance the properties of active pharmaceutical ingredients(APIs)such as solubility,permeability,and bioavailability,all without altering their chemical structure.This approach opens new avenues for developing natural products into effective drugs,especially those previously challenging in formulation.Emodin,an anthraquinone-based natural product,is a notable example due to its diverse biological activities;however,its physicochemical limitations,such as poor solubility and easy sublimation,restricted its clinical application.While various methods have improved emodin's physicochemical properties,research on its bioavailability remains limited.In our study,we summarize cocrystals and salts produced through co-crystallization technology and identify piperazine as a favorable coformer.Conflicting conclusions from computational chemistry and molecular modeling method and machine learning method regarding the formation of an emodin-piperazine cocrystal or salt led us to experimentally validate these possibilities.Ultimately,we successfully obtained the emodin-piperazine cocrystal,which were characterized and evaluated by several in vitro methods and pharmacokinetic studies.In addition,experiments have shown that emodin has a certain therapeutic effect on sepsis,so we also evaluated emodin-piperazine biological activity in a sepsis model.The results demonstrate that co-crystallization significantly enhances emodin's solubility,permeability,and bioavailability.Pharmacodynamic studies indicate that the emodin-piperazine cocrystal improves sepsis symptoms and provides protective effects against liver and kidney damage associated with sepsis.This study offers renewed hope for natural products with broad biological activities yet hindered by physicochemical limitations by advancing co-crystallization as a viable development approach.展开更多
Machine learning(ML)is recognized as a potent tool for the inverse design of environmental functional material,particularly for complex entities like biochar-based catalysts(BCs).Thus,the tailored BCs can have a disti...Machine learning(ML)is recognized as a potent tool for the inverse design of environmental functional material,particularly for complex entities like biochar-based catalysts(BCs).Thus,the tailored BCs can have a distinct ability to trigger the nonradical pathway in advance oxidation processes(AOPs),promising a stable,rapid and selective degradation of persistent contaminants.However,due to the inherent“black box”nature and limitations of input features,results and conclusions derived from ML may not always be intuitively understood or comprehensively validated.To tackle this challenge,we linked the front-point interpretable analysis approaches with back-point density functional theory(DFT)calculations to form a chained learning strategy for deeper sight into the intrinsic activation mechanism of BCs in AOPs.At the front point,we conducted an easy-to-interpret meta-analysis to validate two strategies for enhancing nonradical pathways by increasing oxygen content and specific surface area(SSA),and prepared oxidized biochar(OBC500)and SSA-increased biochar(SBC900)by controlling pyrolysis conditions and modification methods.Subsequently,experimental results showed that OBC500 and SBC900 had distinct dominant degradation pathways for 1O2 generation and electron transfer,respectively.Finally,at the end point,DFT calculations revealed their active sites and degradation mechanisms.This chained learning strategy elucidates fundamental principles for BC inverse design and showcases the exceptional capacity to integrate computational techniques to accelerate catalyst inverse design.展开更多
Artificial intelligence(AI)based models have been used to predict the structural,optical,mechanical,and electrochemical properties of zinc oxide/graphene oxide nanocomposites.Machine learning(ML)models such as Artific...Artificial intelligence(AI)based models have been used to predict the structural,optical,mechanical,and electrochemical properties of zinc oxide/graphene oxide nanocomposites.Machine learning(ML)models such as Artificial Neural Networks(ANN),Support Vector Regression(SVR),Multilayer Perceptron(MLP),and hybrid,along with fuzzy logic tools,were applied to predict the different properties like wavelength at maximum intensity(444 nm),crystallite size(17.50 nm),and optical bandgap(2.85 eV).While some other properties,such as energy density,power density,and charge transfer resistance,were also predicted with the help of datasets of 1000(80:20).In general,the energy parameters were predicted more accurately by hybrid models.The hydrothermal method was used to synthesize graphene oxide(GO)and zinc oxide(ZnO)nanocomposites.The increased surface area,conductivity,and stability of graphene oxide in zinc oxide nanoparticles make the composite an ideal option for energy storage.X-ray diffraction(XRD)confirmed the crystallite size of 17.41 nm for the nanocomposite and the presence of GO(12.8○)peaks.The scanning electron microscope(SEM)showed anchored wrinkled GO sheets on zinc oxide with an average particle size of 2.93μm.Energy-dispersive X-ray spectroscopy(EDX)confirmed the elemental composition,and Fouriertransform infrared spectroscopy(FTIR)revealed the impact of GO on functional groups and electrochemical behavior.Photoluminescence(PL)wavelength of(439 nm)and band gap of(2.81 eV)show that the material is suitable for energy applications in nanocomposites.Smart nanocomposite materials with improved performance in energy storage and related applications were fabricated by combining synthesis,characterization,fuzzy logic,and machine learning in this work.展开更多
Post-kidney transplant rejection is a critical factor influencing transplant success rates and the survival of transplanted organs.With the rapid advancement of artificial intelligence technologies,machine learning(ML...Post-kidney transplant rejection is a critical factor influencing transplant success rates and the survival of transplanted organs.With the rapid advancement of artificial intelligence technologies,machine learning(ML)has emerged as a powerful data analysis tool,widely applied in the prediction,diagnosis,and mechanistic study of kidney transplant rejection.This mini-review systematically summarizes the recent applications of ML techniques in post-kidney transplant rejection,covering areas such as the construction of predictive models,identification of biomarkers,analysis of pathological images,assessment of immune cell infiltration,and formulation of personalized treatment strategies.By integrating multi-omics data and clinical information,ML has significantly enhanced the accuracy of early rejection diagnosis and the capability for prognostic evaluation,driving the development of precision medicine in the field of kidney transplantation.Furthermore,this article discusses the challenges faced in existing research and potential future directions,providing a theoretical basis and technical references for related studies.展开更多
Colorectal cancer is the third most diagnosed cancer worldwide,and immune checkpoint inhibitors have shown promising therapeutic outcomes in selected patient groups.This study performed a comprehensive analysis of mul...Colorectal cancer is the third most diagnosed cancer worldwide,and immune checkpoint inhibitors have shown promising therapeutic outcomes in selected patient groups.This study performed a comprehensive analysis of multi-omics data from The Cancer Genome Atlas colorectal adenocarcinoma cohort(TCGA-COADREAD),accessed through cBioPortal,to develop machine learning models for predicting progression-free survival(PFS)following immunotherapy.The dataset included clinical variables,genomic alterations in Kirsten Rat Sarcoma Viral Oncogene Homolog(KRAS),B-Raf Proto-Oncogene(BRAF),and Neuroblastoma RAS Viral Oncogene Homolog(NRAS),microsatellite instability(MSI)status,tumor mutation burden(TMB),and expression of immune checkpoint genes.Kaplan–Meier analysis showed that KRAS mutations were significantly associated with reduced PFS,while BRAF and NRAS mutations had no significant impact.MSI-high tumors exhibited elevated TMB and increased immune checkpoint expression,reflecting their immunologically active phenotype.We developed both survival and classification models,with the Extra Trees classifier achieving the best performance(accuracy=0.86,precision=0.67,recall=0.70,F1-score=0.68,AUC=0.84).These findings highlight the potential of combining genomic and immune biomarkers with machine learning to improve patient stratification and guide personalized immunotherapy decisions.An interactive web application was also developed to enable clinicians to input patient-specific molecular and clinical data and visualize individualized PFS predictions,supporting timely,data-driven treatment planning.展开更多
Modern intrusion detection systems(MIDS)face persistent challenges in coping with the rapid evolution of cyber threats,high-volume network traffic,and imbalanced datasets.Traditional models often lack the robustness a...Modern intrusion detection systems(MIDS)face persistent challenges in coping with the rapid evolution of cyber threats,high-volume network traffic,and imbalanced datasets.Traditional models often lack the robustness and explainability required to detect novel and sophisticated attacks effectively.This study introduces an advanced,explainable machine learning framework for multi-class IDS using the KDD99 and IDS datasets,which reflects real-world network behavior through a blend of normal and diverse attack classes.The methodology begins with sophisticated data preprocessing,incorporating both RobustScaler and QuantileTransformer to address outliers and skewed feature distributions,ensuring standardized and model-ready inputs.Critical dimensionality reduction is achieved via the Harris Hawks Optimization(HHO)algorithm—a nature-inspired metaheuristic modeled on hawks’hunting strategies.HHO efficiently identifies the most informative features by optimizing a fitness function based on classification performance.Following feature selection,the SMOTE is applied to the training data to resolve class imbalance by synthetically augmenting underrepresented attack types.The stacked architecture is then employed,combining the strengths of XGBoost,SVM,and RF as base learners.This layered approach improves prediction robustness and generalization by balancing bias and variance across diverse classifiers.The model was evaluated using standard classification metrics:precision,recall,F1-score,and overall accuracy.The best overall performance was recorded with an accuracy of 99.44%for UNSW-NB15,demonstrating the model’s effectiveness.After balancing,the model demonstrated a clear improvement in detecting the attacks.We tested the model on four datasets to show the effectiveness of the proposed approach and performed the ablation study to check the effect of each parameter.Also,the proposed model is computationaly efficient.To support transparency and trust in decision-making,explainable AI(XAI)techniques are incorporated that provides both global and local insight into feature contributions,and offers intuitive visualizations for individual predictions.This makes it suitable for practical deployment in cybersecurity environments that demand both precision and accountability.展开更多
Motor imbalance is a critical failure mode in rotating machinery,potentially causing severe equipment damage if undetected.Traditional vibration-based diagnostic methods rely on direct sensor contact,leading to instal...Motor imbalance is a critical failure mode in rotating machinery,potentially causing severe equipment damage if undetected.Traditional vibration-based diagnostic methods rely on direct sensor contact,leading to installation challenges and measurement artifacts that can compromise accuracy.This study presents a novel radar-based framework for non-contact motor imbalance detection using 24 GHz continuous-wave radar.A dataset of 1802 experimental trials was sourced,covering four imbalance levels(0,10,20,30 g)across varying motor speeds(500–1500 rpm)and load torques(0–3 Nm).Dual-channel in-phase and quadrature radar signals were captured at 10,000 samples per second for 30-s intervals,preserving both amplitude and phase information for analysis.A multi-domain feature extraction methodology captured imbalance signatures in time,frequency,and complex signal domains.From 65 initial features,statistical analysis using Kruskal–Wallis tests identified significant descriptors,and recursive feature elimination with Random Forest reduced the feature set to 20 dimensions,achieving 69%dimensionality reduction without loss of performance.Six machine learning algorithms,Random Forest,Extra Trees Classifier,Extreme Gradient Boosting,Categorical Boosting,Support Vector Machine with radial basis function kernel,and k-Nearest Neighbors were evaluated with grid-search hyperparameter optimization and five-fold cross-validation.The Extra Trees Classifier achieved the best performance with 98.52%test accuracy,98%cross-validation accuracy,and minimal variance,maintaining per-class precision and recall above 97%.Its superior performance is attributed to its randomized split selection and full bootstrapping strategy,which reduce variance and overfitting while effectively capturing the nonlinear feature interactions and non-normal distributions present in the dataset.The model’s average inference time of 70 ms enables near real-time deployment.Comparative analysis demonstrates that the radar-based framework matches or exceeds traditional contact-based methods while eliminating their inherent limitations,providing a robust,scalable,and noninvasive solution for industrial motor condition monitoring,particularly in hazardous or space-constrained environments.展开更多
In this paper,we propose a systematic approach for accelerating finite element-type methods by machine learning for the numerical solution of partial differential equations(PDEs).The main idea is to use a neural netwo...In this paper,we propose a systematic approach for accelerating finite element-type methods by machine learning for the numerical solution of partial differential equations(PDEs).The main idea is to use a neural network to learn the solution map of the PDEs and to do so in an element-wise fashion.This map takes input of the element geometry and the PDE’s parameters on that element,and gives output of two operators:(1)the in2out operator for inter-element communication,and(2)the in2sol operator(Green’s function)for element-wise solution recovery.A significant advantage of this approach is that,once trained,this network can be used for the numerical solution of the PDE for any domain geometry and any parameter distribution without retraining.Also,the training is significantly simpler since it is done on the element level instead on the entire domain.We call this approach element learning.This method is closely related to hybridizable discontinuous Galerkin(HDG)methods in the sense that the local solvers of HDG are replaced by machine learning approaches.Numerical tests are presented for an example PDE,the radiative transfer or radiation transport equation,in a variety of scenarios with idealized or realistic cloud fields,with smooth or sharp gradient in the cloud boundary transition.Under a fixed accuracy level of 10^(−3) in the relative L^(2) error,and polynomial degree p=6 in each element,we observe an approximately 5 to 10 times speed-up by element learning compared to a classical finite element-type method.展开更多
BACKGROUND:This study aims to develop and validate a machine learning-based in-hospital mortality predictive model for acute aortic syndrome(AAS)in the emergency department(ED)and to derive a simplifi ed version suita...BACKGROUND:This study aims to develop and validate a machine learning-based in-hospital mortality predictive model for acute aortic syndrome(AAS)in the emergency department(ED)and to derive a simplifi ed version suitable for rapid clinical application.METHODS:In this multi-center retrospective cohort study,AAS patient data from three hospitals were analyzed.The modeling cohort included data from the First Affiliated Hospital of Zhengzhou University and the People’s Hospital of Xinjiang Uygur Autonomous Region,with Peking University Third Hospital data serving as the external test set.Four machine learning algorithms—logistic regression(LR),multilayer perceptron(MLP),Gaussian naive Bayes(GNB),and random forest(RF)—were used to develop predictive models based on 34 early-accessible clinical variables.A simplifi ed model was then derived based on fi ve key variables(Stanford type,pericardial eff usion,asymmetric peripheral arterial pulsation,decreased bowel sounds,and dyspnea)via Least Absolute Shrinkage and Selection Operator(LASSO)regression to improve ED applicability.RESULTS:A total of 929 patients were included in the modeling cohort,and 210 were included in the external test set.Four machine learning models based on 34 clinical variables were developed,achieving internal and external validation AUCs of 0.85-0.90 and 0.73-0.85,respectively.The simplifi ed model incorporating fi ve key variables demonstrated internal and external validation AUCs of 0.71-0.86 and 0.75-0.78,respectively.Both models showed robust calibration and predictive stability across datasets.CONCLUSION:Both kinds of models were built based on machine learning tools,and proved to have certain prediction performance and extrapolation.展开更多
Accurate assessment of snowpack volumetric liquid water content and bulk density is essential for understanding snow hydrology,avalanche risk management,and monitoring cryosphere changes.This study presents a novel du...Accurate assessment of snowpack volumetric liquid water content and bulk density is essential for understanding snow hydrology,avalanche risk management,and monitoring cryosphere changes.This study presents a novel dual-parameter inversion framework that integrates synthetic electromagnetic modelling,dimensionality reduction,and machine learning algorithms to extract relative permittivity and log-resistivity from ground-penetrating radar(GPR)data.Traditional snowpack measurements are invasive,labor-intensive,and limited to point observations.To overcome these limitations,we developed a non-invasive,scalable,and data-driven framework that uses synthetic GPR datasets representing diverse snowpack conditions with variable moisture and density profiles.Synthetic 1D time series reflections(A-scans)are generated using finite-difference time-domain simulations in the state-of-the-art electromagnetic simulator gprMax.Principal component analysis(PCA)is applied to compress each A-scan while preserving key features,which significantly improved and enhanced the model training efficiency.Four machine learning models,including random forest,neural network,support vector machine,and eXtreme gradient boosting,are trained on PCA-reduced features.Among these,the neural network model achieved the best performance,with R^(2)>0.97 for permittivity and R 2>0.92 for resistivity.Gaussian noise(signal-to-noise ratio of 6 dB)is introduced to the synthetic data,and then targeted domain adaptation is employed to enhance generalization to field data.The framework is validated on two contrasting GPR transects in the Altay Mountains of the Chinese mainland,representing moist(T750)and wet(G125)snowpack conditions.The neural network model predictions are most consistent with the GPR derived estimates,Snowfork measurements,and snow pit data,achieving volumetric liquid water content deviation of≤1.5% and bulk density error within the range of 30-84 kg m^(-3).The results demonstrate that machine learning-based inversion,supported by realistic simulations and data augmentation enables scalable,non-invasive snowpack characterization with significant applications in hydrological forecasting,snow monitoring,and water resource management.展开更多
Providing safe and quality food is crucial for every household and is of extreme significance in the growth of any society.It is a complex procedure that deals with all issues focusing on the development of food proce...Providing safe and quality food is crucial for every household and is of extreme significance in the growth of any society.It is a complex procedure that deals with all issues focusing on the development of food processing from seed to harvest,storage,preparation,and consumption.This current paper seeks to demystify the importance of artificial intelligence,machine learning(ML),deep learning(DL),and computer vision(CV)in ensuring food safety and quality.By stressing the importance of these technologies,the audience will feel reassured and confident in their potential.These are very handy for such problems,giving assurance over food safety.CV is incredibly noble in today's generation because it improves food processing quality and positively impacts firms and researchers.Thus,at the present production stage,rich in image processing and computer visioning is incorporated into all facets of food production.In this field,DL and ML are implemented to identify the type of food in addition to quality.Concerning data and result-oriented perceptions,one has found similarities regarding various approaches.As a result,the findings of this study will be helpful for scholars looking for a proper approach to identify the quality of food offered.It helps to indicate which food products have been discussed by other scholars and lets the reader know papers by other scholars inclined to research further.Also,DL is accurately integrated with identifying the quality and safety of foods in the market.This paper describes the current practices and concerns of ML,DL,and probable trends for its future development.展开更多
Accurate prediction of rockburst intensity levels is crucial for ensuring the safety of deep hard rock engineering construction.This paper introduced an expert system for rockburst intensity level prediction that empl...Accurate prediction of rockburst intensity levels is crucial for ensuring the safety of deep hard rock engineering construction.This paper introduced an expert system for rockburst intensity level prediction that employs machine learning algorithms as the basis for its inference rules.The system comprises four modules:a database,a repository,an inference engine,and an interpreter.A database containing 1114 rockburst cases was used to construct 357 datasets that serve as the repository for the expert system.Additionally,19 types of machine learning algorithms were used to establish 6783 micro-models to construct cognitive rules within the inference engine.By integrating probability theory and marginal analysis,a fuzzy scoring method based on the SoftMax function was developed and applied to the interpreter for rockburst intensity level prediction,effectively restoring the continuity of rockburst characteristics.The research results indicate that ensemble algorithms based on decision trees are more effective in capturing the characteristics of rockburst.Key factors for accurate prediction of rockburst intensity include uniaxial compressive strength,elastic energy index,the maximum principal stress,tangential stress,and their composite indicators.The accuracy of the proposed rockburst intensity level prediction expert system was verified using 20 engineering rockburst cases,with predictions aligning closely with the actual rockburst intensity levels.展开更多
Accurate prediction of concrete compressive strength is fundamental for optimizing mix designs,improving material utilization,and ensuring structural safety in modern construction.Traditional empirical methods often f...Accurate prediction of concrete compressive strength is fundamental for optimizing mix designs,improving material utilization,and ensuring structural safety in modern construction.Traditional empirical methods often fail to capture the non-linear relationships among concrete constituents,especially with the growing use of supple-mentary cementitious materials and recycled aggregates.This study presents an integrated machine learning framework for concrete strength prediction,combining advanced regression models—namely CatBoost—with metaheuristic optimization algorithms,with a particular focus on the Somersaulting Spider Optimizer(SSO).A comprehensive dataset encompassing diverse mix proportions and material types was used to evaluate baseline machine learning models,including CatBoost,XGBoost,ExtraTrees,and RandomForest.Among these,CatBoost demonstrated superior accuracy across multiple performance metrics.To further enhance predictive capability,several bio-inspired optimizers were employed for hyperparameter tuning.The SSO-CatBoost hybrid achieved the lowest mean squared error and highest correlation coefficients,outperforming other metaheuristic approaches such as Genetic Algorithm,Particle Swarm Optimization,and Grey Wolf Optimizer.Statistical significance was established through Analysis of Variance and Wilcoxon signed-rank testing,confirming the robustness of the optimized models.The proposed methodology not only delivers improved predictive performance but also offers a transparent framework for mix design optimization,supporting data-driven decision making in sustainable and resilient infrastructure development.展开更多
Accurate land surface temperature(LST)assessment is crucial for comprehending and reducing the impacts of climate change and understanding land use evolution.This study presents an innovative method by utilizing ensem...Accurate land surface temperature(LST)assessment is crucial for comprehending and reducing the impacts of climate change and understanding land use evolution.This study presents an innovative method by utilizing ensemble models,advanced correlation analysis,and trend analysis to investigate its environmental influences.Google Earth Engine(GEE)was utilized to process the datasets from Landsat-7 and Landsat-8 for the five big cities of Punjab,Pakistan,from 2001 to 2023.Results from this study show significant urban warming trends,and a strong correlation between environmental variables and LST was identified.The ensemble-based three machine learning models,including XGBoost,AdaBoost,and random forest(RF),were adopted to improve the accuracy of LST evaluation.Although XGBoost and AdaBoost attained modest levels of accuracy,with R^(2) values of 0.767 and 0.706,respectively,the RF model outperformed them by achieving an exceptional R^(2) of 0.796 and RMSE of 0.476.Moreover,Pearson correlation analysis revealed a negative relationship between LST and normalized difference latent heat index(NDLI)with r=-0.67,normalized difference vegetation index(NDVI)with r=-0.6,and modified normalized difference water index(MNDWI)with the value of r as -0.57.In addition,wavelet analysis showed that vegetation and water offer long-term LST cooling,lasting up to 64 months,while built-up areas and bare soil contribute to short-term warming,lasting 4 to 8 months.Latent heat indicated variable cooling periods,surpassing 60 months in cities.These findings enhance the understanding of LST changes and the impact of climate change on the environment.展开更多
Nasopharyngeal carcinoma(NPC)is a malignant tumor prevalent in southern China and Southeast Asia,where its early detection is crucial for improving patient prognosis and reducing mortality rates.However,existing scree...Nasopharyngeal carcinoma(NPC)is a malignant tumor prevalent in southern China and Southeast Asia,where its early detection is crucial for improving patient prognosis and reducing mortality rates.However,existing screening methods suffer from limitations in accuracy and accessibility,hindering their application in large-scale population screening.In this work,a surface-enhanced Raman spectroscopy(SERS)-based method was established to explore the profiles of different stratified components in saliva from NPC and healthy subjects after fractionation processing.The study findings indicate that all fractionated samples exhibit diseaseassociated molecular signaling differences,where small-molecule(molecular weight cut-offvalue is 10 kDa)demonstrating superior classification capabilities with sensitivity of 90.5%and speci-ficity of 75.6%,area under receiver operating characteristic(ROC)curve of 0:925±0:031.The primary objective of this study was to qualitatively explore patterns in saliva composition across groups.The proposed SERS detection strategy for fractionated saliva offers novel insights for enhancing the sensitivity and reliability of noninvasive NPC screening,laying the foundation for translational application in large-scale clinical settings.展开更多
This study developed a modeling methodology for statistical optimization-based geologic hazard susceptibility assessment,aiming to enhance the comprehensive performance and classification accuracy of the assessment mo...This study developed a modeling methodology for statistical optimization-based geologic hazard susceptibility assessment,aiming to enhance the comprehensive performance and classification accuracy of the assessment models.First,the cumulative probability method revealed that a low probability(15%)of geologic hazards between any two geologic hazard points occurred outside a buffer zone with a radius of 2297 m(i.e.,the distance threshold).The training dataset was established,consisting of negative samples(non-hazard points)randomly generated based on the distance threshold,positive samples(i.e.,historical hazards),and 13 conditioning factors.Then,models were built using five machine learning algorithms,namely random forest(RF),gradient boosting decision tree(GBDT),naive Bayes(NB),logistic regression(LR),and support vector machine(SVM).The comprehensive performance of the models was assessed using the area under the receiver operating characteristic curve(AUC)and overall accuracy(OA)as indicators,revealing that RF exhibited the best performance,with OA and AUC values of 2.7127 and 0.981,respectively.Furthermore,the machine learning models constructed by considering the distance threshold outperformed those built using the unoptimized dataset.The characteristic factors were ranked using the mutual information method,with their scores decreasing in the order of rainfall(0.1616),altitude(0.06),normalized difference vegetation index(NDVI;0.04),and distance from roads(0.03).Finally,the geologic hazard susceptibility classification was assessed using the natural breaks method combined with a clustering algorithm.The results indicate that the clustering algorithm exhibited higher classification accuracy than the natural breaks method.The findings of this study demonstrate that the proposed model optimization scheme can provide a scientific basis for the prevention and control of geologic hazards.展开更多
Sudden wildfires cause significant global ecological damage.While satellite imagery has advanced early fire detection and mitigation,image-based systems face limitations including high false alarm rates,visual obstruc...Sudden wildfires cause significant global ecological damage.While satellite imagery has advanced early fire detection and mitigation,image-based systems face limitations including high false alarm rates,visual obstructions,and substantial computational demands,especially in complex forest terrains.To address these challenges,this study proposes a novel forest fire detection model utilizing audio classification and machine learning.We developed an audio-based pipeline using real-world environmental sound recordings.Sounds were converted into Mel-spectrograms and classified via a Convolutional Neural Network(CNN),enabling the capture of distinctive fire acoustic signatures(e.g.,crackling,roaring)that are minimally impacted by visual or weather conditions.Internet of Things(IoT)sound sensors were crucial for generating complex environmental parameters to optimize feature extraction.The CNN model achieved high performance in stratified 5-fold cross-validation(92.4%±1.6 accuracy,91.2%±1.8 F1-score)and on test data(94.93%accuracy,93.04%F1-score),with 98.44%precision and 88.32%recall,demonstrating reliability across environmental conditions.These results indicate that the audio-based approach not only improves detection reliability but also markedly reduces computational overhead compared to traditional image-based methods.The findings suggest that acoustic sensing integrated with machine learning offers a powerful,low-cost,and efficient solution for real-time forest fire monitoring in complex,dynamic environments.展开更多
Delayed wound healing following radical gastrectomy remains an important yet underappreciated complication that prolongs hospitalization,increases costs,and undermines patient recovery.In An et al’s recent study,the ...Delayed wound healing following radical gastrectomy remains an important yet underappreciated complication that prolongs hospitalization,increases costs,and undermines patient recovery.In An et al’s recent study,the authors present a machine learning-based risk prediction approach using routinely available clinical and laboratory parameters.Among the evaluated algorithms,a decision tree model demonstrated excellent discrimination,achieving an area under the curve of 0.951 in the validation set and notably identifying all true cases of delayed wound healing at the Youden index threshold.The inclusion of variables such as drainage duration,preoperative white blood cell and neutrophil counts,alongside age and sex,highlights the pragmatic appeal of the model for early postoperative monitoring.Nevertheless,several aspects warrant critical reflection,including the reliance on a postoperative variable(drainage duration),internal validation only,and certain reporting inconsistencies.This letter underscores both the promise and the limitations of adopting interpretable machine learning models in perioperative care.We advocate for transparent reporting,external validation,and careful consideration of clinically actionable timepoints before integration into practice.Ultimately,this work represents a valuable step toward precision risk stratification in gastric cancer surgery,and sets the stage for multicenter,prospective evaluations.展开更多
基金Guangzhou Metro Scientific Research Project(No.JT204-100111-23001)Chongqing Municipal Special Project for Technological Innovation and Application Development(No.CSTB2022TIAD-KPX0101)Science and Technology Research and Development Program of China State Railway Group Co.,Ltd.(No.N2023G045)。
文摘The uplift resistance of the soil overlying shield tunnels significantly impacts their anti-floating stability.However,research on uplift resistance concerning special-shaped shield tunnels is limited.This study combines numerical simulation with machine learning techniques to explore this issue.It presents a summary of special-shaped tunnel geometries and introduces a shape coefficient.Through the finite element software,Plaxis3D,the study simulates six key parameters—shape coefficient,burial depth ratio,tunnel’s longest horizontal length,internal friction angle,cohesion,and soil submerged bulk density—that impact uplift resistance across different conditions.Employing XGBoost and ANN methods,the feature importance of each parameter was analyzed based on the numerical simulation results.The findings demonstrate that a tunnel shape more closely resembling a circle leads to reduced uplift resistance in the overlying soil,whereas other parameters exhibit the contrary effects.Furthermore,the study reveals a diminishing trend in the feature importance of buried depth ratio,internal friction angle,tunnel longest horizontal length,cohesion,soil submerged bulk density,and shape coefficient in influencing uplift resistance.
基金Supported by CAS Basic and Interdisciplinary Frontier Scientific Research Pilot Project(XDB1190300,XDB1190302)Youth Innovation Promotion Association CAS(Y2021056)+1 种基金Joint Fund of the Yulin University and the Dalian National Laboratory for Clean Energy(YLU-DNL Fund 2022007)The special fund for Science and Technology Innovation Teams of Shanxi Province(202304051001007)。
文摘Cyclohexene is an important raw material in the production of nylon.Selective hydrogenation of benzene is a key method for preparing cyclohexene.However,the Ru catalysts used in current industrial processes still face challenges,including high metal usage,high process costs,and low cyclohexene yield.This study utilizes existing literature data combined with machine learning methods to analyze the factors influencing benzene conversion,cyclohexene selectivity,and yield in the benzene hydrogenation to cyclohexene reaction.It constructs predictive models based on XGBoost and Random Forest algorithms.After analysis,it was found that reaction time,Ru content,and space velocity are key factors influencing cyclohexene yield,selectivity,and benzene conversion.Shapley Additive Explanations(SHAP)analysis and feature importance analysis further revealed the contribution of each variable to the reaction outcomes.Additionally,we randomly generated one million variable combinations using the Dirichlet distribution to attempt to predict high-yield catalyst formulations.This paper provides new insights into the application of machine learning in heterogeneous catalysis and offers some reference for further research.
基金funded by the National Natural Science Foundation of China(No.22278443)CAMS Innovation Fund for Medical Sciences(No.2022-I2M-1-015)+3 种基金the Key R&D Program of Shandong Province(No.2021ZDSYS26)Xinjiang Uygur Autonomous Region Innovation Environment Construction Special Fund and Technology Innovation Base Construction Key Laboratory Open Project(No.2023D04065)2023 Xinjiang Uygur Autonomous Region Innovation Tianchi Talent Introduction Program for financial supportthe Key Project of Natural Science of Bengbu Medical University(No.2024byzd138).
文摘In the pharmaceutical field,machine learning can play an important role in drug development,production and treatment.Co-crystallization techniques have shown promising potential to enhance the properties of active pharmaceutical ingredients(APIs)such as solubility,permeability,and bioavailability,all without altering their chemical structure.This approach opens new avenues for developing natural products into effective drugs,especially those previously challenging in formulation.Emodin,an anthraquinone-based natural product,is a notable example due to its diverse biological activities;however,its physicochemical limitations,such as poor solubility and easy sublimation,restricted its clinical application.While various methods have improved emodin's physicochemical properties,research on its bioavailability remains limited.In our study,we summarize cocrystals and salts produced through co-crystallization technology and identify piperazine as a favorable coformer.Conflicting conclusions from computational chemistry and molecular modeling method and machine learning method regarding the formation of an emodin-piperazine cocrystal or salt led us to experimentally validate these possibilities.Ultimately,we successfully obtained the emodin-piperazine cocrystal,which were characterized and evaluated by several in vitro methods and pharmacokinetic studies.In addition,experiments have shown that emodin has a certain therapeutic effect on sepsis,so we also evaluated emodin-piperazine biological activity in a sepsis model.The results demonstrate that co-crystallization significantly enhances emodin's solubility,permeability,and bioavailability.Pharmacodynamic studies indicate that the emodin-piperazine cocrystal improves sepsis symptoms and provides protective effects against liver and kidney damage associated with sepsis.This study offers renewed hope for natural products with broad biological activities yet hindered by physicochemical limitations by advancing co-crystallization as a viable development approach.
基金supported by Project of National and Local Joint Engineering Research Center for Biomass Energy Development and Utilization(Harbin Institute of Technology,No.2021A004).
文摘Machine learning(ML)is recognized as a potent tool for the inverse design of environmental functional material,particularly for complex entities like biochar-based catalysts(BCs).Thus,the tailored BCs can have a distinct ability to trigger the nonradical pathway in advance oxidation processes(AOPs),promising a stable,rapid and selective degradation of persistent contaminants.However,due to the inherent“black box”nature and limitations of input features,results and conclusions derived from ML may not always be intuitively understood or comprehensively validated.To tackle this challenge,we linked the front-point interpretable analysis approaches with back-point density functional theory(DFT)calculations to form a chained learning strategy for deeper sight into the intrinsic activation mechanism of BCs in AOPs.At the front point,we conducted an easy-to-interpret meta-analysis to validate two strategies for enhancing nonradical pathways by increasing oxygen content and specific surface area(SSA),and prepared oxidized biochar(OBC500)and SSA-increased biochar(SBC900)by controlling pyrolysis conditions and modification methods.Subsequently,experimental results showed that OBC500 and SBC900 had distinct dominant degradation pathways for 1O2 generation and electron transfer,respectively.Finally,at the end point,DFT calculations revealed their active sites and degradation mechanisms.This chained learning strategy elucidates fundamental principles for BC inverse design and showcases the exceptional capacity to integrate computational techniques to accelerate catalyst inverse design.
基金extend their gratitude to the Deanship of Scientific Research,Vice Presidency for Graduate Studies and Scientific Research,King Faisal University,Saudi Arabia,for funding the publication of this work under the Ambitious Researcher program(Project No.KFU253806).
文摘Artificial intelligence(AI)based models have been used to predict the structural,optical,mechanical,and electrochemical properties of zinc oxide/graphene oxide nanocomposites.Machine learning(ML)models such as Artificial Neural Networks(ANN),Support Vector Regression(SVR),Multilayer Perceptron(MLP),and hybrid,along with fuzzy logic tools,were applied to predict the different properties like wavelength at maximum intensity(444 nm),crystallite size(17.50 nm),and optical bandgap(2.85 eV).While some other properties,such as energy density,power density,and charge transfer resistance,were also predicted with the help of datasets of 1000(80:20).In general,the energy parameters were predicted more accurately by hybrid models.The hydrothermal method was used to synthesize graphene oxide(GO)and zinc oxide(ZnO)nanocomposites.The increased surface area,conductivity,and stability of graphene oxide in zinc oxide nanoparticles make the composite an ideal option for energy storage.X-ray diffraction(XRD)confirmed the crystallite size of 17.41 nm for the nanocomposite and the presence of GO(12.8○)peaks.The scanning electron microscope(SEM)showed anchored wrinkled GO sheets on zinc oxide with an average particle size of 2.93μm.Energy-dispersive X-ray spectroscopy(EDX)confirmed the elemental composition,and Fouriertransform infrared spectroscopy(FTIR)revealed the impact of GO on functional groups and electrochemical behavior.Photoluminescence(PL)wavelength of(439 nm)and band gap of(2.81 eV)show that the material is suitable for energy applications in nanocomposites.Smart nanocomposite materials with improved performance in energy storage and related applications were fabricated by combining synthesis,characterization,fuzzy logic,and machine learning in this work.
文摘Post-kidney transplant rejection is a critical factor influencing transplant success rates and the survival of transplanted organs.With the rapid advancement of artificial intelligence technologies,machine learning(ML)has emerged as a powerful data analysis tool,widely applied in the prediction,diagnosis,and mechanistic study of kidney transplant rejection.This mini-review systematically summarizes the recent applications of ML techniques in post-kidney transplant rejection,covering areas such as the construction of predictive models,identification of biomarkers,analysis of pathological images,assessment of immune cell infiltration,and formulation of personalized treatment strategies.By integrating multi-omics data and clinical information,ML has significantly enhanced the accuracy of early rejection diagnosis and the capability for prognostic evaluation,driving the development of precision medicine in the field of kidney transplantation.Furthermore,this article discusses the challenges faced in existing research and potential future directions,providing a theoretical basis and technical references for related studies.
基金funded by the Research,Development,and Innovation Authority(RDIA)—Kingdom of Saudi Arabia(Grant No.13292-psu-2023-PSNU-R-3-1-EF-).
文摘Colorectal cancer is the third most diagnosed cancer worldwide,and immune checkpoint inhibitors have shown promising therapeutic outcomes in selected patient groups.This study performed a comprehensive analysis of multi-omics data from The Cancer Genome Atlas colorectal adenocarcinoma cohort(TCGA-COADREAD),accessed through cBioPortal,to develop machine learning models for predicting progression-free survival(PFS)following immunotherapy.The dataset included clinical variables,genomic alterations in Kirsten Rat Sarcoma Viral Oncogene Homolog(KRAS),B-Raf Proto-Oncogene(BRAF),and Neuroblastoma RAS Viral Oncogene Homolog(NRAS),microsatellite instability(MSI)status,tumor mutation burden(TMB),and expression of immune checkpoint genes.Kaplan–Meier analysis showed that KRAS mutations were significantly associated with reduced PFS,while BRAF and NRAS mutations had no significant impact.MSI-high tumors exhibited elevated TMB and increased immune checkpoint expression,reflecting their immunologically active phenotype.We developed both survival and classification models,with the Extra Trees classifier achieving the best performance(accuracy=0.86,precision=0.67,recall=0.70,F1-score=0.68,AUC=0.84).These findings highlight the potential of combining genomic and immune biomarkers with machine learning to improve patient stratification and guide personalized immunotherapy decisions.An interactive web application was also developed to enable clinicians to input patient-specific molecular and clinical data and visualize individualized PFS predictions,supporting timely,data-driven treatment planning.
基金funded by Princess Nourah bint Abdulrahman University Researchers Supporting Project number(PNURSP2025R104)Princess Nourah bint Abdulrahman University,Riyadh,Saudi Arabia.
文摘Modern intrusion detection systems(MIDS)face persistent challenges in coping with the rapid evolution of cyber threats,high-volume network traffic,and imbalanced datasets.Traditional models often lack the robustness and explainability required to detect novel and sophisticated attacks effectively.This study introduces an advanced,explainable machine learning framework for multi-class IDS using the KDD99 and IDS datasets,which reflects real-world network behavior through a blend of normal and diverse attack classes.The methodology begins with sophisticated data preprocessing,incorporating both RobustScaler and QuantileTransformer to address outliers and skewed feature distributions,ensuring standardized and model-ready inputs.Critical dimensionality reduction is achieved via the Harris Hawks Optimization(HHO)algorithm—a nature-inspired metaheuristic modeled on hawks’hunting strategies.HHO efficiently identifies the most informative features by optimizing a fitness function based on classification performance.Following feature selection,the SMOTE is applied to the training data to resolve class imbalance by synthetically augmenting underrepresented attack types.The stacked architecture is then employed,combining the strengths of XGBoost,SVM,and RF as base learners.This layered approach improves prediction robustness and generalization by balancing bias and variance across diverse classifiers.The model was evaluated using standard classification metrics:precision,recall,F1-score,and overall accuracy.The best overall performance was recorded with an accuracy of 99.44%for UNSW-NB15,demonstrating the model’s effectiveness.After balancing,the model demonstrated a clear improvement in detecting the attacks.We tested the model on four datasets to show the effectiveness of the proposed approach and performed the ablation study to check the effect of each parameter.Also,the proposed model is computationaly efficient.To support transparency and trust in decision-making,explainable AI(XAI)techniques are incorporated that provides both global and local insight into feature contributions,and offers intuitive visualizations for individual predictions.This makes it suitable for practical deployment in cybersecurity environments that demand both precision and accountability.
基金funded by Princess Nourah bint Abdulrahman University Researchers Support-ing Project number(PNURSP2026R346)Princess Nourah bint Abdulrahman University,Riyadh,Saudi Arabia.
文摘Motor imbalance is a critical failure mode in rotating machinery,potentially causing severe equipment damage if undetected.Traditional vibration-based diagnostic methods rely on direct sensor contact,leading to installation challenges and measurement artifacts that can compromise accuracy.This study presents a novel radar-based framework for non-contact motor imbalance detection using 24 GHz continuous-wave radar.A dataset of 1802 experimental trials was sourced,covering four imbalance levels(0,10,20,30 g)across varying motor speeds(500–1500 rpm)and load torques(0–3 Nm).Dual-channel in-phase and quadrature radar signals were captured at 10,000 samples per second for 30-s intervals,preserving both amplitude and phase information for analysis.A multi-domain feature extraction methodology captured imbalance signatures in time,frequency,and complex signal domains.From 65 initial features,statistical analysis using Kruskal–Wallis tests identified significant descriptors,and recursive feature elimination with Random Forest reduced the feature set to 20 dimensions,achieving 69%dimensionality reduction without loss of performance.Six machine learning algorithms,Random Forest,Extra Trees Classifier,Extreme Gradient Boosting,Categorical Boosting,Support Vector Machine with radial basis function kernel,and k-Nearest Neighbors were evaluated with grid-search hyperparameter optimization and five-fold cross-validation.The Extra Trees Classifier achieved the best performance with 98.52%test accuracy,98%cross-validation accuracy,and minimal variance,maintaining per-class precision and recall above 97%.Its superior performance is attributed to its randomized split selection and full bootstrapping strategy,which reduce variance and overfitting while effectively capturing the nonlinear feature interactions and non-normal distributions present in the dataset.The model’s average inference time of 70 ms enables near real-time deployment.Comparative analysis demonstrates that the radar-based framework matches or exceeds traditional contact-based methods while eliminating their inherent limitations,providing a robust,scalable,and noninvasive solution for industrial motor condition monitoring,particularly in hazardous or space-constrained environments.
基金partially supported by the NSF(Grant No.DMS 2324368)by the Office of the Vice Chancellor for Research and Graduate Education at the University of Wisconsin-Madison with funding from the Wisconsin Alumni Research Foundation.
文摘In this paper,we propose a systematic approach for accelerating finite element-type methods by machine learning for the numerical solution of partial differential equations(PDEs).The main idea is to use a neural network to learn the solution map of the PDEs and to do so in an element-wise fashion.This map takes input of the element geometry and the PDE’s parameters on that element,and gives output of two operators:(1)the in2out operator for inter-element communication,and(2)the in2sol operator(Green’s function)for element-wise solution recovery.A significant advantage of this approach is that,once trained,this network can be used for the numerical solution of the PDE for any domain geometry and any parameter distribution without retraining.Also,the training is significantly simpler since it is done on the element level instead on the entire domain.We call this approach element learning.This method is closely related to hybridizable discontinuous Galerkin(HDG)methods in the sense that the local solvers of HDG are replaced by machine learning approaches.Numerical tests are presented for an example PDE,the radiative transfer or radiation transport equation,in a variety of scenarios with idealized or realistic cloud fields,with smooth or sharp gradient in the cloud boundary transition.Under a fixed accuracy level of 10^(−3) in the relative L^(2) error,and polynomial degree p=6 in each element,we observe an approximately 5 to 10 times speed-up by element learning compared to a classical finite element-type method.
基金supported by the special fund of the National Clinical Key Specialty Construction Program[(2022)301-2305].
文摘BACKGROUND:This study aims to develop and validate a machine learning-based in-hospital mortality predictive model for acute aortic syndrome(AAS)in the emergency department(ED)and to derive a simplifi ed version suitable for rapid clinical application.METHODS:In this multi-center retrospective cohort study,AAS patient data from three hospitals were analyzed.The modeling cohort included data from the First Affiliated Hospital of Zhengzhou University and the People’s Hospital of Xinjiang Uygur Autonomous Region,with Peking University Third Hospital data serving as the external test set.Four machine learning algorithms—logistic regression(LR),multilayer perceptron(MLP),Gaussian naive Bayes(GNB),and random forest(RF)—were used to develop predictive models based on 34 early-accessible clinical variables.A simplifi ed model was then derived based on fi ve key variables(Stanford type,pericardial eff usion,asymmetric peripheral arterial pulsation,decreased bowel sounds,and dyspnea)via Least Absolute Shrinkage and Selection Operator(LASSO)regression to improve ED applicability.RESULTS:A total of 929 patients were included in the modeling cohort,and 210 were included in the external test set.Four machine learning models based on 34 clinical variables were developed,achieving internal and external validation AUCs of 0.85-0.90 and 0.73-0.85,respectively.The simplifi ed model incorporating fi ve key variables demonstrated internal and external validation AUCs of 0.71-0.86 and 0.75-0.78,respectively.Both models showed robust calibration and predictive stability across datasets.CONCLUSION:Both kinds of models were built based on machine learning tools,and proved to have certain prediction performance and extrapolation.
基金supported by the National Key R&D Program of China(Grant Nos.2023YFC3008300&2023YFC3008305)the National Natural Science Foundation of China(Grant No.42172320)+1 种基金the Key Laboratory of Mountain Hazards and Engineering Resilience,Institute of Mountain Hazards and Environment,Chinese Academy of Sciences(Grant Nos.KLMHER-Z06&KLMHER-T07)the Science and Technology Research Program of Institute of Mountain Hazards and Environment,Chinese Academy of Sciences(Grant No.IMHE-CXTD.04).
文摘Accurate assessment of snowpack volumetric liquid water content and bulk density is essential for understanding snow hydrology,avalanche risk management,and monitoring cryosphere changes.This study presents a novel dual-parameter inversion framework that integrates synthetic electromagnetic modelling,dimensionality reduction,and machine learning algorithms to extract relative permittivity and log-resistivity from ground-penetrating radar(GPR)data.Traditional snowpack measurements are invasive,labor-intensive,and limited to point observations.To overcome these limitations,we developed a non-invasive,scalable,and data-driven framework that uses synthetic GPR datasets representing diverse snowpack conditions with variable moisture and density profiles.Synthetic 1D time series reflections(A-scans)are generated using finite-difference time-domain simulations in the state-of-the-art electromagnetic simulator gprMax.Principal component analysis(PCA)is applied to compress each A-scan while preserving key features,which significantly improved and enhanced the model training efficiency.Four machine learning models,including random forest,neural network,support vector machine,and eXtreme gradient boosting,are trained on PCA-reduced features.Among these,the neural network model achieved the best performance,with R^(2)>0.97 for permittivity and R 2>0.92 for resistivity.Gaussian noise(signal-to-noise ratio of 6 dB)is introduced to the synthetic data,and then targeted domain adaptation is employed to enhance generalization to field data.The framework is validated on two contrasting GPR transects in the Altay Mountains of the Chinese mainland,representing moist(T750)and wet(G125)snowpack conditions.The neural network model predictions are most consistent with the GPR derived estimates,Snowfork measurements,and snow pit data,achieving volumetric liquid water content deviation of≤1.5% and bulk density error within the range of 30-84 kg m^(-3).The results demonstrate that machine learning-based inversion,supported by realistic simulations and data augmentation enables scalable,non-invasive snowpack characterization with significant applications in hydrological forecasting,snow monitoring,and water resource management.
文摘Providing safe and quality food is crucial for every household and is of extreme significance in the growth of any society.It is a complex procedure that deals with all issues focusing on the development of food processing from seed to harvest,storage,preparation,and consumption.This current paper seeks to demystify the importance of artificial intelligence,machine learning(ML),deep learning(DL),and computer vision(CV)in ensuring food safety and quality.By stressing the importance of these technologies,the audience will feel reassured and confident in their potential.These are very handy for such problems,giving assurance over food safety.CV is incredibly noble in today's generation because it improves food processing quality and positively impacts firms and researchers.Thus,at the present production stage,rich in image processing and computer visioning is incorporated into all facets of food production.In this field,DL and ML are implemented to identify the type of food in addition to quality.Concerning data and result-oriented perceptions,one has found similarities regarding various approaches.As a result,the findings of this study will be helpful for scholars looking for a proper approach to identify the quality of food offered.It helps to indicate which food products have been discussed by other scholars and lets the reader know papers by other scholars inclined to research further.Also,DL is accurately integrated with identifying the quality and safety of foods in the market.This paper describes the current practices and concerns of ML,DL,and probable trends for its future development.
基金Project(42077244)supported by the National Natural Science Foundation of ChinaProject(2020-05)supported by the Open Research Fund of Guangdong Provincial Key Laboratory of Deep Earth Sciences and Geothermal Energy Exploitation and Utilization,China。
文摘Accurate prediction of rockburst intensity levels is crucial for ensuring the safety of deep hard rock engineering construction.This paper introduced an expert system for rockburst intensity level prediction that employs machine learning algorithms as the basis for its inference rules.The system comprises four modules:a database,a repository,an inference engine,and an interpreter.A database containing 1114 rockburst cases was used to construct 357 datasets that serve as the repository for the expert system.Additionally,19 types of machine learning algorithms were used to establish 6783 micro-models to construct cognitive rules within the inference engine.By integrating probability theory and marginal analysis,a fuzzy scoring method based on the SoftMax function was developed and applied to the interpreter for rockburst intensity level prediction,effectively restoring the continuity of rockburst characteristics.The research results indicate that ensemble algorithms based on decision trees are more effective in capturing the characteristics of rockburst.Key factors for accurate prediction of rockburst intensity include uniaxial compressive strength,elastic energy index,the maximum principal stress,tangential stress,and their composite indicators.The accuracy of the proposed rockburst intensity level prediction expert system was verified using 20 engineering rockburst cases,with predictions aligning closely with the actual rockburst intensity levels.
文摘Accurate prediction of concrete compressive strength is fundamental for optimizing mix designs,improving material utilization,and ensuring structural safety in modern construction.Traditional empirical methods often fail to capture the non-linear relationships among concrete constituents,especially with the growing use of supple-mentary cementitious materials and recycled aggregates.This study presents an integrated machine learning framework for concrete strength prediction,combining advanced regression models—namely CatBoost—with metaheuristic optimization algorithms,with a particular focus on the Somersaulting Spider Optimizer(SSO).A comprehensive dataset encompassing diverse mix proportions and material types was used to evaluate baseline machine learning models,including CatBoost,XGBoost,ExtraTrees,and RandomForest.Among these,CatBoost demonstrated superior accuracy across multiple performance metrics.To further enhance predictive capability,several bio-inspired optimizers were employed for hyperparameter tuning.The SSO-CatBoost hybrid achieved the lowest mean squared error and highest correlation coefficients,outperforming other metaheuristic approaches such as Genetic Algorithm,Particle Swarm Optimization,and Grey Wolf Optimizer.Statistical significance was established through Analysis of Variance and Wilcoxon signed-rank testing,confirming the robustness of the optimized models.The proposed methodology not only delivers improved predictive performance but also offers a transparent framework for mix design optimization,supporting data-driven decision making in sustainable and resilient infrastructure development.
基金supported by the National Natural Science Foundation of China(Grant Nos.52479045,52279042)the Key Research and Development Program in Guangxi(Grant No.AB23026021)the Open Research Fund of Guangxi Key Laboratory of Water Engineering Materials and Structures,Guangxi Institute of Water Resources Research(Grant No.GXHRIWEMS-2022-07).
文摘Accurate land surface temperature(LST)assessment is crucial for comprehending and reducing the impacts of climate change and understanding land use evolution.This study presents an innovative method by utilizing ensemble models,advanced correlation analysis,and trend analysis to investigate its environmental influences.Google Earth Engine(GEE)was utilized to process the datasets from Landsat-7 and Landsat-8 for the five big cities of Punjab,Pakistan,from 2001 to 2023.Results from this study show significant urban warming trends,and a strong correlation between environmental variables and LST was identified.The ensemble-based three machine learning models,including XGBoost,AdaBoost,and random forest(RF),were adopted to improve the accuracy of LST evaluation.Although XGBoost and AdaBoost attained modest levels of accuracy,with R^(2) values of 0.767 and 0.706,respectively,the RF model outperformed them by achieving an exceptional R^(2) of 0.796 and RMSE of 0.476.Moreover,Pearson correlation analysis revealed a negative relationship between LST and normalized difference latent heat index(NDLI)with r=-0.67,normalized difference vegetation index(NDVI)with r=-0.6,and modified normalized difference water index(MNDWI)with the value of r as -0.57.In addition,wavelet analysis showed that vegetation and water offer long-term LST cooling,lasting up to 64 months,while built-up areas and bare soil contribute to short-term warming,lasting 4 to 8 months.Latent heat indicated variable cooling periods,surpassing 60 months in cities.These findings enhance the understanding of LST changes and the impact of climate change on the environment.
基金financially supported by National Natural Science Foundation ofChina(No.12374405)Provincial Science Foundation for Distinguished Young Scholars of Fujian(No.2024J010024)+1 种基金Natural Science Foundation of Fujian Province of China(No.2023J011267)Major Research Projects for Young and Middle-aged Researchers of Fujian Provincial Health Commission(No.2021ZQNZD010).
文摘Nasopharyngeal carcinoma(NPC)is a malignant tumor prevalent in southern China and Southeast Asia,where its early detection is crucial for improving patient prognosis and reducing mortality rates.However,existing screening methods suffer from limitations in accuracy and accessibility,hindering their application in large-scale population screening.In this work,a surface-enhanced Raman spectroscopy(SERS)-based method was established to explore the profiles of different stratified components in saliva from NPC and healthy subjects after fractionation processing.The study findings indicate that all fractionated samples exhibit diseaseassociated molecular signaling differences,where small-molecule(molecular weight cut-offvalue is 10 kDa)demonstrating superior classification capabilities with sensitivity of 90.5%and speci-ficity of 75.6%,area under receiver operating characteristic(ROC)curve of 0:925±0:031.The primary objective of this study was to qualitatively explore patterns in saliva composition across groups.The proposed SERS detection strategy for fractionated saliva offers novel insights for enhancing the sensitivity and reliability of noninvasive NPC screening,laying the foundation for translational application in large-scale clinical settings.
基金supported by a project entitled Loess Plateau Region-Watershed-Slope Geological Hazard Multi-Scale Collaborative Intelligent Early Warning System of the National Key R&D Program of China(2022YFC3003404)a project of the Shaanxi Youth Science and Technology Star(2021KJXX-87)public welfare geological survey projects of Shaanxi Institute of Geologic Survey(20180301,201918,202103,and 202413)。
文摘This study developed a modeling methodology for statistical optimization-based geologic hazard susceptibility assessment,aiming to enhance the comprehensive performance and classification accuracy of the assessment models.First,the cumulative probability method revealed that a low probability(15%)of geologic hazards between any two geologic hazard points occurred outside a buffer zone with a radius of 2297 m(i.e.,the distance threshold).The training dataset was established,consisting of negative samples(non-hazard points)randomly generated based on the distance threshold,positive samples(i.e.,historical hazards),and 13 conditioning factors.Then,models were built using five machine learning algorithms,namely random forest(RF),gradient boosting decision tree(GBDT),naive Bayes(NB),logistic regression(LR),and support vector machine(SVM).The comprehensive performance of the models was assessed using the area under the receiver operating characteristic curve(AUC)and overall accuracy(OA)as indicators,revealing that RF exhibited the best performance,with OA and AUC values of 2.7127 and 0.981,respectively.Furthermore,the machine learning models constructed by considering the distance threshold outperformed those built using the unoptimized dataset.The characteristic factors were ranked using the mutual information method,with their scores decreasing in the order of rainfall(0.1616),altitude(0.06),normalized difference vegetation index(NDVI;0.04),and distance from roads(0.03).Finally,the geologic hazard susceptibility classification was assessed using the natural breaks method combined with a clustering algorithm.The results indicate that the clustering algorithm exhibited higher classification accuracy than the natural breaks method.The findings of this study demonstrate that the proposed model optimization scheme can provide a scientific basis for the prevention and control of geologic hazards.
基金funded by the Directorate of Research and Community Service,Directorate General of Research and Development,Ministry of Higher Education,Science and Technologyin accordance with the Implementation Contract for the Operational Assistance Program for State Universities,Research Program Number:109/C3/DT.05.00/PL/2025.
文摘Sudden wildfires cause significant global ecological damage.While satellite imagery has advanced early fire detection and mitigation,image-based systems face limitations including high false alarm rates,visual obstructions,and substantial computational demands,especially in complex forest terrains.To address these challenges,this study proposes a novel forest fire detection model utilizing audio classification and machine learning.We developed an audio-based pipeline using real-world environmental sound recordings.Sounds were converted into Mel-spectrograms and classified via a Convolutional Neural Network(CNN),enabling the capture of distinctive fire acoustic signatures(e.g.,crackling,roaring)that are minimally impacted by visual or weather conditions.Internet of Things(IoT)sound sensors were crucial for generating complex environmental parameters to optimize feature extraction.The CNN model achieved high performance in stratified 5-fold cross-validation(92.4%±1.6 accuracy,91.2%±1.8 F1-score)and on test data(94.93%accuracy,93.04%F1-score),with 98.44%precision and 88.32%recall,demonstrating reliability across environmental conditions.These results indicate that the audio-based approach not only improves detection reliability but also markedly reduces computational overhead compared to traditional image-based methods.The findings suggest that acoustic sensing integrated with machine learning offers a powerful,low-cost,and efficient solution for real-time forest fire monitoring in complex,dynamic environments.
文摘Delayed wound healing following radical gastrectomy remains an important yet underappreciated complication that prolongs hospitalization,increases costs,and undermines patient recovery.In An et al’s recent study,the authors present a machine learning-based risk prediction approach using routinely available clinical and laboratory parameters.Among the evaluated algorithms,a decision tree model demonstrated excellent discrimination,achieving an area under the curve of 0.951 in the validation set and notably identifying all true cases of delayed wound healing at the Youden index threshold.The inclusion of variables such as drainage duration,preoperative white blood cell and neutrophil counts,alongside age and sex,highlights the pragmatic appeal of the model for early postoperative monitoring.Nevertheless,several aspects warrant critical reflection,including the reliance on a postoperative variable(drainage duration),internal validation only,and certain reporting inconsistencies.This letter underscores both the promise and the limitations of adopting interpretable machine learning models in perioperative care.We advocate for transparent reporting,external validation,and careful consideration of clinically actionable timepoints before integration into practice.Ultimately,this work represents a valuable step toward precision risk stratification in gastric cancer surgery,and sets the stage for multicenter,prospective evaluations.