Modern intrusion detection systems(MIDS)face persistent challenges in coping with the rapid evolution of cyber threats,high-volume network traffic,and imbalanced datasets.Traditional models often lack the robustness a...Modern intrusion detection systems(MIDS)face persistent challenges in coping with the rapid evolution of cyber threats,high-volume network traffic,and imbalanced datasets.Traditional models often lack the robustness and explainability required to detect novel and sophisticated attacks effectively.This study introduces an advanced,explainable machine learning framework for multi-class IDS using the KDD99 and IDS datasets,which reflects real-world network behavior through a blend of normal and diverse attack classes.The methodology begins with sophisticated data preprocessing,incorporating both RobustScaler and QuantileTransformer to address outliers and skewed feature distributions,ensuring standardized and model-ready inputs.Critical dimensionality reduction is achieved via the Harris Hawks Optimization(HHO)algorithm—a nature-inspired metaheuristic modeled on hawks’hunting strategies.HHO efficiently identifies the most informative features by optimizing a fitness function based on classification performance.Following feature selection,the SMOTE is applied to the training data to resolve class imbalance by synthetically augmenting underrepresented attack types.The stacked architecture is then employed,combining the strengths of XGBoost,SVM,and RF as base learners.This layered approach improves prediction robustness and generalization by balancing bias and variance across diverse classifiers.The model was evaluated using standard classification metrics:precision,recall,F1-score,and overall accuracy.The best overall performance was recorded with an accuracy of 99.44%for UNSW-NB15,demonstrating the model’s effectiveness.After balancing,the model demonstrated a clear improvement in detecting the attacks.We tested the model on four datasets to show the effectiveness of the proposed approach and performed the ablation study to check the effect of each parameter.Also,the proposed model is computationaly efficient.To support transparency and trust in decision-making,explainable AI(XAI)techniques are incorporated that provides both global and local insight into feature contributions,and offers intuitive visualizations for individual predictions.This makes it suitable for practical deployment in cybersecurity environments that demand both precision and accountability.展开更多
Landslides pose a formidable natural hazard across the Qinghai-Tibet Plateau(QTP),endangering both ecosystems and human life.Identifying the driving factors behind landslides and accurately assessing susceptibility ar...Landslides pose a formidable natural hazard across the Qinghai-Tibet Plateau(QTP),endangering both ecosystems and human life.Identifying the driving factors behind landslides and accurately assessing susceptibility are key to mitigating disaster risk.This study integrated multi-source historical landslide data with 15 predictive factors and used several machine learning models—Random Forest(RF),Gradient Boosting Regression Trees(GBRT),Extreme Gradient Boosting(XGBoost),and Categorical Boosting(CatBoost)—to generate susceptibility maps.The Shapley additive explanation(SHAP)method was applied to quantify factor importance and explore their nonlinear effects.The results showed that:(1)CatBoost was the best-performing model(CA=0.938,AUC=0.980)in assessing landslide susceptibility,with altitude emerging as the most significant factor,followed by distance to roads and earthquake sites,precipitation,and slope;(2)the SHAP method revealed critical nonlinear thresholds,demonstrating that historical landslides were concentrated at mid-altitudes(1400-4000 m)and decreased markedly above 4000 m,with a parallel reduction in probability beyond 700 m from roads;and(3)landslide-prone areas,comprising 13%of the QTP,were concentrated in the southeastern and northeastern parts of the plateau.By integrating machine learning and SHAP analysis,this study revealed landslide hazard-prone areas and their driving factors,providing insights to support disaster management strategies and sustainable regional planning.展开更多
The advancement of wearable sensing technologies demands multifunctional materials that integrate high sensitivity,environmental resilience,and intelligent signal processing.In this work,a flexible hydrophobic conduct...The advancement of wearable sensing technologies demands multifunctional materials that integrate high sensitivity,environmental resilience,and intelligent signal processing.In this work,a flexible hydrophobic conductive yarn(FCB@SY)featuring a controllable microcrack structure is developed via a synergistic approach combining ultrasonic swelling and non-solvent induced phase separation(NIPS).By embedding a robust conductive network and engineering microcrack morphology,the resulting sensor achieves an ultrahigh gauge factor(GF≈12,670),an ultrabroad working range(0%-547%),a low detection limit(0.5%),rapid response/recovery time(140 ms/140 ms),and outstanding durability over 10,000 cycles.Furthermore,the hydrophobic surface endowed by conductive coatings imparts exceptional chemical stability against acidic and alkaline environments,as well as reliable waterproof performance.This enables consistent functionality under harsh conditions,including underwater operation.Integrated with machine learning algorithms,the FCB@SY-based intelligent sensing system demonstrates dualmode capabilities in human motion tracking and gesture recognition,offering significant potential for applications in wearable electronics,human-machine interfaces,and soft robotics.展开更多
Gastrointestinal(GI)cancers remain a leading cause of cancer-related morbidity and mortality worldwide.Artificial intelligence(AI),particularly machine learning and deep learning(DL),has shown promise in enhancing can...Gastrointestinal(GI)cancers remain a leading cause of cancer-related morbidity and mortality worldwide.Artificial intelligence(AI),particularly machine learning and deep learning(DL),has shown promise in enhancing cancer detection,diagnosis,and prognostication.A narrative review of literature published from January 2015 to march 2025 was conducted using PubMed,Web of Science,and Scopus.Search terms included"gastrointestinal cancer","artificial intelligence","machine learning","deep learning","radiomics","multimodal detection"and"predictive modeling".Studies were included if they focused on clinically relevant AI applications in GI oncology.AI algorithms for GI cancer detection have achieved high performance across imaging modalities,with endoscopic DL systems reporting accuracies of 85%-97%for polyp detection and segmentation.Radiomics-based models have predicted molecular biomarkers such as programmed cell death ligand 2 expression with area under the curves up to 0.92.Large language models applied to radiology reports demonstrated diagnostic accuracy comparable to junior radiologists(78.9%vs 80.0%),though without incremental value when combined with human interpretation.Multimodal AI approaches integrating imaging,pathology,and clinical data show emerging potential for precision oncology.AI in GI oncology has reached clinically relevant accuracy levels in multiple diagnostic tasks,with multimodal approaches and predictive biomarker modeling offering new opportunities for personalized care.However,broader validation,integration into clinical workflows,and attention to ethical,legal,and social implications remain critical for widespread adoption.展开更多
Accurate prediction of flood events is important for flood control and risk management.Machine learning techniques contributed greatly to advances in flood predictions,and existing studies mainly focused on predicting...Accurate prediction of flood events is important for flood control and risk management.Machine learning techniques contributed greatly to advances in flood predictions,and existing studies mainly focused on predicting flood resource variables using single or hybrid machine learning techniques.However,class-based flood predictions have rarely been investigated,which can aid in quickly diagnosing comprehensive flood characteristics and proposing targeted management strategies.This study proposed a prediction approach of flood regime metrics and event classes coupling machine learning algorithms with clustering-deduced membership degrees.Five algorithms were adopted for this exploration.Results showed that the class membership degrees accurately determined event classes with class hit rates up to 100%,compared with the four classes clustered from nine regime metrics.The nonlinear algorithms(Multiple Linear Regression,Random Forest,and least squares-Support Vector Machine)outperformed the linear techniques(Multiple Linear Regression and Stepwise Regression)in predicting flood regime metrics.The proposed approach well predicted flood event classes with average class hit rates of 66.0%-85.4%and 47.2%-76.0%in calibration and validation periods,respectively,particularly for the slow and late flood events.The predictive capability of the proposed prediction approach for flood regime metrics and classes was considerably stronger than that of hydrological modeling approach.展开更多
Delayed wound healing following radical gastrectomy remains an important yet underappreciated complication that prolongs hospitalization,increases costs,and undermines patient recovery.In An et al’s recent study,the ...Delayed wound healing following radical gastrectomy remains an important yet underappreciated complication that prolongs hospitalization,increases costs,and undermines patient recovery.In An et al’s recent study,the authors present a machine learning-based risk prediction approach using routinely available clinical and laboratory parameters.Among the evaluated algorithms,a decision tree model demonstrated excellent discrimination,achieving an area under the curve of 0.951 in the validation set and notably identifying all true cases of delayed wound healing at the Youden index threshold.The inclusion of variables such as drainage duration,preoperative white blood cell and neutrophil counts,alongside age and sex,highlights the pragmatic appeal of the model for early postoperative monitoring.Nevertheless,several aspects warrant critical reflection,including the reliance on a postoperative variable(drainage duration),internal validation only,and certain reporting inconsistencies.This letter underscores both the promise and the limitations of adopting interpretable machine learning models in perioperative care.We advocate for transparent reporting,external validation,and careful consideration of clinically actionable timepoints before integration into practice.Ultimately,this work represents a valuable step toward precision risk stratification in gastric cancer surgery,and sets the stage for multicenter,prospective evaluations.展开更多
Oxide dispersion strengthened(ODS)alloys are extensively used owing to high thermostability and creep strength contributed from uniformly dispersed fine oxides particles.However,the existence of these strengthening pa...Oxide dispersion strengthened(ODS)alloys are extensively used owing to high thermostability and creep strength contributed from uniformly dispersed fine oxides particles.However,the existence of these strengthening particles also deteriorates the processability and it is of great importance to establish accurate processing maps to guide the thermomechanical processes to enhance the formability.In this study,we performed particle swarm optimization-based back propagation artificial neural network model to predict the high temperature flow behavior of 0.25wt%Al2O3 particle-reinforced Cu alloys,and compared the accuracy with that of derived by Arrhenius-type constitutive model and back propagation artificial neural network model.To train these models,we obtained the raw data by fabricating ODS Cu alloys using the internal oxidation and reduction method,and conducting systematic hot compression tests between 400 and800℃with strain rates of 10^(-2)-10 S^(-1).At last,processing maps for ODS Cu alloys were proposed by combining processing parameters,mechanical behavior,microstructure characterization,and the modeling results achieved a coefficient of determination higher than>99%.展开更多
Post-kidney transplant rejection is a critical factor influencing transplant success rates and the survival of transplanted organs.With the rapid advancement of artificial intelligence technologies,machine learning(ML...Post-kidney transplant rejection is a critical factor influencing transplant success rates and the survival of transplanted organs.With the rapid advancement of artificial intelligence technologies,machine learning(ML)has emerged as a powerful data analysis tool,widely applied in the prediction,diagnosis,and mechanistic study of kidney transplant rejection.This mini-review systematically summarizes the recent applications of ML techniques in post-kidney transplant rejection,covering areas such as the construction of predictive models,identification of biomarkers,analysis of pathological images,assessment of immune cell infiltration,and formulation of personalized treatment strategies.By integrating multi-omics data and clinical information,ML has significantly enhanced the accuracy of early rejection diagnosis and the capability for prognostic evaluation,driving the development of precision medicine in the field of kidney transplantation.Furthermore,this article discusses the challenges faced in existing research and potential future directions,providing a theoretical basis and technical references for related studies.展开更多
Sudden wildfires cause significant global ecological damage.While satellite imagery has advanced early fire detection and mitigation,image-based systems face limitations including high false alarm rates,visual obstruc...Sudden wildfires cause significant global ecological damage.While satellite imagery has advanced early fire detection and mitigation,image-based systems face limitations including high false alarm rates,visual obstructions,and substantial computational demands,especially in complex forest terrains.To address these challenges,this study proposes a novel forest fire detection model utilizing audio classification and machine learning.We developed an audio-based pipeline using real-world environmental sound recordings.Sounds were converted into Mel-spectrograms and classified via a Convolutional Neural Network(CNN),enabling the capture of distinctive fire acoustic signatures(e.g.,crackling,roaring)that are minimally impacted by visual or weather conditions.Internet of Things(IoT)sound sensors were crucial for generating complex environmental parameters to optimize feature extraction.The CNN model achieved high performance in stratified 5-fold cross-validation(92.4%±1.6 accuracy,91.2%±1.8 F1-score)and on test data(94.93%accuracy,93.04%F1-score),with 98.44%precision and 88.32%recall,demonstrating reliability across environmental conditions.These results indicate that the audio-based approach not only improves detection reliability but also markedly reduces computational overhead compared to traditional image-based methods.The findings suggest that acoustic sensing integrated with machine learning offers a powerful,low-cost,and efficient solution for real-time forest fire monitoring in complex,dynamic environments.展开更多
To curb the worsening tropospheric ozone(O_(3))pollution problem in China,a rapid and accurate identification of O_(3)-precursor sensitivity(OPS)is a crucial prerequisite for formulating effective contingency O_(3) po...To curb the worsening tropospheric ozone(O_(3))pollution problem in China,a rapid and accurate identification of O_(3)-precursor sensitivity(OPS)is a crucial prerequisite for formulating effective contingency O_(3) pollution control strategies.However,currently widely-used methods,such as statistical models and numerical models,exhibit inherent limitations in identifying OPS in a timely and accurate manner.In this study,we developed a novel approach to identify OPS based on eXtreme Gradient Boosting model,Shapley additive explanation(SHAP)al-gorithm,and volatile organic compound(VOC)photochemical decay adjustment,using the meteorology and speciated pollutant monitoring data as the input.By comparing the difference in SHAP values between base sce-nario and precursor reduction scenario for nitrogen oxides(NO_(x))and VOCs,OPS was divided into NO_(x)-limited,VOCs-limited and transition regime.Using the long-lasting O_(3) pollution episode in the autumn of 2022 at the Guangdong-Hong Kong-Macao Greater Bay Area(GBA)as an example,we demonstrated large spatiotemporal heterogeneities of OPS over the GBA,which were generally shifted from NO_(x)-limited to VOCs-limited from September to October and more inclined to be VOCs-limited at the central and NO_(x)-limited in the peripheral areas.This study developed an innovative OPS identification method by comparing the difference in SHAP value before and after precursor emission reduction.Our method enables the accurate identification of OPS in the time scale of seconds,thereby providing a state-of-the-art tool for the rapid guidance of spatial-specific O_(3) control strategies.展开更多
Together,the heart and lung sound comprise the thoracic cavity sound,which provides informative details that reflect patient conditions,particularly heart failure(HF)patients.However,due to the limitations of human he...Together,the heart and lung sound comprise the thoracic cavity sound,which provides informative details that reflect patient conditions,particularly heart failure(HF)patients.However,due to the limitations of human hearing,a limited amount of information can be auscultated from thoracic cavity sounds.With the aid of artificial intelligence–machine learning,these features can be analyzed and aid in the care of HF patients.Machine learning of thoracic cavity sound data involves sound data pre-processing by denoising,resampling,segmentation,and normalization.Afterwards,the most crucial step is feature extraction and se-lection where relevant features are selected to train the model.The next step is classification and model performance evaluation.This review summarizes the currently available studies that utilized different machine learning models,different feature extraction and selection methods,and different classifiers to generate the desired output.Most studies have analyzed the heart sound component of thoracic cavity sound to distinguish between normal and HF patients.Additionally,some studies have aimed to classify HF patients based on thoracic cavity sounds in their entirety,while others have focused on risk strati-fication and prognostic evaluation of HF patients using thoracic cavity sounds.Overall,the results from these studies demonstrate a promisingly high level of accuracy.Therefore,future prospective studies should incorporate these machine learning models to expedite their integration into daily clinical practice for managing HF patients.展开更多
The advanced driver assistance system(ADAS)primarily serves to assist drivers in monitoring the speed of the car and helps them make the right decision,which leads to fewer fatal accidents and ensures higher safety.In...The advanced driver assistance system(ADAS)primarily serves to assist drivers in monitoring the speed of the car and helps them make the right decision,which leads to fewer fatal accidents and ensures higher safety.In the artificial Intelligence domain,machine learning(ML)was developed to make inferences with a degree of accuracy similar to that of humans;however,enormous amounts of data are required.Machine learning enhances the accuracy of the decisions taken by ADAS,by evaluating all the data received from various vehicle sensors.This study summarizes all the critical algorithms used in ADAS technologies and presents the evolution of ADAS technology.Initially,ADAS technology is introduced,along with its evolution,to understand the objectives of developing this technology.Subsequently,the critical algorithms used in ADAS technology,which include face detection,head-pose estimation,gaze estimation,and link detection are discussed.A further discussion follows on the impact of ML on each algorithm in different environments,leading to increased accuracy at the expense of additional computing,to increase efficiency.The aim of this study was to evaluate all the methods with or without ML for each algorithm.展开更多
Machine learning(ML)has been increasingly adopted to solve engineering problems with performance gauged by accuracy,efficiency,and security.Notably,blockchain technology(BT)has been added to ML when security is a part...Machine learning(ML)has been increasingly adopted to solve engineering problems with performance gauged by accuracy,efficiency,and security.Notably,blockchain technology(BT)has been added to ML when security is a particular concern.Nevertheless,there is a research gap that prevailing solutions focus primarily on data security using blockchain but ignore computational security,making the traditional ML process vulnerable to off-chain risks.Therefore,the research objective is to develop a novel ML on blockchain(MLOB)framework to ensure both the data and computational process security.The central tenet is to place them both on the blockchain,execute them as blockchain smart contracts,and protect the execution records on-chain.The framework is established by developing a prototype and further calibrated using a case study of industrial inspection.It is shown that the MLOB framework,compared with existing ML and BT isolated solutions,is superior in terms of security(successfully defending against corruption on six designed attack scenario),maintaining accuracy(0.01%difference with baseline),albeit with a slightly compromised efficiency(0.231 second latency increased).The key finding is MLOB can significantly enhances the computational security of engineering computing without increasing computing power demands.This finding can alleviate concerns regarding the computational resource requirements of ML-BT integration.With proper adaption,the MLOB framework can inform various novel solutions to achieve computational security in broader engineering challenges.展开更多
Finding materials with specific properties is a hot topic in materials science.Traditional materials design relies on empirical and trial-and-error methods,requiring extensive experiments and time,resulting in high co...Finding materials with specific properties is a hot topic in materials science.Traditional materials design relies on empirical and trial-and-error methods,requiring extensive experiments and time,resulting in high costs.With the development of physics,statistics,computer science,and other fields,machine learning offers opportunities for systematically discovering new materials.Especially through machine learning-based inverse design,machine learning algorithms analyze the mapping relationships between materials and their properties to find materials with desired properties.This paper first outlines the basic concepts of materials inverse design and the challenges faced by machine learning-based approaches to materials inverse design.Then,three main inverse design methods—exploration-based,model-based,and optimization-based—are analyzed in the context of different application scenarios.Finally,the applications of inverse design methods in alloys,optical materials,and acoustic materials are elaborated on,and the prospects for materials inverse design are discussed.The authors hope to accelerate the discovery of new materials and provide new possibilities for advancing materials science and innovative design methods.展开更多
The presence of aluminum(Al^(3+))and fluoride(F^(−))ions in the environment can be harmful to ecosystems and human health,highlighting the need for accurate and efficient monitoring.In this paper,an innovative approac...The presence of aluminum(Al^(3+))and fluoride(F^(−))ions in the environment can be harmful to ecosystems and human health,highlighting the need for accurate and efficient monitoring.In this paper,an innovative approach is presented that leverages the power of machine learning to enhance the accuracy and efficiency of fluorescence-based detection for sequential quantitative analysis of aluminum(Al^(3+))and fluoride(F^(−))ions in aqueous solutions.The proposed method involves the synthesis of sulfur-functionalized carbon dots(C-dots)as fluorescence probes,with fluorescence enhancement upon interaction with Al^(3+)ions,achieving a detection limit of 4.2 nmol/L.Subsequently,in the presence of F^(−)ions,fluorescence is quenched,with a detection limit of 47.6 nmol/L.The fingerprints of fluorescence images are extracted using a cross-platform computer vision library in Python,followed by data preprocessing.Subsequently,the fingerprint data is subjected to cluster analysis using the K-means model from machine learning,and the average Silhouette Coefficient indicates excellent model performance.Finally,a regression analysis based on the principal component analysis method is employed to achieve more precise quantitative analysis of aluminum and fluoride ions.The results demonstrate that the developed model excels in terms of accuracy and sensitivity.This groundbreaking model not only showcases exceptional performance but also addresses the urgent need for effective environmental monitoring and risk assessment,making it a valuable tool for safeguarding our ecosystems and public health.展开更多
Open caissons are widely used in foundation engineering because of their load-bearing efficiency and adaptability in diverse soil conditions.However,accurately predicting their undrained bearing capacity in layered so...Open caissons are widely used in foundation engineering because of their load-bearing efficiency and adaptability in diverse soil conditions.However,accurately predicting their undrained bearing capacity in layered soils remains a complex challenge.This study presents a novel application of five ensemble machine(ML)algorithms-random forest(RF),gradient boosting machine(GBM),extreme gradient boosting(XGBoost),adaptive boosting(AdaBoost),and categorical boosting(CatBoost)-to predict the undrained bearing capacity factor(Nc)of circular open caissons embedded in two-layered clay on the basis of results from finite element limit analysis(FELA).The input dataset consists of 1188 numerical simulations using the Tresca failure criterion,varying in geometrical and soil parameters.The FELA was performed via OptumG2 software with adaptive meshing techniques and verified against existing benchmark studies.The ML models were trained on 70% of the dataset and tested on the remaining 30%.Their performance was evaluated using six statistical metrics:coefficient of determination(R²),mean absolute error(MAE),root mean squared error(RMSE),index of scatter(IOS),RMSE-to-standard deviation ratio(RSR),and variance explained factor(VAF).The results indicate that all the models achieved high accuracy,with R²values exceeding 97.6%and RMSE values below 0.02.Among them,AdaBoost and CatBoost consistently outperformed the other methods across both the training and testing datasets,demonstrating superior generalizability and robustness.The proposed ML framework offers an efficient,accurate,and data-driven alternative to traditional methods for estimating caisson capacity in stratified soils.This approach can aid in reducing computational costs while improving reliability in the early stages of foundation design.展开更多
To better understand the migration behavior of plastic fragments in the environment,development of rapid non-destructive methods for in-situ identification and characterization of plastic fragments is necessary.Howeve...To better understand the migration behavior of plastic fragments in the environment,development of rapid non-destructive methods for in-situ identification and characterization of plastic fragments is necessary.However,most of the studies had focused only on colored plastic fragments,ignoring colorless plastic fragments and the effects of different environmental media(backgrounds),thus underestimating their abundance.To address this issue,the present study used near-infrared spectroscopy to compare the identification of colored and colorless plastic fragments based on partial least squares-discriminant analysis(PLS-DA),extreme gradient boost,support vector machine and random forest classifier.The effects of polymer color,type,thickness,and background on the plastic fragments classification were evaluated.PLS-DA presented the best and most stable outcome,with higher robustness and lower misclassification rate.All models frequently misinterpreted colorless plastic fragments and its background when the fragment thickness was less than 0.1mm.A two-stage modeling method,which first distinguishes the plastic types and then identifies colorless plastic fragments that had been misclassified as background,was proposed.The method presented an accuracy higher than 99%in different backgrounds.In summary,this study developed a novel method for rapid and synchronous identification of colored and colorless plastic fragments under complex environmental backgrounds.展开更多
Carbon emissions resulting from energy consumption have become a pressing issue for governments worldwide.Accurate estimation of carbon emissions using satellite remote sensing data has become a crucial research probl...Carbon emissions resulting from energy consumption have become a pressing issue for governments worldwide.Accurate estimation of carbon emissions using satellite remote sensing data has become a crucial research problem.Previous studies relied on statistical regression models that failed to capture the complex nonlinear relationships between carbon emissions and characteristic variables.In this study,we propose a machine learning algorithm for carbon emissions,a Bayesian optimized XGboost regression model,using multi-year energy carbon emission data and nighttime lights(NTL)remote sensing data from Shaanxi Province,China.Our results demonstrate that the XGboost algorithm outperforms linear regression and four other machine learning models,with an R^(2)of 0.906 and RMSE of 5.687.We observe an annual increase in carbon emissions,with high-emission counties primarily concentrated in northern and central Shaanxi Province,displaying a shift from discrete,sporadic points to contiguous,extended spatial distribution.Spatial autocorrelation clustering reveals predominantly high-high and low-low clustering patterns,with economically developed counties showing high-emission clustering and economically relatively backward counties displaying low-emission clustering.Our findings show that the use of NTL data and the XGboost algorithm can estimate and predict carbon emissionsmore accurately and provide a complementary reference for satellite remote sensing image data to serve carbon emission monitoring and assessment.This research provides an important theoretical basis for formulating practical carbon emission reduction policies and contributes to the development of techniques for accurate carbon emission estimation using remote sensing data.展开更多
With the rapid development of artificial intelligence,magnetocaloric materials as well as other materials are being developed with increased efficiency and enhanced performance.However,most studies do not take phase t...With the rapid development of artificial intelligence,magnetocaloric materials as well as other materials are being developed with increased efficiency and enhanced performance.However,most studies do not take phase transitions into account,and as a result,the predictions are usually not accurate enough.In this context,we have established an explicable relationship between alloy compositions and phase transition by feature imputation.A facile machine learning is proposed to screen candidate NiMn-based Heusler alloys with desired magnetic entropy change and magnetic transition temperature with a high accuracy R^(2)≈0.98.As expected,the measured properties of prepared NiMn-based alloys,including phase transition type,magnetic entropy changes and transition temperature,are all in good agreement with the ML predictions.As well as being the first to demonstrate an explicable relationship between alloy compositions,phase transitions and magnetocaloric properties,our proposed ML model is highly predictive and interpretable,which can provide a strong theoretical foundation for identifying high-performance magnetocaloric materials in the future.展开更多
BACKGROUND Colorectal polyps are precancerous diseases of colorectal cancer.Early detection and resection of colorectal polyps can effectively reduce the mortality of colorectal cancer.Endoscopic mucosal resection(EMR...BACKGROUND Colorectal polyps are precancerous diseases of colorectal cancer.Early detection and resection of colorectal polyps can effectively reduce the mortality of colorectal cancer.Endoscopic mucosal resection(EMR)is a common polypectomy proce-dure in clinical practice,but it has a high postoperative recurrence rate.Currently,there is no predictive model for the recurrence of colorectal polyps after EMR.AIM To construct and validate a machine learning(ML)model for predicting the risk of colorectal polyp recurrence one year after EMR.METHODS This study retrospectively collected data from 1694 patients at three medical centers in Xuzhou.Additionally,a total of 166 patients were collected to form a prospective validation set.Feature variable screening was conducted using uni-variate and multivariate logistic regression analyses,and five ML algorithms were used to construct the predictive models.The optimal models were evaluated based on different performance metrics.Decision curve analysis(DCA)and SHapley Additive exPlanation(SHAP)analysis were performed to assess clinical applicability and predictor importance.RESULTS Multivariate logistic regression analysis identified 8 independent risk factors for colorectal polyp recurrence one year after EMR(P<0.05).Among the models,eXtreme Gradient Boosting(XGBoost)demonstrated the highest area under the curve(AUC)in the training set,internal validation set,and prospective validation set,with AUCs of 0.909(95%CI:0.89-0.92),0.921(95%CI:0.90-0.94),and 0.963(95%CI:0.94-0.99),respectively.DCA indicated favorable clinical utility for the XGBoost model.SHAP analysis identified smoking history,family history,and age as the top three most important predictors in the model.CONCLUSION The XGBoost model has the best predictive performance and can assist clinicians in providing individualized colonoscopy follow-up recommendations.展开更多
基金funded by Princess Nourah bint Abdulrahman University Researchers Supporting Project number(PNURSP2025R104)Princess Nourah bint Abdulrahman University,Riyadh,Saudi Arabia.
文摘Modern intrusion detection systems(MIDS)face persistent challenges in coping with the rapid evolution of cyber threats,high-volume network traffic,and imbalanced datasets.Traditional models often lack the robustness and explainability required to detect novel and sophisticated attacks effectively.This study introduces an advanced,explainable machine learning framework for multi-class IDS using the KDD99 and IDS datasets,which reflects real-world network behavior through a blend of normal and diverse attack classes.The methodology begins with sophisticated data preprocessing,incorporating both RobustScaler and QuantileTransformer to address outliers and skewed feature distributions,ensuring standardized and model-ready inputs.Critical dimensionality reduction is achieved via the Harris Hawks Optimization(HHO)algorithm—a nature-inspired metaheuristic modeled on hawks’hunting strategies.HHO efficiently identifies the most informative features by optimizing a fitness function based on classification performance.Following feature selection,the SMOTE is applied to the training data to resolve class imbalance by synthetically augmenting underrepresented attack types.The stacked architecture is then employed,combining the strengths of XGBoost,SVM,and RF as base learners.This layered approach improves prediction robustness and generalization by balancing bias and variance across diverse classifiers.The model was evaluated using standard classification metrics:precision,recall,F1-score,and overall accuracy.The best overall performance was recorded with an accuracy of 99.44%for UNSW-NB15,demonstrating the model’s effectiveness.After balancing,the model demonstrated a clear improvement in detecting the attacks.We tested the model on four datasets to show the effectiveness of the proposed approach and performed the ablation study to check the effect of each parameter.Also,the proposed model is computationaly efficient.To support transparency and trust in decision-making,explainable AI(XAI)techniques are incorporated that provides both global and local insight into feature contributions,and offers intuitive visualizations for individual predictions.This makes it suitable for practical deployment in cybersecurity environments that demand both precision and accountability.
基金The National Key Research and Development Program of China,No.2023YFC3206601。
文摘Landslides pose a formidable natural hazard across the Qinghai-Tibet Plateau(QTP),endangering both ecosystems and human life.Identifying the driving factors behind landslides and accurately assessing susceptibility are key to mitigating disaster risk.This study integrated multi-source historical landslide data with 15 predictive factors and used several machine learning models—Random Forest(RF),Gradient Boosting Regression Trees(GBRT),Extreme Gradient Boosting(XGBoost),and Categorical Boosting(CatBoost)—to generate susceptibility maps.The Shapley additive explanation(SHAP)method was applied to quantify factor importance and explore their nonlinear effects.The results showed that:(1)CatBoost was the best-performing model(CA=0.938,AUC=0.980)in assessing landslide susceptibility,with altitude emerging as the most significant factor,followed by distance to roads and earthquake sites,precipitation,and slope;(2)the SHAP method revealed critical nonlinear thresholds,demonstrating that historical landslides were concentrated at mid-altitudes(1400-4000 m)and decreased markedly above 4000 m,with a parallel reduction in probability beyond 700 m from roads;and(3)landslide-prone areas,comprising 13%of the QTP,were concentrated in the southeastern and northeastern parts of the plateau.By integrating machine learning and SHAP analysis,this study revealed landslide hazard-prone areas and their driving factors,providing insights to support disaster management strategies and sustainable regional planning.
基金the financial support of this work by the National Natural Science Foundation of China(No.52373093)Excellent Youth Found of Natural Science Foundation of Henan Province(No.242300421062)+1 种基金Central Plains Youth Top notch Talent Program of Henan Provincethe 111 project(No.D18023).
文摘The advancement of wearable sensing technologies demands multifunctional materials that integrate high sensitivity,environmental resilience,and intelligent signal processing.In this work,a flexible hydrophobic conductive yarn(FCB@SY)featuring a controllable microcrack structure is developed via a synergistic approach combining ultrasonic swelling and non-solvent induced phase separation(NIPS).By embedding a robust conductive network and engineering microcrack morphology,the resulting sensor achieves an ultrahigh gauge factor(GF≈12,670),an ultrabroad working range(0%-547%),a low detection limit(0.5%),rapid response/recovery time(140 ms/140 ms),and outstanding durability over 10,000 cycles.Furthermore,the hydrophobic surface endowed by conductive coatings imparts exceptional chemical stability against acidic and alkaline environments,as well as reliable waterproof performance.This enables consistent functionality under harsh conditions,including underwater operation.Integrated with machine learning algorithms,the FCB@SY-based intelligent sensing system demonstrates dualmode capabilities in human motion tracking and gesture recognition,offering significant potential for applications in wearable electronics,human-machine interfaces,and soft robotics.
文摘Gastrointestinal(GI)cancers remain a leading cause of cancer-related morbidity and mortality worldwide.Artificial intelligence(AI),particularly machine learning and deep learning(DL),has shown promise in enhancing cancer detection,diagnosis,and prognostication.A narrative review of literature published from January 2015 to march 2025 was conducted using PubMed,Web of Science,and Scopus.Search terms included"gastrointestinal cancer","artificial intelligence","machine learning","deep learning","radiomics","multimodal detection"and"predictive modeling".Studies were included if they focused on clinically relevant AI applications in GI oncology.AI algorithms for GI cancer detection have achieved high performance across imaging modalities,with endoscopic DL systems reporting accuracies of 85%-97%for polyp detection and segmentation.Radiomics-based models have predicted molecular biomarkers such as programmed cell death ligand 2 expression with area under the curves up to 0.92.Large language models applied to radiology reports demonstrated diagnostic accuracy comparable to junior radiologists(78.9%vs 80.0%),though without incremental value when combined with human interpretation.Multimodal AI approaches integrating imaging,pathology,and clinical data show emerging potential for precision oncology.AI in GI oncology has reached clinically relevant accuracy levels in multiple diagnostic tasks,with multimodal approaches and predictive biomarker modeling offering new opportunities for personalized care.However,broader validation,integration into clinical workflows,and attention to ethical,legal,and social implications remain critical for widespread adoption.
基金National Key Research and Development Program of China,No.2023YFC3006704National Natural Science Foundation of China,No.42171047CAS-CSIRO Partnership Joint Project of 2024,No.177GJHZ2023097MI。
文摘Accurate prediction of flood events is important for flood control and risk management.Machine learning techniques contributed greatly to advances in flood predictions,and existing studies mainly focused on predicting flood resource variables using single or hybrid machine learning techniques.However,class-based flood predictions have rarely been investigated,which can aid in quickly diagnosing comprehensive flood characteristics and proposing targeted management strategies.This study proposed a prediction approach of flood regime metrics and event classes coupling machine learning algorithms with clustering-deduced membership degrees.Five algorithms were adopted for this exploration.Results showed that the class membership degrees accurately determined event classes with class hit rates up to 100%,compared with the four classes clustered from nine regime metrics.The nonlinear algorithms(Multiple Linear Regression,Random Forest,and least squares-Support Vector Machine)outperformed the linear techniques(Multiple Linear Regression and Stepwise Regression)in predicting flood regime metrics.The proposed approach well predicted flood event classes with average class hit rates of 66.0%-85.4%and 47.2%-76.0%in calibration and validation periods,respectively,particularly for the slow and late flood events.The predictive capability of the proposed prediction approach for flood regime metrics and classes was considerably stronger than that of hydrological modeling approach.
文摘Delayed wound healing following radical gastrectomy remains an important yet underappreciated complication that prolongs hospitalization,increases costs,and undermines patient recovery.In An et al’s recent study,the authors present a machine learning-based risk prediction approach using routinely available clinical and laboratory parameters.Among the evaluated algorithms,a decision tree model demonstrated excellent discrimination,achieving an area under the curve of 0.951 in the validation set and notably identifying all true cases of delayed wound healing at the Youden index threshold.The inclusion of variables such as drainage duration,preoperative white blood cell and neutrophil counts,alongside age and sex,highlights the pragmatic appeal of the model for early postoperative monitoring.Nevertheless,several aspects warrant critical reflection,including the reliance on a postoperative variable(drainage duration),internal validation only,and certain reporting inconsistencies.This letter underscores both the promise and the limitations of adopting interpretable machine learning models in perioperative care.We advocate for transparent reporting,external validation,and careful consideration of clinically actionable timepoints before integration into practice.Ultimately,this work represents a valuable step toward precision risk stratification in gastric cancer surgery,and sets the stage for multicenter,prospective evaluations.
基金financial support of the National Natural Science Foundation of China(No.52371103)the Fundamental Research Funds for the Central Universities,China(No.2242023K40028)+1 种基金the Open Research Fund of Jiangsu Key Laboratory for Advanced Metallic Materials,China(No.AMM2023B01).financial support of the Research Fund of Shihezi Key Laboratory of AluminumBased Advanced Materials,China(No.2023PT02)financial support of Guangdong Province Science and Technology Major Project,China(No.2021B0301030005)。
文摘Oxide dispersion strengthened(ODS)alloys are extensively used owing to high thermostability and creep strength contributed from uniformly dispersed fine oxides particles.However,the existence of these strengthening particles also deteriorates the processability and it is of great importance to establish accurate processing maps to guide the thermomechanical processes to enhance the formability.In this study,we performed particle swarm optimization-based back propagation artificial neural network model to predict the high temperature flow behavior of 0.25wt%Al2O3 particle-reinforced Cu alloys,and compared the accuracy with that of derived by Arrhenius-type constitutive model and back propagation artificial neural network model.To train these models,we obtained the raw data by fabricating ODS Cu alloys using the internal oxidation and reduction method,and conducting systematic hot compression tests between 400 and800℃with strain rates of 10^(-2)-10 S^(-1).At last,processing maps for ODS Cu alloys were proposed by combining processing parameters,mechanical behavior,microstructure characterization,and the modeling results achieved a coefficient of determination higher than>99%.
文摘Post-kidney transplant rejection is a critical factor influencing transplant success rates and the survival of transplanted organs.With the rapid advancement of artificial intelligence technologies,machine learning(ML)has emerged as a powerful data analysis tool,widely applied in the prediction,diagnosis,and mechanistic study of kidney transplant rejection.This mini-review systematically summarizes the recent applications of ML techniques in post-kidney transplant rejection,covering areas such as the construction of predictive models,identification of biomarkers,analysis of pathological images,assessment of immune cell infiltration,and formulation of personalized treatment strategies.By integrating multi-omics data and clinical information,ML has significantly enhanced the accuracy of early rejection diagnosis and the capability for prognostic evaluation,driving the development of precision medicine in the field of kidney transplantation.Furthermore,this article discusses the challenges faced in existing research and potential future directions,providing a theoretical basis and technical references for related studies.
基金funded by the Directorate of Research and Community Service,Directorate General of Research and Development,Ministry of Higher Education,Science and Technologyin accordance with the Implementation Contract for the Operational Assistance Program for State Universities,Research Program Number:109/C3/DT.05.00/PL/2025.
文摘Sudden wildfires cause significant global ecological damage.While satellite imagery has advanced early fire detection and mitigation,image-based systems face limitations including high false alarm rates,visual obstructions,and substantial computational demands,especially in complex forest terrains.To address these challenges,this study proposes a novel forest fire detection model utilizing audio classification and machine learning.We developed an audio-based pipeline using real-world environmental sound recordings.Sounds were converted into Mel-spectrograms and classified via a Convolutional Neural Network(CNN),enabling the capture of distinctive fire acoustic signatures(e.g.,crackling,roaring)that are minimally impacted by visual or weather conditions.Internet of Things(IoT)sound sensors were crucial for generating complex environmental parameters to optimize feature extraction.The CNN model achieved high performance in stratified 5-fold cross-validation(92.4%±1.6 accuracy,91.2%±1.8 F1-score)and on test data(94.93%accuracy,93.04%F1-score),with 98.44%precision and 88.32%recall,demonstrating reliability across environmental conditions.These results indicate that the audio-based approach not only improves detection reliability but also markedly reduces computational overhead compared to traditional image-based methods.The findings suggest that acoustic sensing integrated with machine learning offers a powerful,low-cost,and efficient solution for real-time forest fire monitoring in complex,dynamic environments.
基金supported by the Key-Area Research and Development Program of Guangdong Province(No.2020B1111360003)the National Natural Science Foundation of China(Nos.42465008 and 42105164)+2 种基金Yunnan Science and Technology Department Project(No.202501AT070239)Yunnan Science and Technology Department Youth Project(No.202401AU070202)Xianyang Rapid Response Decision Support Project for Ozone(No.YZ2024-ZB019).
文摘To curb the worsening tropospheric ozone(O_(3))pollution problem in China,a rapid and accurate identification of O_(3)-precursor sensitivity(OPS)is a crucial prerequisite for formulating effective contingency O_(3) pollution control strategies.However,currently widely-used methods,such as statistical models and numerical models,exhibit inherent limitations in identifying OPS in a timely and accurate manner.In this study,we developed a novel approach to identify OPS based on eXtreme Gradient Boosting model,Shapley additive explanation(SHAP)al-gorithm,and volatile organic compound(VOC)photochemical decay adjustment,using the meteorology and speciated pollutant monitoring data as the input.By comparing the difference in SHAP values between base sce-nario and precursor reduction scenario for nitrogen oxides(NO_(x))and VOCs,OPS was divided into NO_(x)-limited,VOCs-limited and transition regime.Using the long-lasting O_(3) pollution episode in the autumn of 2022 at the Guangdong-Hong Kong-Macao Greater Bay Area(GBA)as an example,we demonstrated large spatiotemporal heterogeneities of OPS over the GBA,which were generally shifted from NO_(x)-limited to VOCs-limited from September to October and more inclined to be VOCs-limited at the central and NO_(x)-limited in the peripheral areas.This study developed an innovative OPS identification method by comparing the difference in SHAP value before and after precursor emission reduction.Our method enables the accurate identification of OPS in the time scale of seconds,thereby providing a state-of-the-art tool for the rapid guidance of spatial-specific O_(3) control strategies.
文摘Together,the heart and lung sound comprise the thoracic cavity sound,which provides informative details that reflect patient conditions,particularly heart failure(HF)patients.However,due to the limitations of human hearing,a limited amount of information can be auscultated from thoracic cavity sounds.With the aid of artificial intelligence–machine learning,these features can be analyzed and aid in the care of HF patients.Machine learning of thoracic cavity sound data involves sound data pre-processing by denoising,resampling,segmentation,and normalization.Afterwards,the most crucial step is feature extraction and se-lection where relevant features are selected to train the model.The next step is classification and model performance evaluation.This review summarizes the currently available studies that utilized different machine learning models,different feature extraction and selection methods,and different classifiers to generate the desired output.Most studies have analyzed the heart sound component of thoracic cavity sound to distinguish between normal and HF patients.Additionally,some studies have aimed to classify HF patients based on thoracic cavity sounds in their entirety,while others have focused on risk strati-fication and prognostic evaluation of HF patients using thoracic cavity sounds.Overall,the results from these studies demonstrate a promisingly high level of accuracy.Therefore,future prospective studies should incorporate these machine learning models to expedite their integration into daily clinical practice for managing HF patients.
文摘The advanced driver assistance system(ADAS)primarily serves to assist drivers in monitoring the speed of the car and helps them make the right decision,which leads to fewer fatal accidents and ensures higher safety.In the artificial Intelligence domain,machine learning(ML)was developed to make inferences with a degree of accuracy similar to that of humans;however,enormous amounts of data are required.Machine learning enhances the accuracy of the decisions taken by ADAS,by evaluating all the data received from various vehicle sensors.This study summarizes all the critical algorithms used in ADAS technologies and presents the evolution of ADAS technology.Initially,ADAS technology is introduced,along with its evolution,to understand the objectives of developing this technology.Subsequently,the critical algorithms used in ADAS technology,which include face detection,head-pose estimation,gaze estimation,and link detection are discussed.A further discussion follows on the impact of ML on each algorithm in different environments,leading to increased accuracy at the expense of additional computing,to increase efficiency.The aim of this study was to evaluate all the methods with or without ML for each algorithm.
文摘Machine learning(ML)has been increasingly adopted to solve engineering problems with performance gauged by accuracy,efficiency,and security.Notably,blockchain technology(BT)has been added to ML when security is a particular concern.Nevertheless,there is a research gap that prevailing solutions focus primarily on data security using blockchain but ignore computational security,making the traditional ML process vulnerable to off-chain risks.Therefore,the research objective is to develop a novel ML on blockchain(MLOB)framework to ensure both the data and computational process security.The central tenet is to place them both on the blockchain,execute them as blockchain smart contracts,and protect the execution records on-chain.The framework is established by developing a prototype and further calibrated using a case study of industrial inspection.It is shown that the MLOB framework,compared with existing ML and BT isolated solutions,is superior in terms of security(successfully defending against corruption on six designed attack scenario),maintaining accuracy(0.01%difference with baseline),albeit with a slightly compromised efficiency(0.231 second latency increased).The key finding is MLOB can significantly enhances the computational security of engineering computing without increasing computing power demands.This finding can alleviate concerns regarding the computational resource requirements of ML-BT integration.With proper adaption,the MLOB framework can inform various novel solutions to achieve computational security in broader engineering challenges.
基金funded by theNationalNatural Science Foundation of China(52061020)Major Science and Technology Projects in Yunnan Province(202302AG050009)Yunnan Fundamental Research Projects(202301AV070003).
文摘Finding materials with specific properties is a hot topic in materials science.Traditional materials design relies on empirical and trial-and-error methods,requiring extensive experiments and time,resulting in high costs.With the development of physics,statistics,computer science,and other fields,machine learning offers opportunities for systematically discovering new materials.Especially through machine learning-based inverse design,machine learning algorithms analyze the mapping relationships between materials and their properties to find materials with desired properties.This paper first outlines the basic concepts of materials inverse design and the challenges faced by machine learning-based approaches to materials inverse design.Then,three main inverse design methods—exploration-based,model-based,and optimization-based—are analyzed in the context of different application scenarios.Finally,the applications of inverse design methods in alloys,optical materials,and acoustic materials are elaborated on,and the prospects for materials inverse design are discussed.The authors hope to accelerate the discovery of new materials and provide new possibilities for advancing materials science and innovative design methods.
基金supported by the National Natural Science Foundation of China(No.U21A20290)Guangdong Basic and Applied Basic Research Foundation(No.2022A1515011656)+2 种基金the Projects of Talents Recruitment of GDUPT(No.2023rcyj1003)the 2022“Sail Plan”Project of Maoming Green Chemical Industry Research Institute(No.MMGCIRI2022YFJH-Y-024)Maoming Science and Technology Project(No.2023382).
文摘The presence of aluminum(Al^(3+))and fluoride(F^(−))ions in the environment can be harmful to ecosystems and human health,highlighting the need for accurate and efficient monitoring.In this paper,an innovative approach is presented that leverages the power of machine learning to enhance the accuracy and efficiency of fluorescence-based detection for sequential quantitative analysis of aluminum(Al^(3+))and fluoride(F^(−))ions in aqueous solutions.The proposed method involves the synthesis of sulfur-functionalized carbon dots(C-dots)as fluorescence probes,with fluorescence enhancement upon interaction with Al^(3+)ions,achieving a detection limit of 4.2 nmol/L.Subsequently,in the presence of F^(−)ions,fluorescence is quenched,with a detection limit of 47.6 nmol/L.The fingerprints of fluorescence images are extracted using a cross-platform computer vision library in Python,followed by data preprocessing.Subsequently,the fingerprint data is subjected to cluster analysis using the K-means model from machine learning,and the average Silhouette Coefficient indicates excellent model performance.Finally,a regression analysis based on the principal component analysis method is employed to achieve more precise quantitative analysis of aluminum and fluoride ions.The results demonstrate that the developed model excels in terms of accuracy and sensitivity.This groundbreaking model not only showcases exceptional performance but also addresses the urgent need for effective environmental monitoring and risk assessment,making it a valuable tool for safeguarding our ecosystems and public health.
文摘Open caissons are widely used in foundation engineering because of their load-bearing efficiency and adaptability in diverse soil conditions.However,accurately predicting their undrained bearing capacity in layered soils remains a complex challenge.This study presents a novel application of five ensemble machine(ML)algorithms-random forest(RF),gradient boosting machine(GBM),extreme gradient boosting(XGBoost),adaptive boosting(AdaBoost),and categorical boosting(CatBoost)-to predict the undrained bearing capacity factor(Nc)of circular open caissons embedded in two-layered clay on the basis of results from finite element limit analysis(FELA).The input dataset consists of 1188 numerical simulations using the Tresca failure criterion,varying in geometrical and soil parameters.The FELA was performed via OptumG2 software with adaptive meshing techniques and verified against existing benchmark studies.The ML models were trained on 70% of the dataset and tested on the remaining 30%.Their performance was evaluated using six statistical metrics:coefficient of determination(R²),mean absolute error(MAE),root mean squared error(RMSE),index of scatter(IOS),RMSE-to-standard deviation ratio(RSR),and variance explained factor(VAF).The results indicate that all the models achieved high accuracy,with R²values exceeding 97.6%and RMSE values below 0.02.Among them,AdaBoost and CatBoost consistently outperformed the other methods across both the training and testing datasets,demonstrating superior generalizability and robustness.The proposed ML framework offers an efficient,accurate,and data-driven alternative to traditional methods for estimating caisson capacity in stratified soils.This approach can aid in reducing computational costs while improving reliability in the early stages of foundation design.
基金supported by the National Natural Science Foundation of China(No.22276139)the Shanghai’s Municipal State-owned Assets Supervision and Administration Commission(No.2022028).
文摘To better understand the migration behavior of plastic fragments in the environment,development of rapid non-destructive methods for in-situ identification and characterization of plastic fragments is necessary.However,most of the studies had focused only on colored plastic fragments,ignoring colorless plastic fragments and the effects of different environmental media(backgrounds),thus underestimating their abundance.To address this issue,the present study used near-infrared spectroscopy to compare the identification of colored and colorless plastic fragments based on partial least squares-discriminant analysis(PLS-DA),extreme gradient boost,support vector machine and random forest classifier.The effects of polymer color,type,thickness,and background on the plastic fragments classification were evaluated.PLS-DA presented the best and most stable outcome,with higher robustness and lower misclassification rate.All models frequently misinterpreted colorless plastic fragments and its background when the fragment thickness was less than 0.1mm.A two-stage modeling method,which first distinguishes the plastic types and then identifies colorless plastic fragments that had been misclassified as background,was proposed.The method presented an accuracy higher than 99%in different backgrounds.In summary,this study developed a novel method for rapid and synchronous identification of colored and colorless plastic fragments under complex environmental backgrounds.
基金supported by the Key Research and Development Program in Shaanxi Province,China(No.2022ZDLSF07-05)the Fundamental Research Funds for the Central Universities,CHD(No.300102352901)。
文摘Carbon emissions resulting from energy consumption have become a pressing issue for governments worldwide.Accurate estimation of carbon emissions using satellite remote sensing data has become a crucial research problem.Previous studies relied on statistical regression models that failed to capture the complex nonlinear relationships between carbon emissions and characteristic variables.In this study,we propose a machine learning algorithm for carbon emissions,a Bayesian optimized XGboost regression model,using multi-year energy carbon emission data and nighttime lights(NTL)remote sensing data from Shaanxi Province,China.Our results demonstrate that the XGboost algorithm outperforms linear regression and four other machine learning models,with an R^(2)of 0.906 and RMSE of 5.687.We observe an annual increase in carbon emissions,with high-emission counties primarily concentrated in northern and central Shaanxi Province,displaying a shift from discrete,sporadic points to contiguous,extended spatial distribution.Spatial autocorrelation clustering reveals predominantly high-high and low-low clustering patterns,with economically developed counties showing high-emission clustering and economically relatively backward counties displaying low-emission clustering.Our findings show that the use of NTL data and the XGboost algorithm can estimate and predict carbon emissionsmore accurately and provide a complementary reference for satellite remote sensing image data to serve carbon emission monitoring and assessment.This research provides an important theoretical basis for formulating practical carbon emission reduction policies and contributes to the development of techniques for accurate carbon emission estimation using remote sensing data.
基金supported by the National Key R&D Program of China(No.2022YFE0109500)the National Natural Science Foundation of China(Nos.52071255,52301250,52171190 and 12304027)+2 种基金the Key R&D Project of Shaanxi Province(No.2022GXLH-01-07)the Fundamental Research Funds for the Central Universities(China)the World-Class Universities(Disciplines)and the Characteristic Development Guidance Funds for the Central Universities.
文摘With the rapid development of artificial intelligence,magnetocaloric materials as well as other materials are being developed with increased efficiency and enhanced performance.However,most studies do not take phase transitions into account,and as a result,the predictions are usually not accurate enough.In this context,we have established an explicable relationship between alloy compositions and phase transition by feature imputation.A facile machine learning is proposed to screen candidate NiMn-based Heusler alloys with desired magnetic entropy change and magnetic transition temperature with a high accuracy R^(2)≈0.98.As expected,the measured properties of prepared NiMn-based alloys,including phase transition type,magnetic entropy changes and transition temperature,are all in good agreement with the ML predictions.As well as being the first to demonstrate an explicable relationship between alloy compositions,phase transitions and magnetocaloric properties,our proposed ML model is highly predictive and interpretable,which can provide a strong theoretical foundation for identifying high-performance magnetocaloric materials in the future.
文摘BACKGROUND Colorectal polyps are precancerous diseases of colorectal cancer.Early detection and resection of colorectal polyps can effectively reduce the mortality of colorectal cancer.Endoscopic mucosal resection(EMR)is a common polypectomy proce-dure in clinical practice,but it has a high postoperative recurrence rate.Currently,there is no predictive model for the recurrence of colorectal polyps after EMR.AIM To construct and validate a machine learning(ML)model for predicting the risk of colorectal polyp recurrence one year after EMR.METHODS This study retrospectively collected data from 1694 patients at three medical centers in Xuzhou.Additionally,a total of 166 patients were collected to form a prospective validation set.Feature variable screening was conducted using uni-variate and multivariate logistic regression analyses,and five ML algorithms were used to construct the predictive models.The optimal models were evaluated based on different performance metrics.Decision curve analysis(DCA)and SHapley Additive exPlanation(SHAP)analysis were performed to assess clinical applicability and predictor importance.RESULTS Multivariate logistic regression analysis identified 8 independent risk factors for colorectal polyp recurrence one year after EMR(P<0.05).Among the models,eXtreme Gradient Boosting(XGBoost)demonstrated the highest area under the curve(AUC)in the training set,internal validation set,and prospective validation set,with AUCs of 0.909(95%CI:0.89-0.92),0.921(95%CI:0.90-0.94),and 0.963(95%CI:0.94-0.99),respectively.DCA indicated favorable clinical utility for the XGBoost model.SHAP analysis identified smoking history,family history,and age as the top three most important predictors in the model.CONCLUSION The XGBoost model has the best predictive performance and can assist clinicians in providing individualized colonoscopy follow-up recommendations.