This study provides an in-depth comparative evaluation of landslide susceptibility using two distinct spatial units:and slope units(SUs)and hydrological response units(HRUs),within Goesan County,South Korea.Leveraging...This study provides an in-depth comparative evaluation of landslide susceptibility using two distinct spatial units:and slope units(SUs)and hydrological response units(HRUs),within Goesan County,South Korea.Leveraging the capabilities of the extreme gradient boosting(XGB)algorithm combined with Shapley Additive Explanations(SHAP),this work assesses the precision and clarity with which each unit predicts areas vulnerable to landslides.SUs focus on the geomorphological features like ridges and valleys,focusing on slope stability and landslide triggers.Conversely,HRUs are established based on a variety of hydrological factors,including land cover,soil type and slope gradients,to encapsulate the dynamic water processes of the region.The methodological framework includes the systematic gathering,preparation and analysis of data,ranging from historical landslide occurrences to topographical and environmental variables like elevation,slope angle and land curvature etc.The XGB algorithm used to construct the Landslide Susceptibility Model(LSM)was combined with SHAP for model interpretation and the results were evaluated using Random Cross-validation(RCV)to ensure accuracy and reliability.To ensure optimal model performance,the XGB algorithm’s hyperparameters were tuned using Differential Evolution,considering multicollinearity-free variables.The results show that SU and HRU are effective for LSM,but their effectiveness varies depending on landscape characteristics.The XGB algorithm demonstrates strong predictive power and SHAP enhances model transparency of the influential variables involved.This work underscores the importance of selecting appropriate assessment units tailored to specific landscape characteristics for accurate LSM.The integration of advanced machine learning techniques with interpretative tools offers a robust framework for landslide susceptibility assessment,improving both predictive capabilities and model interpretability.Future research should integrate broader data sets and explore hybrid analytical models to strengthen the generalizability of these findings across varied geographical settings.展开更多
The methods of network attacks have become increasingly sophisticated,rendering traditional cybersecurity defense mechanisms insufficient to address novel and complex threats effectively.In recent years,artificial int...The methods of network attacks have become increasingly sophisticated,rendering traditional cybersecurity defense mechanisms insufficient to address novel and complex threats effectively.In recent years,artificial intelligence has achieved significant progress in the field of network security.However,many challenges and issues remain,particularly regarding the interpretability of deep learning and ensemble learning algorithms.To address the challenge of enhancing the interpretability of network attack prediction models,this paper proposes a method that combines Light Gradient Boosting Machine(LGBM)and SHapley Additive exPlanations(SHAP).LGBM is employed to model anomalous fluctuations in various network indicators,enabling the rapid and accurate identification and prediction of potential network attack types,thereby facilitating the implementation of timely defense measures,the model achieved an accuracy of 0.977,precision of 0.985,recall of 0.975,and an F1 score of 0.979,demonstrating better performance compared to other models in the domain of network attack prediction.SHAP is utilized to analyze the black-box decision-making process of the model,providing interpretability by quantifying the contribution of each feature to the prediction results and elucidating the relationships between features.The experimental results demonstrate that the network attack predictionmodel based on LGBM exhibits superior accuracy and outstanding predictive capabilities.Moreover,the SHAP-based interpretability analysis significantly improves the model’s transparency and interpretability.展开更多
Predicting molecular properties is essential for advancing for advancing drug discovery and design. Recently, Graph Neural Networks (GNNs) have gained prominence due to their ability to capture the complex structural ...Predicting molecular properties is essential for advancing for advancing drug discovery and design. Recently, Graph Neural Networks (GNNs) have gained prominence due to their ability to capture the complex structural and relational information inherent in molecular graphs. Despite their effectiveness, the “black-box” nature of GNNs remains a significant obstacle to their widespread adoption in chemistry, as it hinders interpretability and trust. In this context, several explanation methods based on factual reasoning have emerged. These methods aim to interpret the predictions made by GNNs by analyzing the key features contributing to the prediction. However, these approaches fail to answer critical questions: “How to ensure that the structure-property mapping learned by GNNs is consistent with established domain knowledge”. In this paper, we propose MMGCF, a novel counterfactual explanation framework designed specifically for the prediction of GNN-based molecular properties. MMGCF constructs a hierarchical tree structure on molecular motifs, enabling the systematic generation of counterfactuals through motif perturbations. This framework identifies causally significant motifs and elucidates their impact on model predictions, offering insights into the relationship between structural modifications and predicted properties. Our method demonstrates its effectiveness through comprehensive quantitative and qualitative evaluations of four real-world molecular datasets.展开更多
Accurate prediction of shield tunneling-induced settlement is a complex problem that requires consideration of many influential parameters.Recent studies reveal that machine learning(ML)algorithms can predict the sett...Accurate prediction of shield tunneling-induced settlement is a complex problem that requires consideration of many influential parameters.Recent studies reveal that machine learning(ML)algorithms can predict the settlement caused by tunneling.However,well-performing ML models are usually less interpretable.Irrelevant input features decrease the performance and interpretability of an ML model.Nonetheless,feature selection,a critical step in the ML pipeline,is usually ignored in most studies that focused on predicting tunneling-induced settlement.This study applies four techniques,i.e.Pearson correlation method,sequential forward selection(SFS),sequential backward selection(SBS)and Boruta algorithm,to investigate the effect of feature selection on the model’s performance when predicting the tunneling-induced maximum surface settlement(S_(max)).The data set used in this study was compiled from two metro tunnel projects excavated in Hangzhou,China using earth pressure balance(EPB)shields and consists of 14 input features and a single output(i.e.S_(max)).The ML model that is trained on features selected from the Boruta algorithm demonstrates the best performance in both the training and testing phases.The relevant features chosen from the Boruta algorithm further indicate that tunneling-induced settlement is affected by parameters related to tunnel geometry,geological conditions and shield operation.The recently proposed Shapley additive explanations(SHAP)method explores how the input features contribute to the output of a complex ML model.It is observed that the larger settlements are induced during shield tunneling in silty clay.Moreover,the SHAP analysis reveals that the low magnitudes of face pressure at the top of the shield increase the model’s output。展开更多
This paper takes a microanalytic perspective on the speech and gestures used by one teacher of ESL (English as a Second Language) in an intensive English program classroom. Videotaped excerpts from her intermediate-...This paper takes a microanalytic perspective on the speech and gestures used by one teacher of ESL (English as a Second Language) in an intensive English program classroom. Videotaped excerpts from her intermediate-level grammar course were transcribed to represent the speech, gesture and other non-verbal behavior that accompanied unplanned explanations of vocabulary that arose during three focus-on-form lessons. The gesture classification system of McNeill (1992), which delineates different types of hand movements (iconics metaphorics, deictics, beats), was used to understand the role the gestures played in these explanations. Results suggest that gestures and other non-verbal behavior are forms of input to classroom second language learners that must be considered a salient factor in classroom-based SLA (Second Language Acquisition) research展开更多
In this paper I examine the following claims by William Eaton in his monograph Boyle on Fire: (i) that Boyle's religious convictions led him to believe that the world was not completely explicable, and this shows ...In this paper I examine the following claims by William Eaton in his monograph Boyle on Fire: (i) that Boyle's religious convictions led him to believe that the world was not completely explicable, and this shows that there is a shortcoming in the power of mechanical explanations; (ii) that mechanical explanations offer only sufficient, not necessary explanations, and this too was taken by Boyle to be a limit in the explanatory power of mechanical explanations; (iii) that the mature Boyle thought that there could be more intelligible explanatory models than mechanism; and (iv) that what Boyle says at any point in his career is incompatible with the statement of Maria Boas-Hall, i.e., that the mechanical hypothesis can explicate all natural phenomena. Since all four of these claims are part of Eaton's developmental argument, my rejection of them will not only show how the particular developmental story Eaton diagnoses is inaccurate, but will also explain what limits there actually are in Boyle's account of the intelligibility of mechanical explanations. My account will also show why important philosophers like Locke and Leibniz should be interested in Boyle's philosophical work.展开更多
Wind power forecasting(WPF)is important for safe,stable,and reliable integration of new energy technologies into power systems.Machine learning(ML)algorithms have recently attracted increasing attention in the field o...Wind power forecasting(WPF)is important for safe,stable,and reliable integration of new energy technologies into power systems.Machine learning(ML)algorithms have recently attracted increasing attention in the field of WPF.However,opaque decisions and lack of trustworthiness of black-box models for WPF could cause scheduling risks.This study develops a method for identifying risky models in practical applications and avoiding the risks.First,a local interpretable model-agnostic explanations algorithm is introduced and improved for WPF model analysis.On that basis,a novel index is presented to quantify the level at which neural networks or other black-box models can trust features involved in training.Then,by revealing the operational mechanism for local samples,human interpretability of the black-box model is examined under different accuracies,time horizons,and seasons.This interpretability provides a basis for several technical routes for WPF from the viewpoint of the forecasting model.Moreover,further improvements in accuracy of WPF are explored by evaluating possibilities of using interpretable ML models that use multi-horizons global trust modeling and multi-seasons interpretable feature selection methods.Experimental results from a wind farm in China show that error can be robustly reduced.展开更多
Generative Artificial Intelligence(GAI)refers to a class of AI systems capable of creating novel,coherent,and contextually relevant content—such as text,images,audio,and video—based on patterns learned from extensiv...Generative Artificial Intelligence(GAI)refers to a class of AI systems capable of creating novel,coherent,and contextually relevant content—such as text,images,audio,and video—based on patterns learned from extensive training datasets.The public release and rapid refinement of large language models(LLMs)like ChatGPT have accelerated the adoption of GAI across various medical specialties,offering new tools for education,clinical simulation,and research.Dermatology training,which heavily relies on visual pattern recognition and requires extensive exposure to diverse morphological presentations,faces persistent challenges such as uneven distribu-tion of educational resources,limited patient exposure for rare conditions,and variability in teaching quality.Exploring the integration of GAI into pedagogical frameworks offers innovative approaches to address these challenges,potentially enhancing the quality,standardization,scalability,and accessibility of dermatology ed-ucation.This comprehensive review examines the core concepts and technical foundations of GAI,highlights its specific applications within dermatology teaching and learning—including simulated case generation,per-sonalized learning pathways,and academic support—and discusses the current limitations,practical challenges,and ethical considerations surrounding its use.The aim is to provide a balanced perspective on the significant potential of GAI for transforming dermatology education and to offer evidence-based insights to guide future exploration,implementation,and policy development.展开更多
Heart disease remains a leading cause of mortality worldwide,emphasizing the urgent need for reliable and interpretable predictive models to support early diagnosis and timely intervention.However,existing Deep Learni...Heart disease remains a leading cause of mortality worldwide,emphasizing the urgent need for reliable and interpretable predictive models to support early diagnosis and timely intervention.However,existing Deep Learning(DL)approaches often face several limitations,including inefficient feature extraction,class imbalance,suboptimal classification performance,and limited interpretability,which collectively hinder their deployment in clinical settings.To address these challenges,we propose a novel DL framework for heart disease prediction that integrates a comprehensive preprocessing pipeline with an advanced classification architecture.The preprocessing stage involves label encoding and feature scaling.To address the issue of class imbalance inherent in the personal key indicators of the heart disease dataset,the localized random affine shadowsampling technique is employed,which enhances minority class representation while minimizing overfitting.At the core of the framework lies the Deep Residual Network(DeepResNet),which employs hierarchical residual transformations to facilitate efficient feature extraction and capture complex,non-linear relationships in the data.Experimental results demonstrate that the proposed model significantly outperforms existing techniques,achieving improvements of 3.26%in accuracy,3.16%in area under the receiver operating characteristics,1.09%in recall,and 1.07%in F1-score.Furthermore,robustness is validated using 10-fold crossvalidation,confirming the model’s generalizability across diverse data distributions.Moreover,model interpretability is ensured through the integration of Shapley additive explanations and local interpretable model-agnostic explanations,offering valuable insights into the contribution of individual features to model predictions.Overall,the proposed DL framework presents a robust,interpretable,and clinically applicable solution for heart disease prediction.展开更多
本文聚焦我国青海湖流域的水文过程,基于多年气象和水文动态数据,发展了一种融合概念性水文模型FLEX (FluxExchange)和门控循环单元(Gated Recurrent Unit, GRU)的混合模型对流域内最大支流布哈河的逐日径流进行了模拟和预测。在构建混...本文聚焦我国青海湖流域的水文过程,基于多年气象和水文动态数据,发展了一种融合概念性水文模型FLEX (FluxExchange)和门控循环单元(Gated Recurrent Unit, GRU)的混合模型对流域内最大支流布哈河的逐日径流进行了模拟和预测。在构建混合模型中,采用了三种策略提升模拟精度:引入差分进化自适应算法DREAM(zs)反演水文参数优化FLEX模型;采用变分模态分解(VMD)提取径流数据的信息和特征;利用麻雀搜索算法(SSA)优化深度学习GRU的参数。研究将FLEX模型的模拟结果连同气象数据一起作为神经网络的输入,从而构建了FLEX-VMD-SSA-GRU混合模型。同时,探讨了不同的气象输入条件对模拟结果的影响和贡献:基于7个主要气象要素,由少及多设置了14组输入情景模拟。最后通过SHAP对深度学习方法的结果进行分析,揭示了气象变量对径流长期趋势的贡献和重要度。展开更多
Contemporary demands necessitate the swift and accurate detection of cracks in critical infrastructures,including tunnels and pavements.This study proposed a transfer learning-based encoder-decoder method with visual ...Contemporary demands necessitate the swift and accurate detection of cracks in critical infrastructures,including tunnels and pavements.This study proposed a transfer learning-based encoder-decoder method with visual explanations for infrastructure crack segmentation.Firstly,a vast dataset containing 7089 images was developed,comprising diverse conditions—simple and complex crack patterns as well as clean and rough backgrounds.Secondly,leveraging transfer learning,an encoder-decoder model with visual explanations was formulated,utilizing varied pre-trained convolutional neural network(CNN)as the encoder.Visual explanations were achieved through gradient-weighted class activation mapping(Grad-CAM)to interpret the CNN segmentation model.Thirdly,accuracy,complexity(computation and model),and memory usage assessed CNN feasibility in practical engineering.Model performance was gauged via prediction and visual explanation.The investigation encompassed hyperparameters,data augmentation,deep learning from scratch vs.transfer learning,segmentation model architectures,segmentation model encoders,and encoder pre-training strategies.Results underscored transfer learning’s potency in enhancing CNN accuracy for crack segmentation,surpassing deep learning from scratch.Notably,encoder classification accuracy bore no significant correlation with CNN segmentation accuracy.Among all tested models,UNet-EfficientNet_B7 excelled in crack segmentation,harmonizing accuracy,complexity,memory usage,prediction,and visual explanation.展开更多
This study presents an enhanced convolutional neural network(CNN)model integrated with Explainable Artificial Intelligence(XAI)techniques for accurate prediction and interpretation of wheat crop diseases.The aim is to...This study presents an enhanced convolutional neural network(CNN)model integrated with Explainable Artificial Intelligence(XAI)techniques for accurate prediction and interpretation of wheat crop diseases.The aim is to streamline the detection process while offering transparent insights into the model’s decision-making to support effective disease management.To evaluate the model,a dataset was collected from wheat fields in Kotli,Azad Kashmir,Pakistan,and tested across multiple data splits.The proposed model demonstrates improved stability,faster conver-gence,and higher classification accuracy.The results show significant improvements in prediction accuracy and stability compared to prior works,achieving up to 100%accuracy in certain configurations.In addition,XAI methods such as Local Interpretable Model-agnostic Explanations(LIME)and Shapley Additive Explanations(SHAP)were employed to explain the model’s predictions,highlighting the most influential features contributing to classification decisions.The combined use of CNN and XAI offers a dual benefit:strong predictive performance and clear interpretability of outcomes,which is especially critical in real-world agricultural applications.These findings underscore the potential of integrating deep learning models with XAI to advance automated plant disease detection.The study offers a precise,reliable,and interpretable solution for improving wheat production and promoting agricultural sustainability.Future extensions of this work may include scaling the dataset across broader regions and incorporating additional modalities such as environmental data to enhance model robustness and generalization.展开更多
Mechanical properties are critical to the quality of hot-rolled steel pipe products.Accurately understanding the relationship between rolling parameters and mechanical properties is crucial for effective prediction an...Mechanical properties are critical to the quality of hot-rolled steel pipe products.Accurately understanding the relationship between rolling parameters and mechanical properties is crucial for effective prediction and control.To address this,an industrial big data platform was developed to collect and process multi-source heterogeneous data from the entire production process,providing a complete dataset for mechanical property prediction.The adaptive bandwidth kernel density estimation(ABKDE)method was proposed to adjust bandwidth dynamically based on data density.Combining long short-term memory neural networks with ABKDE offers robust prediction interval capabilities for mechanical properties.The proposed method was deployed in a large-scale steel plant,which demonstrated superior prediction interval performance compared to lower upper bound estimation,mean variance estimation,and extreme learning machine-adaptive bandwidth kernel density estimation,achieving a prediction interval normalized average width of 0.37,a prediction interval coverage probability of 0.94,and the lowest coverage width-based criterion of 1.35.Notably,shapley additive explanations-based explanations significantly improved the proposed model’s credibility by providing a clear analysis of feature impacts.展开更多
BACKGROUND Despite the promising prospects of using artificial intelligence and machine learning(ML)for disease classification and prediction purposes,the complexity and lack of explainability of this method make it d...BACKGROUND Despite the promising prospects of using artificial intelligence and machine learning(ML)for disease classification and prediction purposes,the complexity and lack of explainability of this method make it difficult to apply the constructed models in clinical practice.We developed and validated an interpretable ML model based on magnetic resonance imaging(MRI)radiomics and clinical features for the preoperative prediction of the pathological grades of hepatocellular carcinomas(HCCs).This model will help clinicians better understand the situation and develop personalized treatment plans.AIM To develop and validate an interpretable ML model for preoperative pathological grade prediction in HCC patients via a combination of multisequence MRI radiomics and clinical features.METHODS MRI and clinical data derived from 125 patients with HCCs confirmed by postoperative pathological examinations were retrospectively analyzed.The patients were randomly split into training and validation groups(7:3 ratio).Univariate and multivariate logistic regression analyses were performed to identify independent clinical predictors.The tumor lesions observed on axial fatsuppressed T2-weighted imaging(FS-T2WI),arterial phase(AP),and portal venous phase(PVP)images were delineated in a slice-by-slice manner using 3D-slicer to generate volumetric regions of interest,and radiomic features were extracted.Interclass correlation coefficients were calculated,and least absolute selection and shrinkage operator regression were conducted for feature selection purposes.Six predictive models were subsequently developed for pathological grade prediction:FS-T2WI,AP,PVP,integrated radiomics,clinical,and combined radiomics-clinical(RC)models.The effectiveness of these models was assessed by calculating their area under the receiver operating characteristic curve(AUC)values.The clinical applicability of the models was evaluated via decision curve analysis.Finally,the contributions of the different features contained in the model with optimal performance were interpreted via a SHapley Additive exPlanations analysis.RESULTS Among the 125 patients,87 were assigned to the training group,and 38 were assigned to the validation group.The maximum tumor diameter,hepatitis B virus status,and monocyte count were identified as independent predictors of pathological grade.Twelve optimal radiomic features were ultimately selected.The AUC values obtained for the FS-T2WI model,AP model,PVP model,radiomics model,clinical model,and combined RC model in the training group were 0.761[95%confidence interval(CI):0.562-0.857],0.870(95%CI:0.714-0.918),0.868(95%CI:0.714-0.959),0.917(95%CI:0.857-0.959),0.869(95%CI:0.643-0.973),and 0.941(95%CI:0.857-0.945),respectively;in the validation group,the AUC values were 0.724(95%CI:0.625-0.833),0.802(95%CI:0.686-1.000),0.797(95%CI:0.688-1.000),0.901(95%CI:0.833-0.906),0.865(95%CI:0.594-1.000),and 0.932(95%CI:0.812-1.000),respectively.The combined RC model demonstrated the best performance.Additionally,the decision curve analysis revealed that the combined RC model had satisfactory prediction efficiency,and the SHapley Additive exPlanations value analysis revealed that the“FS-T2WI-wavelet-HLL_gldm_Large Dependence High Gray Level Emphasis”feature contributed the most to the model,exhibiting a positive effect.CONCLUSION An interpretable ML model based on MRI radiomics provides a noninvasive tool for predicting the pathological grade of HCCs,which will help clinicians develop personalized treatment plans.展开更多
The study by Huang et al,published in the World Journal of Gastroenterology,advances intrahepatic cholangiocarcinoma(ICC)management by developing a machine-learning model to predict textbook outcomes(TO)based on preop...The study by Huang et al,published in the World Journal of Gastroenterology,advances intrahepatic cholangiocarcinoma(ICC)management by developing a machine-learning model to predict textbook outcomes(TO)based on preoperative factors.By analyzing data from 376 patients across four Chinese medical centers,the researchers identified key variables influencing TO,including Child-Pugh classification,Eastern Cooperative Oncology Group score,hepatitis B status,and tumor size.The model,created using logistic regression and the extreme gradient boosting algorithm,demonstrated high predictive accuracy,with area under the curve values of 0.8825 for internal validation and 0.8346 for external validation.The integration of the Shapley additive explanation technique enhances the interpretability of the model,which is crucial for clinical decision-making.This research highlights the potential of machine learning to improve surgical planning and patient outcomes in ICC,opening possibilities for personalized treatment approaches based on individual patient characteristics and risk factors.展开更多
Machine learning-based Debris Flow Susceptibility Mapping(DFSM)has emerged as an effective approach for assessing debris flow likelihood,yet its application faces three critical challenges:insufficient reliability of ...Machine learning-based Debris Flow Susceptibility Mapping(DFSM)has emerged as an effective approach for assessing debris flow likelihood,yet its application faces three critical challenges:insufficient reliability of training samples caused by biased negative sampling,opaque decision-making mechanisms in models,and subjective susceptibility mapping methods that lack quantitative evaluation criteria.This study focuses on the Yalong River basin.By integrating high-resolution remote sensing interpretation and field surveys,we established a refined sample database that includes 1,736 debris flow gullies.To address spatial bias in traditional random negative sampling,we developed a semi-supervised optimization strategy based on iterative confidence screening.Comparative experiments with four treebased models(XGBoost,CatBoost,LGBM,and Random Forest)reveal that the optimized sampling strategy improved overall model performance by 8%-12%,with XGBoost achieving the highest accuracy(AUC=0.882)and RF performing the lowest(AUC=0.820).SHAP-based global-local interpretability analysis(applicable to all tree models)identifies elevation and short-duration rainfall as dominant controlling factors.Furthermore,among the tested tree-based models,XGBoost optimized with semisupervised sampling demonstrates the highest reliability in debris flow susceptibility mapping(DFSM),achieving a comprehensive accuracy of 83.64%due to its optimal generalization-stability equilibrium.展开更多
基金supported by a National Research Foundation of Korea(NRF)grant funded by the Korean government(MSIT)(RS-2023-00222536).
文摘This study provides an in-depth comparative evaluation of landslide susceptibility using two distinct spatial units:and slope units(SUs)and hydrological response units(HRUs),within Goesan County,South Korea.Leveraging the capabilities of the extreme gradient boosting(XGB)algorithm combined with Shapley Additive Explanations(SHAP),this work assesses the precision and clarity with which each unit predicts areas vulnerable to landslides.SUs focus on the geomorphological features like ridges and valleys,focusing on slope stability and landslide triggers.Conversely,HRUs are established based on a variety of hydrological factors,including land cover,soil type and slope gradients,to encapsulate the dynamic water processes of the region.The methodological framework includes the systematic gathering,preparation and analysis of data,ranging from historical landslide occurrences to topographical and environmental variables like elevation,slope angle and land curvature etc.The XGB algorithm used to construct the Landslide Susceptibility Model(LSM)was combined with SHAP for model interpretation and the results were evaluated using Random Cross-validation(RCV)to ensure accuracy and reliability.To ensure optimal model performance,the XGB algorithm’s hyperparameters were tuned using Differential Evolution,considering multicollinearity-free variables.The results show that SU and HRU are effective for LSM,but their effectiveness varies depending on landscape characteristics.The XGB algorithm demonstrates strong predictive power and SHAP enhances model transparency of the influential variables involved.This work underscores the importance of selecting appropriate assessment units tailored to specific landscape characteristics for accurate LSM.The integration of advanced machine learning techniques with interpretative tools offers a robust framework for landslide susceptibility assessment,improving both predictive capabilities and model interpretability.Future research should integrate broader data sets and explore hybrid analytical models to strengthen the generalizability of these findings across varied geographical settings.
基金supported by the National Natural Science Foundation of China Project(No.62302540)please visit their website at https://www.nsfc.gov.cn/(accessed on 18 June 2024).
文摘The methods of network attacks have become increasingly sophisticated,rendering traditional cybersecurity defense mechanisms insufficient to address novel and complex threats effectively.In recent years,artificial intelligence has achieved significant progress in the field of network security.However,many challenges and issues remain,particularly regarding the interpretability of deep learning and ensemble learning algorithms.To address the challenge of enhancing the interpretability of network attack prediction models,this paper proposes a method that combines Light Gradient Boosting Machine(LGBM)and SHapley Additive exPlanations(SHAP).LGBM is employed to model anomalous fluctuations in various network indicators,enabling the rapid and accurate identification and prediction of potential network attack types,thereby facilitating the implementation of timely defense measures,the model achieved an accuracy of 0.977,precision of 0.985,recall of 0.975,and an F1 score of 0.979,demonstrating better performance compared to other models in the domain of network attack prediction.SHAP is utilized to analyze the black-box decision-making process of the model,providing interpretability by quantifying the contribution of each feature to the prediction results and elucidating the relationships between features.The experimental results demonstrate that the network attack predictionmodel based on LGBM exhibits superior accuracy and outstanding predictive capabilities.Moreover,the SHAP-based interpretability analysis significantly improves the model’s transparency and interpretability.
文摘Predicting molecular properties is essential for advancing for advancing drug discovery and design. Recently, Graph Neural Networks (GNNs) have gained prominence due to their ability to capture the complex structural and relational information inherent in molecular graphs. Despite their effectiveness, the “black-box” nature of GNNs remains a significant obstacle to their widespread adoption in chemistry, as it hinders interpretability and trust. In this context, several explanation methods based on factual reasoning have emerged. These methods aim to interpret the predictions made by GNNs by analyzing the key features contributing to the prediction. However, these approaches fail to answer critical questions: “How to ensure that the structure-property mapping learned by GNNs is consistent with established domain knowledge”. In this paper, we propose MMGCF, a novel counterfactual explanation framework designed specifically for the prediction of GNN-based molecular properties. MMGCF constructs a hierarchical tree structure on molecular motifs, enabling the systematic generation of counterfactuals through motif perturbations. This framework identifies causally significant motifs and elucidates their impact on model predictions, offering insights into the relationship between structural modifications and predicted properties. Our method demonstrates its effectiveness through comprehensive quantitative and qualitative evaluations of four real-world molecular datasets.
基金support provided by The Science and Technology Development Fund,Macao SAR,China(File Nos.0057/2020/AGJ and SKL-IOTSC-2021-2023)Science and Technology Program of Guangdong Province,China(Grant No.2021A0505080009).
文摘Accurate prediction of shield tunneling-induced settlement is a complex problem that requires consideration of many influential parameters.Recent studies reveal that machine learning(ML)algorithms can predict the settlement caused by tunneling.However,well-performing ML models are usually less interpretable.Irrelevant input features decrease the performance and interpretability of an ML model.Nonetheless,feature selection,a critical step in the ML pipeline,is usually ignored in most studies that focused on predicting tunneling-induced settlement.This study applies four techniques,i.e.Pearson correlation method,sequential forward selection(SFS),sequential backward selection(SBS)and Boruta algorithm,to investigate the effect of feature selection on the model’s performance when predicting the tunneling-induced maximum surface settlement(S_(max)).The data set used in this study was compiled from two metro tunnel projects excavated in Hangzhou,China using earth pressure balance(EPB)shields and consists of 14 input features and a single output(i.e.S_(max)).The ML model that is trained on features selected from the Boruta algorithm demonstrates the best performance in both the training and testing phases.The relevant features chosen from the Boruta algorithm further indicate that tunneling-induced settlement is affected by parameters related to tunnel geometry,geological conditions and shield operation.The recently proposed Shapley additive explanations(SHAP)method explores how the input features contribute to the output of a complex ML model.It is observed that the larger settlements are induced during shield tunneling in silty clay.Moreover,the SHAP analysis reveals that the low magnitudes of face pressure at the top of the shield increase the model’s output。
文摘This paper takes a microanalytic perspective on the speech and gestures used by one teacher of ESL (English as a Second Language) in an intensive English program classroom. Videotaped excerpts from her intermediate-level grammar course were transcribed to represent the speech, gesture and other non-verbal behavior that accompanied unplanned explanations of vocabulary that arose during three focus-on-form lessons. The gesture classification system of McNeill (1992), which delineates different types of hand movements (iconics metaphorics, deictics, beats), was used to understand the role the gestures played in these explanations. Results suggest that gestures and other non-verbal behavior are forms of input to classroom second language learners that must be considered a salient factor in classroom-based SLA (Second Language Acquisition) research
文摘In this paper I examine the following claims by William Eaton in his monograph Boyle on Fire: (i) that Boyle's religious convictions led him to believe that the world was not completely explicable, and this shows that there is a shortcoming in the power of mechanical explanations; (ii) that mechanical explanations offer only sufficient, not necessary explanations, and this too was taken by Boyle to be a limit in the explanatory power of mechanical explanations; (iii) that the mature Boyle thought that there could be more intelligible explanatory models than mechanism; and (iv) that what Boyle says at any point in his career is incompatible with the statement of Maria Boas-Hall, i.e., that the mechanical hypothesis can explicate all natural phenomena. Since all four of these claims are part of Eaton's developmental argument, my rejection of them will not only show how the particular developmental story Eaton diagnoses is inaccurate, but will also explain what limits there actually are in Boyle's account of the intelligibility of mechanical explanations. My account will also show why important philosophers like Locke and Leibniz should be interested in Boyle's philosophical work.
基金supported by the National Key R&D Program of China(Technology and application of wind power/photovoltaic power prediction for promoting renewable energy consumption)under Grant(2018YFB0904200).
文摘Wind power forecasting(WPF)is important for safe,stable,and reliable integration of new energy technologies into power systems.Machine learning(ML)algorithms have recently attracted increasing attention in the field of WPF.However,opaque decisions and lack of trustworthiness of black-box models for WPF could cause scheduling risks.This study develops a method for identifying risky models in practical applications and avoiding the risks.First,a local interpretable model-agnostic explanations algorithm is introduced and improved for WPF model analysis.On that basis,a novel index is presented to quantify the level at which neural networks or other black-box models can trust features involved in training.Then,by revealing the operational mechanism for local samples,human interpretability of the black-box model is examined under different accuracies,time horizons,and seasons.This interpretability provides a basis for several technical routes for WPF from the viewpoint of the forecasting model.Moreover,further improvements in accuracy of WPF are explored by evaluating possibilities of using interpretable ML models that use multi-horizons global trust modeling and multi-seasons interpretable feature selection methods.Experimental results from a wind farm in China show that error can be robustly reduced.
文摘Generative Artificial Intelligence(GAI)refers to a class of AI systems capable of creating novel,coherent,and contextually relevant content—such as text,images,audio,and video—based on patterns learned from extensive training datasets.The public release and rapid refinement of large language models(LLMs)like ChatGPT have accelerated the adoption of GAI across various medical specialties,offering new tools for education,clinical simulation,and research.Dermatology training,which heavily relies on visual pattern recognition and requires extensive exposure to diverse morphological presentations,faces persistent challenges such as uneven distribu-tion of educational resources,limited patient exposure for rare conditions,and variability in teaching quality.Exploring the integration of GAI into pedagogical frameworks offers innovative approaches to address these challenges,potentially enhancing the quality,standardization,scalability,and accessibility of dermatology ed-ucation.This comprehensive review examines the core concepts and technical foundations of GAI,highlights its specific applications within dermatology teaching and learning—including simulated case generation,per-sonalized learning pathways,and academic support—and discusses the current limitations,practical challenges,and ethical considerations surrounding its use.The aim is to provide a balanced perspective on the significant potential of GAI for transforming dermatology education and to offer evidence-based insights to guide future exploration,implementation,and policy development.
基金funded by Ongoing Research Funding Program for Project number(ORF-2025-648),King Saud University,Riyadh,Saudi Arabia.
文摘Heart disease remains a leading cause of mortality worldwide,emphasizing the urgent need for reliable and interpretable predictive models to support early diagnosis and timely intervention.However,existing Deep Learning(DL)approaches often face several limitations,including inefficient feature extraction,class imbalance,suboptimal classification performance,and limited interpretability,which collectively hinder their deployment in clinical settings.To address these challenges,we propose a novel DL framework for heart disease prediction that integrates a comprehensive preprocessing pipeline with an advanced classification architecture.The preprocessing stage involves label encoding and feature scaling.To address the issue of class imbalance inherent in the personal key indicators of the heart disease dataset,the localized random affine shadowsampling technique is employed,which enhances minority class representation while minimizing overfitting.At the core of the framework lies the Deep Residual Network(DeepResNet),which employs hierarchical residual transformations to facilitate efficient feature extraction and capture complex,non-linear relationships in the data.Experimental results demonstrate that the proposed model significantly outperforms existing techniques,achieving improvements of 3.26%in accuracy,3.16%in area under the receiver operating characteristics,1.09%in recall,and 1.07%in F1-score.Furthermore,robustness is validated using 10-fold crossvalidation,confirming the model’s generalizability across diverse data distributions.Moreover,model interpretability is ensured through the integration of Shapley additive explanations and local interpretable model-agnostic explanations,offering valuable insights into the contribution of individual features to model predictions.Overall,the proposed DL framework presents a robust,interpretable,and clinically applicable solution for heart disease prediction.
文摘本文聚焦我国青海湖流域的水文过程,基于多年气象和水文动态数据,发展了一种融合概念性水文模型FLEX (FluxExchange)和门控循环单元(Gated Recurrent Unit, GRU)的混合模型对流域内最大支流布哈河的逐日径流进行了模拟和预测。在构建混合模型中,采用了三种策略提升模拟精度:引入差分进化自适应算法DREAM(zs)反演水文参数优化FLEX模型;采用变分模态分解(VMD)提取径流数据的信息和特征;利用麻雀搜索算法(SSA)优化深度学习GRU的参数。研究将FLEX模型的模拟结果连同气象数据一起作为神经网络的输入,从而构建了FLEX-VMD-SSA-GRU混合模型。同时,探讨了不同的气象输入条件对模拟结果的影响和贡献:基于7个主要气象要素,由少及多设置了14组输入情景模拟。最后通过SHAP对深度学习方法的结果进行分析,揭示了气象变量对径流长期趋势的贡献和重要度。
基金the National Natural Science Foundation of China(Grant Nos.52090083 and 52378405)Key Technology R&D Plan of Yunnan Provincial Department of Science and Technology(Grant No.202303AA080003)for their financial support.
文摘Contemporary demands necessitate the swift and accurate detection of cracks in critical infrastructures,including tunnels and pavements.This study proposed a transfer learning-based encoder-decoder method with visual explanations for infrastructure crack segmentation.Firstly,a vast dataset containing 7089 images was developed,comprising diverse conditions—simple and complex crack patterns as well as clean and rough backgrounds.Secondly,leveraging transfer learning,an encoder-decoder model with visual explanations was formulated,utilizing varied pre-trained convolutional neural network(CNN)as the encoder.Visual explanations were achieved through gradient-weighted class activation mapping(Grad-CAM)to interpret the CNN segmentation model.Thirdly,accuracy,complexity(computation and model),and memory usage assessed CNN feasibility in practical engineering.Model performance was gauged via prediction and visual explanation.The investigation encompassed hyperparameters,data augmentation,deep learning from scratch vs.transfer learning,segmentation model architectures,segmentation model encoders,and encoder pre-training strategies.Results underscored transfer learning’s potency in enhancing CNN accuracy for crack segmentation,surpassing deep learning from scratch.Notably,encoder classification accuracy bore no significant correlation with CNN segmentation accuracy.Among all tested models,UNet-EfficientNet_B7 excelled in crack segmentation,harmonizing accuracy,complexity,memory usage,prediction,and visual explanation.
文摘This study presents an enhanced convolutional neural network(CNN)model integrated with Explainable Artificial Intelligence(XAI)techniques for accurate prediction and interpretation of wheat crop diseases.The aim is to streamline the detection process while offering transparent insights into the model’s decision-making to support effective disease management.To evaluate the model,a dataset was collected from wheat fields in Kotli,Azad Kashmir,Pakistan,and tested across multiple data splits.The proposed model demonstrates improved stability,faster conver-gence,and higher classification accuracy.The results show significant improvements in prediction accuracy and stability compared to prior works,achieving up to 100%accuracy in certain configurations.In addition,XAI methods such as Local Interpretable Model-agnostic Explanations(LIME)and Shapley Additive Explanations(SHAP)were employed to explain the model’s predictions,highlighting the most influential features contributing to classification decisions.The combined use of CNN and XAI offers a dual benefit:strong predictive performance and clear interpretability of outcomes,which is especially critical in real-world agricultural applications.These findings underscore the potential of integrating deep learning models with XAI to advance automated plant disease detection.The study offers a precise,reliable,and interpretable solution for improving wheat production and promoting agricultural sustainability.Future extensions of this work may include scaling the dataset across broader regions and incorporating additional modalities such as environmental data to enhance model robustness and generalization.
基金supported by the National Key Research and Development Plan(Grant No.2023YFB3712400)the National Key Research and Development Plan(Grant No.2020YFB1713600).
文摘Mechanical properties are critical to the quality of hot-rolled steel pipe products.Accurately understanding the relationship between rolling parameters and mechanical properties is crucial for effective prediction and control.To address this,an industrial big data platform was developed to collect and process multi-source heterogeneous data from the entire production process,providing a complete dataset for mechanical property prediction.The adaptive bandwidth kernel density estimation(ABKDE)method was proposed to adjust bandwidth dynamically based on data density.Combining long short-term memory neural networks with ABKDE offers robust prediction interval capabilities for mechanical properties.The proposed method was deployed in a large-scale steel plant,which demonstrated superior prediction interval performance compared to lower upper bound estimation,mean variance estimation,and extreme learning machine-adaptive bandwidth kernel density estimation,achieving a prediction interval normalized average width of 0.37,a prediction interval coverage probability of 0.94,and the lowest coverage width-based criterion of 1.35.Notably,shapley additive explanations-based explanations significantly improved the proposed model’s credibility by providing a clear analysis of feature impacts.
文摘BACKGROUND Despite the promising prospects of using artificial intelligence and machine learning(ML)for disease classification and prediction purposes,the complexity and lack of explainability of this method make it difficult to apply the constructed models in clinical practice.We developed and validated an interpretable ML model based on magnetic resonance imaging(MRI)radiomics and clinical features for the preoperative prediction of the pathological grades of hepatocellular carcinomas(HCCs).This model will help clinicians better understand the situation and develop personalized treatment plans.AIM To develop and validate an interpretable ML model for preoperative pathological grade prediction in HCC patients via a combination of multisequence MRI radiomics and clinical features.METHODS MRI and clinical data derived from 125 patients with HCCs confirmed by postoperative pathological examinations were retrospectively analyzed.The patients were randomly split into training and validation groups(7:3 ratio).Univariate and multivariate logistic regression analyses were performed to identify independent clinical predictors.The tumor lesions observed on axial fatsuppressed T2-weighted imaging(FS-T2WI),arterial phase(AP),and portal venous phase(PVP)images were delineated in a slice-by-slice manner using 3D-slicer to generate volumetric regions of interest,and radiomic features were extracted.Interclass correlation coefficients were calculated,and least absolute selection and shrinkage operator regression were conducted for feature selection purposes.Six predictive models were subsequently developed for pathological grade prediction:FS-T2WI,AP,PVP,integrated radiomics,clinical,and combined radiomics-clinical(RC)models.The effectiveness of these models was assessed by calculating their area under the receiver operating characteristic curve(AUC)values.The clinical applicability of the models was evaluated via decision curve analysis.Finally,the contributions of the different features contained in the model with optimal performance were interpreted via a SHapley Additive exPlanations analysis.RESULTS Among the 125 patients,87 were assigned to the training group,and 38 were assigned to the validation group.The maximum tumor diameter,hepatitis B virus status,and monocyte count were identified as independent predictors of pathological grade.Twelve optimal radiomic features were ultimately selected.The AUC values obtained for the FS-T2WI model,AP model,PVP model,radiomics model,clinical model,and combined RC model in the training group were 0.761[95%confidence interval(CI):0.562-0.857],0.870(95%CI:0.714-0.918),0.868(95%CI:0.714-0.959),0.917(95%CI:0.857-0.959),0.869(95%CI:0.643-0.973),and 0.941(95%CI:0.857-0.945),respectively;in the validation group,the AUC values were 0.724(95%CI:0.625-0.833),0.802(95%CI:0.686-1.000),0.797(95%CI:0.688-1.000),0.901(95%CI:0.833-0.906),0.865(95%CI:0.594-1.000),and 0.932(95%CI:0.812-1.000),respectively.The combined RC model demonstrated the best performance.Additionally,the decision curve analysis revealed that the combined RC model had satisfactory prediction efficiency,and the SHapley Additive exPlanations value analysis revealed that the“FS-T2WI-wavelet-HLL_gldm_Large Dependence High Gray Level Emphasis”feature contributed the most to the model,exhibiting a positive effect.CONCLUSION An interpretable ML model based on MRI radiomics provides a noninvasive tool for predicting the pathological grade of HCCs,which will help clinicians develop personalized treatment plans.
文摘The study by Huang et al,published in the World Journal of Gastroenterology,advances intrahepatic cholangiocarcinoma(ICC)management by developing a machine-learning model to predict textbook outcomes(TO)based on preoperative factors.By analyzing data from 376 patients across four Chinese medical centers,the researchers identified key variables influencing TO,including Child-Pugh classification,Eastern Cooperative Oncology Group score,hepatitis B status,and tumor size.The model,created using logistic regression and the extreme gradient boosting algorithm,demonstrated high predictive accuracy,with area under the curve values of 0.8825 for internal validation and 0.8346 for external validation.The integration of the Shapley additive explanation technique enhances the interpretability of the model,which is crucial for clinical decision-making.This research highlights the potential of machine learning to improve surgical planning and patient outcomes in ICC,opening possibilities for personalized treatment approaches based on individual patient characteristics and risk factors.
基金funded by the Second Tibetan Plateau Scientific Expedition and Research,Ministry of Science and Technology(Project No.2019QZKK0902)the West Light Foundation of the Chinese Academy of Sciences(Project No.E3R2120)the Research Programme of Institute of Mountain Hazards and Environment,Chinese Academy of Sciences(Project No.IMHE-ZDRW-01).
文摘Machine learning-based Debris Flow Susceptibility Mapping(DFSM)has emerged as an effective approach for assessing debris flow likelihood,yet its application faces three critical challenges:insufficient reliability of training samples caused by biased negative sampling,opaque decision-making mechanisms in models,and subjective susceptibility mapping methods that lack quantitative evaluation criteria.This study focuses on the Yalong River basin.By integrating high-resolution remote sensing interpretation and field surveys,we established a refined sample database that includes 1,736 debris flow gullies.To address spatial bias in traditional random negative sampling,we developed a semi-supervised optimization strategy based on iterative confidence screening.Comparative experiments with four treebased models(XGBoost,CatBoost,LGBM,and Random Forest)reveal that the optimized sampling strategy improved overall model performance by 8%-12%,with XGBoost achieving the highest accuracy(AUC=0.882)and RF performing the lowest(AUC=0.820).SHAP-based global-local interpretability analysis(applicable to all tree models)identifies elevation and short-duration rainfall as dominant controlling factors.Furthermore,among the tested tree-based models,XGBoost optimized with semisupervised sampling demonstrates the highest reliability in debris flow susceptibility mapping(DFSM),achieving a comprehensive accuracy of 83.64%due to its optimal generalization-stability equilibrium.