Black-box models have demonstrated remarkable accuracy in forecasting building energy loads.However,they usually lack interpretability and do not incorporate domain knowledge,making it difficult for users to trust the...Black-box models have demonstrated remarkable accuracy in forecasting building energy loads.However,they usually lack interpretability and do not incorporate domain knowledge,making it difficult for users to trust their predictions in practical applications.One important and interesting question remains unanswered:is it possible to use intrinsically interpretable models to achieve accuracy comparable to that of black-box models?With an aim of answering this question,this study proposes an intrinsically interpretable machine learning-based method to forecast building energy loads.It creatively combines two intrinsically interpretable machine learning algorithms:clustering decision trees and adaptive multiple linear regression.Clustering decision trees aim to automatically identify various building operation conditions,allowing for the training of multiple models tailored to each condition.It can reduce the complexity of model training data,leading to higher accuracy.Adaptive multiple linear regression is an improved regression algorithm tailored to building energy load prediction.It can adaptively modify regression coefficients according to building operations,enhancing the non-linear fitting capability of multiple linear regression.The proposed method is evaluated utilizing the operational data from an office building.The results indicate that the proposed method exhibits comparable accuracy to both random forests and extreme gradient boosting.Furthermore,it shows significantly superior accuracy,with an average improvement of 10.2%,compared with some popular black-box algorithms such as artificial neural networks,support vector regression,and classification and regression trees.As for model interpretability,the proposed method reveals that historical cooling loads are the most crucial for predicting building cooling loads under most conditions.Additionally,outdoor air temperature has a significant contribution to building cooling load prediction during the daytime on weekdays in summer and transition seasons.In the future,it will be valuable to explore integrating the laws of physics into the proposed method to further enhance its interpretability.展开更多
Deep learning has become integral to robotics,particularly in tasks such as robotic grasping,where objects often exhibit diverse shapes,textures,and physical properties.In robotic grasping tasks,due to the diverse cha...Deep learning has become integral to robotics,particularly in tasks such as robotic grasping,where objects often exhibit diverse shapes,textures,and physical properties.In robotic grasping tasks,due to the diverse characteristics of the targets,frequent adjustments to the network architecture and parameters are required to avoid a decrease in model accuracy,which presents a significant challenge for non-experts.Neural Architecture Search(NAS)provides a compelling method through the automated generation of network architectures,enabling the discovery of models that achieve high accuracy through efficient search algorithms.Compared to manually designed networks,NAS methods can significantly reduce design costs,time expenditure,and improve model performance.However,such methods often involve complex topological connections,and these redundant structures can severely reduce computational efficiency.To overcome this challenge,this work puts forward a robotic grasp detection framework founded on NAS.The method automatically designs a lightweight network with high accuracy and low topological complexity,effectively adapting to the target object to generate the optimal grasp pose,thereby significantly improving the success rate of robotic grasping.Additionally,we use Class Activation Mapping(CAM)as an interpretability tool,which captures sensitive information during the perception process through visualized results.The searched model achieved competitive,and in some cases superior,performance on the Cornell and Jacquard public datasets,achieving accuracies of 98.3%and 96.8%,respectively,while sustaining a detection speed of 89 frames per second with only 0.41 million parameters.To further validate its effectiveness beyond benchmark evaluations,we conducted real-world grasping experiments on a UR5 robotic arm,where the model demonstrated reliable performance across diverse objects and high grasp success rates,thereby confirming its practical applicability in robotic manipulation tasks.展开更多
Objective To develop a prognostic prediction model for early-stage triple-negative breast cancer(TNBC)using H&E-stained pathological images and to investigate its underlying biological interpretability.Methods A d...Objective To develop a prognostic prediction model for early-stage triple-negative breast cancer(TNBC)using H&E-stained pathological images and to investigate its underlying biological interpretability.Methods A deep learning model was trained on 340 WSIs and externally validated using 81 TCGA cases.Image-derived features extracted through convolutional neural networks were integrated with clinicopathological variables.Model performance was assessed using ROC curve analysis,and interpretability was evaluated by correlating image features with mRNA-seq data and characteristics of the immune microenvironment.Results The model achieved AUCs of 0.86 and 0.75 in the training and validation cohorts,respectively.Analysis using HoVer-Net indicated that lymphocyte abundance was associated with recurrence risk.Texture-related features showed significant correlations with immune cell infiltration and prognostic gene expression profiles.Conclusion This study demonstrates that deep learning can enable accurate prognostic prediction in early-stage TNBC,with interpretable image features that reflect the tumor immune microenvironment and gene expression profiles.展开更多
Dear Editor,This letter presents a new approach to developing interpretable and reliable soft sensors for Industry 5.0 applications.Although sophisticated machine learning methods have made remarkable strides in soft-...Dear Editor,This letter presents a new approach to developing interpretable and reliable soft sensors for Industry 5.0 applications.Although sophisticated machine learning methods have made remarkable strides in soft-sensor predictive accuracy,ensuring interpretability and reliable performance across varying industrial operating conditions remains a challenge[1]–[4].This is precisely what Industry 5.0,proposed by the European Commission in 2021,advocates[5],[6].It integrates various cutting-edge technologies,such as human-machine interaction,digital twins,cybersecurity and artificial intelligence,to facilitate the development of better soft sensors.展开更多
Landslide susceptibility mapping(LSM)is an essential tool for mitigating the escalating global risk of landslides.However,challenges such as the heterogeneity of different landslide triggers,extensive engineering acti...Landslide susceptibility mapping(LSM)is an essential tool for mitigating the escalating global risk of landslides.However,challenges such as the heterogeneity of different landslide triggers,extensive engineering activities exacerbated reactivation,and the interpretability of data-driven models have hindered the practical application of LSM.This work proposes a novel framework for enhancing LSM considering different triggers for accumulation and rock landslides,leveraging interpretable machine learning and Multi-temporal Interferometric Synthetic Aperture Radar(MT-InSAR)technology.Initially,a refined fieldinvestigation was conducted to delineate the accumulation and rock area according to landslide types,leading to the identificationof relevant contributing factors.Deformation along the slope was then combined with time-series analysis to derive a landslide activity level(AL)index to recognize the likelihood of reactivation or dormancy.The SHapley Additive exPlanation(SHAP)technique facilitated the interpretation of factors and the identificationof determinants in high susceptibility areas.The results indicate that random forest(RF)outperformed other models in both accumulation and rock areas.Key factors including thickness and weak intercalation were identifiedfor accumulation and rock landslides.The introduction of AL substantially enhanced the predictive capability of the LSM and outperformed models that neglect movement trends or deformation rates with an average ratio of 81.23%in high susceptibility zones.Besides,the fieldvalidation confirmedthat 83.8%of newly identifiedlandslides were correctly upgraded.Given its efficiencyand operational simplicity,the proposed hybrid model opens new avenues for the feasibility of enhancement in LSM at urban settlements worldwide.展开更多
Ground water is a crucial ecological resource and source of drinking water to a great percentage of theworld population.The quality of groundwater in an area with industrial emission and air pollution is an especially...Ground water is a crucial ecological resource and source of drinking water to a great percentage of theworld population.The quality of groundwater in an area with industrial emission and air pollution is an especiallyimportant issue that requires proper evaluation.This paper introduces a spatiotemporal deep learning model thatincorporates the use of metaheuristic optimization in predicting groundwater quality in various pollution contexts.Thegiven method is a combination of the Spatial-Temporal-Assisted Deep Belief Network(StaDBN)and a hybrid WhaleOptimization Algorithm and Tiki-Taka Algorithms(WOA-TTA)that would model intricate patterns of contamination.Historical ground water data sets with the hydrochemical data and time are preprocessed and pertinent and nonredundant features are determined with the Addax Optimization Algorithm(AOA).Spatial and temporal dependenciesare explicitly integrated in StaDBN architecture to facilitate representation learning,and network hyperparametersare optimized by the WOA-TTA module to increase the training efficiency and predictive performance.The modelwas coded in Python and tested based on common statistical measures,such as root mean square error(RMSE),Nash Sutcliffe efficiency(NSE),mean absolute error(MAE),and the correlation coefficient(R).The proposedGWQP-StaDBN-WOA-TTA framework demonstrates superior predictive performance and interpretability comparedto conventional machine learning and deep learning models,achieving higher correlation(R=0.963),improvedNash-Sutcliffe efficiency(NSE=0.84),and substantially lower prediction errors(MAE=0.29,RMSE=0.48),therebyvalidating its effectiveness for groundwater quality assessment under industrial and atmospheric pollution scenarios.展开更多
Geological prospecting and the identification of adverse geological features are essential in tunnel construction,providing critical information to ensure safety and guide engineering decisions.As tunnel projects exte...Geological prospecting and the identification of adverse geological features are essential in tunnel construction,providing critical information to ensure safety and guide engineering decisions.As tunnel projects extend into deeper and more mountainous terrains,engineers face increasingly complex geological conditions,including high water pressure,intense geo-stress,elevated geothermal gradients,and active fault zones.These conditions pose substantial risks such as high-pressure water inrush,largescale collapses,and tunnel boring machine(TBM)blockages.Addressing these challenges requires advanced detection technologies capable of long-distance,high-precision,and intelligent assessments of adverse geology.This paper presents a comprehensive review of recent advancements in tunnel geological ahead prospecting methods.It summarizes the fundamental principles,technical maturity,key challenges,development trends,and real-world applications of various detection techniques.Airborne and semi-airborne geophysical methods enable large-scale reconnaissance for initial surveys in complex terrain.Tunnel-and borehole-based approaches offer high-resolution detection during excavation,including seismic ahead prospecting(SAP),TBM rock-breaking source seismic methods,fulltime-domain tunnel induced polarization(TIP),borehole electrical resistivity,and ground penetrating radar(GPR).To address scenarios involving multiple,coexisting adverse geologies,intelligent inversion and geological identification methods have been developed based on multi-source data fusion and artificial intelligence(AI)techniques.Overall,these advances significantly improve detection range,resolution,and geological characterization capabilities.The methods demonstrate strong adaptability to complex environments and provide reliable subsurface information,supporting safer and more efficient tunnel construction.展开更多
Artificial Intelligence(AI)is changing healthcare by helping with diagnosis.However,for doctors to trust AI tools,they need to be both accurate and easy to understand.In this study,we created a new machine learning sy...Artificial Intelligence(AI)is changing healthcare by helping with diagnosis.However,for doctors to trust AI tools,they need to be both accurate and easy to understand.In this study,we created a new machine learning system for the early detection of Autism Spectrum Disorder(ASD)in children.Our main goal was to build a model that is not only good at predicting ASD but also clear in its reasoning.For this,we combined several different models,including Random Forest,XGBoost,and Neural Networks,into a single,more powerful framework.We used two different types of datasets:(i)a standard behavioral dataset and(ii)a more complex multimodal dataset with images,audio,and physiological information.The datasets were carefully preprocessed for missing values,redundant features,and dataset imbalance to ensure fair learning.The results outperformed the state-of-the-art with a Regularized Neural Network,achieving 97.6%accuracy on behavioral data.Whereas,on the multimodal data,the accuracy is 98.2%.Other models also did well with accuracies consistently above 96%.We also used SHAP and LIME on a behavioral dataset for models’explainability.展开更多
Dear Editor,I am writing in response to Jamil's letter,"Interpretative Challenges of the Missing Perilymph'Sign in PLF Diagnosis."I concur with the author's emphasis on the necessity for cautious...Dear Editor,I am writing in response to Jamil's letter,"Interpretative Challenges of the Missing Perilymph'Sign in PLF Diagnosis."I concur with the author's emphasis on the necessity for cautious interpretation of low-signal areas as evidence of active perilymph leakage,requiring correlation with clinical findings,surgical confirmation,and longitudinal imaging changes.展开更多
Deep Learning(DL)model has been widely used in the field of Synthetic Aperture Radar Automatic Target Recognition(SAR-ATR)and has achieved excellent performance.However,the black-box nature of DL models has been the f...Deep Learning(DL)model has been widely used in the field of Synthetic Aperture Radar Automatic Target Recognition(SAR-ATR)and has achieved excellent performance.However,the black-box nature of DL models has been the focus of criticism,especially in the application of SARATR,which is closely associated with the national defense and security domain.To address these issues,a new interpretable recognition model Physics-Guided BagNet(PGBN)is proposed in this article.The model adopts an interpretable convolutional neural network framework and uses time–frequency analysis to extract physical scattering features in SAR images.Based on the physical scattering features,an unsupervised segmentation method is proposed to distinguish targets from the background in SAR images.On the basis of the segmentation result,a structure is designed,which constrains the model's spatial attention to focus more on the targets themselves rather than the background,thereby making the model's decision-making more in line with physical principles.In contrast to previous interpretable research methods,this model combines interpretable structure with physical interpretability,further reducing the model's risk of error recognition.Experiments on the MSTAR dataset verify that the PGBN model exhibits excellent interpretability and recognition performance,and comparative experiments with heatmaps indicate that the physical feature guidance module presented in this article can constrain the model to focus more on the target itself rather than the background.展开更多
Based on 1,003 articles about empirical research on interpreting teaching from 2002 to 2022 retrieved from China National Knowledge Internet,this paper extracts three main research methods,uncovering common problems i...Based on 1,003 articles about empirical research on interpreting teaching from 2002 to 2022 retrieved from China National Knowledge Internet,this paper extracts three main research methods,uncovering common problems in interpreting education and practical teaching suggestions:(1)Corpus-based researches collect numerous audios to study typical mistakes made by interpreting learners,particularly pause and self-repair,and suggest interpreting teaching improve learners’ability to use language chunks and encourage students to interpret smoothly;(2)Questionnaire surveys help understand requirements for professional interpreters and how interpreting teaching meets market demands;(3)Teaching experiments last for one to two semesters,addressing issues like outdated teaching materials and modes,and show how teaching materials and modes integrate modern technology.But empirical researches need to build new corpora,professional interpreters’corpora and address problems that haven’t been adequately discussed.This paper is helpful for improving interpreting education in China and other countries and for making clear tasks to be fulfilled in empirical research on interpreting education.展开更多
Nowadays, college students, as a main part of interpreters, have engaged more and more in interpreting practices in formsof foreign vips’reception, telephone interpreting, escort interpreting and consecutive interp...Nowadays, college students, as a main part of interpreters, have engaged more and more in interpreting practices in formsof foreign vips’reception, telephone interpreting, escort interpreting and consecutive interpreting, etc. However, these practicesstill remain a lot of problems, such as low quality, disordered management and improper resource utilization, which are in urgentneed of systematic interpreting project management. Combined with the features of students’interpreting practice, students-oriented interpreting project can better manage these problems above and build a standardized and effective language servicegroup. After summarizing the features of students’interpreting practice, this paper will focus on the concrete application of interpreting project management in college students’interpreting practice. Furthermore, this paper will also provide specific work flowand methods in students-oriented interpreting project.展开更多
BACKGROUND To investigate the preoperative factors influencing textbook outcomes(TO)in Intrahepatic cholangiocarcinoma(ICC)patients and evaluate the feasibility of an interpretable machine learning model for preoperat...BACKGROUND To investigate the preoperative factors influencing textbook outcomes(TO)in Intrahepatic cholangiocarcinoma(ICC)patients and evaluate the feasibility of an interpretable machine learning model for preoperative prediction of TO,we developed a machine learning model for preoperative prediction of TO and used the SHapley Additive exPlanations(SHAP)technique to illustrate the prediction process.AIM To analyze the factors influencing textbook outcomes before surgery and to establish interpretable machine learning models for preoperative prediction.METHODS A total of 376 patients diagnosed with ICC were retrospectively collected from four major medical institutions in China,covering the period from 2011 to 2017.Logistic regression analysis was conducted to identify preoperative variables associated with achieving TO.Based on these variables,an EXtreme Gradient Boosting(XGBoost)machine learning prediction model was constructed using the XGBoost package.The SHAP(package:Shapviz)algorithm was employed to visualize each variable's contribution to the model's predictions.Kaplan-Meier survival analysis was performed to compare the prognostic differences between the TO-achieving and non-TO-achieving groups.RESULTS Among 376 patients,287 were included in the training group and 89 in the validation group.Logistic regression identified the following preoperative variables influencing TO:Child-Pugh classification,Eastern Cooperative Oncology Group(ECOG)score,hepatitis B,and tumor size.The XGBoost prediction model demonstrated high accuracy in internal validation(AUC=0.8825)and external validation(AUC=0.8346).Survival analysis revealed that the disease-free survival rates for patients achieving TO at 1,2,and 3 years were 64.2%,56.8%,and 43.4%,respectively.CONCLUSION Child-Pugh classification,ECOG score,hepatitis B,and tumor size are preoperative predictors of TO.In both the training group and the validation group,the machine learning model had certain effectiveness in predicting TO before surgery.The SHAP algorithm provided intuitive visualization of the machine learning prediction process,enhancing its interpretability.展开更多
The application of machine learning in alloy design is increasingly widespread,yet traditional models still face challenges when dealing with limited datasets and complex nonlinear relationships.This work proposes an ...The application of machine learning in alloy design is increasingly widespread,yet traditional models still face challenges when dealing with limited datasets and complex nonlinear relationships.This work proposes an interpretable machine learning method based on data augmentation and reconstruction,excavating high-performance low-alloyed magnesium(Mg)alloys.The data augmentation technique expands the original dataset through Gaussian noise.The data reconstruction method reorganizes and transforms the original data to extract more representative features,significantly improving the model's generalization ability and prediction accuracy,with a coefficient of determination(R^(2))of 95.9%for the ultimate tensile strength(UTS)model and a R^(2)of 95.3%for the elongation-to-failure(EL)model.The correlation coefficient assisted screening(CCAS)method is proposed to filter low-alloyed target alloys.A new Mg-2.2Mn-0.4Zn-0.2Al-0.2Ca(MZAX2000,wt%)alloy is designed and extruded into bar at given processing parameters,achieving room-temperature strength-ductility synergy showing an excellent UTS of 395 MPa and a high EL of 17.9%.This is closely related to its hetero-structured characteristic in the as-extruded MZAX2000 alloy consisting of coarse grains(16%),fine grains(75%),and fiber regions(9%).Therefore,this work offers new insights into optimizing alloy compositions and processing parameters for attaining new high strong and ductile low-alloyed Mg alloys.展开更多
Forecasting landslide deformation is challenging due to influence of various internal and external factors on the occurrence of systemic and localized heterogeneities.Despite the potential to improve landslide predict...Forecasting landslide deformation is challenging due to influence of various internal and external factors on the occurrence of systemic and localized heterogeneities.Despite the potential to improve landslide predictability,deep learning has yet to be sufficiently explored for complex deformation patterns associated with landslides and is inherently opaque.Herein,we developed a holistic landslide deformation forecasting method that considers spatiotemporal correlations of landslide deformation by integrating domain knowledge into interpretable deep learning.By spatially capturing the interconnections between multiple deformations from different observation points,our method contributes to the understanding and forecasting of landslide systematic behavior.By integrating specific domain knowledge relevant to each observation point and merging internal properties with external variables,the local heterogeneity is considered in our method,identifying deformation temporal patterns in different landslide zones.Case studies involving reservoir-induced landslides and creeping landslides demonstrated that our approach(1)enhances the accuracy of landslide deformation forecasting,(2)identifies significant contributing factors and their influence on spatiotemporal deformation characteristics,and(3)demonstrates how identifying these factors and patterns facilitates landslide forecasting.Our research offers a promising and pragmatic pathway toward a deeper understanding and forecasting of complex landslide behaviors.展开更多
Despite significant progress in the Prognostics and Health Management(PHM)domain using pattern learning systems from data,machine learning(ML)still faces challenges related to limited generalization and weak interpret...Despite significant progress in the Prognostics and Health Management(PHM)domain using pattern learning systems from data,machine learning(ML)still faces challenges related to limited generalization and weak interpretability.A promising approach to overcoming these challenges is to embed domain knowledge into the ML pipeline,enhancing the model with additional pattern information.In this paper,we review the latest developments in PHM,encapsulated under the concept of Knowledge Driven Machine Learning(KDML).We propose a hierarchical framework to define KDML in PHM,which includes scientific paradigms,knowledge sources,knowledge representations,and knowledge embedding methods.Using this framework,we examine current research to demonstrate how various forms of knowledge can be integrated into the ML pipeline and provide roadmap to specific usage.Furthermore,we present several case studies that illustrate specific implementations of KDML in the PHM domain,including inductive experience,physical model,and signal processing.We analyze the improvements in generalization capability and interpretability that KDML can achieve.Finally,we discuss the challenges,potential applications,and usage recommendations of KDML in PHM,with a particular focus on the critical need for interpretability to ensure trustworthy deployment of artificial intelligence in PHM.展开更多
Intelligent fault diagnosis technology plays an indispensable role in ensuring the safety,stability,and efficiency of railway operations.However,existing studies have the following limitations.1)They are typical black-...Intelligent fault diagnosis technology plays an indispensable role in ensuring the safety,stability,and efficiency of railway operations.However,existing studies have the following limitations.1)They are typical black-box models that lacks interpretability as well as they fuse features by simply stacking them,overlooking the discrepancies in the importance of different features,which reduces the credibility and diagnosis accuracy of the models.2)They ignore the effects of potentially mistaken labels in the training datasets disrupting the ability of the models to learn the true data distribution,which degrades the generalization performance of intelligent diagnosis models,especially when the training samples are limited.To address the above items,an interpretable few-shot framework for fault diagnosis with noisy labels is proposed for train transmission systems.In the proposed framework,a feature extractor is constructed by stacked frequency band focus modules,which can capture signal features in different frequency bands and further adaptively concentrate on the features corresponding to the potential fault characteristic frequency.Then,according to prototypical network,a novel metric-based classifier is developed that is tolerant to mislabeled support samples in the case of limited samples.Besides,a new loss function is designed to decrease the impact of label mistakes in query datasets.Finally,fault simulation experiments of subway train transmission systems are designed and conducted,and the effectiveness as well as superiority of the proposed method are proved by ablation experiments and comparison with the existing methods.展开更多
Topographic maps,as essential tools and sources of information for geographic research,contain precise spatial locations and rich map features,and they illustrate spatio-temporal information on the distribution and di...Topographic maps,as essential tools and sources of information for geographic research,contain precise spatial locations and rich map features,and they illustrate spatio-temporal information on the distribution and differences of various surface features.Currently,topographic maps are mainly stored in raster and vector formats.Extraction of the spatio-temporal knowledge in the maps—such as spatial distribution patterns,feature relationships,and dynamic evolution—still primarily relies on manual interpretation.However,manual interpretation is time-consuming and laborious,especially for large-scale,long-term map knowledge extraction and application.With the development of artificial intelligence technology,it is possible to improve the automation level of map knowledge interpretation.Therefore,the present study proposes an automatic interpretation method for raster topographic map knowledge based on deep learning.To address the limitations of current data-driven intelligent technology in learning map spatial relations and cognitive logic,we establish a formal description of map knowledge by mapping the relationship between map knowledge and features,thereby ensuring interpretation accuracy.Subsequently,deep learning techniques are employed to extract map features automatically,and the spatio-temporal knowledge is constructed by combining formal descriptions of geographic feature knowledge.Validation experiments demonstrate that the proposed method effectively achieves automatic interpretation of spatio-temporal knowledge of geographic features in maps,with an accuracy exceeding 80%.The findings of the present study contribute to machine understanding of spatio-temporal differences in map knowledge and advances the intelligent interpretation and utilization of cartographic information.展开更多
In recent years,deeps learning has been widely applied in synthetic aperture radar(SAR)image processing.However,the collection of large-scale labeled SAR images is challenging and costly,and the classification accurac...In recent years,deeps learning has been widely applied in synthetic aperture radar(SAR)image processing.However,the collection of large-scale labeled SAR images is challenging and costly,and the classification accuracy is often poor when only limited SAR images are available.To address this issue,we propose a novel framework for sparse SAR target classification under few-shot cases,termed the transfer learning-based interpretable lightweight convolutional neural network(TL-IL-CNN).Additionally,we employ enhanced gradient-weighted class activation mapping(Grad-CAM)to mitigate the“black box”effect often associated with deep learning models and to explore the mechanisms by which a CNN classifies various sparse SAR targets.Initially,we apply a novel bidirectional iterative soft thresholding(BiIST)algorithm to generate sparse images of superior quality compared to those produced by traditional matched filtering(MF)techniques.Subsequently,we pretrain multiple shallow CNNs on a simulated SAR image dataset.Using the sparse SAR dataset as input for the CNNs,we assess the efficacy of transfer learning in sparse SAR target classification and suggest the integration of TL-IL-CNN to enhance the classification accuracy further.Finally,Grad-CAM is utilized to provide visual explanations for the predictions made by the classification framework.The experimental results on the MSTAR dataset reveal that the proposed TL-IL-CNN achieves nearly 90%classification accuracy with only 20%of the training data required under standard operating conditions(SOC),surpassing typical deep learning methods such as vision Transformer(ViT)in the context of small samples.Remarkably,it even presents better performance under extended operating conditions(EOC).Furthermore,the application of Grad-CAM elucidates the CNN’s differentiation process among various sparse SAR targets.The experiments indicate that the model focuses on the target and the background can differ among target classes.The study contributes to an enhanced understanding of the interpretability of such results and enables us to infer the classification outcomes for each category more accurately.展开更多
To enhance the prediction accuracy of landslides in in Longyan City,China,this study developed a methodology for geologic hazard susceptibility assessment based on a coupled model composed of a Geographic Information ...To enhance the prediction accuracy of landslides in in Longyan City,China,this study developed a methodology for geologic hazard susceptibility assessment based on a coupled model composed of a Geographic Information System(GIS)with integrated spatial data,a frequency ratio(FR)model,and a random forest(RF)model(also referred to as the coupled FR-RF model).The coupled FR-RF model was constructed based on the analysis of nine influential factors,including distance from roads,normalized difference vegetation index(NDVI),and slope.The performance of the coupled FR-RF model was assessed using metrics such as Receiver Operating Characteristic(ROC)and Precision-Recall(PR)curves,yielding Area Under the Curve(AUC)values of 0.93 and 0.95,which indicate high predictive accuracy and reliability for geological hazard forecasting.Based on the model predictions,five susceptibility levels were determined in the study area,providing crucial spatial information for geologic hazard prevention and control.The contributions of various influential factors to landslide susceptibility were determined using SHapley Additive exPlanations(SHAP)analysis and the Gini index,enhancing the model interpretability and transparency.Additionally,this study discussed the limitations of the coupled FR-RF model and the prospects for its improvement using new technologies.This study provides an innovative method and theoretical support for geologic hazard prediction and management,holding promising prospects for application.展开更多
基金supported by the National Natural Science Foundation of China(No.52161135202).
文摘Black-box models have demonstrated remarkable accuracy in forecasting building energy loads.However,they usually lack interpretability and do not incorporate domain knowledge,making it difficult for users to trust their predictions in practical applications.One important and interesting question remains unanswered:is it possible to use intrinsically interpretable models to achieve accuracy comparable to that of black-box models?With an aim of answering this question,this study proposes an intrinsically interpretable machine learning-based method to forecast building energy loads.It creatively combines two intrinsically interpretable machine learning algorithms:clustering decision trees and adaptive multiple linear regression.Clustering decision trees aim to automatically identify various building operation conditions,allowing for the training of multiple models tailored to each condition.It can reduce the complexity of model training data,leading to higher accuracy.Adaptive multiple linear regression is an improved regression algorithm tailored to building energy load prediction.It can adaptively modify regression coefficients according to building operations,enhancing the non-linear fitting capability of multiple linear regression.The proposed method is evaluated utilizing the operational data from an office building.The results indicate that the proposed method exhibits comparable accuracy to both random forests and extreme gradient boosting.Furthermore,it shows significantly superior accuracy,with an average improvement of 10.2%,compared with some popular black-box algorithms such as artificial neural networks,support vector regression,and classification and regression trees.As for model interpretability,the proposed method reveals that historical cooling loads are the most crucial for predicting building cooling loads under most conditions.Additionally,outdoor air temperature has a significant contribution to building cooling load prediction during the daytime on weekdays in summer and transition seasons.In the future,it will be valuable to explore integrating the laws of physics into the proposed method to further enhance its interpretability.
基金funded by Guangdong Basic and Applied Basic Research Foundation(2023B1515120064)National Natural Science Foundation of China(62273097).
文摘Deep learning has become integral to robotics,particularly in tasks such as robotic grasping,where objects often exhibit diverse shapes,textures,and physical properties.In robotic grasping tasks,due to the diverse characteristics of the targets,frequent adjustments to the network architecture and parameters are required to avoid a decrease in model accuracy,which presents a significant challenge for non-experts.Neural Architecture Search(NAS)provides a compelling method through the automated generation of network architectures,enabling the discovery of models that achieve high accuracy through efficient search algorithms.Compared to manually designed networks,NAS methods can significantly reduce design costs,time expenditure,and improve model performance.However,such methods often involve complex topological connections,and these redundant structures can severely reduce computational efficiency.To overcome this challenge,this work puts forward a robotic grasp detection framework founded on NAS.The method automatically designs a lightweight network with high accuracy and low topological complexity,effectively adapting to the target object to generate the optimal grasp pose,thereby significantly improving the success rate of robotic grasping.Additionally,we use Class Activation Mapping(CAM)as an interpretability tool,which captures sensitive information during the perception process through visualized results.The searched model achieved competitive,and in some cases superior,performance on the Cornell and Jacquard public datasets,achieving accuracies of 98.3%and 96.8%,respectively,while sustaining a detection speed of 89 frames per second with only 0.41 million parameters.To further validate its effectiveness beyond benchmark evaluations,we conducted real-world grasping experiments on a UR5 robotic arm,where the model demonstrated reliable performance across diverse objects and high grasp success rates,thereby confirming its practical applicability in robotic manipulation tasks.
基金Supported by Capital’s Funds for Health Improvement and Research(CFH2024-1-4021)。
文摘Objective To develop a prognostic prediction model for early-stage triple-negative breast cancer(TNBC)using H&E-stained pathological images and to investigate its underlying biological interpretability.Methods A deep learning model was trained on 340 WSIs and externally validated using 81 TCGA cases.Image-derived features extracted through convolutional neural networks were integrated with clinicopathological variables.Model performance was assessed using ROC curve analysis,and interpretability was evaluated by correlating image features with mRNA-seq data and characteristics of the immune microenvironment.Results The model achieved AUCs of 0.86 and 0.75 in the training and validation cohorts,respectively.Analysis using HoVer-Net indicated that lymphocyte abundance was associated with recurrence risk.Texture-related features showed significant correlations with immune cell infiltration and prognostic gene expression profiles.Conclusion This study demonstrates that deep learning can enable accurate prognostic prediction in early-stage TNBC,with interpretable image features that reflect the tumor immune microenvironment and gene expression profiles.
文摘Dear Editor,This letter presents a new approach to developing interpretable and reliable soft sensors for Industry 5.0 applications.Although sophisticated machine learning methods have made remarkable strides in soft-sensor predictive accuracy,ensuring interpretability and reliable performance across varying industrial operating conditions remains a challenge[1]–[4].This is precisely what Industry 5.0,proposed by the European Commission in 2021,advocates[5],[6].It integrates various cutting-edge technologies,such as human-machine interaction,digital twins,cybersecurity and artificial intelligence,to facilitate the development of better soft sensors.
基金supported by the National Key R&D Program of China(Grant No.2023YFC3007201)the National Natural Science Foundation of China(Grant No.42377161)the Opening Fund of Key Laboratory of Geological Survey and Evaluation of Ministry of Education(Grant No.GLAB 2024ZR03).
文摘Landslide susceptibility mapping(LSM)is an essential tool for mitigating the escalating global risk of landslides.However,challenges such as the heterogeneity of different landslide triggers,extensive engineering activities exacerbated reactivation,and the interpretability of data-driven models have hindered the practical application of LSM.This work proposes a novel framework for enhancing LSM considering different triggers for accumulation and rock landslides,leveraging interpretable machine learning and Multi-temporal Interferometric Synthetic Aperture Radar(MT-InSAR)technology.Initially,a refined fieldinvestigation was conducted to delineate the accumulation and rock area according to landslide types,leading to the identificationof relevant contributing factors.Deformation along the slope was then combined with time-series analysis to derive a landslide activity level(AL)index to recognize the likelihood of reactivation or dormancy.The SHapley Additive exPlanation(SHAP)technique facilitated the interpretation of factors and the identificationof determinants in high susceptibility areas.The results indicate that random forest(RF)outperformed other models in both accumulation and rock areas.Key factors including thickness and weak intercalation were identifiedfor accumulation and rock landslides.The introduction of AL substantially enhanced the predictive capability of the LSM and outperformed models that neglect movement trends or deformation rates with an average ratio of 81.23%in high susceptibility zones.Besides,the fieldvalidation confirmedthat 83.8%of newly identifiedlandslides were correctly upgraded.Given its efficiencyand operational simplicity,the proposed hybrid model opens new avenues for the feasibility of enhancement in LSM at urban settlements worldwide.
基金Fund for funding this research work under Research Support Program for Central labs at King Khalid University through the project number CL/CO/B/6.
文摘Ground water is a crucial ecological resource and source of drinking water to a great percentage of theworld population.The quality of groundwater in an area with industrial emission and air pollution is an especiallyimportant issue that requires proper evaluation.This paper introduces a spatiotemporal deep learning model thatincorporates the use of metaheuristic optimization in predicting groundwater quality in various pollution contexts.Thegiven method is a combination of the Spatial-Temporal-Assisted Deep Belief Network(StaDBN)and a hybrid WhaleOptimization Algorithm and Tiki-Taka Algorithms(WOA-TTA)that would model intricate patterns of contamination.Historical ground water data sets with the hydrochemical data and time are preprocessed and pertinent and nonredundant features are determined with the Addax Optimization Algorithm(AOA).Spatial and temporal dependenciesare explicitly integrated in StaDBN architecture to facilitate representation learning,and network hyperparametersare optimized by the WOA-TTA module to increase the training efficiency and predictive performance.The modelwas coded in Python and tested based on common statistical measures,such as root mean square error(RMSE),Nash Sutcliffe efficiency(NSE),mean absolute error(MAE),and the correlation coefficient(R).The proposedGWQP-StaDBN-WOA-TTA framework demonstrates superior predictive performance and interpretability comparedto conventional machine learning and deep learning models,achieving higher correlation(R=0.963),improvedNash-Sutcliffe efficiency(NSE=0.84),and substantially lower prediction errors(MAE=0.29,RMSE=0.48),therebyvalidating its effectiveness for groundwater quality assessment under industrial and atmospheric pollution scenarios.
基金supported by the National Natural Science Foundation of China(Grant Nos.52021005,52325904,and 51991391)。
文摘Geological prospecting and the identification of adverse geological features are essential in tunnel construction,providing critical information to ensure safety and guide engineering decisions.As tunnel projects extend into deeper and more mountainous terrains,engineers face increasingly complex geological conditions,including high water pressure,intense geo-stress,elevated geothermal gradients,and active fault zones.These conditions pose substantial risks such as high-pressure water inrush,largescale collapses,and tunnel boring machine(TBM)blockages.Addressing these challenges requires advanced detection technologies capable of long-distance,high-precision,and intelligent assessments of adverse geology.This paper presents a comprehensive review of recent advancements in tunnel geological ahead prospecting methods.It summarizes the fundamental principles,technical maturity,key challenges,development trends,and real-world applications of various detection techniques.Airborne and semi-airborne geophysical methods enable large-scale reconnaissance for initial surveys in complex terrain.Tunnel-and borehole-based approaches offer high-resolution detection during excavation,including seismic ahead prospecting(SAP),TBM rock-breaking source seismic methods,fulltime-domain tunnel induced polarization(TIP),borehole electrical resistivity,and ground penetrating radar(GPR).To address scenarios involving multiple,coexisting adverse geologies,intelligent inversion and geological identification methods have been developed based on multi-source data fusion and artificial intelligence(AI)techniques.Overall,these advances significantly improve detection range,resolution,and geological characterization capabilities.The methods demonstrate strong adaptability to complex environments and provide reliable subsurface information,supporting safer and more efficient tunnel construction.
基金the King Salman center for Disability Research for funding this work through Research Group No.KSRG-2024-050.
文摘Artificial Intelligence(AI)is changing healthcare by helping with diagnosis.However,for doctors to trust AI tools,they need to be both accurate and easy to understand.In this study,we created a new machine learning system for the early detection of Autism Spectrum Disorder(ASD)in children.Our main goal was to build a model that is not only good at predicting ASD but also clear in its reasoning.For this,we combined several different models,including Random Forest,XGBoost,and Neural Networks,into a single,more powerful framework.We used two different types of datasets:(i)a standard behavioral dataset and(ii)a more complex multimodal dataset with images,audio,and physiological information.The datasets were carefully preprocessed for missing values,redundant features,and dataset imbalance to ensure fair learning.The results outperformed the state-of-the-art with a Regularized Neural Network,achieving 97.6%accuracy on behavioral data.Whereas,on the multimodal data,the accuracy is 98.2%.Other models also did well with accuracies consistently above 96%.We also used SHAP and LIME on a behavioral dataset for models’explainability.
文摘Dear Editor,I am writing in response to Jamil's letter,"Interpretative Challenges of the Missing Perilymph'Sign in PLF Diagnosis."I concur with the author's emphasis on the necessity for cautious interpretation of low-signal areas as evidence of active perilymph leakage,requiring correlation with clinical findings,surgical confirmation,and longitudinal imaging changes.
基金co-supported by the National Natural Science Foundation of China(No.62001507)the Youth Talent Lifting Project of the China Association for Science and Technology(No.2021-JCJQ-QT-018)+1 种基金the Program of the Youth Innovation Team of Shaanxi Universitiesthe Natural Science Basic Research Plan in Shaanxi Province of China(No.2023-JC-YB-491)。
文摘Deep Learning(DL)model has been widely used in the field of Synthetic Aperture Radar Automatic Target Recognition(SAR-ATR)and has achieved excellent performance.However,the black-box nature of DL models has been the focus of criticism,especially in the application of SARATR,which is closely associated with the national defense and security domain.To address these issues,a new interpretable recognition model Physics-Guided BagNet(PGBN)is proposed in this article.The model adopts an interpretable convolutional neural network framework and uses time–frequency analysis to extract physical scattering features in SAR images.Based on the physical scattering features,an unsupervised segmentation method is proposed to distinguish targets from the background in SAR images.On the basis of the segmentation result,a structure is designed,which constrains the model's spatial attention to focus more on the targets themselves rather than the background,thereby making the model's decision-making more in line with physical principles.In contrast to previous interpretable research methods,this model combines interpretable structure with physical interpretability,further reducing the model's risk of error recognition.Experiments on the MSTAR dataset verify that the PGBN model exhibits excellent interpretability and recognition performance,and comparative experiments with heatmaps indicate that the physical feature guidance module presented in this article can constrain the model to focus more on the target itself rather than the background.
基金USST Construction Project of English-taught Courses for International Students in 2024Key Course Construction Project in Universities of Shanghai in 2024USST Teaching Achievement Award(postgraduate)Cultivation Project in 2024。
文摘Based on 1,003 articles about empirical research on interpreting teaching from 2002 to 2022 retrieved from China National Knowledge Internet,this paper extracts three main research methods,uncovering common problems in interpreting education and practical teaching suggestions:(1)Corpus-based researches collect numerous audios to study typical mistakes made by interpreting learners,particularly pause and self-repair,and suggest interpreting teaching improve learners’ability to use language chunks and encourage students to interpret smoothly;(2)Questionnaire surveys help understand requirements for professional interpreters and how interpreting teaching meets market demands;(3)Teaching experiments last for one to two semesters,addressing issues like outdated teaching materials and modes,and show how teaching materials and modes integrate modern technology.But empirical researches need to build new corpora,professional interpreters’corpora and address problems that haven’t been adequately discussed.This paper is helpful for improving interpreting education in China and other countries and for making clear tasks to be fulfilled in empirical research on interpreting education.
文摘Nowadays, college students, as a main part of interpreters, have engaged more and more in interpreting practices in formsof foreign vips’reception, telephone interpreting, escort interpreting and consecutive interpreting, etc. However, these practicesstill remain a lot of problems, such as low quality, disordered management and improper resource utilization, which are in urgentneed of systematic interpreting project management. Combined with the features of students’interpreting practice, students-oriented interpreting project can better manage these problems above and build a standardized and effective language servicegroup. After summarizing the features of students’interpreting practice, this paper will focus on the concrete application of interpreting project management in college students’interpreting practice. Furthermore, this paper will also provide specific work flowand methods in students-oriented interpreting project.
基金Supported by National Key Research and Development Program,No.2022YFC2407304Major Research Project for Middle-Aged and Young Scientists of Fujian Provincial Health Commission,No.2021ZQNZD013+2 种基金The National Natural Science Foundation of China,No.62275050Fujian Province Science and Technology Innovation Joint Fund Project,No.2019Y9108Major Science and Technology Projects of Fujian Province,No.2021YZ036017.
文摘BACKGROUND To investigate the preoperative factors influencing textbook outcomes(TO)in Intrahepatic cholangiocarcinoma(ICC)patients and evaluate the feasibility of an interpretable machine learning model for preoperative prediction of TO,we developed a machine learning model for preoperative prediction of TO and used the SHapley Additive exPlanations(SHAP)technique to illustrate the prediction process.AIM To analyze the factors influencing textbook outcomes before surgery and to establish interpretable machine learning models for preoperative prediction.METHODS A total of 376 patients diagnosed with ICC were retrospectively collected from four major medical institutions in China,covering the period from 2011 to 2017.Logistic regression analysis was conducted to identify preoperative variables associated with achieving TO.Based on these variables,an EXtreme Gradient Boosting(XGBoost)machine learning prediction model was constructed using the XGBoost package.The SHAP(package:Shapviz)algorithm was employed to visualize each variable's contribution to the model's predictions.Kaplan-Meier survival analysis was performed to compare the prognostic differences between the TO-achieving and non-TO-achieving groups.RESULTS Among 376 patients,287 were included in the training group and 89 in the validation group.Logistic regression identified the following preoperative variables influencing TO:Child-Pugh classification,Eastern Cooperative Oncology Group(ECOG)score,hepatitis B,and tumor size.The XGBoost prediction model demonstrated high accuracy in internal validation(AUC=0.8825)and external validation(AUC=0.8346).Survival analysis revealed that the disease-free survival rates for patients achieving TO at 1,2,and 3 years were 64.2%,56.8%,and 43.4%,respectively.CONCLUSION Child-Pugh classification,ECOG score,hepatitis B,and tumor size are preoperative predictors of TO.In both the training group and the validation group,the machine learning model had certain effectiveness in predicting TO before surgery.The SHAP algorithm provided intuitive visualization of the machine learning prediction process,enhancing its interpretability.
基金funded by the National Natural Science Foundation of China(No.52204407)the Natural Science Foundation of Jiangsu Province(No.BK20220595)+1 种基金the China Postdoctoral Science Foundation(No.2022M723689)the Industrial Collaborative Innovation Project of Shanghai(No.XTCX-KJ-2022-2-11)。
文摘The application of machine learning in alloy design is increasingly widespread,yet traditional models still face challenges when dealing with limited datasets and complex nonlinear relationships.This work proposes an interpretable machine learning method based on data augmentation and reconstruction,excavating high-performance low-alloyed magnesium(Mg)alloys.The data augmentation technique expands the original dataset through Gaussian noise.The data reconstruction method reorganizes and transforms the original data to extract more representative features,significantly improving the model's generalization ability and prediction accuracy,with a coefficient of determination(R^(2))of 95.9%for the ultimate tensile strength(UTS)model and a R^(2)of 95.3%for the elongation-to-failure(EL)model.The correlation coefficient assisted screening(CCAS)method is proposed to filter low-alloyed target alloys.A new Mg-2.2Mn-0.4Zn-0.2Al-0.2Ca(MZAX2000,wt%)alloy is designed and extruded into bar at given processing parameters,achieving room-temperature strength-ductility synergy showing an excellent UTS of 395 MPa and a high EL of 17.9%.This is closely related to its hetero-structured characteristic in the as-extruded MZAX2000 alloy consisting of coarse grains(16%),fine grains(75%),and fiber regions(9%).Therefore,this work offers new insights into optimizing alloy compositions and processing parameters for attaining new high strong and ductile low-alloyed Mg alloys.
基金supported by the Postdoctoral Fellowship Program of CPSF(Grant No.GZB20230685)the National Science Foundation of China(Grant No.42277161).
文摘Forecasting landslide deformation is challenging due to influence of various internal and external factors on the occurrence of systemic and localized heterogeneities.Despite the potential to improve landslide predictability,deep learning has yet to be sufficiently explored for complex deformation patterns associated with landslides and is inherently opaque.Herein,we developed a holistic landslide deformation forecasting method that considers spatiotemporal correlations of landslide deformation by integrating domain knowledge into interpretable deep learning.By spatially capturing the interconnections between multiple deformations from different observation points,our method contributes to the understanding and forecasting of landslide systematic behavior.By integrating specific domain knowledge relevant to each observation point and merging internal properties with external variables,the local heterogeneity is considered in our method,identifying deformation temporal patterns in different landslide zones.Case studies involving reservoir-induced landslides and creeping landslides demonstrated that our approach(1)enhances the accuracy of landslide deformation forecasting,(2)identifies significant contributing factors and their influence on spatiotemporal deformation characteristics,and(3)demonstrates how identifying these factors and patterns facilitates landslide forecasting.Our research offers a promising and pragmatic pathway toward a deeper understanding and forecasting of complex landslide behaviors.
基金Supported in part by Science Center for Gas Turbine Project(Project No.P2022-DC-I-003-001)National Natural Science Foundation of China(Grant No.52275130).
文摘Despite significant progress in the Prognostics and Health Management(PHM)domain using pattern learning systems from data,machine learning(ML)still faces challenges related to limited generalization and weak interpretability.A promising approach to overcoming these challenges is to embed domain knowledge into the ML pipeline,enhancing the model with additional pattern information.In this paper,we review the latest developments in PHM,encapsulated under the concept of Knowledge Driven Machine Learning(KDML).We propose a hierarchical framework to define KDML in PHM,which includes scientific paradigms,knowledge sources,knowledge representations,and knowledge embedding methods.Using this framework,we examine current research to demonstrate how various forms of knowledge can be integrated into the ML pipeline and provide roadmap to specific usage.Furthermore,we present several case studies that illustrate specific implementations of KDML in the PHM domain,including inductive experience,physical model,and signal processing.We analyze the improvements in generalization capability and interpretability that KDML can achieve.Finally,we discuss the challenges,potential applications,and usage recommendations of KDML in PHM,with a particular focus on the critical need for interpretability to ensure trustworthy deployment of artificial intelligence in PHM.
基金supported in part by the National Key R&D Program of China under Grant 2022YFB4300601in part by the State Key Laboratory of Advanced Rail Autonomous Operation under Grant RAO2023ZZ003.
文摘Intelligent fault diagnosis technology plays an indispensable role in ensuring the safety,stability,and efficiency of railway operations.However,existing studies have the following limitations.1)They are typical black-box models that lacks interpretability as well as they fuse features by simply stacking them,overlooking the discrepancies in the importance of different features,which reduces the credibility and diagnosis accuracy of the models.2)They ignore the effects of potentially mistaken labels in the training datasets disrupting the ability of the models to learn the true data distribution,which degrades the generalization performance of intelligent diagnosis models,especially when the training samples are limited.To address the above items,an interpretable few-shot framework for fault diagnosis with noisy labels is proposed for train transmission systems.In the proposed framework,a feature extractor is constructed by stacked frequency band focus modules,which can capture signal features in different frequency bands and further adaptively concentrate on the features corresponding to the potential fault characteristic frequency.Then,according to prototypical network,a novel metric-based classifier is developed that is tolerant to mislabeled support samples in the case of limited samples.Besides,a new loss function is designed to decrease the impact of label mistakes in query datasets.Finally,fault simulation experiments of subway train transmission systems are designed and conducted,and the effectiveness as well as superiority of the proposed method are proved by ablation experiments and comparison with the existing methods.
基金Deep-time Digital Earth(DDE)Big Science Program(No.GJ-C03-SGF-2025-004)National Natural Science Foundation of China(No.42394063)Sichuan Science and Technology Program(No.2025ZNSFSC0325).
文摘Topographic maps,as essential tools and sources of information for geographic research,contain precise spatial locations and rich map features,and they illustrate spatio-temporal information on the distribution and differences of various surface features.Currently,topographic maps are mainly stored in raster and vector formats.Extraction of the spatio-temporal knowledge in the maps—such as spatial distribution patterns,feature relationships,and dynamic evolution—still primarily relies on manual interpretation.However,manual interpretation is time-consuming and laborious,especially for large-scale,long-term map knowledge extraction and application.With the development of artificial intelligence technology,it is possible to improve the automation level of map knowledge interpretation.Therefore,the present study proposes an automatic interpretation method for raster topographic map knowledge based on deep learning.To address the limitations of current data-driven intelligent technology in learning map spatial relations and cognitive logic,we establish a formal description of map knowledge by mapping the relationship between map knowledge and features,thereby ensuring interpretation accuracy.Subsequently,deep learning techniques are employed to extract map features automatically,and the spatio-temporal knowledge is constructed by combining formal descriptions of geographic feature knowledge.Validation experiments demonstrate that the proposed method effectively achieves automatic interpretation of spatio-temporal knowledge of geographic features in maps,with an accuracy exceeding 80%.The findings of the present study contribute to machine understanding of spatio-temporal differences in map knowledge and advances the intelligent interpretation and utilization of cartographic information.
基金supported in part by the National Natural Science Foundation(Nos.62271248,62401256)in part by the Natural Science Foundation of Ji-angsu Province(Nos.BK20230090,BK20241384)in part by the Key Laboratory of Land Satellite Remote Sens-ing Application,Ministry of Natural Resources of China(No.KLSMNR-K202303)。
文摘In recent years,deeps learning has been widely applied in synthetic aperture radar(SAR)image processing.However,the collection of large-scale labeled SAR images is challenging and costly,and the classification accuracy is often poor when only limited SAR images are available.To address this issue,we propose a novel framework for sparse SAR target classification under few-shot cases,termed the transfer learning-based interpretable lightweight convolutional neural network(TL-IL-CNN).Additionally,we employ enhanced gradient-weighted class activation mapping(Grad-CAM)to mitigate the“black box”effect often associated with deep learning models and to explore the mechanisms by which a CNN classifies various sparse SAR targets.Initially,we apply a novel bidirectional iterative soft thresholding(BiIST)algorithm to generate sparse images of superior quality compared to those produced by traditional matched filtering(MF)techniques.Subsequently,we pretrain multiple shallow CNNs on a simulated SAR image dataset.Using the sparse SAR dataset as input for the CNNs,we assess the efficacy of transfer learning in sparse SAR target classification and suggest the integration of TL-IL-CNN to enhance the classification accuracy further.Finally,Grad-CAM is utilized to provide visual explanations for the predictions made by the classification framework.The experimental results on the MSTAR dataset reveal that the proposed TL-IL-CNN achieves nearly 90%classification accuracy with only 20%of the training data required under standard operating conditions(SOC),surpassing typical deep learning methods such as vision Transformer(ViT)in the context of small samples.Remarkably,it even presents better performance under extended operating conditions(EOC).Furthermore,the application of Grad-CAM elucidates the CNN’s differentiation process among various sparse SAR targets.The experiments indicate that the model focuses on the target and the background can differ among target classes.The study contributes to an enhanced understanding of the interpretability of such results and enables us to infer the classification outcomes for each category more accurately.
基金supported by the project of the China Geological Survey(DD20230591).
文摘To enhance the prediction accuracy of landslides in in Longyan City,China,this study developed a methodology for geologic hazard susceptibility assessment based on a coupled model composed of a Geographic Information System(GIS)with integrated spatial data,a frequency ratio(FR)model,and a random forest(RF)model(also referred to as the coupled FR-RF model).The coupled FR-RF model was constructed based on the analysis of nine influential factors,including distance from roads,normalized difference vegetation index(NDVI),and slope.The performance of the coupled FR-RF model was assessed using metrics such as Receiver Operating Characteristic(ROC)and Precision-Recall(PR)curves,yielding Area Under the Curve(AUC)values of 0.93 and 0.95,which indicate high predictive accuracy and reliability for geological hazard forecasting.Based on the model predictions,five susceptibility levels were determined in the study area,providing crucial spatial information for geologic hazard prevention and control.The contributions of various influential factors to landslide susceptibility were determined using SHapley Additive exPlanations(SHAP)analysis and the Gini index,enhancing the model interpretability and transparency.Additionally,this study discussed the limitations of the coupled FR-RF model and the prospects for its improvement using new technologies.This study provides an innovative method and theoretical support for geologic hazard prediction and management,holding promising prospects for application.