Since the publication of Sons and Lovers,it has inspired a wide range of critical interpretation, which testifies to its enduring status as a masterpiece of twentieth-century literature. Most critics analyze and evalu...Since the publication of Sons and Lovers,it has inspired a wide range of critical interpretation, which testifies to its enduring status as a masterpiece of twentieth-century literature. Most critics analyze and evaluate Sons and Lovers by adopting a psychoanalytical or social approach. The discussion either just searches for Oedipus Complex or is confined to the content analysis. This essay attempts to integrate all the theoretical analysis which attach to the novel, Sons and Lovers.展开更多
Improving the accuracy of digital elevation is essential for reducing hydro-topographic derivation errors pertaining to, e.g., flow direction, basin borders, channel networks, depressions, flood forecasting, and soil ...Improving the accuracy of digital elevation is essential for reducing hydro-topographic derivation errors pertaining to, e.g., flow direction, basin borders, channel networks, depressions, flood forecasting, and soil drainage. This article demonstrates how a gain in this accuracy is improved through digital elevation model (DEM) fusion, and using LiDAR-derived elevation layers for conformance testing and validation. This demonstration is done for the Province of New Brunswick (NB, Canada), using five province-wide DEM sources (SRTM 90 m;SRTM 30 m;ASTER 30 m;CDED 22 m;NB-DEM 10 m) and a five-stage process that guides the re-projection of these DEMs while minimizing their elevational differences relative to LiDAR-captured bare-earth DEMs, through calibration and validation. This effort decreased the resulting non-LiDAR to LiDAR elevation differences by a factor of two, reduced the minimum distance conformance between the non-LiDAR and LiDAR-derived flow channels to ± 10 m at 8.5 times out of 10, and dropped the non-LiDAR wet-area percentages of false positives from 59% to 49%, and of false negatives from 14% to 7%. While these reductions are modest, they are nevertheless not only consistent with already existing hydrographic data layers informing about stream and wet-area locations, they also extend these data layers across the province by comprehensively locating previously unmapped flow channels and wet areas.展开更多
Objective To investigate distinctive features in drug-resistant mutations (DRMs) and interpretations for reverse transcriptase inhibitors (RTIs) between proviral DNA and paired viral RNA in HIV-l-infected patients...Objective To investigate distinctive features in drug-resistant mutations (DRMs) and interpretations for reverse transcriptase inhibitors (RTIs) between proviral DNA and paired viral RNA in HIV-l-infected patients. Methods Forty-three HIV-l-infected individuals receiving first-line antiretroviral therapy were recruited to participate in a multicenter AIDS Cohort Study in Anhui and Henan Provinces in China in 2004. Drug resistance genotyping was performed by bulk sequencing and deep sequencing on the plasma and whole blood of 77 samples, respectively. Drug-resistance interpretation was compared between viral RNA and paired proviral DNA. Results Compared with bulk sequencing, deep sequencing could detect more DRMs and samples with DRMs in both viral RNA and proviral DNA. The mutations M1841 and M2301 were more prevalent in proviral DNA than in viral RNA (Fisher's exact test, P〈0.05). Considering 'majority resistant variants', 15 samples (19.48%) showed differences in drug resistance interpretation between viral RNA and proviral DNA, and 5 of these samples with different DRMs between proviral DNA and paired viral RNA showed a higher level of drug resistance to the first-line drugs. Considering 'minority resistant variants', 22 samples (28.57%) were associated with a higher level of drug resistance to the tested RTIs for proviral DNA when compared with paired viral RNA. Conclusion Compared with viral RNA, the distinctive information of DRMs and drug resistance interpretations for proviral DNA could be obtained by deep sequencing, which could provide more detailed and precise information for drug resistance monitoring and the rational design of optimal antiretroviral therapy regimens.展开更多
The Appellate Body report in January 2012 had supported the decision of Panel in the"China-measures related to the exportation of various raw materials"case(WT/DS394,395,398)and affirmed that China's res...The Appellate Body report in January 2012 had supported the decision of Panel in the"China-measures related to the exportation of various raw materials"case(WT/DS394,395,398)and affirmed that China's restrictions(such as tariffs and quota measures)on the exportation of raw materials violated rules put forth by the WTO,which were required to be modified.In this case China's right to invoke Article 20 of GATT1994("general exception")to justify its exemption from the guidelines in Article 11.3 of the WTO Accession Protocol was denied by the Panel and the Appellate Body.This was due to the fact that the phrasing in Article 11.3 of Protocol failed to mention"GATT."This was the consequence of the two interpretation approaches the Dispute Settlement Body(DSB)adopted-a narrow textual interpretation and a subjective presumption of"legislative silence."The inappropriate use of the two methods of interpretation lead to an imbalance between the right and obligation of China under the additional obligations that were imposed upon China by the WTO,which create a negative impact on China's rare earth case and the protection of domestic natural resources.展开更多
In this paper, the author focuses on the ecourbarchitectonic physical structures created after year 2000, whose artistic-esthetic value has an iconological character. An entirely new approach in formation of the facad...In this paper, the author focuses on the ecourbarchitectonic physical structures created after year 2000, whose artistic-esthetic value has an iconological character. An entirely new approach in formation of the facade and roof planes as well as of the forms of structures whose appearance resemble sculptural creations has been analyzed. The buildings from all over the world, with different functions contents, indicate a tendency of a different understanding of interpretation of physical structures and correlation with natural and artifact environment. Water surfaces and vegetative material contribute to an effective, cultural, majestic impression of engineering-technological philosophy of city building. The examples in the paper suggest the obvious need of radical changing of the way of thinking in the application of the design strategy in conceptualization of urban agglomerations, and essentially important, conceptually inspired metabolic of relationships among the spatial structures. The world entered new non-globalization trends of creation of the city memory, of the new iconically, symbolically strong, non-cliché, non-standard forms which define the contemporary cultural-artistic and historical identity of macro-ambient entities. This is a good and encouraging sign.展开更多
This paper addresses the problem of the interpretation of the stochastic differential equations (SDE). Even if from a theoretical point of view, there are infinite ways of interpreting them, in practice only Stratonov...This paper addresses the problem of the interpretation of the stochastic differential equations (SDE). Even if from a theoretical point of view, there are infinite ways of interpreting them, in practice only Stratonovich’s and Itô’s interpretations and the kinetic form are important. Restricting the attention to the first two, they give rise to two different Fokker-Planck-Kolmogorov equations for the transition probability density function (PDF) of the solution. According to Stratonovich’s interpretation, there is one more term in the drift, which is not present in the physical equation, the so-called spurious drift. This term is not present in Itô’s interpretation so that the transition PDF’s of the two interpretations are different. Several examples are shown in which the two solutions are strongly different. Thus, caution is needed when a physical phenomenon is modelled by a SDE. However, the meaning of the spurious drift remains unclear.展开更多
Linear and circular interpretation structure maps of different relative depths are obtained by processing 1:200000 aeromagnetic data to the pole in Ailaoshan region,interpreting upward extension of 4 heights,extractin...Linear and circular interpretation structure maps of different relative depths are obtained by processing 1:200000 aeromagnetic data to the pole in Ailaoshan region,interpreting upward extension of 4 heights,extracting a vertical second derivative line of 0 value and a series of calculations. Concealed boundary of deep magnetic rocks can be delineated according to the maps. On the basis of the conclusions above,a set of economical and practical methods to graph the deep structure are summarized. In addition,the relationship between deep structure and mineralization positions is discussed.展开更多
The method and theoretical system of well logging geology have been widely used in the fields of basic geology,petroleum geology and engineering geology,but the different response sensitivity of different well logging...The method and theoretical system of well logging geology have been widely used in the fields of basic geology,petroleum geology and engineering geology,but the different response sensitivity of different well logging series to geological information and the mismatching between geophysical properties of multiple well logs and geological genesis of rocks frequently result in misunderstandings in the research process of well logging geology.Therefore,it is in an urgent need to analyze the typical misunderstanding cases in the research of well logging geology and explore the corresponding scientific ideas and countermeasures.After analyzing the typical misunderstandings in the research of well logging geology,this paper investigates vertical resolution scale of various logging series and its contradiction with detection depth and illustrates the importance of the integration of different scales of data.In addition,the factor inducing“fake logging data”and its influence on interpretation evaluation are clarified and a set of ideas for well logging evaluation of geological interpretation is put forward.And the following research results are obtained.First,the typical misunderstandings in the research of well logging geology can be classified into two categories,namely geological body interpretation misunderstanding and reservoir property parameter calculation misunderstanding.Second,special geological phenomena,such as high-density and high-resistivity mudstone can lead to logging data ambiguity,so attention shall be paid to petrophysical response mechanisms during geological logging interpretation.Third,to carry out well logging evaluation of unconventional oil and gas,it is necessary to integrate new technologies of electric imaging logging,dipole acoustic logging and nuclear magnetic resonance logging,and the calibration of core data and the integration of geological ideas can improve the interpretation accuracy.Fourth,In the process of borehole structural logging analysis,sedimentary response,geostress evaluation and fracture identification,geological ideas shall be integrated to avoid the logging interpretation misunderstanding caused by the same response of different geological phenomena in well logs.In conclusion,the dialectical and systematic thinking from geology to logging and then to geology,from practice to recognition and then to practice and from“a narrow view”to“a broad view”can provide a scientific ideas for the comprehensive research of well logging geology.展开更多
Environmentalmonitoring systems based on remote sensing technology have a wider monitoringrange and longer timeliness, which makes them widely used in the detection andmanagement of pollution sources. However, haze we...Environmentalmonitoring systems based on remote sensing technology have a wider monitoringrange and longer timeliness, which makes them widely used in the detection andmanagement of pollution sources. However, haze weather conditions degrade image qualityand reduce the precision of environmental monitoring systems. To address this problem,this research proposes a remote sensing image dehazingmethod based on the atmosphericscattering model and a dark channel prior constrained network. The method consists ofa dehazing network, a dark channel information injection network (DCIIN), and a transmissionmap network. Within the dehazing network, the branch fusion module optimizesfeature weights to enhance the dehazing effect. By leveraging dark channel information,the DCIIN enables high-quality estimation of the atmospheric veil. To ensure the outputof the deep learning model aligns with physical laws, we reconstruct the haze image usingthe prediction results from the three networks. Subsequently, we apply the traditionalloss function and dark channel loss function between the reconstructed haze image and theoriginal haze image. This approach enhances interpretability and reliabilitywhile maintainingadherence to physical principles. Furthermore, the network is trained on a synthesizednon-homogeneous haze remote sensing dataset using dark channel information from cloudmaps. The experimental results show that the proposed network can achieve better imagedehazing on both synthetic and real remote sensing images with non-homogeneous hazedistribution. This research provides a new idea for solving the problem of decreased accuracyof environmental monitoring systems under haze weather conditions and has strongpracticability.展开更多
BACKGROUND To investigate the preoperative factors influencing textbook outcomes(TO)in Intrahepatic cholangiocarcinoma(ICC)patients and evaluate the feasibility of an interpretable machine learning model for preoperat...BACKGROUND To investigate the preoperative factors influencing textbook outcomes(TO)in Intrahepatic cholangiocarcinoma(ICC)patients and evaluate the feasibility of an interpretable machine learning model for preoperative prediction of TO,we developed a machine learning model for preoperative prediction of TO and used the SHapley Additive exPlanations(SHAP)technique to illustrate the prediction process.AIM To analyze the factors influencing textbook outcomes before surgery and to establish interpretable machine learning models for preoperative prediction.METHODS A total of 376 patients diagnosed with ICC were retrospectively collected from four major medical institutions in China,covering the period from 2011 to 2017.Logistic regression analysis was conducted to identify preoperative variables associated with achieving TO.Based on these variables,an EXtreme Gradient Boosting(XGBoost)machine learning prediction model was constructed using the XGBoost package.The SHAP(package:Shapviz)algorithm was employed to visualize each variable's contribution to the model's predictions.Kaplan-Meier survival analysis was performed to compare the prognostic differences between the TO-achieving and non-TO-achieving groups.RESULTS Among 376 patients,287 were included in the training group and 89 in the validation group.Logistic regression identified the following preoperative variables influencing TO:Child-Pugh classification,Eastern Cooperative Oncology Group(ECOG)score,hepatitis B,and tumor size.The XGBoost prediction model demonstrated high accuracy in internal validation(AUC=0.8825)and external validation(AUC=0.8346).Survival analysis revealed that the disease-free survival rates for patients achieving TO at 1,2,and 3 years were 64.2%,56.8%,and 43.4%,respectively.CONCLUSION Child-Pugh classification,ECOG score,hepatitis B,and tumor size are preoperative predictors of TO.In both the training group and the validation group,the machine learning model had certain effectiveness in predicting TO before surgery.The SHAP algorithm provided intuitive visualization of the machine learning prediction process,enhancing its interpretability.展开更多
The application of machine learning in alloy design is increasingly widespread,yet traditional models still face challenges when dealing with limited datasets and complex nonlinear relationships.This work proposes an ...The application of machine learning in alloy design is increasingly widespread,yet traditional models still face challenges when dealing with limited datasets and complex nonlinear relationships.This work proposes an interpretable machine learning method based on data augmentation and reconstruction,excavating high-performance low-alloyed magnesium(Mg)alloys.The data augmentation technique expands the original dataset through Gaussian noise.The data reconstruction method reorganizes and transforms the original data to extract more representative features,significantly improving the model's generalization ability and prediction accuracy,with a coefficient of determination(R^(2))of 95.9%for the ultimate tensile strength(UTS)model and a R^(2)of 95.3%for the elongation-to-failure(EL)model.The correlation coefficient assisted screening(CCAS)method is proposed to filter low-alloyed target alloys.A new Mg-2.2Mn-0.4Zn-0.2Al-0.2Ca(MZAX2000,wt%)alloy is designed and extruded into bar at given processing parameters,achieving room-temperature strength-ductility synergy showing an excellent UTS of 395 MPa and a high EL of 17.9%.This is closely related to its hetero-structured characteristic in the as-extruded MZAX2000 alloy consisting of coarse grains(16%),fine grains(75%),and fiber regions(9%).Therefore,this work offers new insights into optimizing alloy compositions and processing parameters for attaining new high strong and ductile low-alloyed Mg alloys.展开更多
As batteries become increasingly essential for energy storage technologies,battery prognosis,and diagnosis remain central to ensure reliable operation and effective management,as well as to aid the in-depth investigat...As batteries become increasingly essential for energy storage technologies,battery prognosis,and diagnosis remain central to ensure reliable operation and effective management,as well as to aid the in-depth investigation of degradation mechanisms.However,dynamic operating conditions,cell-to-cell inconsistencies,and limited availability of labeled data have posed significant challenges to accurate and robust prognosis and diagnosis.Herein,we introduce a time-series-decomposition-based ensembled lightweight learning model(TELL-Me),which employs a synergistic dual-module framework to facilitate accurate and reliable forecasting.The feature module formulates features with physical implications and sheds light on battery aging mechanisms,while the gradient module monitors capacity degradation rates and captures aging trend.TELL-Me achieves high accuracy in end-of-life prediction using minimal historical data from a single battery without requiring offline training dataset,and demonstrates impressive generality and robustness across various operating conditions and battery types.Additionally,by correlating feature contributions with degradation mechanisms across different datasets,TELL-Me is endowed with the diagnostic ability that not only enhances prediction reliability but also provides critical insights into the design and optimization of next-generation batteries.展开更多
Topographic maps,as essential tools and sources of information for geographic research,contain precise spatial locations and rich map features,and they illustrate spatio-temporal information on the distribution and di...Topographic maps,as essential tools and sources of information for geographic research,contain precise spatial locations and rich map features,and they illustrate spatio-temporal information on the distribution and differences of various surface features.Currently,topographic maps are mainly stored in raster and vector formats.Extraction of the spatio-temporal knowledge in the maps—such as spatial distribution patterns,feature relationships,and dynamic evolution—still primarily relies on manual interpretation.However,manual interpretation is time-consuming and laborious,especially for large-scale,long-term map knowledge extraction and application.With the development of artificial intelligence technology,it is possible to improve the automation level of map knowledge interpretation.Therefore,the present study proposes an automatic interpretation method for raster topographic map knowledge based on deep learning.To address the limitations of current data-driven intelligent technology in learning map spatial relations and cognitive logic,we establish a formal description of map knowledge by mapping the relationship between map knowledge and features,thereby ensuring interpretation accuracy.Subsequently,deep learning techniques are employed to extract map features automatically,and the spatio-temporal knowledge is constructed by combining formal descriptions of geographic feature knowledge.Validation experiments demonstrate that the proposed method effectively achieves automatic interpretation of spatio-temporal knowledge of geographic features in maps,with an accuracy exceeding 80%.The findings of the present study contribute to machine understanding of spatio-temporal differences in map knowledge and advances the intelligent interpretation and utilization of cartographic information.展开更多
Deep Learning(DL)model has been widely used in the field of Synthetic Aperture Radar Automatic Target Recognition(SAR-ATR)and has achieved excellent performance.However,the black-box nature of DL models has been the f...Deep Learning(DL)model has been widely used in the field of Synthetic Aperture Radar Automatic Target Recognition(SAR-ATR)and has achieved excellent performance.However,the black-box nature of DL models has been the focus of criticism,especially in the application of SARATR,which is closely associated with the national defense and security domain.To address these issues,a new interpretable recognition model Physics-Guided BagNet(PGBN)is proposed in this article.The model adopts an interpretable convolutional neural network framework and uses time–frequency analysis to extract physical scattering features in SAR images.Based on the physical scattering features,an unsupervised segmentation method is proposed to distinguish targets from the background in SAR images.On the basis of the segmentation result,a structure is designed,which constrains the model's spatial attention to focus more on the targets themselves rather than the background,thereby making the model's decision-making more in line with physical principles.In contrast to previous interpretable research methods,this model combines interpretable structure with physical interpretability,further reducing the model's risk of error recognition.Experiments on the MSTAR dataset verify that the PGBN model exhibits excellent interpretability and recognition performance,and comparative experiments with heatmaps indicate that the physical feature guidance module presented in this article can constrain the model to focus more on the target itself rather than the background.展开更多
This paper offers an analysis of the approaches employed in the three interpretations of the Basic Law of the Special Administrative Region of Hong Kong by the Standing Committee of the National People's Congress (...This paper offers an analysis of the approaches employed in the three interpretations of the Basic Law of the Special Administrative Region of Hong Kong by the Standing Committee of the National People's Congress (NPC) after the return of Hong Kong to China, including textualism, structural reading and originalism. The paper stresses the application of jurisprudential theory in the skilful employment of these methods in the NPC interpretations. In the case of "the right of abode" in Hong Kong the differences between the interpretations by the Court of Final Appeal of Hong Kong and by the NPC rest mainly in whether a formalist procedural review or a substantivist presumption of intent should be adopted in the process of determining an authoritative text that embodies the original intention of the legislation. That is not just a difference of legal interpretation but also one of jurisprudential theory and political stance. Based on the above considerations, this paper criticizes the common misconception that it is not appropriate for legislators to undertake legal interpretation, and calls for an understanding of the Basic Law in the framework of Chinese constitutional government.展开更多
Based on 1,003 articles about empirical research on interpreting teaching from 2002 to 2022 retrieved from China National Knowledge Internet,this paper extracts three main research methods,uncovering common problems i...Based on 1,003 articles about empirical research on interpreting teaching from 2002 to 2022 retrieved from China National Knowledge Internet,this paper extracts three main research methods,uncovering common problems in interpreting education and practical teaching suggestions:(1)Corpus-based researches collect numerous audios to study typical mistakes made by interpreting learners,particularly pause and self-repair,and suggest interpreting teaching improve learners’ability to use language chunks and encourage students to interpret smoothly;(2)Questionnaire surveys help understand requirements for professional interpreters and how interpreting teaching meets market demands;(3)Teaching experiments last for one to two semesters,addressing issues like outdated teaching materials and modes,and show how teaching materials and modes integrate modern technology.But empirical researches need to build new corpora,professional interpreters’corpora and address problems that haven’t been adequately discussed.This paper is helpful for improving interpreting education in China and other countries and for making clear tasks to be fulfilled in empirical research on interpreting education.展开更多
The potential toxicity of ionic liquids(ILs)affects their applications;how to control the toxicity is one of the key issues in their applications.To understand its toxicity structure relationship and promote its green...The potential toxicity of ionic liquids(ILs)affects their applications;how to control the toxicity is one of the key issues in their applications.To understand its toxicity structure relationship and promote its greener application,six different machine learning algorithms,including Bagging,Adaptive Boosting(AdaBoost),Gradient Boosting(GBoost),Stacking,Voting and Categorical Boosting(CatBoost),are established to model the toxicity of ILs on four distinct datasets including Leukemia rat cell line IPC-81(IPC-81),Acetylcholinesterase(AChE),Escherichia coli(E.coli)and Vibrio fischeri.Molecular descriptors obtained from the simplified molecular input line entry system(SMILES)are used to characterize ILs.All models are assessed by the mean square error(MSE),root mean square error(RMSE),mean absolute error(MAE)and correlation coefficient(R^(2)).Additionally,an interpretation model based on SHapley Additive exPlanations(SHAP)is built to determine the positive and negative effects of each molecular feature on toxicity.With additional parameters and complexity,the Catboost model outperforms the other models,making it a more reliable model for ILs'toxicity prediction.The results of the model's interpretation indicate that the most significant positive features,SMR_VSA5,PEOE_VSA8,Kappa2,PEOE_VSA6,SMR_VSA5,PEOE_VSA6 and EState_VSA1,can increase the toxicity of ILs as their levels rise,while the most significant negative features,VSA_EState7,EState_VSA8,PEOE_VSA9 and FpDensityMorgan1,can decrease the toxicity as their levels rise.Also,an IL's toxicity will grow as its average molecular weight and number of pyridine rings increase,whereas its toxicity will decrease as its hydrogen bond acceptors increase.This finding offers a theoretical foundation for rapid screening and synthesis of environmentally-benign ILs.展开更多
Developing machine learning frameworks with predictive power,interpretability,and transferability is crucial,yet it faces challenges in the field of electrocatalysis.To achieve this,we employed rigorous feature engine...Developing machine learning frameworks with predictive power,interpretability,and transferability is crucial,yet it faces challenges in the field of electrocatalysis.To achieve this,we employed rigorous feature engineering to establish a finely tuned gradient boosting regressor(GBR)model,which adeptly captures the physical complexity from feature space to target variables.We demonstrated that environmental electron effects and atomic number significantly govern the success of the mapping process via global and local explanations.The finely tuned GBR model exhibits exceptional robustness in predicting CO adsorption energies(R_(ave)^(2)=0.937,RMSE=0.153 eV).Moreover,the model demonstrated remarkable transfer learning ability,showing excellent predictive power for OH,NO,and N_(2) adsorption.Importantly,the GBR model exhibits exceptional predictive capability across an extensive search space,thereby demonstrating profound adaptability and versatility.Our research framework significantly enhances the interpretability and transferability of machine learning in electrocatalysis,offering vital insights for further advancements.展开更多
Low-temperature hydrogenation of silicon tetrachloride(STC)is an essential step in polysilicon production.The addition of CuCl to silicon powder is currently a commonly used catalytic method and the silicon powder act...Low-temperature hydrogenation of silicon tetrachloride(STC)is an essential step in polysilicon production.The addition of CuCl to silicon powder is currently a commonly used catalytic method and the silicon powder acts as both a reactant and a catalyst.However,the reaction mechanism and the structure-activity relationship of this process have not been fully elucidated.In this work,a comprehensive study of the reaction mechanism in the presence of Si and Cu_(3)Si was carried out using density functional theory(DFT)combined with experiments,respectively.The results indicated that the ratedetermining step(RDS)in the presence of Si is the phase transition of Si atom,meanwhile,the RDS in the presence of Cu_(3)Si is the TCS-generation process.The activation barrier of the latter is smaller,highlighting that the interaction of Si with the bulk phase is the pivotal factor influencing the catalytic activity.The feasibility of transition metal doping to facilitate this step was further investigated.The Si disengage energy(E_(d))was used as a quantitative parameter to assess the catalytic activity of the catalysts,and the optimal descriptor was determined through interpretable machine learning.It was demonstrated that d-band center and electron transfer play a crucial role in regulating the level of Ed.This work reveals the mechanism and structure-activity relationship for the low-temperature hydrogenation reaction of STC,and provides a basis for the rational design of catalysts.展开更多
In recent years,with the rapid development of software systems,the continuous expansion of software scale and the increasing complexity of systems have led to the emergence of a growing number of software metrics.Defe...In recent years,with the rapid development of software systems,the continuous expansion of software scale and the increasing complexity of systems have led to the emergence of a growing number of software metrics.Defect prediction methods based on software metric elements highly rely on software metric data.However,redundant software metric data is not conducive to efficient defect prediction,posing severe challenges to current software defect prediction tasks.To address these issues,this paper focuses on the rational clustering of software metric data.Firstly,multiple software projects are evaluated to determine the preset number of clusters for software metrics,and various clustering methods are employed to cluster the metric elements.Subsequently,a co-occurrence matrix is designed to comprehensively quantify the number of times that metrics appear in the same category.Based on the comprehensive results,the software metric data are divided into two semantic views containing different metrics,thereby analyzing the semantic information behind the software metrics.On this basis,this paper also conducts an in-depth analysis of the impact of different semantic view of metrics on defect prediction results,as well as the performance of various classification models under these semantic views.Experiments show that the joint use of the two semantic views can significantly improve the performance of models in software defect prediction,providing a new understanding and approach at the semantic view level for defect prediction research based on software metrics.展开更多
文摘Since the publication of Sons and Lovers,it has inspired a wide range of critical interpretation, which testifies to its enduring status as a masterpiece of twentieth-century literature. Most critics analyze and evaluate Sons and Lovers by adopting a psychoanalytical or social approach. The discussion either just searches for Oedipus Complex or is confined to the content analysis. This essay attempts to integrate all the theoretical analysis which attach to the novel, Sons and Lovers.
文摘Improving the accuracy of digital elevation is essential for reducing hydro-topographic derivation errors pertaining to, e.g., flow direction, basin borders, channel networks, depressions, flood forecasting, and soil drainage. This article demonstrates how a gain in this accuracy is improved through digital elevation model (DEM) fusion, and using LiDAR-derived elevation layers for conformance testing and validation. This demonstration is done for the Province of New Brunswick (NB, Canada), using five province-wide DEM sources (SRTM 90 m;SRTM 30 m;ASTER 30 m;CDED 22 m;NB-DEM 10 m) and a five-stage process that guides the re-projection of these DEMs while minimizing their elevational differences relative to LiDAR-captured bare-earth DEMs, through calibration and validation. This effort decreased the resulting non-LiDAR to LiDAR elevation differences by a factor of two, reduced the minimum distance conformance between the non-LiDAR and LiDAR-derived flow channels to ± 10 m at 8.5 times out of 10, and dropped the non-LiDAR wet-area percentages of false positives from 59% to 49%, and of false negatives from 14% to 7%. While these reductions are modest, they are nevertheless not only consistent with already existing hydrographic data layers informing about stream and wet-area locations, they also extend these data layers across the province by comprehensively locating previously unmapped flow channels and wet areas.
基金supported by grants from the State Key Laboratory of Infectious Disease Prevention and Control(2011SKLID102)the National Nature Science Foundation of China(81172733 and 81561128006)the 12th Five-Year National Science and Technology Major Project(2013ZX10001-006)
文摘Objective To investigate distinctive features in drug-resistant mutations (DRMs) and interpretations for reverse transcriptase inhibitors (RTIs) between proviral DNA and paired viral RNA in HIV-l-infected patients. Methods Forty-three HIV-l-infected individuals receiving first-line antiretroviral therapy were recruited to participate in a multicenter AIDS Cohort Study in Anhui and Henan Provinces in China in 2004. Drug resistance genotyping was performed by bulk sequencing and deep sequencing on the plasma and whole blood of 77 samples, respectively. Drug-resistance interpretation was compared between viral RNA and paired proviral DNA. Results Compared with bulk sequencing, deep sequencing could detect more DRMs and samples with DRMs in both viral RNA and proviral DNA. The mutations M1841 and M2301 were more prevalent in proviral DNA than in viral RNA (Fisher's exact test, P〈0.05). Considering 'majority resistant variants', 15 samples (19.48%) showed differences in drug resistance interpretation between viral RNA and proviral DNA, and 5 of these samples with different DRMs between proviral DNA and paired viral RNA showed a higher level of drug resistance to the first-line drugs. Considering 'minority resistant variants', 22 samples (28.57%) were associated with a higher level of drug resistance to the tested RTIs for proviral DNA when compared with paired viral RNA. Conclusion Compared with viral RNA, the distinctive information of DRMs and drug resistance interpretations for proviral DNA could be obtained by deep sequencing, which could provide more detailed and precise information for drug resistance monitoring and the rational design of optimal antiretroviral therapy regimens.
文摘The Appellate Body report in January 2012 had supported the decision of Panel in the"China-measures related to the exportation of various raw materials"case(WT/DS394,395,398)and affirmed that China's restrictions(such as tariffs and quota measures)on the exportation of raw materials violated rules put forth by the WTO,which were required to be modified.In this case China's right to invoke Article 20 of GATT1994("general exception")to justify its exemption from the guidelines in Article 11.3 of the WTO Accession Protocol was denied by the Panel and the Appellate Body.This was due to the fact that the phrasing in Article 11.3 of Protocol failed to mention"GATT."This was the consequence of the two interpretation approaches the Dispute Settlement Body(DSB)adopted-a narrow textual interpretation and a subjective presumption of"legislative silence."The inappropriate use of the two methods of interpretation lead to an imbalance between the right and obligation of China under the additional obligations that were imposed upon China by the WTO,which create a negative impact on China's rare earth case and the protection of domestic natural resources.
文摘In this paper, the author focuses on the ecourbarchitectonic physical structures created after year 2000, whose artistic-esthetic value has an iconological character. An entirely new approach in formation of the facade and roof planes as well as of the forms of structures whose appearance resemble sculptural creations has been analyzed. The buildings from all over the world, with different functions contents, indicate a tendency of a different understanding of interpretation of physical structures and correlation with natural and artifact environment. Water surfaces and vegetative material contribute to an effective, cultural, majestic impression of engineering-technological philosophy of city building. The examples in the paper suggest the obvious need of radical changing of the way of thinking in the application of the design strategy in conceptualization of urban agglomerations, and essentially important, conceptually inspired metabolic of relationships among the spatial structures. The world entered new non-globalization trends of creation of the city memory, of the new iconically, symbolically strong, non-cliché, non-standard forms which define the contemporary cultural-artistic and historical identity of macro-ambient entities. This is a good and encouraging sign.
文摘This paper addresses the problem of the interpretation of the stochastic differential equations (SDE). Even if from a theoretical point of view, there are infinite ways of interpreting them, in practice only Stratonovich’s and Itô’s interpretations and the kinetic form are important. Restricting the attention to the first two, they give rise to two different Fokker-Planck-Kolmogorov equations for the transition probability density function (PDF) of the solution. According to Stratonovich’s interpretation, there is one more term in the drift, which is not present in the physical equation, the so-called spurious drift. This term is not present in Itô’s interpretation so that the transition PDF’s of the two interpretations are different. Several examples are shown in which the two solutions are strongly different. Thus, caution is needed when a physical phenomenon is modelled by a SDE. However, the meaning of the spurious drift remains unclear.
基金Project supported by National Key Technology R &D Program (No.2006BAB01B10)
文摘Linear and circular interpretation structure maps of different relative depths are obtained by processing 1:200000 aeromagnetic data to the pole in Ailaoshan region,interpreting upward extension of 4 heights,extracting a vertical second derivative line of 0 value and a series of calculations. Concealed boundary of deep magnetic rocks can be delineated according to the maps. On the basis of the conclusions above,a set of economical and practical methods to graph the deep structure are summarized. In addition,the relationship between deep structure and mineralization positions is discussed.
文摘The method and theoretical system of well logging geology have been widely used in the fields of basic geology,petroleum geology and engineering geology,but the different response sensitivity of different well logging series to geological information and the mismatching between geophysical properties of multiple well logs and geological genesis of rocks frequently result in misunderstandings in the research process of well logging geology.Therefore,it is in an urgent need to analyze the typical misunderstanding cases in the research of well logging geology and explore the corresponding scientific ideas and countermeasures.After analyzing the typical misunderstandings in the research of well logging geology,this paper investigates vertical resolution scale of various logging series and its contradiction with detection depth and illustrates the importance of the integration of different scales of data.In addition,the factor inducing“fake logging data”and its influence on interpretation evaluation are clarified and a set of ideas for well logging evaluation of geological interpretation is put forward.And the following research results are obtained.First,the typical misunderstandings in the research of well logging geology can be classified into two categories,namely geological body interpretation misunderstanding and reservoir property parameter calculation misunderstanding.Second,special geological phenomena,such as high-density and high-resistivity mudstone can lead to logging data ambiguity,so attention shall be paid to petrophysical response mechanisms during geological logging interpretation.Third,to carry out well logging evaluation of unconventional oil and gas,it is necessary to integrate new technologies of electric imaging logging,dipole acoustic logging and nuclear magnetic resonance logging,and the calibration of core data and the integration of geological ideas can improve the interpretation accuracy.Fourth,In the process of borehole structural logging analysis,sedimentary response,geostress evaluation and fracture identification,geological ideas shall be integrated to avoid the logging interpretation misunderstanding caused by the same response of different geological phenomena in well logs.In conclusion,the dialectical and systematic thinking from geology to logging and then to geology,from practice to recognition and then to practice and from“a narrow view”to“a broad view”can provide a scientific ideas for the comprehensive research of well logging geology.
基金supported by the National Natural Science Foundation of China(No.51605054).
文摘Environmentalmonitoring systems based on remote sensing technology have a wider monitoringrange and longer timeliness, which makes them widely used in the detection andmanagement of pollution sources. However, haze weather conditions degrade image qualityand reduce the precision of environmental monitoring systems. To address this problem,this research proposes a remote sensing image dehazingmethod based on the atmosphericscattering model and a dark channel prior constrained network. The method consists ofa dehazing network, a dark channel information injection network (DCIIN), and a transmissionmap network. Within the dehazing network, the branch fusion module optimizesfeature weights to enhance the dehazing effect. By leveraging dark channel information,the DCIIN enables high-quality estimation of the atmospheric veil. To ensure the outputof the deep learning model aligns with physical laws, we reconstruct the haze image usingthe prediction results from the three networks. Subsequently, we apply the traditionalloss function and dark channel loss function between the reconstructed haze image and theoriginal haze image. This approach enhances interpretability and reliabilitywhile maintainingadherence to physical principles. Furthermore, the network is trained on a synthesizednon-homogeneous haze remote sensing dataset using dark channel information from cloudmaps. The experimental results show that the proposed network can achieve better imagedehazing on both synthetic and real remote sensing images with non-homogeneous hazedistribution. This research provides a new idea for solving the problem of decreased accuracyof environmental monitoring systems under haze weather conditions and has strongpracticability.
基金Supported by National Key Research and Development Program,No.2022YFC2407304Major Research Project for Middle-Aged and Young Scientists of Fujian Provincial Health Commission,No.2021ZQNZD013+2 种基金The National Natural Science Foundation of China,No.62275050Fujian Province Science and Technology Innovation Joint Fund Project,No.2019Y9108Major Science and Technology Projects of Fujian Province,No.2021YZ036017.
文摘BACKGROUND To investigate the preoperative factors influencing textbook outcomes(TO)in Intrahepatic cholangiocarcinoma(ICC)patients and evaluate the feasibility of an interpretable machine learning model for preoperative prediction of TO,we developed a machine learning model for preoperative prediction of TO and used the SHapley Additive exPlanations(SHAP)technique to illustrate the prediction process.AIM To analyze the factors influencing textbook outcomes before surgery and to establish interpretable machine learning models for preoperative prediction.METHODS A total of 376 patients diagnosed with ICC were retrospectively collected from four major medical institutions in China,covering the period from 2011 to 2017.Logistic regression analysis was conducted to identify preoperative variables associated with achieving TO.Based on these variables,an EXtreme Gradient Boosting(XGBoost)machine learning prediction model was constructed using the XGBoost package.The SHAP(package:Shapviz)algorithm was employed to visualize each variable's contribution to the model's predictions.Kaplan-Meier survival analysis was performed to compare the prognostic differences between the TO-achieving and non-TO-achieving groups.RESULTS Among 376 patients,287 were included in the training group and 89 in the validation group.Logistic regression identified the following preoperative variables influencing TO:Child-Pugh classification,Eastern Cooperative Oncology Group(ECOG)score,hepatitis B,and tumor size.The XGBoost prediction model demonstrated high accuracy in internal validation(AUC=0.8825)and external validation(AUC=0.8346).Survival analysis revealed that the disease-free survival rates for patients achieving TO at 1,2,and 3 years were 64.2%,56.8%,and 43.4%,respectively.CONCLUSION Child-Pugh classification,ECOG score,hepatitis B,and tumor size are preoperative predictors of TO.In both the training group and the validation group,the machine learning model had certain effectiveness in predicting TO before surgery.The SHAP algorithm provided intuitive visualization of the machine learning prediction process,enhancing its interpretability.
基金funded by the National Natural Science Foundation of China(No.52204407)the Natural Science Foundation of Jiangsu Province(No.BK20220595)+1 种基金the China Postdoctoral Science Foundation(No.2022M723689)the Industrial Collaborative Innovation Project of Shanghai(No.XTCX-KJ-2022-2-11)。
文摘The application of machine learning in alloy design is increasingly widespread,yet traditional models still face challenges when dealing with limited datasets and complex nonlinear relationships.This work proposes an interpretable machine learning method based on data augmentation and reconstruction,excavating high-performance low-alloyed magnesium(Mg)alloys.The data augmentation technique expands the original dataset through Gaussian noise.The data reconstruction method reorganizes and transforms the original data to extract more representative features,significantly improving the model's generalization ability and prediction accuracy,with a coefficient of determination(R^(2))of 95.9%for the ultimate tensile strength(UTS)model and a R^(2)of 95.3%for the elongation-to-failure(EL)model.The correlation coefficient assisted screening(CCAS)method is proposed to filter low-alloyed target alloys.A new Mg-2.2Mn-0.4Zn-0.2Al-0.2Ca(MZAX2000,wt%)alloy is designed and extruded into bar at given processing parameters,achieving room-temperature strength-ductility synergy showing an excellent UTS of 395 MPa and a high EL of 17.9%.This is closely related to its hetero-structured characteristic in the as-extruded MZAX2000 alloy consisting of coarse grains(16%),fine grains(75%),and fiber regions(9%).Therefore,this work offers new insights into optimizing alloy compositions and processing parameters for attaining new high strong and ductile low-alloyed Mg alloys.
基金supported by the National Natural Science Foundation of China(22379021 and 22479021)。
文摘As batteries become increasingly essential for energy storage technologies,battery prognosis,and diagnosis remain central to ensure reliable operation and effective management,as well as to aid the in-depth investigation of degradation mechanisms.However,dynamic operating conditions,cell-to-cell inconsistencies,and limited availability of labeled data have posed significant challenges to accurate and robust prognosis and diagnosis.Herein,we introduce a time-series-decomposition-based ensembled lightweight learning model(TELL-Me),which employs a synergistic dual-module framework to facilitate accurate and reliable forecasting.The feature module formulates features with physical implications and sheds light on battery aging mechanisms,while the gradient module monitors capacity degradation rates and captures aging trend.TELL-Me achieves high accuracy in end-of-life prediction using minimal historical data from a single battery without requiring offline training dataset,and demonstrates impressive generality and robustness across various operating conditions and battery types.Additionally,by correlating feature contributions with degradation mechanisms across different datasets,TELL-Me is endowed with the diagnostic ability that not only enhances prediction reliability but also provides critical insights into the design and optimization of next-generation batteries.
基金Deep-time Digital Earth(DDE)Big Science Program(No.GJ-C03-SGF-2025-004)National Natural Science Foundation of China(No.42394063)Sichuan Science and Technology Program(No.2025ZNSFSC0325).
文摘Topographic maps,as essential tools and sources of information for geographic research,contain precise spatial locations and rich map features,and they illustrate spatio-temporal information on the distribution and differences of various surface features.Currently,topographic maps are mainly stored in raster and vector formats.Extraction of the spatio-temporal knowledge in the maps—such as spatial distribution patterns,feature relationships,and dynamic evolution—still primarily relies on manual interpretation.However,manual interpretation is time-consuming and laborious,especially for large-scale,long-term map knowledge extraction and application.With the development of artificial intelligence technology,it is possible to improve the automation level of map knowledge interpretation.Therefore,the present study proposes an automatic interpretation method for raster topographic map knowledge based on deep learning.To address the limitations of current data-driven intelligent technology in learning map spatial relations and cognitive logic,we establish a formal description of map knowledge by mapping the relationship between map knowledge and features,thereby ensuring interpretation accuracy.Subsequently,deep learning techniques are employed to extract map features automatically,and the spatio-temporal knowledge is constructed by combining formal descriptions of geographic feature knowledge.Validation experiments demonstrate that the proposed method effectively achieves automatic interpretation of spatio-temporal knowledge of geographic features in maps,with an accuracy exceeding 80%.The findings of the present study contribute to machine understanding of spatio-temporal differences in map knowledge and advances the intelligent interpretation and utilization of cartographic information.
基金co-supported by the National Natural Science Foundation of China(No.62001507)the Youth Talent Lifting Project of the China Association for Science and Technology(No.2021-JCJQ-QT-018)+1 种基金the Program of the Youth Innovation Team of Shaanxi Universitiesthe Natural Science Basic Research Plan in Shaanxi Province of China(No.2023-JC-YB-491)。
文摘Deep Learning(DL)model has been widely used in the field of Synthetic Aperture Radar Automatic Target Recognition(SAR-ATR)and has achieved excellent performance.However,the black-box nature of DL models has been the focus of criticism,especially in the application of SARATR,which is closely associated with the national defense and security domain.To address these issues,a new interpretable recognition model Physics-Guided BagNet(PGBN)is proposed in this article.The model adopts an interpretable convolutional neural network framework and uses time–frequency analysis to extract physical scattering features in SAR images.Based on the physical scattering features,an unsupervised segmentation method is proposed to distinguish targets from the background in SAR images.On the basis of the segmentation result,a structure is designed,which constrains the model's spatial attention to focus more on the targets themselves rather than the background,thereby making the model's decision-making more in line with physical principles.In contrast to previous interpretable research methods,this model combines interpretable structure with physical interpretability,further reducing the model's risk of error recognition.Experiments on the MSTAR dataset verify that the PGBN model exhibits excellent interpretability and recognition performance,and comparative experiments with heatmaps indicate that the physical feature guidance module presented in this article can constrain the model to focus more on the target itself rather than the background.
文摘This paper offers an analysis of the approaches employed in the three interpretations of the Basic Law of the Special Administrative Region of Hong Kong by the Standing Committee of the National People's Congress (NPC) after the return of Hong Kong to China, including textualism, structural reading and originalism. The paper stresses the application of jurisprudential theory in the skilful employment of these methods in the NPC interpretations. In the case of "the right of abode" in Hong Kong the differences between the interpretations by the Court of Final Appeal of Hong Kong and by the NPC rest mainly in whether a formalist procedural review or a substantivist presumption of intent should be adopted in the process of determining an authoritative text that embodies the original intention of the legislation. That is not just a difference of legal interpretation but also one of jurisprudential theory and political stance. Based on the above considerations, this paper criticizes the common misconception that it is not appropriate for legislators to undertake legal interpretation, and calls for an understanding of the Basic Law in the framework of Chinese constitutional government.
基金USST Construction Project of English-taught Courses for International Students in 2024Key Course Construction Project in Universities of Shanghai in 2024USST Teaching Achievement Award(postgraduate)Cultivation Project in 2024。
文摘Based on 1,003 articles about empirical research on interpreting teaching from 2002 to 2022 retrieved from China National Knowledge Internet,this paper extracts three main research methods,uncovering common problems in interpreting education and practical teaching suggestions:(1)Corpus-based researches collect numerous audios to study typical mistakes made by interpreting learners,particularly pause and self-repair,and suggest interpreting teaching improve learners’ability to use language chunks and encourage students to interpret smoothly;(2)Questionnaire surveys help understand requirements for professional interpreters and how interpreting teaching meets market demands;(3)Teaching experiments last for one to two semesters,addressing issues like outdated teaching materials and modes,and show how teaching materials and modes integrate modern technology.But empirical researches need to build new corpora,professional interpreters’corpora and address problems that haven’t been adequately discussed.This paper is helpful for improving interpreting education in China and other countries and for making clear tasks to be fulfilled in empirical research on interpreting education.
基金funded by Research Platforms and Projects for Higher Education Institutions of Department of Education of Guangdong Province in 2024(2024KTSCX256)2023 Guangdong Province Higher Vocational Education Teaching Quality and Teaching Reform Project(2023JG080).
文摘The potential toxicity of ionic liquids(ILs)affects their applications;how to control the toxicity is one of the key issues in their applications.To understand its toxicity structure relationship and promote its greener application,six different machine learning algorithms,including Bagging,Adaptive Boosting(AdaBoost),Gradient Boosting(GBoost),Stacking,Voting and Categorical Boosting(CatBoost),are established to model the toxicity of ILs on four distinct datasets including Leukemia rat cell line IPC-81(IPC-81),Acetylcholinesterase(AChE),Escherichia coli(E.coli)and Vibrio fischeri.Molecular descriptors obtained from the simplified molecular input line entry system(SMILES)are used to characterize ILs.All models are assessed by the mean square error(MSE),root mean square error(RMSE),mean absolute error(MAE)and correlation coefficient(R^(2)).Additionally,an interpretation model based on SHapley Additive exPlanations(SHAP)is built to determine the positive and negative effects of each molecular feature on toxicity.With additional parameters and complexity,the Catboost model outperforms the other models,making it a more reliable model for ILs'toxicity prediction.The results of the model's interpretation indicate that the most significant positive features,SMR_VSA5,PEOE_VSA8,Kappa2,PEOE_VSA6,SMR_VSA5,PEOE_VSA6 and EState_VSA1,can increase the toxicity of ILs as their levels rise,while the most significant negative features,VSA_EState7,EState_VSA8,PEOE_VSA9 and FpDensityMorgan1,can decrease the toxicity as their levels rise.Also,an IL's toxicity will grow as its average molecular weight and number of pyridine rings increase,whereas its toxicity will decrease as its hydrogen bond acceptors increase.This finding offers a theoretical foundation for rapid screening and synthesis of environmentally-benign ILs.
基金supported by the Research Grants Council of Hong Kong(CityU 11305919 and 11308620)and NSFC/RGC Joint Research Scheme N_CityU104/19Hong Kong Research Grant Council Collaborative Research Fund:C1002-21G and C1017-22Gsupported by the Hong Kong Research Grant Council Collaborative Research Fund:C6021-19E.
文摘Developing machine learning frameworks with predictive power,interpretability,and transferability is crucial,yet it faces challenges in the field of electrocatalysis.To achieve this,we employed rigorous feature engineering to establish a finely tuned gradient boosting regressor(GBR)model,which adeptly captures the physical complexity from feature space to target variables.We demonstrated that environmental electron effects and atomic number significantly govern the success of the mapping process via global and local explanations.The finely tuned GBR model exhibits exceptional robustness in predicting CO adsorption energies(R_(ave)^(2)=0.937,RMSE=0.153 eV).Moreover,the model demonstrated remarkable transfer learning ability,showing excellent predictive power for OH,NO,and N_(2) adsorption.Importantly,the GBR model exhibits exceptional predictive capability across an extensive search space,thereby demonstrating profound adaptability and versatility.Our research framework significantly enhances the interpretability and transferability of machine learning in electrocatalysis,offering vital insights for further advancements.
基金supported by Hubei Three Gorges Laboratory Open Innovation Fund Project(SC231002)CFD Simulation to Explore the Mass and Heat Transfer Laws of Thermal Decomposition of Mixed Salt Organic Compounds Project(2021YFC 3201404).
文摘Low-temperature hydrogenation of silicon tetrachloride(STC)is an essential step in polysilicon production.The addition of CuCl to silicon powder is currently a commonly used catalytic method and the silicon powder acts as both a reactant and a catalyst.However,the reaction mechanism and the structure-activity relationship of this process have not been fully elucidated.In this work,a comprehensive study of the reaction mechanism in the presence of Si and Cu_(3)Si was carried out using density functional theory(DFT)combined with experiments,respectively.The results indicated that the ratedetermining step(RDS)in the presence of Si is the phase transition of Si atom,meanwhile,the RDS in the presence of Cu_(3)Si is the TCS-generation process.The activation barrier of the latter is smaller,highlighting that the interaction of Si with the bulk phase is the pivotal factor influencing the catalytic activity.The feasibility of transition metal doping to facilitate this step was further investigated.The Si disengage energy(E_(d))was used as a quantitative parameter to assess the catalytic activity of the catalysts,and the optimal descriptor was determined through interpretable machine learning.It was demonstrated that d-band center and electron transfer play a crucial role in regulating the level of Ed.This work reveals the mechanism and structure-activity relationship for the low-temperature hydrogenation reaction of STC,and provides a basis for the rational design of catalysts.
基金supported by the CCF-NSFOCUS‘Kunpeng’Research Fund(CCF-NSFOCUS2024012).
文摘In recent years,with the rapid development of software systems,the continuous expansion of software scale and the increasing complexity of systems have led to the emergence of a growing number of software metrics.Defect prediction methods based on software metric elements highly rely on software metric data.However,redundant software metric data is not conducive to efficient defect prediction,posing severe challenges to current software defect prediction tasks.To address these issues,this paper focuses on the rational clustering of software metric data.Firstly,multiple software projects are evaluated to determine the preset number of clusters for software metrics,and various clustering methods are employed to cluster the metric elements.Subsequently,a co-occurrence matrix is designed to comprehensively quantify the number of times that metrics appear in the same category.Based on the comprehensive results,the software metric data are divided into two semantic views containing different metrics,thereby analyzing the semantic information behind the software metrics.On this basis,this paper also conducts an in-depth analysis of the impact of different semantic view of metrics on defect prediction results,as well as the performance of various classification models under these semantic views.Experiments show that the joint use of the two semantic views can significantly improve the performance of models in software defect prediction,providing a new understanding and approach at the semantic view level for defect prediction research based on software metrics.