Pine wilt disease(PWD)is currently one of the main causes of large-scale forest destruction.To control the spread of PWD,it is essential to detect affected pine trees quickly.This study investigated the feasibility of...Pine wilt disease(PWD)is currently one of the main causes of large-scale forest destruction.To control the spread of PWD,it is essential to detect affected pine trees quickly.This study investigated the feasibility of using the object-oriented multi-scale segmentation algorithm to identify trees discolored by PWD.We used an unmanned aerial vehicle(UAV)platform equipped with an RGB digital camera to obtain high spatial resolution images,and multiscale segmentation was applied to delineate the tree crown,coupling the use of object-oriented classification to classify trees discolored by PWD.Then,the optimal segmentation scale was implemented using the estimation of scale parameter(ESP2)plug-in.The feature space of the segmentation results was optimized,and appropriate features were selected for classification.The results showed that the optimal scale,shape,and compactness values of the tree crown segmentation algorithm were 56,0.5,and 0.8,respectively.The producer’s accuracy(PA),user’s accuracy(UA),and F1 score were 0.722,0.605,and 0.658,respectively.There were no significant classification errors in the final classification results,and the low accuracy was attributed to the low number of objects count caused by incorrect segmentation.The multi-scale segmentation and object-oriented classification method could accurately identify trees discolored by PWD with a straightforward and rapid processing.This study provides a technical method for monitoring the occurrence of PWD and identifying the discolored trees of disease using UAV-based high-resolution images.展开更多
UAV remote sensing images have the advantages of high spatial resolution,fast speed,strong real-time performance,and convenient operation,etc.,and have become a recently developed,vital means of acquiring surface info...UAV remote sensing images have the advantages of high spatial resolution,fast speed,strong real-time performance,and convenient operation,etc.,and have become a recently developed,vital means of acquiring surface information.It is an important research task for precision agriculture to make full use of the spectrum,texture,color and other characteristic information of crops,especially the spatial arrangement and structure information of features,to explore effective methods for the classification of multiple varieties of crops.In order to explore the applicability of the object-oriented method to achieve accurate classification of UAV high-resolution images,the paper used the object-oriented classification method in ENVI to classify the UAV high-resolution remote sensing image obtained from the orderly structured 28 species of crops in the test field,which mainly includes image segmentation and object classification.The results showed that the plots obtained after classification were continuous and complete,basically in line with the actual situation,and the overall accuracy of crop classification was 91.73%,with Kappa coefficient of 0.87.Compared with the crop planting area based on remote sensing interpretation and field survey,the area error of 17 species of crops in this study was controlled within 15%,which provides a basis for object-oriented crop classification of UAV remote sensing images.展开更多
In recent decades,the proliferation of email communication has markedly escalated,resulting in a concomitant surge in spam emails that congest networks and presenting security risks.This study introduces an innovative...In recent decades,the proliferation of email communication has markedly escalated,resulting in a concomitant surge in spam emails that congest networks and presenting security risks.This study introduces an innovative spam detection method utilizing the Horse Herd Optimization Algorithm(HHOA),designed for binary classification within multi⁃objective framework.The method proficiently identifies essential features,minimizing redundancy and improving classification precision.The suggested HHOA attained an impressive accuracy of 97.21%on the Kaggle email dataset,with precision of 94.30%,recall of 90.50%,and F1⁃score of 92.80%.Compared to conventional techniques,such as Support Vector Machine(93.89%accuracy),Random Forest(96.14%accuracy),and K⁃Nearest Neighbours(92.08%accuracy),HHOA exhibited enhanced performance with reduced computing complexity.The suggested method demonstrated enhanced feature selection efficiency,decreasing the number of selected features while maintaining high classification accuracy.The results underscore the efficacy of HHOA in spam identification and indicate its potential for further applications in practical email filtering systems.展开更多
Near-Earth objects are important not only in studying the early formation of the Solar System,but also because they pose a serious hazard to humanity when they make close approaches to the Earth.Study of their physica...Near-Earth objects are important not only in studying the early formation of the Solar System,but also because they pose a serious hazard to humanity when they make close approaches to the Earth.Study of their physical properties can provide useful information on their origin,evolution,and hazard to human beings.However,it remains challenging to investigate small,newly discovered,near-Earth objects because of our limited observational window.This investigation seeks to determine the visible colors of near-Earth asteroids(NEAs),perform an initial taxonomic classification based on visible colors and analyze possible correlations between the distribution of taxonomic classification and asteroid size or orbital parameters.Observations were performed in the broadband BVRI Johnson−Cousins photometric system,applied to images from the Yaoan High Precision Telescope and the 1.88 m telescope at the Kottamia Astronomical Observatory.We present new photometric observations of 84 near-Earth asteroids,and classify 80 of them taxonomically,based on their photometric colors.We find that nearly half(46.3%)of the objects in our sample can be classified as S-complex,26.3%as C-complex,6%as D-complex,and 15.0%as X-complex;the remaining belong to the A-or V-types.Additionally,we identify three P-type NEAs in our sample,according to the Tholen scheme.The fractional abundances of the C/X-complex members with absolute magnitude H≥17.0 were more than twice as large as those with H<17.0.However,the fractions of C-and S-complex members with diameters≤1 km and>1 km are nearly equal,while X-complex members tend to have sub-kilometer diameters.In our sample,the C/D-complex objects are predominant among those with a Jovian Tisserand parameter of T_(J)<3.1.These bodies could have a cometary origin.C-and S-complex members account for a considerable proportion of the asteroids that are potentially hazardous.展开更多
Graph Neural Networks(GNNs)have proven highly effective for graph classification across diverse fields such as social networks,bioinformatics,and finance,due to their capability to learn complex graph structures.Howev...Graph Neural Networks(GNNs)have proven highly effective for graph classification across diverse fields such as social networks,bioinformatics,and finance,due to their capability to learn complex graph structures.However,despite their success,GNNs remain vulnerable to adversarial attacks that can significantly degrade their classification accuracy.Existing adversarial attack strategies primarily rely on label information to guide the attacks,which limits their applicability in scenarios where such information is scarce or unavailable.This paper introduces an innovative unsupervised attack method for graph classification,which operates without relying on label information,thereby enhancing its applicability in a broad range of scenarios.Specifically,our method first leverages a graph contrastive learning loss to learn high-quality graph embeddings by comparing different stochastic augmented views of the graphs.To effectively perturb the graphs,we then introduce an implicit estimator that measures the impact of various modifications on graph structures.The proposed strategy identifies and flips edges with the top-K highest scores,determined by the estimator,to maximize the degradation of the model’s performance.In addition,to defend against such attack,we propose a lightweight regularization-based defense mechanism that is specifically tailored to mitigate the structural perturbations introduced by our attack strategy.It enhances model robustness by enforcing embedding consistency and edge-level smoothness during training.We conduct experiments on six public TU graph classification datasets:NCI1,NCI109,Mutagenicity,ENZYMES,COLLAB,and DBLP_v1,to evaluate the effectiveness of our attack and defense strategies.Under an attack budget of 3,the maximum reduction in model accuracy reaches 6.67%on the Graph Convolutional Network(GCN)and 11.67%on the Graph Attention Network(GAT)across different datasets,indicating that our unsupervised method induces degradation comparable to state-of-the-art supervised attacks.Meanwhile,our defense achieves the highest accuracy recovery of 3.89%(GCN)and 5.00%(GAT),demonstrating improved robustness against structural perturbations.展开更多
Skin diseases affect millions worldwide.Early detection is key to preventing disfigurement,lifelong disability,or death.Dermoscopic images acquired in primary-care settings show high intra-class visual similarity and ...Skin diseases affect millions worldwide.Early detection is key to preventing disfigurement,lifelong disability,or death.Dermoscopic images acquired in primary-care settings show high intra-class visual similarity and severe class imbalance,and occasional imaging artifacts can create ambiguity for state-of-the-art convolutional neural networks(CNNs).We frame skin lesion recognition as graph-based reasoning and,to ensure fair evaluation and avoid data leakage,adopt a strict lesion-level partitioning strategy.Each image is first over-segmented using SLIC(Simple Linear Iterative Clustering)to produce perceptually homogeneous superpixels.These superpixels form the nodes of a region-adjacency graph whose edges encode spatial continuity.Node attributes are 1280-dimensional embeddings extracted with a lightweight yet expressive EfficientNet-B0 backbone,providing strong representational power at modest computational cost.The resulting graphs are processed by a five-layer Graph Attention Network(GAT)that learns to weight inter-node relationships dynamically and aggregates multi-hop context before classifying lesions into seven classes with a log-softmax output.Extensive experiments on the DermaMNIST benchmark show the proposed pipeline achieves 88.35%accuracy and 98.04%AUC,outperforming contemporary CNNs,AutoML approaches,and alternative graph neural networks.An ablation study indicates EfficientNet-B0 produces superior node descriptors compared with ResNet-18 and DenseNet,and that roughly five GAT layers strike a good balance between being too shallow and over-deep while avoiding oversmoothing.The method requires no data augmentation or external metadata,making it a drop-in upgrade for clinical computer-aided diagnosis systems.展开更多
This systematic review aims to comprehensively examine and compare deep learning methods for brain tumor segmentation and classification using MRI and other imaging modalities,focusing on recent trends from 2022 to 20...This systematic review aims to comprehensively examine and compare deep learning methods for brain tumor segmentation and classification using MRI and other imaging modalities,focusing on recent trends from 2022 to 2025.The primary objective is to evaluate methodological advancements,model performance,dataset usage,and existing challenges in developing clinically robust AI systems.We included peer-reviewed journal articles and highimpact conference papers published between 2022 and 2025,written in English,that proposed or evaluated deep learning methods for brain tumor segmentation and/or classification.Excluded were non-open-access publications,books,and non-English articles.A structured search was conducted across Scopus,Google Scholar,Wiley,and Taylor&Francis,with the last search performed in August 2025.Risk of bias was not formally quantified but considered during full-text screening based on dataset diversity,validation methods,and availability of performance metrics.We used narrative synthesis and tabular benchmarking to compare performance metrics(e.g.,accuracy,Dice score)across model types(CNN,Transformer,Hybrid),imaging modalities,and datasets.A total of 49 studies were included(43 journal articles and 6 conference papers).These studies spanned over 9 public datasets(e.g.,BraTS,Figshare,REMBRANDT,MOLAB)and utilized a range of imaging modalities,predominantly MRI.Hybrid models,especially ResViT and UNetFormer,consistently achieved high performance,with classification accuracy exceeding 98%and segmentation Dice scores above 0.90 across multiple studies.Transformers and hybrid architectures showed increasing adoption post2023.Many studies lacked external validation and were evaluated only on a few benchmark datasets,raising concerns about generalizability and dataset bias.Few studies addressed clinical interpretability or uncertainty quantification.Despite promising results,particularly for hybrid deep learning models,widespread clinical adoption remains limited due to lack of validation,interpretability concerns,and real-world deployment barriers.展开更多
Honeycombing Lung(HCL)is a chronic lung condition marked by advanced fibrosis,resulting in enlarged air spaces with thick fibrotic walls,which are visible on Computed Tomography(CT)scans.Differentiating between normal...Honeycombing Lung(HCL)is a chronic lung condition marked by advanced fibrosis,resulting in enlarged air spaces with thick fibrotic walls,which are visible on Computed Tomography(CT)scans.Differentiating between normal lung tissue,honeycombing lungs,and Ground Glass Opacity(GGO)in CT images is often challenging for radiologists and may lead to misinterpretations.Although earlier studies have proposed models to detect and classify HCL,many faced limitations such as high computational demands,lower accuracy,and difficulty distinguishing between HCL and GGO.CT images are highly effective for lung classification due to their high resolution,3D visualization,and sensitivity to tissue density variations.This study introduces Honeycombing Lungs Network(HCL Net),a novel classification algorithm inspired by ResNet50V2 and enhanced to overcome the shortcomings of previous approaches.HCL Net incorporates additional residual blocks,refined preprocessing techniques,and selective parameter tuning to improve classification performance.The dataset,sourced from the University Malaya Medical Centre(UMMC)and verified by expert radiologists,consists of CT images of normal,honeycombing,and GGO lungs.Experimental evaluations across five assessments demonstrated that HCL Net achieved an outstanding classification accuracy of approximately 99.97%.It also recorded strong performance in other metrics,achieving 93%precision,100%sensitivity,89%specificity,and an AUC-ROC score of 97%.Comparative analysis with baseline feature engineering methods confirmed the superior efficacy of HCL Net.The model significantly reduces misclassification,particularly between honeycombing and GGO lungs,enhancing diagnostic precision and reliability in lung image analysis.展开更多
Sudden wildfires cause significant global ecological damage.While satellite imagery has advanced early fire detection and mitigation,image-based systems face limitations including high false alarm rates,visual obstruc...Sudden wildfires cause significant global ecological damage.While satellite imagery has advanced early fire detection and mitigation,image-based systems face limitations including high false alarm rates,visual obstructions,and substantial computational demands,especially in complex forest terrains.To address these challenges,this study proposes a novel forest fire detection model utilizing audio classification and machine learning.We developed an audio-based pipeline using real-world environmental sound recordings.Sounds were converted into Mel-spectrograms and classified via a Convolutional Neural Network(CNN),enabling the capture of distinctive fire acoustic signatures(e.g.,crackling,roaring)that are minimally impacted by visual or weather conditions.Internet of Things(IoT)sound sensors were crucial for generating complex environmental parameters to optimize feature extraction.The CNN model achieved high performance in stratified 5-fold cross-validation(92.4%±1.6 accuracy,91.2%±1.8 F1-score)and on test data(94.93%accuracy,93.04%F1-score),with 98.44%precision and 88.32%recall,demonstrating reliability across environmental conditions.These results indicate that the audio-based approach not only improves detection reliability but also markedly reduces computational overhead compared to traditional image-based methods.The findings suggest that acoustic sensing integrated with machine learning offers a powerful,low-cost,and efficient solution for real-time forest fire monitoring in complex,dynamic environments.展开更多
Accurate extraction of surface water extent is a fundamental prerequisite for monitoring its dynamic changes.Although machine learning algorithms have been widely applied to surface water mapping,most studies focus pr...Accurate extraction of surface water extent is a fundamental prerequisite for monitoring its dynamic changes.Although machine learning algorithms have been widely applied to surface water mapping,most studies focus primarily on algorithmic outputs,with limited systematic evaluation of their applicability and constrained classification accuracy.In this study,we focused on the Songnen Plain in Northeast China and employed Sentinel-2 imagery acquired during 2020-2021 via the Google Earth Engine(GEE)platform to evaluate the performance of Classification and Regression Trees(CART),Random Forest(RF),and Support Vector Machine(SVM)for surface water classification.The classification process was optimized by incorporating automated training sample selection and integration of time series features.Validation with independent samples demonstrated the feasibility of automatic sample selection,yielding mean overall accuracies of 91.16%,90.99%,and 90.76%for RF,SVM,and CART,respectively.After integrating time series features,the mean overall accuracies of the three algorithms improved by 4.51%,5.45%,and 6.36%,respectively.In addition,spectral features such as MNDWI(Modified Normalized Difference Water Index),SWIR(Short Wave Infrared),and NDVI(Normalized Difference Vegetation Index)were identified as more important for surface water classification.This study establishes a more consistent framework for surface water mapping,offering new perspectives for improving and automating classification processes in the era of big and open data.展开更多
Alzheimer’s Disease(AD)is a progressive neurodegenerative disorder that significantly affects cognitive function,making early and accurate diagnosis essential.Traditional Deep Learning(DL)-based approaches often stru...Alzheimer’s Disease(AD)is a progressive neurodegenerative disorder that significantly affects cognitive function,making early and accurate diagnosis essential.Traditional Deep Learning(DL)-based approaches often struggle with low-contrast MRI images,class imbalance,and suboptimal feature extraction.This paper develops a Hybrid DL system that unites MobileNetV2 with adaptive classification methods to boost Alzheimer’s diagnosis by processing MRI scans.Image enhancement is done using Contrast-Limited Adaptive Histogram Equalization(CLAHE)and Enhanced Super-Resolution Generative Adversarial Networks(ESRGAN).A classification robustness enhancement system integrates class weighting techniques and a Matthews Correlation Coefficient(MCC)-based evaluation method into the design.The trained and validated model gives a 98.88%accuracy rate and 0.9614 MCC score.We also performed a 10-fold cross-validation experiment with an average accuracy of 96.52%(±1.51),a loss of 0.1671,and an MCC score of 0.9429 across folds.The proposed framework outperforms the state-of-the-art models with a 98%weighted F1-score while decreasing misdiagnosis results for every AD stage.The model demonstrates apparent separation abilities between AD progression stages according to the results of the confusion matrix analysis.These results validate the effectiveness of hybrid DL models with adaptive preprocessing for early and reliable Alzheimer’s diagnosis,contributing to improved computer-aided diagnosis(CAD)systems in clinical practice.展开更多
With the wide use of high-resolution remotely sensed imagery, the object-oriented remotely sensed informa- tion classification pattern has been intensively studied. Starting with the definition of object-oriented remo...With the wide use of high-resolution remotely sensed imagery, the object-oriented remotely sensed informa- tion classification pattern has been intensively studied. Starting with the definition of object-oriented remotely sensed information classification pattern and a literature review of related research progress, this paper sums up 4 developing phases of object-oriented classification pattern during the past 20 years. Then, we discuss the three aspects of method- ology in detail, namely remotely sensed imagery segmentation, feature analysis and feature selection, and classification rule generation, through comparing them with remotely sensed information classification method based on per-pixel. At last, this paper presents several points that need to be paid attention to in the future studies on object-oriented RS in- formation classification pattern: 1) developing robust and highly effective image segmentation algorithm for multi-spectral RS imagery; 2) improving the feature-set including edge, spatial-adjacent and temporal characteristics; 3) discussing the classification rule generation classifier based on the decision tree; 4) presenting evaluation methods for classification result by object-oriented classification pattern.展开更多
Benthic habitat mapping is an emerging discipline in the international marine field in recent years,providing an effective tool for marine spatial planning,marine ecological management,and decision-making applications...Benthic habitat mapping is an emerging discipline in the international marine field in recent years,providing an effective tool for marine spatial planning,marine ecological management,and decision-making applications.Seabed sediment classification is one of the main contents of seabed habitat mapping.In response to the impact of remote sensing imaging quality and the limitations of acoustic measurement range,where a single data source does not fully reflect the substrate type,we proposed a high-precision seabed habitat sediment classification method that integrates data from multiple sources.Based on WorldView-2 multi-spectral remote sensing image data and multibeam bathymetry data,constructed a random forests(RF)classifier with optimal feature selection.A seabed sediment classification experiment integrating optical remote sensing and acoustic remote sensing data was carried out in the shallow water area of Wuzhizhou Island,Hainan,South China.Different seabed sediment types,such as sand,seagrass,and coral reefs were effectively identified,with an overall classification accuracy of 92%.Experimental results show that RF matrix optimized by fusing multi-source remote sensing data for feature selection were better than the classification results of simple combinations of data sources,which improved the accuracy of seabed sediment classification.Therefore,the method proposed in this paper can be effectively applied to high-precision seabed sediment classification and habitat mapping around islands and reefs.展开更多
This paper proposed to use double polarization synthetic aperture radar (SAR) image to classify surface feature, based on DEM. It takes fully use of the polarization information and external information. This pa-per u...This paper proposed to use double polarization synthetic aperture radar (SAR) image to classify surface feature, based on DEM. It takes fully use of the polarization information and external information. This pa-per utilizes ENVISAT ASAR APP double-polarization data of Poyang lake area in Jiangxi Province. Com-pared with traditional pixel-based classification, this paper fully uses object features (color, shape, hierarchy) and accessorial DEM information. The classification accuracy improves from the original 73.7% to 91.84%. The result shows that object-oriented classification technology is suitable for double polarization SAR’s high precision classification.展开更多
With the development of remote sensing technology,the spatial resolution,spectral resolution and time resolution of remote sensing data are greatly improved.How to efficiently process and interpret the massive high re...With the development of remote sensing technology,the spatial resolution,spectral resolution and time resolution of remote sensing data are greatly improved.How to efficiently process and interpret the massive high resolution remote sensing image data for ground objects,which are of spatial geometry and texture information,has become the focus and key issue in the field of remote sensing research.A new method of the classification(OCRC(object-oriented and class rule classification))of remote sensing,which is of object-oriented and rule,has been presented in this paper,that is,through the discovery and mining the knowledge of spectrum and spatial characteristics of high-resolution remote sensing image,establish a multi-level network image object segmentation and classification structure of remote sensing image to achieve classification and accuracy assessment accurately and quickly for ground targets.Selected the worldview-2 image data in the Zangnan area as a study object,using the OCRC to verify the experiment which was a combination of the mean variance method,the maximum area method and the accuracy comparison to analysis selected three kinds of optimal segmentation scale and established a multi-level image object network hierarchy for image classification experiments.The results show that the OCRC can enable the high resolution image classification results similar to the visual interpretation of the results,and has higher classification accuracy.The overall accuracy and Kappa coefficient of the object-oriented rule classification method are 97.38%,0.9673;compared with object-oriented SVM method,respectively higher than 6.23%,0.078;compared with object-oriented KNN method,respectively more than 7.96%,0.0996.The extraction precision and user accuracy of the buildings compared with object-oriented SVM method,respectively higher than 18.39%,3.98%,respectively better than the object-oriented KNN(K-Nearest Neighbor)method 21.27%,14.97%.展开更多
As one of the main geographical elements in urban areas,buildings are closely related to the development of the city.Therefore,how to quickly and accurately extract building information from remote sensing images is o...As one of the main geographical elements in urban areas,buildings are closely related to the development of the city.Therefore,how to quickly and accurately extract building information from remote sensing images is of great significance for urban map updating,urban planning and construction,etc.Extracting building information around power facilities,especially obtaining this information from high-resolution images,has become one of the current hot topics in remote sensing technology research.This study made full use of the characteristics of GF-2 satellite remote sensing images,adopted an object-oriented classification method,combined with multi-scale segmentation technology and CART classification algorithm,and successfully extracted the buildings in the study area.The research results showed that the overall classification accuracy reached 89.5%and the Kappa coefficient was 0.86.Using the object-oriented CART classification algorithm for building extraction could be closer to actual ground objects and had higher accuracy.The extraction of buildings in the city contributed to urban development planning and provided decision support for management.展开更多
This paper investigates the contribution of multi-temporal enhanced vegetation index(EVI)data to the improvement of object-based classification accuracy using multi-spectral moderate resolution imaging spectral-radiom...This paper investigates the contribution of multi-temporal enhanced vegetation index(EVI)data to the improvement of object-based classification accuracy using multi-spectral moderate resolution imaging spectral-radiometer(MODIS)imagery.In object-oriented classification,similar pixels are firstly grouped together and then classified;the produced result does not suffer the speckled appearance and closer to human vision.EVI data are from the MODIS sensor aboard Terra spacecraft.69 EVI data(scenes)were collected during the period of three years(20012003)in a mountainous vegetated area.These data sets were used to study the phenology of the land cover types.Different land cover types show distinct fluctuations over time in EVI values and this information might be used to improve object-oriented land cover classification.Two experiments were carried out:one was only with single date MODIS multispectral data,and the other one including also the 69 EVI images.Eight classes were distinguished:temperate forest,tropical dry forest,grassland,irrigated agriculture,rain-fed agriculture,orchards,lava flows and human settlement.The two classifications were evaluated with independent verification data,and the results showed that with multitemporal EVI data,the classification accuracy was improved 5.2%.Evaluated by McNemar’s test,this improved was significant,with significance level p=0.01.展开更多
Urban tree species provide various essential ecosystem services in cities,such as regulating urban temperatures,reducing noise,capturing carbon,and mitigating the urban heat island effect.The quality of these services...Urban tree species provide various essential ecosystem services in cities,such as regulating urban temperatures,reducing noise,capturing carbon,and mitigating the urban heat island effect.The quality of these services is influenced by species diversity,tree health,and the distribution and the composition of trees.Traditionally,data on urban trees has been collected through field surveys and manual interpretation of remote sensing images.In this study,we evaluated the effectiveness of multispectral airborne laser scanning(ALS)data in classifying 24 common urban roadside tree species in Espoo,Finland.Tree crown structure information,intensity features,and spectral data were used for classification.Eight different machine learning algorithms were tested,with the extra trees(ET)algorithm performing the best,achieving an overall accuracy of 71.7%using multispectral LiDAR data.This result highlights that integrating structural and spectral information within a single framework can improve the classification accuracy.Future research will focus on identifying the most important features for species classification and developing algorithms with greater efficiency and accuracy.展开更多
Purpose:Interdisciplinary research has become a critical approach to addressing complex societal,economic,technological,and environmental challenges,driving innovation and integrating scientific knowledge.While interd...Purpose:Interdisciplinary research has become a critical approach to addressing complex societal,economic,technological,and environmental challenges,driving innovation and integrating scientific knowledge.While interdisciplinarity indicators are widely used to evaluate research performance,the impact of classification granularity on these assessments remains underexplored.Design/methodology/approach:This study investigates how different levels of classification granularity-macro,meso,and micro-affect the evaluation of interdisciplinarity in research institutes.Using a dataset of 262 institutes from four major German non-university organizations(FHG,HGF,MPG,WGL)from 2018 to 2022,we examine inconsistencies in interdisciplinarity across levels,analyze ranking changes,and explore the influence of institutional fields and research focus(applied vs.basic).Findings:Our findings reveal significant inconsistencies in interdisciplinarity across classification levels,with rankings varying substantially.Notably,the Fraunhofer Society(FHG),which performs well at the macro level,experiences significant ranking declines at meso and micro levels.Normalizing interdisciplinarity by research field confirmed that these declines persist.The research focus of institutes,whether applied,basic,or mixed,does not significantly explain the observed ranking dynamics.Research limitations:This study has only considered the publication-based dimension of institutional interdisciplinarity and has not explored other aspects.Practical implications:The findings provide insights for policymakers,research managers,and scholars to better interpret interdisciplinarity metrics and support interdisciplinary research effectively.Originality/value:This study underscores the critical role of classification granularity in interdisciplinarity assessment and emphasizes the need for standardized approaches to ensure robust and fair evaluations.展开更多
Breast cancer is one of the major causes of deaths in women.However,the early diagnosis is important for screening and control the mortality rate.Thus for the diagnosis of breast cancer at the early stage,a computer-a...Breast cancer is one of the major causes of deaths in women.However,the early diagnosis is important for screening and control the mortality rate.Thus for the diagnosis of breast cancer at the early stage,a computer-aided diagnosis system is highly required.Ultrasound is an important examination technique for breast cancer diagnosis due to its low cost.Recently,many learning-based techniques have been introduced to classify breast cancer using breast ultrasound imaging dataset(BUSI)datasets;however,the manual handling is not an easy process and time consuming.The authors propose an EfficientNet-integrated ResNet deep network and XAI-based framework for accurately classifying breast cancer(malignant and benign).In the initial step,data augmentation is performed to increase the number of training samples.For this purpose,three-pixel flip mathematical equations are introduced:horizontal,vertical,and 90°.Later,two pretrained deep learning models were employed,skipped some layers,and fine-tuned.Both fine-tuned models are later trained using a deep transfer learning process and extracted features from the deeper layer.Explainable artificial intelligence-based analysed the performance of trained models.After that,a new feature selection technique is proposed based on the cuckoo search algorithm called cuckoo search controlled standard error mean.This technique selects the best features and fuses using a new parallel zeropadding maximum correlated coefficient features.In the end,the selection algorithm is applied again to the fused feature vector and classified using machine learning algorithms.The experimental process of the proposed framework is conducted on a publicly available BUSI and obtained 98.4%and 98%accuracy in two different experiments.Comparing the proposed framework is also conducted with recent techniques and shows improved accuracy.In addition,the proposed framework was executed less than the original deep learning models.展开更多
基金supported by the National Natural Science Foundation of China(No.31870620)the National Technology Extension Fund of Forestry([2019]06)the Fundamental Research Funds for the Central Universities(No.PTYX202107)。
文摘Pine wilt disease(PWD)is currently one of the main causes of large-scale forest destruction.To control the spread of PWD,it is essential to detect affected pine trees quickly.This study investigated the feasibility of using the object-oriented multi-scale segmentation algorithm to identify trees discolored by PWD.We used an unmanned aerial vehicle(UAV)platform equipped with an RGB digital camera to obtain high spatial resolution images,and multiscale segmentation was applied to delineate the tree crown,coupling the use of object-oriented classification to classify trees discolored by PWD.Then,the optimal segmentation scale was implemented using the estimation of scale parameter(ESP2)plug-in.The feature space of the segmentation results was optimized,and appropriate features were selected for classification.The results showed that the optimal scale,shape,and compactness values of the tree crown segmentation algorithm were 56,0.5,and 0.8,respectively.The producer’s accuracy(PA),user’s accuracy(UA),and F1 score were 0.722,0.605,and 0.658,respectively.There were no significant classification errors in the final classification results,and the low accuracy was attributed to the low number of objects count caused by incorrect segmentation.The multi-scale segmentation and object-oriented classification method could accurately identify trees discolored by PWD with a straightforward and rapid processing.This study provides a technical method for monitoring the occurrence of PWD and identifying the discolored trees of disease using UAV-based high-resolution images.
基金Supported by College Students Innovation and Entrepreneurship Training Program of Jilin University(No.202010183695)。
文摘UAV remote sensing images have the advantages of high spatial resolution,fast speed,strong real-time performance,and convenient operation,etc.,and have become a recently developed,vital means of acquiring surface information.It is an important research task for precision agriculture to make full use of the spectrum,texture,color and other characteristic information of crops,especially the spatial arrangement and structure information of features,to explore effective methods for the classification of multiple varieties of crops.In order to explore the applicability of the object-oriented method to achieve accurate classification of UAV high-resolution images,the paper used the object-oriented classification method in ENVI to classify the UAV high-resolution remote sensing image obtained from the orderly structured 28 species of crops in the test field,which mainly includes image segmentation and object classification.The results showed that the plots obtained after classification were continuous and complete,basically in line with the actual situation,and the overall accuracy of crop classification was 91.73%,with Kappa coefficient of 0.87.Compared with the crop planting area based on remote sensing interpretation and field survey,the area error of 17 species of crops in this study was controlled within 15%,which provides a basis for object-oriented crop classification of UAV remote sensing images.
文摘In recent decades,the proliferation of email communication has markedly escalated,resulting in a concomitant surge in spam emails that congest networks and presenting security risks.This study introduces an innovative spam detection method utilizing the Horse Herd Optimization Algorithm(HHOA),designed for binary classification within multi⁃objective framework.The method proficiently identifies essential features,minimizing redundancy and improving classification precision.The suggested HHOA attained an impressive accuracy of 97.21%on the Kaggle email dataset,with precision of 94.30%,recall of 90.50%,and F1⁃score of 92.80%.Compared to conventional techniques,such as Support Vector Machine(93.89%accuracy),Random Forest(96.14%accuracy),and K⁃Nearest Neighbours(92.08%accuracy),HHOA exhibited enhanced performance with reduced computing complexity.The suggested method demonstrated enhanced feature selection efficiency,decreasing the number of selected features while maintaining high classification accuracy.The results underscore the efficacy of HHOA in spam identification and indicate its potential for further applications in practical email filtering systems.
基金funded by the China National Space Administration(KJSP2023020105)supported by the National Key R&D Program of China(Grant No.2023YFA1608100)+2 种基金the NSFC(Grant No.62227901)the Minor Planet Foundationsupported by the Egyptian Science,Technology&Innovation Funding Authority(STDF)under Grant No.48102.
文摘Near-Earth objects are important not only in studying the early formation of the Solar System,but also because they pose a serious hazard to humanity when they make close approaches to the Earth.Study of their physical properties can provide useful information on their origin,evolution,and hazard to human beings.However,it remains challenging to investigate small,newly discovered,near-Earth objects because of our limited observational window.This investigation seeks to determine the visible colors of near-Earth asteroids(NEAs),perform an initial taxonomic classification based on visible colors and analyze possible correlations between the distribution of taxonomic classification and asteroid size or orbital parameters.Observations were performed in the broadband BVRI Johnson−Cousins photometric system,applied to images from the Yaoan High Precision Telescope and the 1.88 m telescope at the Kottamia Astronomical Observatory.We present new photometric observations of 84 near-Earth asteroids,and classify 80 of them taxonomically,based on their photometric colors.We find that nearly half(46.3%)of the objects in our sample can be classified as S-complex,26.3%as C-complex,6%as D-complex,and 15.0%as X-complex;the remaining belong to the A-or V-types.Additionally,we identify three P-type NEAs in our sample,according to the Tholen scheme.The fractional abundances of the C/X-complex members with absolute magnitude H≥17.0 were more than twice as large as those with H<17.0.However,the fractions of C-and S-complex members with diameters≤1 km and>1 km are nearly equal,while X-complex members tend to have sub-kilometer diameters.In our sample,the C/D-complex objects are predominant among those with a Jovian Tisserand parameter of T_(J)<3.1.These bodies could have a cometary origin.C-and S-complex members account for a considerable proportion of the asteroids that are potentially hazardous.
基金funded by the National Key Research and Development Program of China(Grant No.2024YFE0209000)the NSFC(Grant No.U23B2019).
文摘Graph Neural Networks(GNNs)have proven highly effective for graph classification across diverse fields such as social networks,bioinformatics,and finance,due to their capability to learn complex graph structures.However,despite their success,GNNs remain vulnerable to adversarial attacks that can significantly degrade their classification accuracy.Existing adversarial attack strategies primarily rely on label information to guide the attacks,which limits their applicability in scenarios where such information is scarce or unavailable.This paper introduces an innovative unsupervised attack method for graph classification,which operates without relying on label information,thereby enhancing its applicability in a broad range of scenarios.Specifically,our method first leverages a graph contrastive learning loss to learn high-quality graph embeddings by comparing different stochastic augmented views of the graphs.To effectively perturb the graphs,we then introduce an implicit estimator that measures the impact of various modifications on graph structures.The proposed strategy identifies and flips edges with the top-K highest scores,determined by the estimator,to maximize the degradation of the model’s performance.In addition,to defend against such attack,we propose a lightweight regularization-based defense mechanism that is specifically tailored to mitigate the structural perturbations introduced by our attack strategy.It enhances model robustness by enforcing embedding consistency and edge-level smoothness during training.We conduct experiments on six public TU graph classification datasets:NCI1,NCI109,Mutagenicity,ENZYMES,COLLAB,and DBLP_v1,to evaluate the effectiveness of our attack and defense strategies.Under an attack budget of 3,the maximum reduction in model accuracy reaches 6.67%on the Graph Convolutional Network(GCN)and 11.67%on the Graph Attention Network(GAT)across different datasets,indicating that our unsupervised method induces degradation comparable to state-of-the-art supervised attacks.Meanwhile,our defense achieves the highest accuracy recovery of 3.89%(GCN)and 5.00%(GAT),demonstrating improved robustness against structural perturbations.
基金funded by the Deanship of Graduate Studies and Scientific Research at Jouf University under grant No.(DGSSR-2025-02-01296).
文摘Skin diseases affect millions worldwide.Early detection is key to preventing disfigurement,lifelong disability,or death.Dermoscopic images acquired in primary-care settings show high intra-class visual similarity and severe class imbalance,and occasional imaging artifacts can create ambiguity for state-of-the-art convolutional neural networks(CNNs).We frame skin lesion recognition as graph-based reasoning and,to ensure fair evaluation and avoid data leakage,adopt a strict lesion-level partitioning strategy.Each image is first over-segmented using SLIC(Simple Linear Iterative Clustering)to produce perceptually homogeneous superpixels.These superpixels form the nodes of a region-adjacency graph whose edges encode spatial continuity.Node attributes are 1280-dimensional embeddings extracted with a lightweight yet expressive EfficientNet-B0 backbone,providing strong representational power at modest computational cost.The resulting graphs are processed by a five-layer Graph Attention Network(GAT)that learns to weight inter-node relationships dynamically and aggregates multi-hop context before classifying lesions into seven classes with a log-softmax output.Extensive experiments on the DermaMNIST benchmark show the proposed pipeline achieves 88.35%accuracy and 98.04%AUC,outperforming contemporary CNNs,AutoML approaches,and alternative graph neural networks.An ablation study indicates EfficientNet-B0 produces superior node descriptors compared with ResNet-18 and DenseNet,and that roughly five GAT layers strike a good balance between being too shallow and over-deep while avoiding oversmoothing.The method requires no data augmentation or external metadata,making it a drop-in upgrade for clinical computer-aided diagnosis systems.
文摘This systematic review aims to comprehensively examine and compare deep learning methods for brain tumor segmentation and classification using MRI and other imaging modalities,focusing on recent trends from 2022 to 2025.The primary objective is to evaluate methodological advancements,model performance,dataset usage,and existing challenges in developing clinically robust AI systems.We included peer-reviewed journal articles and highimpact conference papers published between 2022 and 2025,written in English,that proposed or evaluated deep learning methods for brain tumor segmentation and/or classification.Excluded were non-open-access publications,books,and non-English articles.A structured search was conducted across Scopus,Google Scholar,Wiley,and Taylor&Francis,with the last search performed in August 2025.Risk of bias was not formally quantified but considered during full-text screening based on dataset diversity,validation methods,and availability of performance metrics.We used narrative synthesis and tabular benchmarking to compare performance metrics(e.g.,accuracy,Dice score)across model types(CNN,Transformer,Hybrid),imaging modalities,and datasets.A total of 49 studies were included(43 journal articles and 6 conference papers).These studies spanned over 9 public datasets(e.g.,BraTS,Figshare,REMBRANDT,MOLAB)and utilized a range of imaging modalities,predominantly MRI.Hybrid models,especially ResViT and UNetFormer,consistently achieved high performance,with classification accuracy exceeding 98%and segmentation Dice scores above 0.90 across multiple studies.Transformers and hybrid architectures showed increasing adoption post2023.Many studies lacked external validation and were evaluated only on a few benchmark datasets,raising concerns about generalizability and dataset bias.Few studies addressed clinical interpretability or uncertainty quantification.Despite promising results,particularly for hybrid deep learning models,widespread clinical adoption remains limited due to lack of validation,interpretability concerns,and real-world deployment barriers.
文摘Honeycombing Lung(HCL)is a chronic lung condition marked by advanced fibrosis,resulting in enlarged air spaces with thick fibrotic walls,which are visible on Computed Tomography(CT)scans.Differentiating between normal lung tissue,honeycombing lungs,and Ground Glass Opacity(GGO)in CT images is often challenging for radiologists and may lead to misinterpretations.Although earlier studies have proposed models to detect and classify HCL,many faced limitations such as high computational demands,lower accuracy,and difficulty distinguishing between HCL and GGO.CT images are highly effective for lung classification due to their high resolution,3D visualization,and sensitivity to tissue density variations.This study introduces Honeycombing Lungs Network(HCL Net),a novel classification algorithm inspired by ResNet50V2 and enhanced to overcome the shortcomings of previous approaches.HCL Net incorporates additional residual blocks,refined preprocessing techniques,and selective parameter tuning to improve classification performance.The dataset,sourced from the University Malaya Medical Centre(UMMC)and verified by expert radiologists,consists of CT images of normal,honeycombing,and GGO lungs.Experimental evaluations across five assessments demonstrated that HCL Net achieved an outstanding classification accuracy of approximately 99.97%.It also recorded strong performance in other metrics,achieving 93%precision,100%sensitivity,89%specificity,and an AUC-ROC score of 97%.Comparative analysis with baseline feature engineering methods confirmed the superior efficacy of HCL Net.The model significantly reduces misclassification,particularly between honeycombing and GGO lungs,enhancing diagnostic precision and reliability in lung image analysis.
基金funded by the Directorate of Research and Community Service,Directorate General of Research and Development,Ministry of Higher Education,Science and Technologyin accordance with the Implementation Contract for the Operational Assistance Program for State Universities,Research Program Number:109/C3/DT.05.00/PL/2025.
文摘Sudden wildfires cause significant global ecological damage.While satellite imagery has advanced early fire detection and mitigation,image-based systems face limitations including high false alarm rates,visual obstructions,and substantial computational demands,especially in complex forest terrains.To address these challenges,this study proposes a novel forest fire detection model utilizing audio classification and machine learning.We developed an audio-based pipeline using real-world environmental sound recordings.Sounds were converted into Mel-spectrograms and classified via a Convolutional Neural Network(CNN),enabling the capture of distinctive fire acoustic signatures(e.g.,crackling,roaring)that are minimally impacted by visual or weather conditions.Internet of Things(IoT)sound sensors were crucial for generating complex environmental parameters to optimize feature extraction.The CNN model achieved high performance in stratified 5-fold cross-validation(92.4%±1.6 accuracy,91.2%±1.8 F1-score)and on test data(94.93%accuracy,93.04%F1-score),with 98.44%precision and 88.32%recall,demonstrating reliability across environmental conditions.These results indicate that the audio-based approach not only improves detection reliability but also markedly reduces computational overhead compared to traditional image-based methods.The findings suggest that acoustic sensing integrated with machine learning offers a powerful,low-cost,and efficient solution for real-time forest fire monitoring in complex,dynamic environments.
基金Under the auspices of National Key R&D Program of China(No.2024YFF1306405)。
文摘Accurate extraction of surface water extent is a fundamental prerequisite for monitoring its dynamic changes.Although machine learning algorithms have been widely applied to surface water mapping,most studies focus primarily on algorithmic outputs,with limited systematic evaluation of their applicability and constrained classification accuracy.In this study,we focused on the Songnen Plain in Northeast China and employed Sentinel-2 imagery acquired during 2020-2021 via the Google Earth Engine(GEE)platform to evaluate the performance of Classification and Regression Trees(CART),Random Forest(RF),and Support Vector Machine(SVM)for surface water classification.The classification process was optimized by incorporating automated training sample selection and integration of time series features.Validation with independent samples demonstrated the feasibility of automatic sample selection,yielding mean overall accuracies of 91.16%,90.99%,and 90.76%for RF,SVM,and CART,respectively.After integrating time series features,the mean overall accuracies of the three algorithms improved by 4.51%,5.45%,and 6.36%,respectively.In addition,spectral features such as MNDWI(Modified Normalized Difference Water Index),SWIR(Short Wave Infrared),and NDVI(Normalized Difference Vegetation Index)were identified as more important for surface water classification.This study establishes a more consistent framework for surface water mapping,offering new perspectives for improving and automating classification processes in the era of big and open data.
基金funded by the Deanship of Graduate Studies and Scientific Research at Jouf University under grant No.(DGSSR-2025-02-01295).
文摘Alzheimer’s Disease(AD)is a progressive neurodegenerative disorder that significantly affects cognitive function,making early and accurate diagnosis essential.Traditional Deep Learning(DL)-based approaches often struggle with low-contrast MRI images,class imbalance,and suboptimal feature extraction.This paper develops a Hybrid DL system that unites MobileNetV2 with adaptive classification methods to boost Alzheimer’s diagnosis by processing MRI scans.Image enhancement is done using Contrast-Limited Adaptive Histogram Equalization(CLAHE)and Enhanced Super-Resolution Generative Adversarial Networks(ESRGAN).A classification robustness enhancement system integrates class weighting techniques and a Matthews Correlation Coefficient(MCC)-based evaluation method into the design.The trained and validated model gives a 98.88%accuracy rate and 0.9614 MCC score.We also performed a 10-fold cross-validation experiment with an average accuracy of 96.52%(±1.51),a loss of 0.1671,and an MCC score of 0.9429 across folds.The proposed framework outperforms the state-of-the-art models with a 98%weighted F1-score while decreasing misdiagnosis results for every AD stage.The model demonstrates apparent separation abilities between AD progression stages according to the results of the confusion matrix analysis.These results validate the effectiveness of hybrid DL models with adaptive preprocessing for early and reliable Alzheimer’s diagnosis,contributing to improved computer-aided diagnosis(CAD)systems in clinical practice.
基金Under the auspices of the National Natural Science Foundation of China (No. 40301038), Talents Recruitment Foun-dation of Nanjing University
文摘With the wide use of high-resolution remotely sensed imagery, the object-oriented remotely sensed informa- tion classification pattern has been intensively studied. Starting with the definition of object-oriented remotely sensed information classification pattern and a literature review of related research progress, this paper sums up 4 developing phases of object-oriented classification pattern during the past 20 years. Then, we discuss the three aspects of method- ology in detail, namely remotely sensed imagery segmentation, feature analysis and feature selection, and classification rule generation, through comparing them with remotely sensed information classification method based on per-pixel. At last, this paper presents several points that need to be paid attention to in the future studies on object-oriented RS in- formation classification pattern: 1) developing robust and highly effective image segmentation algorithm for multi-spectral RS imagery; 2) improving the feature-set including edge, spatial-adjacent and temporal characteristics; 3) discussing the classification rule generation classifier based on the decision tree; 4) presenting evaluation methods for classification result by object-oriented classification pattern.
基金Supported by the National Natural Science Foundation of China(Nos.42376185,41876111)the Shandong Provincial Natural Science Foundation(No.ZR2023MD073)。
文摘Benthic habitat mapping is an emerging discipline in the international marine field in recent years,providing an effective tool for marine spatial planning,marine ecological management,and decision-making applications.Seabed sediment classification is one of the main contents of seabed habitat mapping.In response to the impact of remote sensing imaging quality and the limitations of acoustic measurement range,where a single data source does not fully reflect the substrate type,we proposed a high-precision seabed habitat sediment classification method that integrates data from multiple sources.Based on WorldView-2 multi-spectral remote sensing image data and multibeam bathymetry data,constructed a random forests(RF)classifier with optimal feature selection.A seabed sediment classification experiment integrating optical remote sensing and acoustic remote sensing data was carried out in the shallow water area of Wuzhizhou Island,Hainan,South China.Different seabed sediment types,such as sand,seagrass,and coral reefs were effectively identified,with an overall classification accuracy of 92%.Experimental results show that RF matrix optimized by fusing multi-source remote sensing data for feature selection were better than the classification results of simple combinations of data sources,which improved the accuracy of seabed sediment classification.Therefore,the method proposed in this paper can be effectively applied to high-precision seabed sediment classification and habitat mapping around islands and reefs.
文摘This paper proposed to use double polarization synthetic aperture radar (SAR) image to classify surface feature, based on DEM. It takes fully use of the polarization information and external information. This pa-per utilizes ENVISAT ASAR APP double-polarization data of Poyang lake area in Jiangxi Province. Com-pared with traditional pixel-based classification, this paper fully uses object features (color, shape, hierarchy) and accessorial DEM information. The classification accuracy improves from the original 73.7% to 91.84%. The result shows that object-oriented classification technology is suitable for double polarization SAR’s high precision classification.
基金This paper is funded jointly by projects of the National Natural Science Foundation of China(41571374),the key research project of Hunan Education Ministry(No.16A070),Nature Science Joint Funding of Hunan province and Xiangtan Local(No.2017JJ4037).
文摘With the development of remote sensing technology,the spatial resolution,spectral resolution and time resolution of remote sensing data are greatly improved.How to efficiently process and interpret the massive high resolution remote sensing image data for ground objects,which are of spatial geometry and texture information,has become the focus and key issue in the field of remote sensing research.A new method of the classification(OCRC(object-oriented and class rule classification))of remote sensing,which is of object-oriented and rule,has been presented in this paper,that is,through the discovery and mining the knowledge of spectrum and spatial characteristics of high-resolution remote sensing image,establish a multi-level network image object segmentation and classification structure of remote sensing image to achieve classification and accuracy assessment accurately and quickly for ground targets.Selected the worldview-2 image data in the Zangnan area as a study object,using the OCRC to verify the experiment which was a combination of the mean variance method,the maximum area method and the accuracy comparison to analysis selected three kinds of optimal segmentation scale and established a multi-level image object network hierarchy for image classification experiments.The results show that the OCRC can enable the high resolution image classification results similar to the visual interpretation of the results,and has higher classification accuracy.The overall accuracy and Kappa coefficient of the object-oriented rule classification method are 97.38%,0.9673;compared with object-oriented SVM method,respectively higher than 6.23%,0.078;compared with object-oriented KNN method,respectively more than 7.96%,0.0996.The extraction precision and user accuracy of the buildings compared with object-oriented SVM method,respectively higher than 18.39%,3.98%,respectively better than the object-oriented KNN(K-Nearest Neighbor)method 21.27%,14.97%.
基金Research on Algorithm Model for Monitoring and Evaluating Typical Disaster Situations of Electric Power Equipment Based on Remote Sensing Imaging Technology of Heaven and Earth,South Grid Guangxi Power Grid Company Science and Technology Project(GXKJXM20222160).
文摘As one of the main geographical elements in urban areas,buildings are closely related to the development of the city.Therefore,how to quickly and accurately extract building information from remote sensing images is of great significance for urban map updating,urban planning and construction,etc.Extracting building information around power facilities,especially obtaining this information from high-resolution images,has become one of the current hot topics in remote sensing technology research.This study made full use of the characteristics of GF-2 satellite remote sensing images,adopted an object-oriented classification method,combined with multi-scale segmentation technology and CART classification algorithm,and successfully extracted the buildings in the study area.The research results showed that the overall classification accuracy reached 89.5%and the Kappa coefficient was 0.86.Using the object-oriented CART classification algorithm for building extraction could be closer to actual ground objects and had higher accuracy.The extraction of buildings in the city contributed to urban development planning and provided decision support for management.
文摘This paper investigates the contribution of multi-temporal enhanced vegetation index(EVI)data to the improvement of object-based classification accuracy using multi-spectral moderate resolution imaging spectral-radiometer(MODIS)imagery.In object-oriented classification,similar pixels are firstly grouped together and then classified;the produced result does not suffer the speckled appearance and closer to human vision.EVI data are from the MODIS sensor aboard Terra spacecraft.69 EVI data(scenes)were collected during the period of three years(20012003)in a mountainous vegetated area.These data sets were used to study the phenology of the land cover types.Different land cover types show distinct fluctuations over time in EVI values and this information might be used to improve object-oriented land cover classification.Two experiments were carried out:one was only with single date MODIS multispectral data,and the other one including also the 69 EVI images.Eight classes were distinguished:temperate forest,tropical dry forest,grassland,irrigated agriculture,rain-fed agriculture,orchards,lava flows and human settlement.The two classifications were evaluated with independent verification data,and the results showed that with multitemporal EVI data,the classification accuracy was improved 5.2%.Evaluated by McNemar’s test,this improved was significant,with significance level p=0.01.
文摘Urban tree species provide various essential ecosystem services in cities,such as regulating urban temperatures,reducing noise,capturing carbon,and mitigating the urban heat island effect.The quality of these services is influenced by species diversity,tree health,and the distribution and the composition of trees.Traditionally,data on urban trees has been collected through field surveys and manual interpretation of remote sensing images.In this study,we evaluated the effectiveness of multispectral airborne laser scanning(ALS)data in classifying 24 common urban roadside tree species in Espoo,Finland.Tree crown structure information,intensity features,and spectral data were used for classification.Eight different machine learning algorithms were tested,with the extra trees(ET)algorithm performing the best,achieving an overall accuracy of 71.7%using multispectral LiDAR data.This result highlights that integrating structural and spectral information within a single framework can improve the classification accuracy.Future research will focus on identifying the most important features for species classification and developing algorithms with greater efficiency and accuracy.
文摘Purpose:Interdisciplinary research has become a critical approach to addressing complex societal,economic,technological,and environmental challenges,driving innovation and integrating scientific knowledge.While interdisciplinarity indicators are widely used to evaluate research performance,the impact of classification granularity on these assessments remains underexplored.Design/methodology/approach:This study investigates how different levels of classification granularity-macro,meso,and micro-affect the evaluation of interdisciplinarity in research institutes.Using a dataset of 262 institutes from four major German non-university organizations(FHG,HGF,MPG,WGL)from 2018 to 2022,we examine inconsistencies in interdisciplinarity across levels,analyze ranking changes,and explore the influence of institutional fields and research focus(applied vs.basic).Findings:Our findings reveal significant inconsistencies in interdisciplinarity across classification levels,with rankings varying substantially.Notably,the Fraunhofer Society(FHG),which performs well at the macro level,experiences significant ranking declines at meso and micro levels.Normalizing interdisciplinarity by research field confirmed that these declines persist.The research focus of institutes,whether applied,basic,or mixed,does not significantly explain the observed ranking dynamics.Research limitations:This study has only considered the publication-based dimension of institutional interdisciplinarity and has not explored other aspects.Practical implications:The findings provide insights for policymakers,research managers,and scholars to better interpret interdisciplinarity metrics and support interdisciplinary research effectively.Originality/value:This study underscores the critical role of classification granularity in interdisciplinarity assessment and emphasizes the need for standardized approaches to ensure robust and fair evaluations.
文摘Breast cancer is one of the major causes of deaths in women.However,the early diagnosis is important for screening and control the mortality rate.Thus for the diagnosis of breast cancer at the early stage,a computer-aided diagnosis system is highly required.Ultrasound is an important examination technique for breast cancer diagnosis due to its low cost.Recently,many learning-based techniques have been introduced to classify breast cancer using breast ultrasound imaging dataset(BUSI)datasets;however,the manual handling is not an easy process and time consuming.The authors propose an EfficientNet-integrated ResNet deep network and XAI-based framework for accurately classifying breast cancer(malignant and benign).In the initial step,data augmentation is performed to increase the number of training samples.For this purpose,three-pixel flip mathematical equations are introduced:horizontal,vertical,and 90°.Later,two pretrained deep learning models were employed,skipped some layers,and fine-tuned.Both fine-tuned models are later trained using a deep transfer learning process and extracted features from the deeper layer.Explainable artificial intelligence-based analysed the performance of trained models.After that,a new feature selection technique is proposed based on the cuckoo search algorithm called cuckoo search controlled standard error mean.This technique selects the best features and fuses using a new parallel zeropadding maximum correlated coefficient features.In the end,the selection algorithm is applied again to the fused feature vector and classified using machine learning algorithms.The experimental process of the proposed framework is conducted on a publicly available BUSI and obtained 98.4%and 98%accuracy in two different experiments.Comparing the proposed framework is also conducted with recent techniques and shows improved accuracy.In addition,the proposed framework was executed less than the original deep learning models.