Near-Earth objects are important not only in studying the early formation of the Solar System,but also because they pose a serious hazard to humanity when they make close approaches to the Earth.Study of their physica...Near-Earth objects are important not only in studying the early formation of the Solar System,but also because they pose a serious hazard to humanity when they make close approaches to the Earth.Study of their physical properties can provide useful information on their origin,evolution,and hazard to human beings.However,it remains challenging to investigate small,newly discovered,near-Earth objects because of our limited observational window.This investigation seeks to determine the visible colors of near-Earth asteroids(NEAs),perform an initial taxonomic classification based on visible colors and analyze possible correlations between the distribution of taxonomic classification and asteroid size or orbital parameters.Observations were performed in the broadband BVRI Johnson−Cousins photometric system,applied to images from the Yaoan High Precision Telescope and the 1.88 m telescope at the Kottamia Astronomical Observatory.We present new photometric observations of 84 near-Earth asteroids,and classify 80 of them taxonomically,based on their photometric colors.We find that nearly half(46.3%)of the objects in our sample can be classified as S-complex,26.3%as C-complex,6%as D-complex,and 15.0%as X-complex;the remaining belong to the A-or V-types.Additionally,we identify three P-type NEAs in our sample,according to the Tholen scheme.The fractional abundances of the C/X-complex members with absolute magnitude H≥17.0 were more than twice as large as those with H<17.0.However,the fractions of C-and S-complex members with diameters≤1 km and>1 km are nearly equal,while X-complex members tend to have sub-kilometer diameters.In our sample,the C/D-complex objects are predominant among those with a Jovian Tisserand parameter of T_(J)<3.1.These bodies could have a cometary origin.C-and S-complex members account for a considerable proportion of the asteroids that are potentially hazardous.展开更多
Graph Neural Networks(GNNs)have proven highly effective for graph classification across diverse fields such as social networks,bioinformatics,and finance,due to their capability to learn complex graph structures.Howev...Graph Neural Networks(GNNs)have proven highly effective for graph classification across diverse fields such as social networks,bioinformatics,and finance,due to their capability to learn complex graph structures.However,despite their success,GNNs remain vulnerable to adversarial attacks that can significantly degrade their classification accuracy.Existing adversarial attack strategies primarily rely on label information to guide the attacks,which limits their applicability in scenarios where such information is scarce or unavailable.This paper introduces an innovative unsupervised attack method for graph classification,which operates without relying on label information,thereby enhancing its applicability in a broad range of scenarios.Specifically,our method first leverages a graph contrastive learning loss to learn high-quality graph embeddings by comparing different stochastic augmented views of the graphs.To effectively perturb the graphs,we then introduce an implicit estimator that measures the impact of various modifications on graph structures.The proposed strategy identifies and flips edges with the top-K highest scores,determined by the estimator,to maximize the degradation of the model’s performance.In addition,to defend against such attack,we propose a lightweight regularization-based defense mechanism that is specifically tailored to mitigate the structural perturbations introduced by our attack strategy.It enhances model robustness by enforcing embedding consistency and edge-level smoothness during training.We conduct experiments on six public TU graph classification datasets:NCI1,NCI109,Mutagenicity,ENZYMES,COLLAB,and DBLP_v1,to evaluate the effectiveness of our attack and defense strategies.Under an attack budget of 3,the maximum reduction in model accuracy reaches 6.67%on the Graph Convolutional Network(GCN)and 11.67%on the Graph Attention Network(GAT)across different datasets,indicating that our unsupervised method induces degradation comparable to state-of-the-art supervised attacks.Meanwhile,our defense achieves the highest accuracy recovery of 3.89%(GCN)and 5.00%(GAT),demonstrating improved robustness against structural perturbations.展开更多
Skin diseases affect millions worldwide.Early detection is key to preventing disfigurement,lifelong disability,or death.Dermoscopic images acquired in primary-care settings show high intra-class visual similarity and ...Skin diseases affect millions worldwide.Early detection is key to preventing disfigurement,lifelong disability,or death.Dermoscopic images acquired in primary-care settings show high intra-class visual similarity and severe class imbalance,and occasional imaging artifacts can create ambiguity for state-of-the-art convolutional neural networks(CNNs).We frame skin lesion recognition as graph-based reasoning and,to ensure fair evaluation and avoid data leakage,adopt a strict lesion-level partitioning strategy.Each image is first over-segmented using SLIC(Simple Linear Iterative Clustering)to produce perceptually homogeneous superpixels.These superpixels form the nodes of a region-adjacency graph whose edges encode spatial continuity.Node attributes are 1280-dimensional embeddings extracted with a lightweight yet expressive EfficientNet-B0 backbone,providing strong representational power at modest computational cost.The resulting graphs are processed by a five-layer Graph Attention Network(GAT)that learns to weight inter-node relationships dynamically and aggregates multi-hop context before classifying lesions into seven classes with a log-softmax output.Extensive experiments on the DermaMNIST benchmark show the proposed pipeline achieves 88.35%accuracy and 98.04%AUC,outperforming contemporary CNNs,AutoML approaches,and alternative graph neural networks.An ablation study indicates EfficientNet-B0 produces superior node descriptors compared with ResNet-18 and DenseNet,and that roughly five GAT layers strike a good balance between being too shallow and over-deep while avoiding oversmoothing.The method requires no data augmentation or external metadata,making it a drop-in upgrade for clinical computer-aided diagnosis systems.展开更多
This systematic review aims to comprehensively examine and compare deep learning methods for brain tumor segmentation and classification using MRI and other imaging modalities,focusing on recent trends from 2022 to 20...This systematic review aims to comprehensively examine and compare deep learning methods for brain tumor segmentation and classification using MRI and other imaging modalities,focusing on recent trends from 2022 to 2025.The primary objective is to evaluate methodological advancements,model performance,dataset usage,and existing challenges in developing clinically robust AI systems.We included peer-reviewed journal articles and highimpact conference papers published between 2022 and 2025,written in English,that proposed or evaluated deep learning methods for brain tumor segmentation and/or classification.Excluded were non-open-access publications,books,and non-English articles.A structured search was conducted across Scopus,Google Scholar,Wiley,and Taylor&Francis,with the last search performed in August 2025.Risk of bias was not formally quantified but considered during full-text screening based on dataset diversity,validation methods,and availability of performance metrics.We used narrative synthesis and tabular benchmarking to compare performance metrics(e.g.,accuracy,Dice score)across model types(CNN,Transformer,Hybrid),imaging modalities,and datasets.A total of 49 studies were included(43 journal articles and 6 conference papers).These studies spanned over 9 public datasets(e.g.,BraTS,Figshare,REMBRANDT,MOLAB)and utilized a range of imaging modalities,predominantly MRI.Hybrid models,especially ResViT and UNetFormer,consistently achieved high performance,with classification accuracy exceeding 98%and segmentation Dice scores above 0.90 across multiple studies.Transformers and hybrid architectures showed increasing adoption post2023.Many studies lacked external validation and were evaluated only on a few benchmark datasets,raising concerns about generalizability and dataset bias.Few studies addressed clinical interpretability or uncertainty quantification.Despite promising results,particularly for hybrid deep learning models,widespread clinical adoption remains limited due to lack of validation,interpretability concerns,and real-world deployment barriers.展开更多
Honeycombing Lung(HCL)is a chronic lung condition marked by advanced fibrosis,resulting in enlarged air spaces with thick fibrotic walls,which are visible on Computed Tomography(CT)scans.Differentiating between normal...Honeycombing Lung(HCL)is a chronic lung condition marked by advanced fibrosis,resulting in enlarged air spaces with thick fibrotic walls,which are visible on Computed Tomography(CT)scans.Differentiating between normal lung tissue,honeycombing lungs,and Ground Glass Opacity(GGO)in CT images is often challenging for radiologists and may lead to misinterpretations.Although earlier studies have proposed models to detect and classify HCL,many faced limitations such as high computational demands,lower accuracy,and difficulty distinguishing between HCL and GGO.CT images are highly effective for lung classification due to their high resolution,3D visualization,and sensitivity to tissue density variations.This study introduces Honeycombing Lungs Network(HCL Net),a novel classification algorithm inspired by ResNet50V2 and enhanced to overcome the shortcomings of previous approaches.HCL Net incorporates additional residual blocks,refined preprocessing techniques,and selective parameter tuning to improve classification performance.The dataset,sourced from the University Malaya Medical Centre(UMMC)and verified by expert radiologists,consists of CT images of normal,honeycombing,and GGO lungs.Experimental evaluations across five assessments demonstrated that HCL Net achieved an outstanding classification accuracy of approximately 99.97%.It also recorded strong performance in other metrics,achieving 93%precision,100%sensitivity,89%specificity,and an AUC-ROC score of 97%.Comparative analysis with baseline feature engineering methods confirmed the superior efficacy of HCL Net.The model significantly reduces misclassification,particularly between honeycombing and GGO lungs,enhancing diagnostic precision and reliability in lung image analysis.展开更多
Sudden wildfires cause significant global ecological damage.While satellite imagery has advanced early fire detection and mitigation,image-based systems face limitations including high false alarm rates,visual obstruc...Sudden wildfires cause significant global ecological damage.While satellite imagery has advanced early fire detection and mitigation,image-based systems face limitations including high false alarm rates,visual obstructions,and substantial computational demands,especially in complex forest terrains.To address these challenges,this study proposes a novel forest fire detection model utilizing audio classification and machine learning.We developed an audio-based pipeline using real-world environmental sound recordings.Sounds were converted into Mel-spectrograms and classified via a Convolutional Neural Network(CNN),enabling the capture of distinctive fire acoustic signatures(e.g.,crackling,roaring)that are minimally impacted by visual or weather conditions.Internet of Things(IoT)sound sensors were crucial for generating complex environmental parameters to optimize feature extraction.The CNN model achieved high performance in stratified 5-fold cross-validation(92.4%±1.6 accuracy,91.2%±1.8 F1-score)and on test data(94.93%accuracy,93.04%F1-score),with 98.44%precision and 88.32%recall,demonstrating reliability across environmental conditions.These results indicate that the audio-based approach not only improves detection reliability but also markedly reduces computational overhead compared to traditional image-based methods.The findings suggest that acoustic sensing integrated with machine learning offers a powerful,low-cost,and efficient solution for real-time forest fire monitoring in complex,dynamic environments.展开更多
Alzheimer’s Disease(AD)is a progressive neurodegenerative disorder that significantly affects cognitive function,making early and accurate diagnosis essential.Traditional Deep Learning(DL)-based approaches often stru...Alzheimer’s Disease(AD)is a progressive neurodegenerative disorder that significantly affects cognitive function,making early and accurate diagnosis essential.Traditional Deep Learning(DL)-based approaches often struggle with low-contrast MRI images,class imbalance,and suboptimal feature extraction.This paper develops a Hybrid DL system that unites MobileNetV2 with adaptive classification methods to boost Alzheimer’s diagnosis by processing MRI scans.Image enhancement is done using Contrast-Limited Adaptive Histogram Equalization(CLAHE)and Enhanced Super-Resolution Generative Adversarial Networks(ESRGAN).A classification robustness enhancement system integrates class weighting techniques and a Matthews Correlation Coefficient(MCC)-based evaluation method into the design.The trained and validated model gives a 98.88%accuracy rate and 0.9614 MCC score.We also performed a 10-fold cross-validation experiment with an average accuracy of 96.52%(±1.51),a loss of 0.1671,and an MCC score of 0.9429 across folds.The proposed framework outperforms the state-of-the-art models with a 98%weighted F1-score while decreasing misdiagnosis results for every AD stage.The model demonstrates apparent separation abilities between AD progression stages according to the results of the confusion matrix analysis.These results validate the effectiveness of hybrid DL models with adaptive preprocessing for early and reliable Alzheimer’s diagnosis,contributing to improved computer-aided diagnosis(CAD)systems in clinical practice.展开更多
Objective:The World Health Organization(WHO)grading based on histopathology cannot always accurately predict tumor behavior of meningiomas.To overcome the limitations of the WHO grading,the study aims to propose a nov...Objective:The World Health Organization(WHO)grading based on histopathology cannot always accurately predict tumor behavior of meningiomas.To overcome the limitations of the WHO grading,the study aims to propose a novel oxidative stress-based molecular classification for WHO grade 2/3 meningiomas.Methods:Differentially expressed oxidative stress-related genes were analyzed between 86 WHO grade 1(low grade)meningiomas and 99 grade 2/3(high grade)meningiomas.An oxidative stress-based molecular classification was developed in high-grade meningiomas through consensus clustering analysis.Immune microenvironment features,responses to immunotherapy and chemotherapy,and targeted drugs were evaluated.Three machine learning models:logistic regression,support vector machine,and random forest,were built for differentiating the classification.Key oxidative stress-related geneswere verified in humanmeningeal cells(HMC)and two meningioma cells(CH-157MN and IOMMLee)via reverse transcription quantitative polymerase chain reaction(RT-qPCR)and western blot.After knockdown of Forkhead Box M1(FOXM1)or Prion Protein(PRNP),cell growth,migration,and reactive oxygen species(ROS)levels were measured through cell counting kit-8(CCK-8),transwell,and immunofluorescence,respectively.Results:We classified high-grade meningiomas into two oxidative stress-based clusters,termed cluster 1 and cluster 2.Cluster 1 exhibited higher infiltrations of immune and stromal cells and higher expression of classic immune checkpoints:Cluster of Differentiation 86(CD86),Programmed Cell Death 1(PDCD1),and Leukocyte-Associated Immunoglobulin-Like Receptor 1(LAIR1),indicating that cluster 1 meningiomas might respond to immunotherapy.Drug sensitivity was heterogeneous between the two clusters.Three classifiers were established,which could accurately differentiate this molecular classification.FOXM1 and PRNP were experimentally evidenced to be highly expressed inmeningioma cells,and their knockdown hindered cell growth and migration and triggered ROS accumulation.Conclusion:In summary,our findings established a novel oxidative stress-based molecular classification and identified potential treatment vulnerabilities in high-grade meningiomas,which might assist personalized clinical management.展开更多
A modified hydrometeor classification algorithm (HCA) is developed in this study for Chinese polarimetric radars. This algorithm is based on the U.S. operational HCA. Meanwhile, the methodology of statistics-based o...A modified hydrometeor classification algorithm (HCA) is developed in this study for Chinese polarimetric radars. This algorithm is based on the U.S. operational HCA. Meanwhile, the methodology of statistics-based optimization is proposed including calibration checking, datasets selection, membership functions modification, computation thresholds modification, and effect verification. Zhuhai radar, the first operational polarimetric radar in South China, applies these procedures. The systematic bias of calibration is corrected, the reliability of radar measurements deteriorates when the signal-to-noise ratio is low, and correlation coefficient within the melting layer is usually lower than that of the U.S. WSR-88D radar. Through modification based on statistical analysis of polarimetric variables, the localized HCA especially for Zhuhai is obtained, and it performs well over a one-month test through comparison with sounding and surface observations. The algorithm is then utilized for analysis of a squall line process on 11 May 2014 and is found to provide reasonable details with respect to horizontal and vertical structures, and the HCA results---especially in the mixed rain-hail region--can reflect the life cycle of the squall line. In addition, the kinematic and microphysical processes of cloud evolution and the differences between radar- detected hail and surface observations are also analyzed. The results of this study provide evidence for the improvement of this HCA developed specifically for China.展开更多
Urban tree species provide various essential ecosystem services in cities,such as regulating urban temperatures,reducing noise,capturing carbon,and mitigating the urban heat island effect.The quality of these services...Urban tree species provide various essential ecosystem services in cities,such as regulating urban temperatures,reducing noise,capturing carbon,and mitigating the urban heat island effect.The quality of these services is influenced by species diversity,tree health,and the distribution and the composition of trees.Traditionally,data on urban trees has been collected through field surveys and manual interpretation of remote sensing images.In this study,we evaluated the effectiveness of multispectral airborne laser scanning(ALS)data in classifying 24 common urban roadside tree species in Espoo,Finland.Tree crown structure information,intensity features,and spectral data were used for classification.Eight different machine learning algorithms were tested,with the extra trees(ET)algorithm performing the best,achieving an overall accuracy of 71.7%using multispectral LiDAR data.This result highlights that integrating structural and spectral information within a single framework can improve the classification accuracy.Future research will focus on identifying the most important features for species classification and developing algorithms with greater efficiency and accuracy.展开更多
A small scale red soil resources information system (RSRIS) with applied mathematical models wasdeveloped and applied in red soil resources (RSR) classification and evaluation, taking Zhejiang Province,a typical distr...A small scale red soil resources information system (RSRIS) with applied mathematical models wasdeveloped and applied in red soil resources (RSR) classification and evaluation, taking Zhejiang Province,a typical distribution area of red soil, as the study area. Computer-aided overlay was conducted to classifyRSR types. The evaluation was carried out by using three methods, i.e., index summation, square root ofindex multiplication and fuzzy comprehensive assessment, with almost identical results. The result of indexsummation could represent the basic qualitative condition of RSR, that of square root of index multiplicationreflected the real condition of RSR qualitative rank, while fuzzy comprehensive assessment could satisfactorilyhandle the relationship between the evaluation factors and the qualitative rank of RSR, and therefore it is afeasible method for RSR evaluation.展开更多
Purpose:Interdisciplinary research has become a critical approach to addressing complex societal,economic,technological,and environmental challenges,driving innovation and integrating scientific knowledge.While interd...Purpose:Interdisciplinary research has become a critical approach to addressing complex societal,economic,technological,and environmental challenges,driving innovation and integrating scientific knowledge.While interdisciplinarity indicators are widely used to evaluate research performance,the impact of classification granularity on these assessments remains underexplored.Design/methodology/approach:This study investigates how different levels of classification granularity-macro,meso,and micro-affect the evaluation of interdisciplinarity in research institutes.Using a dataset of 262 institutes from four major German non-university organizations(FHG,HGF,MPG,WGL)from 2018 to 2022,we examine inconsistencies in interdisciplinarity across levels,analyze ranking changes,and explore the influence of institutional fields and research focus(applied vs.basic).Findings:Our findings reveal significant inconsistencies in interdisciplinarity across classification levels,with rankings varying substantially.Notably,the Fraunhofer Society(FHG),which performs well at the macro level,experiences significant ranking declines at meso and micro levels.Normalizing interdisciplinarity by research field confirmed that these declines persist.The research focus of institutes,whether applied,basic,or mixed,does not significantly explain the observed ranking dynamics.Research limitations:This study has only considered the publication-based dimension of institutional interdisciplinarity and has not explored other aspects.Practical implications:The findings provide insights for policymakers,research managers,and scholars to better interpret interdisciplinarity metrics and support interdisciplinary research effectively.Originality/value:This study underscores the critical role of classification granularity in interdisciplinarity assessment and emphasizes the need for standardized approaches to ensure robust and fair evaluations.展开更多
Support Vector Machines (SVM) is a new general machine-learning tool based on structural risk minimization principle. This characteristic is very signific ant for the fault diagnostics when the number of fault sampl...Support Vector Machines (SVM) is a new general machine-learning tool based on structural risk minimization principle. This characteristic is very signific ant for the fault diagnostics when the number of fault samples is limited. Considering that SVM theory is originally designed for a two-class classification, a hybrid SVM scheme is proposed for multi-fault classification of rotating machinery in our paper. Two SVM strategies, 1-v-1 (one versus one) and 1-v-r (one versus rest), are respectively adopted at different classification levels. At the parallel classification level, using l-v-1 strategy, the fault features extracted by various signal analysis methods are transferred into the multiple parallel SVM and the local classification results are obtained. At the serial classification level, these local results values are fused by one serial SVM based on 1-v-r strategy. The hybrid SVM scheme introduced in our paper not only generalizes the performance of signal binary SVMs but improves the precision and reliability of the fault classification results. The actually testing results show the availability suitability of this new method.展开更多
Preservation of the crops depends on early and accurate detection of pests on crops as they cause several diseases decreasing crop production and quality. Several deep-learning techniques have been applied to overcome...Preservation of the crops depends on early and accurate detection of pests on crops as they cause several diseases decreasing crop production and quality. Several deep-learning techniques have been applied to overcome the issue of pest detection on crops. We have developed the YOLOCSP-PEST model for Pest localization and classification. With the Cross Stage Partial Network (CSPNET) backbone, the proposed model is a modified version of You Only Look Once Version 7 (YOLOv7) that is intended primarily for pest localization and classification. Our proposed model gives exceptionally good results under conditions that are very challenging for any other comparable models especially conditions where we have issues with the luminance and the orientation of the images. It helps farmers working out on their crops in distant areas to determine any infestation quickly and accurately on their crops which helps in the quality and quantity of the production yield. The model has been trained and tested on 2 datasets namely the IP102 data set and a local crop data set on both of which it has shown exceptional results. It gave us a mean average precision (mAP) of 88.40% along with a precision of 85.55% and a recall of 84.25% on the IP102 dataset meanwhile giving a mAP of 97.18% on the local data set along with a recall of 94.88% and a precision of 97.50%. These findings demonstrate that the proposed model is very effective in detecting real-life scenarios and can help in the production of crops improving the yield quality and quantity at the same time.展开更多
Myocardial perfusion imaging(MPI),which uses single-photon emission computed tomography(SPECT),is a well-known estimating tool for medical diagnosis,employing the classification of images to show situations in coronar...Myocardial perfusion imaging(MPI),which uses single-photon emission computed tomography(SPECT),is a well-known estimating tool for medical diagnosis,employing the classification of images to show situations in coronary artery disease(CAD).The automatic classification of SPECT images for different techniques has achieved near-optimal accuracy when using convolutional neural networks(CNNs).This paper uses a SPECT classification framework with three steps:1)Image denoising,2)Attenuation correction,and 3)Image classification.Image denoising is done by a U-Net architecture that ensures effective image denoising.Attenuation correction is implemented by a convolution neural network model that can remove the attenuation that affects the feature extraction process of classification.Finally,a novel multi-scale diluted convolution(MSDC)network is proposed.It merges the features extracted in different scales and makes the model learn the features more efficiently.Three scales of filters with size 3×3 are used to extract features.All three steps are compared with state-of-the-art methods.The proposed denoising architecture ensures a high-quality image with the highest peak signal-to-noise ratio(PSNR)value of 39.7.The proposed classification method is compared with the five different CNN models,and the proposed method ensures better classification with an accuracy of 96%,precision of 87%,sensitivity of 87%,specificity of 89%,and F1-score of 87%.To demonstrate the importance of preprocessing,the classification model was analyzed without denoising and attenuation correction.展开更多
The cleanliness of seed cotton plays a critical role in the pre-treatment of cotton textiles,and the removal of impurity during the harvesting process directly determines the quality and market value of cotton textile...The cleanliness of seed cotton plays a critical role in the pre-treatment of cotton textiles,and the removal of impurity during the harvesting process directly determines the quality and market value of cotton textiles.By fusing band combination optimization with deep learning,this study aims to achieve more efficient and accurate detection of film impurities in seed cotton on the production line.By applying hyperspectral imaging and a one-dimensional deep learning algorithm,we detect and classify impurities in seed cotton after harvest.The main categories detected include pure cotton,conveyor belt,film covering seed cotton,and film adhered to the conveyor belt.The proposed method achieves an impurity detection rate of 99.698%.To further ensure the feasibility and practical application potential of this strategy,we compare our results against existing mainstream methods.In addition,the model shows excellent recognition performance on pseudo-color images of real samples.With a processing time of 11.764μs per pixel from experimental data,it shows a much improved speed requirement while maintaining the accuracy of real production lines.This strategy provides an accurate and efficient method for removing impurities during cotton processing.展开更多
This paper proposes a solution to localization and classification of rice grains in an image.All existing related works rely on conventional based machine learning approaches.However,those techniques do not do well fo...This paper proposes a solution to localization and classification of rice grains in an image.All existing related works rely on conventional based machine learning approaches.However,those techniques do not do well for the problem designed in this paper,due to the high similarities between different types of rice grains.The deep learning based solution is developed in the proposed solution.It contains pre-processing steps of data annotation using the watershed algorithm,auto-alignment using the major axis orientation,and image enhancement using the contrast-limited adaptive histogram equalization(CLAHE)technique.Then,the mask region-based convolutional neural networks(R-CNN)is trained to localize and classify rice grains in an input image.The performance is enhanced by using the transfer learning and the dropout regularization for overfitting prevention.The proposed method is validated using many scenarios of experiments,reported in the forms of mean average precision(mAP)and a confusion matrix.It achieves above 80%mAP for main scenarios in the experiments.It is also shown to perform outstanding,when compared to human experts.展开更多
Diagnosing cardiac diseases relies heavily on electrocardiogram(ECG)analysis,but detecting myocardial infarction-related arrhythmias remains challenging due to irregular heartbeats and signal variations.Despite advanc...Diagnosing cardiac diseases relies heavily on electrocardiogram(ECG)analysis,but detecting myocardial infarction-related arrhythmias remains challenging due to irregular heartbeats and signal variations.Despite advancements in machine learning,achieving both high accuracy and low computational cost for arrhythmia classification remains a critical issue.Computer-aided diagnosis systems can play a key role in early detection,reducing mortality rates associated with cardiac disorders.This study proposes a fully automated approach for ECG arrhythmia classification using deep learning and machine learning techniques to improve diagnostic accuracy while minimizing processing time.The methodology consists of three stages:1)preprocessing,where ECG signals undergo noise reduction and feature extraction;2)feature Identification,where deep convolutional neural network(CNN)blocks,combined with data augmentation and transfer learning,extract key parameters;3)classification,where a hybrid CNN-SVM model is employed for arrhythmia recognition.CNN-extracted features were fed into a binary support vector machine(SVM)classifier,and model performance was assessed using five-fold cross-validation.Experimental findings demonstrated that the CNN2 model achieved 85.52%accuracy,while the hybrid CNN2-SVM approach significantly improved accuracy to 97.33%,outperforming conventional methods.This model enhances classification efficiency while reducing computational complexity.The proposed approach bridges the gap between accuracy and processing speed in ECG arrhythmia classification,offering a promising solution for real-time clinical applications.Its superior performance compared to nonlinear classifiers highlights its potential for improving automated cardiac diagnosis.展开更多
In the era of precision medicine,the classification of diabetes mellitus has evolved beyond the traditional categories.Various classification methods now account for a multitude of factors,including variations in spec...In the era of precision medicine,the classification of diabetes mellitus has evolved beyond the traditional categories.Various classification methods now account for a multitude of factors,including variations in specific genes,type ofβ-cell impairment,degree of insulin resistance,and clinical characteristics of metabolic profiles.Improved classification methods enable healthcare providers to formulate blood glucose management strategies more precisely.Applying these updated classification systems,will assist clinicians in further optimising treatment plans,including targeted drug therapies,personalized dietary advice,and specific exercise plans.Ultimately,this will facilitate stricter blood glucose control,minimize the risks of hypoglycaemia and hyperglycaemia,and reduce long-term complications associated with diabetes.展开更多
Lunar impact glasses have been identified as crucial indicators of geochemical information regarding their source regions. Impact glasses can be categorized as either local or exotic. Those preserving geochemical sign...Lunar impact glasses have been identified as crucial indicators of geochemical information regarding their source regions. Impact glasses can be categorized as either local or exotic. Those preserving geochemical signatures matching local lithologies (e.g., mare basalts or their single minerals) or regolith bulk soil compositions are classified as “local”. Otherwise, they could be defined as “exotic”. The analysis of exotic glasses provides the opportunity to explore previously unsampled lunar areas. This study focuses on the identification of exotic glasses within the Chang’e-5 (CE-5) soil sample by analyzing the trace elements of 28 impact glasses with distinct major element compositions in comparison with the CE-5 bulk soil. However, the results indicate that 18 of the analyzed glasses exhibit trace element compositions comparable to those of the local CE-5 materials. In particular, some of them could match the local single mineral component in major and trace elements, suggesting a local origin. Therefore, it is recommended that the investigation be expanded from using major elements to including nonvolatile trace elements, with a view to enhancing our understanding on the provenance of lunar impact glasses. To achieve a more accurate identification of exotic glasses within the CE-5 soil sample, a novel classification plot of Mg# versus La is proposed. The remaining 10 glasses, which exhibit diverse trace element variations, were identified as exotic. A comparative analysis of their chemical characteristics with remote sensing data indicates that they may have originated from the Aristarchus, Mairan, Sharp, or Pythagoras craters. This study elucidates the classification and possible provenance of exotic materials within the CE-5 soil sample, thereby providing constraints for the enhanced identification of local and exotic components at the CE-5 landing site.展开更多
基金funded by the China National Space Administration(KJSP2023020105)supported by the National Key R&D Program of China(Grant No.2023YFA1608100)+2 种基金the NSFC(Grant No.62227901)the Minor Planet Foundationsupported by the Egyptian Science,Technology&Innovation Funding Authority(STDF)under Grant No.48102.
文摘Near-Earth objects are important not only in studying the early formation of the Solar System,but also because they pose a serious hazard to humanity when they make close approaches to the Earth.Study of their physical properties can provide useful information on their origin,evolution,and hazard to human beings.However,it remains challenging to investigate small,newly discovered,near-Earth objects because of our limited observational window.This investigation seeks to determine the visible colors of near-Earth asteroids(NEAs),perform an initial taxonomic classification based on visible colors and analyze possible correlations between the distribution of taxonomic classification and asteroid size or orbital parameters.Observations were performed in the broadband BVRI Johnson−Cousins photometric system,applied to images from the Yaoan High Precision Telescope and the 1.88 m telescope at the Kottamia Astronomical Observatory.We present new photometric observations of 84 near-Earth asteroids,and classify 80 of them taxonomically,based on their photometric colors.We find that nearly half(46.3%)of the objects in our sample can be classified as S-complex,26.3%as C-complex,6%as D-complex,and 15.0%as X-complex;the remaining belong to the A-or V-types.Additionally,we identify three P-type NEAs in our sample,according to the Tholen scheme.The fractional abundances of the C/X-complex members with absolute magnitude H≥17.0 were more than twice as large as those with H<17.0.However,the fractions of C-and S-complex members with diameters≤1 km and>1 km are nearly equal,while X-complex members tend to have sub-kilometer diameters.In our sample,the C/D-complex objects are predominant among those with a Jovian Tisserand parameter of T_(J)<3.1.These bodies could have a cometary origin.C-and S-complex members account for a considerable proportion of the asteroids that are potentially hazardous.
基金funded by the National Key Research and Development Program of China(Grant No.2024YFE0209000)the NSFC(Grant No.U23B2019).
文摘Graph Neural Networks(GNNs)have proven highly effective for graph classification across diverse fields such as social networks,bioinformatics,and finance,due to their capability to learn complex graph structures.However,despite their success,GNNs remain vulnerable to adversarial attacks that can significantly degrade their classification accuracy.Existing adversarial attack strategies primarily rely on label information to guide the attacks,which limits their applicability in scenarios where such information is scarce or unavailable.This paper introduces an innovative unsupervised attack method for graph classification,which operates without relying on label information,thereby enhancing its applicability in a broad range of scenarios.Specifically,our method first leverages a graph contrastive learning loss to learn high-quality graph embeddings by comparing different stochastic augmented views of the graphs.To effectively perturb the graphs,we then introduce an implicit estimator that measures the impact of various modifications on graph structures.The proposed strategy identifies and flips edges with the top-K highest scores,determined by the estimator,to maximize the degradation of the model’s performance.In addition,to defend against such attack,we propose a lightweight regularization-based defense mechanism that is specifically tailored to mitigate the structural perturbations introduced by our attack strategy.It enhances model robustness by enforcing embedding consistency and edge-level smoothness during training.We conduct experiments on six public TU graph classification datasets:NCI1,NCI109,Mutagenicity,ENZYMES,COLLAB,and DBLP_v1,to evaluate the effectiveness of our attack and defense strategies.Under an attack budget of 3,the maximum reduction in model accuracy reaches 6.67%on the Graph Convolutional Network(GCN)and 11.67%on the Graph Attention Network(GAT)across different datasets,indicating that our unsupervised method induces degradation comparable to state-of-the-art supervised attacks.Meanwhile,our defense achieves the highest accuracy recovery of 3.89%(GCN)and 5.00%(GAT),demonstrating improved robustness against structural perturbations.
基金funded by the Deanship of Graduate Studies and Scientific Research at Jouf University under grant No.(DGSSR-2025-02-01296).
文摘Skin diseases affect millions worldwide.Early detection is key to preventing disfigurement,lifelong disability,or death.Dermoscopic images acquired in primary-care settings show high intra-class visual similarity and severe class imbalance,and occasional imaging artifacts can create ambiguity for state-of-the-art convolutional neural networks(CNNs).We frame skin lesion recognition as graph-based reasoning and,to ensure fair evaluation and avoid data leakage,adopt a strict lesion-level partitioning strategy.Each image is first over-segmented using SLIC(Simple Linear Iterative Clustering)to produce perceptually homogeneous superpixels.These superpixels form the nodes of a region-adjacency graph whose edges encode spatial continuity.Node attributes are 1280-dimensional embeddings extracted with a lightweight yet expressive EfficientNet-B0 backbone,providing strong representational power at modest computational cost.The resulting graphs are processed by a five-layer Graph Attention Network(GAT)that learns to weight inter-node relationships dynamically and aggregates multi-hop context before classifying lesions into seven classes with a log-softmax output.Extensive experiments on the DermaMNIST benchmark show the proposed pipeline achieves 88.35%accuracy and 98.04%AUC,outperforming contemporary CNNs,AutoML approaches,and alternative graph neural networks.An ablation study indicates EfficientNet-B0 produces superior node descriptors compared with ResNet-18 and DenseNet,and that roughly five GAT layers strike a good balance between being too shallow and over-deep while avoiding oversmoothing.The method requires no data augmentation or external metadata,making it a drop-in upgrade for clinical computer-aided diagnosis systems.
文摘This systematic review aims to comprehensively examine and compare deep learning methods for brain tumor segmentation and classification using MRI and other imaging modalities,focusing on recent trends from 2022 to 2025.The primary objective is to evaluate methodological advancements,model performance,dataset usage,and existing challenges in developing clinically robust AI systems.We included peer-reviewed journal articles and highimpact conference papers published between 2022 and 2025,written in English,that proposed or evaluated deep learning methods for brain tumor segmentation and/or classification.Excluded were non-open-access publications,books,and non-English articles.A structured search was conducted across Scopus,Google Scholar,Wiley,and Taylor&Francis,with the last search performed in August 2025.Risk of bias was not formally quantified but considered during full-text screening based on dataset diversity,validation methods,and availability of performance metrics.We used narrative synthesis and tabular benchmarking to compare performance metrics(e.g.,accuracy,Dice score)across model types(CNN,Transformer,Hybrid),imaging modalities,and datasets.A total of 49 studies were included(43 journal articles and 6 conference papers).These studies spanned over 9 public datasets(e.g.,BraTS,Figshare,REMBRANDT,MOLAB)and utilized a range of imaging modalities,predominantly MRI.Hybrid models,especially ResViT and UNetFormer,consistently achieved high performance,with classification accuracy exceeding 98%and segmentation Dice scores above 0.90 across multiple studies.Transformers and hybrid architectures showed increasing adoption post2023.Many studies lacked external validation and were evaluated only on a few benchmark datasets,raising concerns about generalizability and dataset bias.Few studies addressed clinical interpretability or uncertainty quantification.Despite promising results,particularly for hybrid deep learning models,widespread clinical adoption remains limited due to lack of validation,interpretability concerns,and real-world deployment barriers.
文摘Honeycombing Lung(HCL)is a chronic lung condition marked by advanced fibrosis,resulting in enlarged air spaces with thick fibrotic walls,which are visible on Computed Tomography(CT)scans.Differentiating between normal lung tissue,honeycombing lungs,and Ground Glass Opacity(GGO)in CT images is often challenging for radiologists and may lead to misinterpretations.Although earlier studies have proposed models to detect and classify HCL,many faced limitations such as high computational demands,lower accuracy,and difficulty distinguishing between HCL and GGO.CT images are highly effective for lung classification due to their high resolution,3D visualization,and sensitivity to tissue density variations.This study introduces Honeycombing Lungs Network(HCL Net),a novel classification algorithm inspired by ResNet50V2 and enhanced to overcome the shortcomings of previous approaches.HCL Net incorporates additional residual blocks,refined preprocessing techniques,and selective parameter tuning to improve classification performance.The dataset,sourced from the University Malaya Medical Centre(UMMC)and verified by expert radiologists,consists of CT images of normal,honeycombing,and GGO lungs.Experimental evaluations across five assessments demonstrated that HCL Net achieved an outstanding classification accuracy of approximately 99.97%.It also recorded strong performance in other metrics,achieving 93%precision,100%sensitivity,89%specificity,and an AUC-ROC score of 97%.Comparative analysis with baseline feature engineering methods confirmed the superior efficacy of HCL Net.The model significantly reduces misclassification,particularly between honeycombing and GGO lungs,enhancing diagnostic precision and reliability in lung image analysis.
基金funded by the Directorate of Research and Community Service,Directorate General of Research and Development,Ministry of Higher Education,Science and Technologyin accordance with the Implementation Contract for the Operational Assistance Program for State Universities,Research Program Number:109/C3/DT.05.00/PL/2025.
文摘Sudden wildfires cause significant global ecological damage.While satellite imagery has advanced early fire detection and mitigation,image-based systems face limitations including high false alarm rates,visual obstructions,and substantial computational demands,especially in complex forest terrains.To address these challenges,this study proposes a novel forest fire detection model utilizing audio classification and machine learning.We developed an audio-based pipeline using real-world environmental sound recordings.Sounds were converted into Mel-spectrograms and classified via a Convolutional Neural Network(CNN),enabling the capture of distinctive fire acoustic signatures(e.g.,crackling,roaring)that are minimally impacted by visual or weather conditions.Internet of Things(IoT)sound sensors were crucial for generating complex environmental parameters to optimize feature extraction.The CNN model achieved high performance in stratified 5-fold cross-validation(92.4%±1.6 accuracy,91.2%±1.8 F1-score)and on test data(94.93%accuracy,93.04%F1-score),with 98.44%precision and 88.32%recall,demonstrating reliability across environmental conditions.These results indicate that the audio-based approach not only improves detection reliability but also markedly reduces computational overhead compared to traditional image-based methods.The findings suggest that acoustic sensing integrated with machine learning offers a powerful,low-cost,and efficient solution for real-time forest fire monitoring in complex,dynamic environments.
基金funded by the Deanship of Graduate Studies and Scientific Research at Jouf University under grant No.(DGSSR-2025-02-01295).
文摘Alzheimer’s Disease(AD)is a progressive neurodegenerative disorder that significantly affects cognitive function,making early and accurate diagnosis essential.Traditional Deep Learning(DL)-based approaches often struggle with low-contrast MRI images,class imbalance,and suboptimal feature extraction.This paper develops a Hybrid DL system that unites MobileNetV2 with adaptive classification methods to boost Alzheimer’s diagnosis by processing MRI scans.Image enhancement is done using Contrast-Limited Adaptive Histogram Equalization(CLAHE)and Enhanced Super-Resolution Generative Adversarial Networks(ESRGAN).A classification robustness enhancement system integrates class weighting techniques and a Matthews Correlation Coefficient(MCC)-based evaluation method into the design.The trained and validated model gives a 98.88%accuracy rate and 0.9614 MCC score.We also performed a 10-fold cross-validation experiment with an average accuracy of 96.52%(±1.51),a loss of 0.1671,and an MCC score of 0.9429 across folds.The proposed framework outperforms the state-of-the-art models with a 98%weighted F1-score while decreasing misdiagnosis results for every AD stage.The model demonstrates apparent separation abilities between AD progression stages according to the results of the confusion matrix analysis.These results validate the effectiveness of hybrid DL models with adaptive preprocessing for early and reliable Alzheimer’s diagnosis,contributing to improved computer-aided diagnosis(CAD)systems in clinical practice.
基金supported by Hubei Provincial Natural Science Foundation of China(grants 2023AFB208)the Chinese Primary Health Care Foundation(Grant No.cphcf-2022-222)+2 种基金2025 Hubei Provincial Natural Science Foundation Innovation and Development Joint Fund Project:(JCZRLH202500457)Shanghai Foundation for Anti-Cancer&Cancer Prevention Development Phase II Exploration Oncology Research Fund Project:“Study on the Mechanism of ANO9-Mediated Cetuximab Resistance in Head and Neck Squamous Cell Carcinoma”(No Grant Number)Qingdao Sheci Public Welfare Relief Center Pan-Cancer Treatment Research Fund Project:(QD-HN30008).
文摘Objective:The World Health Organization(WHO)grading based on histopathology cannot always accurately predict tumor behavior of meningiomas.To overcome the limitations of the WHO grading,the study aims to propose a novel oxidative stress-based molecular classification for WHO grade 2/3 meningiomas.Methods:Differentially expressed oxidative stress-related genes were analyzed between 86 WHO grade 1(low grade)meningiomas and 99 grade 2/3(high grade)meningiomas.An oxidative stress-based molecular classification was developed in high-grade meningiomas through consensus clustering analysis.Immune microenvironment features,responses to immunotherapy and chemotherapy,and targeted drugs were evaluated.Three machine learning models:logistic regression,support vector machine,and random forest,were built for differentiating the classification.Key oxidative stress-related geneswere verified in humanmeningeal cells(HMC)and two meningioma cells(CH-157MN and IOMMLee)via reverse transcription quantitative polymerase chain reaction(RT-qPCR)and western blot.After knockdown of Forkhead Box M1(FOXM1)or Prion Protein(PRNP),cell growth,migration,and reactive oxygen species(ROS)levels were measured through cell counting kit-8(CCK-8),transwell,and immunofluorescence,respectively.Results:We classified high-grade meningiomas into two oxidative stress-based clusters,termed cluster 1 and cluster 2.Cluster 1 exhibited higher infiltrations of immune and stromal cells and higher expression of classic immune checkpoints:Cluster of Differentiation 86(CD86),Programmed Cell Death 1(PDCD1),and Leukocyte-Associated Immunoglobulin-Like Receptor 1(LAIR1),indicating that cluster 1 meningiomas might respond to immunotherapy.Drug sensitivity was heterogeneous between the two clusters.Three classifiers were established,which could accurately differentiate this molecular classification.FOXM1 and PRNP were experimentally evidenced to be highly expressed inmeningioma cells,and their knockdown hindered cell growth and migration and triggered ROS accumulation.Conclusion:In summary,our findings established a novel oxidative stress-based molecular classification and identified potential treatment vulnerabilities in high-grade meningiomas,which might assist personalized clinical management.
基金jointly funded by the National Natural Science Foundation of China (Grant Nos. 41675023, 91337103, 91437101 and 41675029)the Scientific Research Projects of CAMS (Grant Nos. 2016Z005 and 2016LASW-B12)the Postgraduate Research & Practice Innovation Program of Jiangsu Province (Grant Nos. KYCX17 0880)
文摘A modified hydrometeor classification algorithm (HCA) is developed in this study for Chinese polarimetric radars. This algorithm is based on the U.S. operational HCA. Meanwhile, the methodology of statistics-based optimization is proposed including calibration checking, datasets selection, membership functions modification, computation thresholds modification, and effect verification. Zhuhai radar, the first operational polarimetric radar in South China, applies these procedures. The systematic bias of calibration is corrected, the reliability of radar measurements deteriorates when the signal-to-noise ratio is low, and correlation coefficient within the melting layer is usually lower than that of the U.S. WSR-88D radar. Through modification based on statistical analysis of polarimetric variables, the localized HCA especially for Zhuhai is obtained, and it performs well over a one-month test through comparison with sounding and surface observations. The algorithm is then utilized for analysis of a squall line process on 11 May 2014 and is found to provide reasonable details with respect to horizontal and vertical structures, and the HCA results---especially in the mixed rain-hail region--can reflect the life cycle of the squall line. In addition, the kinematic and microphysical processes of cloud evolution and the differences between radar- detected hail and surface observations are also analyzed. The results of this study provide evidence for the improvement of this HCA developed specifically for China.
文摘Urban tree species provide various essential ecosystem services in cities,such as regulating urban temperatures,reducing noise,capturing carbon,and mitigating the urban heat island effect.The quality of these services is influenced by species diversity,tree health,and the distribution and the composition of trees.Traditionally,data on urban trees has been collected through field surveys and manual interpretation of remote sensing images.In this study,we evaluated the effectiveness of multispectral airborne laser scanning(ALS)data in classifying 24 common urban roadside tree species in Espoo,Finland.Tree crown structure information,intensity features,and spectral data were used for classification.Eight different machine learning algorithms were tested,with the extra trees(ET)algorithm performing the best,achieving an overall accuracy of 71.7%using multispectral LiDAR data.This result highlights that integrating structural and spectral information within a single framework can improve the classification accuracy.Future research will focus on identifying the most important features for species classification and developing algorithms with greater efficiency and accuracy.
文摘A small scale red soil resources information system (RSRIS) with applied mathematical models wasdeveloped and applied in red soil resources (RSR) classification and evaluation, taking Zhejiang Province,a typical distribution area of red soil, as the study area. Computer-aided overlay was conducted to classifyRSR types. The evaluation was carried out by using three methods, i.e., index summation, square root ofindex multiplication and fuzzy comprehensive assessment, with almost identical results. The result of indexsummation could represent the basic qualitative condition of RSR, that of square root of index multiplicationreflected the real condition of RSR qualitative rank, while fuzzy comprehensive assessment could satisfactorilyhandle the relationship between the evaluation factors and the qualitative rank of RSR, and therefore it is afeasible method for RSR evaluation.
文摘Purpose:Interdisciplinary research has become a critical approach to addressing complex societal,economic,technological,and environmental challenges,driving innovation and integrating scientific knowledge.While interdisciplinarity indicators are widely used to evaluate research performance,the impact of classification granularity on these assessments remains underexplored.Design/methodology/approach:This study investigates how different levels of classification granularity-macro,meso,and micro-affect the evaluation of interdisciplinarity in research institutes.Using a dataset of 262 institutes from four major German non-university organizations(FHG,HGF,MPG,WGL)from 2018 to 2022,we examine inconsistencies in interdisciplinarity across levels,analyze ranking changes,and explore the influence of institutional fields and research focus(applied vs.basic).Findings:Our findings reveal significant inconsistencies in interdisciplinarity across classification levels,with rankings varying substantially.Notably,the Fraunhofer Society(FHG),which performs well at the macro level,experiences significant ranking declines at meso and micro levels.Normalizing interdisciplinarity by research field confirmed that these declines persist.The research focus of institutes,whether applied,basic,or mixed,does not significantly explain the observed ranking dynamics.Research limitations:This study has only considered the publication-based dimension of institutional interdisciplinarity and has not explored other aspects.Practical implications:The findings provide insights for policymakers,research managers,and scholars to better interpret interdisciplinarity metrics and support interdisciplinary research effectively.Originality/value:This study underscores the critical role of classification granularity in interdisciplinarity assessment and emphasizes the need for standardized approaches to ensure robust and fair evaluations.
文摘Support Vector Machines (SVM) is a new general machine-learning tool based on structural risk minimization principle. This characteristic is very signific ant for the fault diagnostics when the number of fault samples is limited. Considering that SVM theory is originally designed for a two-class classification, a hybrid SVM scheme is proposed for multi-fault classification of rotating machinery in our paper. Two SVM strategies, 1-v-1 (one versus one) and 1-v-r (one versus rest), are respectively adopted at different classification levels. At the parallel classification level, using l-v-1 strategy, the fault features extracted by various signal analysis methods are transferred into the multiple parallel SVM and the local classification results are obtained. At the serial classification level, these local results values are fused by one serial SVM based on 1-v-r strategy. The hybrid SVM scheme introduced in our paper not only generalizes the performance of signal binary SVMs but improves the precision and reliability of the fault classification results. The actually testing results show the availability suitability of this new method.
基金supported by King Saud University,Riyadh,Saudi Arabia,through the Researchers Supporting Project under Grant RSPD2025R697.
文摘Preservation of the crops depends on early and accurate detection of pests on crops as they cause several diseases decreasing crop production and quality. Several deep-learning techniques have been applied to overcome the issue of pest detection on crops. We have developed the YOLOCSP-PEST model for Pest localization and classification. With the Cross Stage Partial Network (CSPNET) backbone, the proposed model is a modified version of You Only Look Once Version 7 (YOLOv7) that is intended primarily for pest localization and classification. Our proposed model gives exceptionally good results under conditions that are very challenging for any other comparable models especially conditions where we have issues with the luminance and the orientation of the images. It helps farmers working out on their crops in distant areas to determine any infestation quickly and accurately on their crops which helps in the quality and quantity of the production yield. The model has been trained and tested on 2 datasets namely the IP102 data set and a local crop data set on both of which it has shown exceptional results. It gave us a mean average precision (mAP) of 88.40% along with a precision of 85.55% and a recall of 84.25% on the IP102 dataset meanwhile giving a mAP of 97.18% on the local data set along with a recall of 94.88% and a precision of 97.50%. These findings demonstrate that the proposed model is very effective in detecting real-life scenarios and can help in the production of crops improving the yield quality and quantity at the same time.
基金the Research Grant of Kwangwoon University in 2024.
文摘Myocardial perfusion imaging(MPI),which uses single-photon emission computed tomography(SPECT),is a well-known estimating tool for medical diagnosis,employing the classification of images to show situations in coronary artery disease(CAD).The automatic classification of SPECT images for different techniques has achieved near-optimal accuracy when using convolutional neural networks(CNNs).This paper uses a SPECT classification framework with three steps:1)Image denoising,2)Attenuation correction,and 3)Image classification.Image denoising is done by a U-Net architecture that ensures effective image denoising.Attenuation correction is implemented by a convolution neural network model that can remove the attenuation that affects the feature extraction process of classification.Finally,a novel multi-scale diluted convolution(MSDC)network is proposed.It merges the features extracted in different scales and makes the model learn the features more efficiently.Three scales of filters with size 3×3 are used to extract features.All three steps are compared with state-of-the-art methods.The proposed denoising architecture ensures a high-quality image with the highest peak signal-to-noise ratio(PSNR)value of 39.7.The proposed classification method is compared with the five different CNN models,and the proposed method ensures better classification with an accuracy of 96%,precision of 87%,sensitivity of 87%,specificity of 89%,and F1-score of 87%.To demonstrate the importance of preprocessing,the classification model was analyzed without denoising and attenuation correction.
基金supported in part by the Six Talent Peaks Project in Jiangsu Province under Grant 013040315in part by the China Textile Industry Federation Science and Technology Guidance Project under Grant 2017107+1 种基金in part by the National Natural Science Foundation of China under Grant 31570714in part by the China Scholarship Council under Grant 202108320290。
文摘The cleanliness of seed cotton plays a critical role in the pre-treatment of cotton textiles,and the removal of impurity during the harvesting process directly determines the quality and market value of cotton textiles.By fusing band combination optimization with deep learning,this study aims to achieve more efficient and accurate detection of film impurities in seed cotton on the production line.By applying hyperspectral imaging and a one-dimensional deep learning algorithm,we detect and classify impurities in seed cotton after harvest.The main categories detected include pure cotton,conveyor belt,film covering seed cotton,and film adhered to the conveyor belt.The proposed method achieves an impurity detection rate of 99.698%.To further ensure the feasibility and practical application potential of this strategy,we compare our results against existing mainstream methods.In addition,the model shows excellent recognition performance on pseudo-color images of real samples.With a processing time of 11.764μs per pixel from experimental data,it shows a much improved speed requirement while maintaining the accuracy of real production lines.This strategy provides an accurate and efficient method for removing impurities during cotton processing.
文摘This paper proposes a solution to localization and classification of rice grains in an image.All existing related works rely on conventional based machine learning approaches.However,those techniques do not do well for the problem designed in this paper,due to the high similarities between different types of rice grains.The deep learning based solution is developed in the proposed solution.It contains pre-processing steps of data annotation using the watershed algorithm,auto-alignment using the major axis orientation,and image enhancement using the contrast-limited adaptive histogram equalization(CLAHE)technique.Then,the mask region-based convolutional neural networks(R-CNN)is trained to localize and classify rice grains in an input image.The performance is enhanced by using the transfer learning and the dropout regularization for overfitting prevention.The proposed method is validated using many scenarios of experiments,reported in the forms of mean average precision(mAP)and a confusion matrix.It achieves above 80%mAP for main scenarios in the experiments.It is also shown to perform outstanding,when compared to human experts.
文摘Diagnosing cardiac diseases relies heavily on electrocardiogram(ECG)analysis,but detecting myocardial infarction-related arrhythmias remains challenging due to irregular heartbeats and signal variations.Despite advancements in machine learning,achieving both high accuracy and low computational cost for arrhythmia classification remains a critical issue.Computer-aided diagnosis systems can play a key role in early detection,reducing mortality rates associated with cardiac disorders.This study proposes a fully automated approach for ECG arrhythmia classification using deep learning and machine learning techniques to improve diagnostic accuracy while minimizing processing time.The methodology consists of three stages:1)preprocessing,where ECG signals undergo noise reduction and feature extraction;2)feature Identification,where deep convolutional neural network(CNN)blocks,combined with data augmentation and transfer learning,extract key parameters;3)classification,where a hybrid CNN-SVM model is employed for arrhythmia recognition.CNN-extracted features were fed into a binary support vector machine(SVM)classifier,and model performance was assessed using five-fold cross-validation.Experimental findings demonstrated that the CNN2 model achieved 85.52%accuracy,while the hybrid CNN2-SVM approach significantly improved accuracy to 97.33%,outperforming conventional methods.This model enhances classification efficiency while reducing computational complexity.The proposed approach bridges the gap between accuracy and processing speed in ECG arrhythmia classification,offering a promising solution for real-time clinical applications.Its superior performance compared to nonlinear classifiers highlights its potential for improving automated cardiac diagnosis.
文摘In the era of precision medicine,the classification of diabetes mellitus has evolved beyond the traditional categories.Various classification methods now account for a multitude of factors,including variations in specific genes,type ofβ-cell impairment,degree of insulin resistance,and clinical characteristics of metabolic profiles.Improved classification methods enable healthcare providers to formulate blood glucose management strategies more precisely.Applying these updated classification systems,will assist clinicians in further optimising treatment plans,including targeted drug therapies,personalized dietary advice,and specific exercise plans.Ultimately,this will facilitate stricter blood glucose control,minimize the risks of hypoglycaemia and hyperglycaemia,and reduce long-term complications associated with diabetes.
基金funded by the National Natural Science Foundation of China (Grant Nos. 42241103 and 62227901)the Key Research Program of the Institute of Geology and Geophysics, Chinese Academy of Sciences (Grant Nos. IGGCAS-202101 and IGGCAS-202401)
文摘Lunar impact glasses have been identified as crucial indicators of geochemical information regarding their source regions. Impact glasses can be categorized as either local or exotic. Those preserving geochemical signatures matching local lithologies (e.g., mare basalts or their single minerals) or regolith bulk soil compositions are classified as “local”. Otherwise, they could be defined as “exotic”. The analysis of exotic glasses provides the opportunity to explore previously unsampled lunar areas. This study focuses on the identification of exotic glasses within the Chang’e-5 (CE-5) soil sample by analyzing the trace elements of 28 impact glasses with distinct major element compositions in comparison with the CE-5 bulk soil. However, the results indicate that 18 of the analyzed glasses exhibit trace element compositions comparable to those of the local CE-5 materials. In particular, some of them could match the local single mineral component in major and trace elements, suggesting a local origin. Therefore, it is recommended that the investigation be expanded from using major elements to including nonvolatile trace elements, with a view to enhancing our understanding on the provenance of lunar impact glasses. To achieve a more accurate identification of exotic glasses within the CE-5 soil sample, a novel classification plot of Mg# versus La is proposed. The remaining 10 glasses, which exhibit diverse trace element variations, were identified as exotic. A comparative analysis of their chemical characteristics with remote sensing data indicates that they may have originated from the Aristarchus, Mairan, Sharp, or Pythagoras craters. This study elucidates the classification and possible provenance of exotic materials within the CE-5 soil sample, thereby providing constraints for the enhanced identification of local and exotic components at the CE-5 landing site.