Desertification has become a global threat and caused a crisis,especially in Middle Eastern countries,such as Saudi Arabia.Makkah is one of the most important cities in Saudi Arabia that needs to be protected from des...Desertification has become a global threat and caused a crisis,especially in Middle Eastern countries,such as Saudi Arabia.Makkah is one of the most important cities in Saudi Arabia that needs to be protected from desertification.The vegetation area in Makkah has been damaged because of desertification through wind,floods,overgrazing,and global climate change.The damage caused by desertification can be recovered provided urgent action is taken to prevent further degradation of the vegetation area.In this paper,we propose an automatic desertification detection system based on Deep Learning techniques.Aerial images are classified using Convolutional Neural Networks(CNN)to detect land state variation in real-time.CNNs have been widely used for computer vision applications,such as image classification,image segmentation,and quality enhancement.The proposed CNN model was trained and evaluated on the Arial Image Dataset(AID).Compared to state-of-the-art methods,the proposed model has better performance while being suitable for embedded implementation.It has achieved high efficiency with 96.47% accuracy.In light of the current research,we assert the appropriateness of the proposed CNN model in detecting desertification from aerial images.展开更多
We redesign the parameterized quantum circuit in the quantum deep neural network, construct a three-layer structure as the hidden layer, and then use classical optimization algorithms to train the parameterized quantu...We redesign the parameterized quantum circuit in the quantum deep neural network, construct a three-layer structure as the hidden layer, and then use classical optimization algorithms to train the parameterized quantum circuit, thereby propose a novel hybrid quantum deep neural network(HQDNN) used for image classification. After bilinear interpolation reduces the original image to a suitable size, an improved novel enhanced quantum representation(INEQR) is used to encode it into quantum states as the input of the HQDNN. Multi-layer parameterized quantum circuits are used as the main structure to implement feature extraction and classification. The output results of parameterized quantum circuits are converted into classical data through quantum measurements and then optimized on a classical computer. To verify the performance of the HQDNN, we conduct binary classification and three classification experiments on the MNIST(Modified National Institute of Standards and Technology) data set. In the first binary classification, the accuracy of 0 and 4 exceeds98%. Then we compare the performance of three classification with other algorithms, the results on two datasets show that the classification accuracy is higher than that of quantum deep neural network and general quantum convolutional neural network.展开更多
To study the scene classification in the Synthetic Aperture Radar (SAR) image, a novel method based on kernel estimate, with the Maxkov context and Dempster-Shafer evidence theory is proposed. Initially, a nonpaxame...To study the scene classification in the Synthetic Aperture Radar (SAR) image, a novel method based on kernel estimate, with the Maxkov context and Dempster-Shafer evidence theory is proposed. Initially, a nonpaxametric Probability Density Function (PDF) estimate method is introduced, to describe the scene of SAR images. And then under the Maxkov context, both the determinate PDF and the kernel estimate method axe adopted respectively, to form a primary classification. Next, the primary classification results are fused using the evidence theory in an unsupervised way to get the scene classification. Finally, a regularization step is used, in which an iterated maximum selecting approach is introduced to control the fragments and modify the errors of the classification. Use of the kernel estimate and evidence theory can describe the complicated scenes with little prior knowledge and eliminate the ambiguities of the primary classification results. Experimental results on real SAR images illustrate a rather impressive performance.展开更多
Skin cancer is among the most common malignancies worldwide,but its mortality burden is largely driven by aggressive subtypes such as melanoma,with outcomes varying across regions and healthcare settings.These variati...Skin cancer is among the most common malignancies worldwide,but its mortality burden is largely driven by aggressive subtypes such as melanoma,with outcomes varying across regions and healthcare settings.These variations emphasize the importance of reliable diagnostic technologies that support clinicians in detecting skin malignancies with higher accuracy.Traditional diagnostic methods often rely on subjective visual assessments,which can lead to misdiagnosis.This study addresses these challenges by developing HybridFusionNet,a novel model that integrates Convolutional Neural Networks(CNN)with 1D feature extraction techniques to enhance diagnostic accuracy.Utilizing two extensive datasets,BCN20000 and HAM10000,the methodology includes data preprocessing,application of Synthetic Minority Oversampling Technique combined with Edited Nearest Neighbors(SMOTEENN)for data balancing,and optimization of feature selection using the Tree-based Pipeline Optimization Tool(TPOT).The results demonstrate significant performance improvements over traditional CNN models,achieving an accuracy of 0.9693 on the BCN20000 dataset and 0.9909 on the HAM10000 dataset.The HybridFusionNet model not only outperforms conventionalmethods but also effectively addresses class imbalance.To enhance transparency,it integrates post-hoc explanation techniques such as LIME,which highlight the features influencing predictions.These findings highlight the potential of HybridFusionNet to support real-world applications,including physician-assist systems,teledermatology,and large-scale skin cancer screening programs.By improving diagnostic efficiency and enabling access to expert-level analysis,the modelmay enhance patient outcomes and foster greater trust in artificial intelligence(AI)-assisted clinical decision-making.展开更多
The multi-modal characteristics of mineral particles play a pivotal role in enhancing the classification accuracy,which is critical for obtaining a profound understanding of the Earth's composition and ensuring ef...The multi-modal characteristics of mineral particles play a pivotal role in enhancing the classification accuracy,which is critical for obtaining a profound understanding of the Earth's composition and ensuring effective exploitation utilization of its resources.However,the existing methods for classifying mineral particles do not fully utilize these multi-modal features,thereby limiting the classification accuracy.Furthermore,when conventional multi-modal image classification methods are applied to planepolarized and cross-polarized sequence images of mineral particles,they encounter issues such as information loss,misaligned features,and challenges in spatiotemporal feature extraction.To address these challenges,we propose a multi-modal mineral particle polarization image classification network(MMGC-Net)for precise mineral particle classification.Initially,MMGC-Net employs a two-dimensional(2D)backbone network with shared parameters to extract features from two types of polarized images to ensure feature alignment.Subsequently,a cross-polarized intra-modal feature fusion module is designed to refine the spatiotemporal features from the extracted features of the cross-polarized sequence images.Ultimately,the inter-modal feature fusion module integrates the two types of modal features to enhance the classification precision.Quantitative and qualitative experimental results indicate that when compared with the current state-of-the-art multi-modal image classification methods,MMGC-Net demonstrates marked superiority in terms of mineral particle multi-modal feature learning and four classification evaluation metrics.It also demonstrates better stability than the existing models.展开更多
Skin diseases affect millions worldwide.Early detection is key to preventing disfigurement,lifelong disability,or death.Dermoscopic images acquired in primary-care settings show high intra-class visual similarity and ...Skin diseases affect millions worldwide.Early detection is key to preventing disfigurement,lifelong disability,or death.Dermoscopic images acquired in primary-care settings show high intra-class visual similarity and severe class imbalance,and occasional imaging artifacts can create ambiguity for state-of-the-art convolutional neural networks(CNNs).We frame skin lesion recognition as graph-based reasoning and,to ensure fair evaluation and avoid data leakage,adopt a strict lesion-level partitioning strategy.Each image is first over-segmented using SLIC(Simple Linear Iterative Clustering)to produce perceptually homogeneous superpixels.These superpixels form the nodes of a region-adjacency graph whose edges encode spatial continuity.Node attributes are 1280-dimensional embeddings extracted with a lightweight yet expressive EfficientNet-B0 backbone,providing strong representational power at modest computational cost.The resulting graphs are processed by a five-layer Graph Attention Network(GAT)that learns to weight inter-node relationships dynamically and aggregates multi-hop context before classifying lesions into seven classes with a log-softmax output.Extensive experiments on the DermaMNIST benchmark show the proposed pipeline achieves 88.35%accuracy and 98.04%AUC,outperforming contemporary CNNs,AutoML approaches,and alternative graph neural networks.An ablation study indicates EfficientNet-B0 produces superior node descriptors compared with ResNet-18 and DenseNet,and that roughly five GAT layers strike a good balance between being too shallow and over-deep while avoiding oversmoothing.The method requires no data augmentation or external metadata,making it a drop-in upgrade for clinical computer-aided diagnosis systems.展开更多
AIM:To conduct a classification study of high myopic maculopathy(HMM)using limited datasets,including tessellated fundus,diffuse chorioretinal atrophy,patchy chorioretinal atrophy,and macular atrophy,and minimize anno...AIM:To conduct a classification study of high myopic maculopathy(HMM)using limited datasets,including tessellated fundus,diffuse chorioretinal atrophy,patchy chorioretinal atrophy,and macular atrophy,and minimize annotation costs,and to optimize the ALFA-Mix active learning algorithm and apply it to HMM classification.METHODS:The optimized ALFA-Mix algorithm(ALFAMix+)was compared with five algorithms,including ALFA-Mix.Four models,including Res Net18,were established.Each algorithm was combined with four models for experiments on the HMM dataset.Each experiment consisted of 20 active learning rounds,with 100 images selected per round.The algorithm was evaluated by comparing the number of rounds in which ALFA-Mix+outperformed other algorithms.Finally,this study employed six models,including Efficient Former,to classify HMM.The best-performing model among these models was selected as the baseline model and combined with the ALFA-Mix+algorithm to achieve satisfactor y classification results with a small dataset.RESULTS:ALFA-Mix+outperforms other algorithms with an average superiority of 16.6,14.75,16.8,and 16.7 rounds in terms of accuracy,sensitivity,specificity,and Kappa value,respectively.This study conducted experiments on classifying HMM using several advanced deep learning models with a complete training set of 4252 images.The Efficient Former achieved the best results with an accuracy,sensitivity,specificity,and Kappa value of 0.8821,0.8334,0.9693,and 0.8339,respectively.Therefore,by combining ALFA-Mix+with Efficient Former,this study achieved results with an accuracy,sensitivity,specificity,and Kappa value of 0.8964,0.8643,0.9721,and 0.8537,respectively.CONCLUSION:The ALFA-Mix+algorithm reduces the required samples without compromising accuracy.Compared to other algorithms,ALFA-Mix+outperforms in more rounds of experiments.It effectively selects valuable samples compared to other algorithms.In HMM classification,combining ALFA-Mix+with Efficient Former enhances model performance,further demonstrating the effectiveness of ALFA-Mix+.展开更多
Recently,the convolutional neural network(CNN)has been dom-inant in studies on interpreting remote sensing images(RSI).However,it appears that training optimization strategies have received less attention in relevant ...Recently,the convolutional neural network(CNN)has been dom-inant in studies on interpreting remote sensing images(RSI).However,it appears that training optimization strategies have received less attention in relevant research.To evaluate this problem,the author proposes a novel algo-rithm named the Fast Training CNN(FST-CNN).To verify the algorithm’s effectiveness,twenty methods,including six classic models and thirty archi-tectures from previous studies,are included in a performance comparison.The overall accuracy(OA)trained by the FST-CNN algorithm on the same model architecture and dataset is treated as an evaluation baseline.Results show that there is a maximal OA gap of 8.35%between the FST-CNN and those methods in the literature,which means a 10%margin in performance.Meanwhile,all those complex roadmaps,e.g.,deep feature fusion,model combination,model ensembles,and human feature engineering,are not as effective as expected.It reveals that there was systemic suboptimal perfor-mance in the previous studies.Most of the CNN-based methods proposed in the previous studies show a consistent mistake,which has made the model’s accuracy lower than its potential value.The most important reasons seem to be the inappropriate training strategy and the shift in data distribution introduced by data augmentation(DA).As a result,most of the performance evaluation was conducted based on an inaccurate,suboptimal,and unfair result.It has made most of the previous research findings questionable to some extent.However,all these confusing results also exactly demonstrate the effectiveness of FST-CNN.This novel algorithm is model-agnostic and can be employed on any image classification model to potentially boost performance.In addition,the results also show that a standardized training strategy is indeed very meaningful for the research tasks of the RSI-SC.展开更多
The results of the development of the new fast-speed method of classification images using a structural approach are presented.The method is based on the system of hierarchical features,based on the bitwise data distr...The results of the development of the new fast-speed method of classification images using a structural approach are presented.The method is based on the system of hierarchical features,based on the bitwise data distribution for the set of descriptors of image description.The article also proposes the use of the spatial data processing apparatus,which simplifies and accelerates the classification process.Experiments have shown that the time of calculation of the relevance for two descriptions according to their distributions is about 1000 times less than for the traditional voting procedure,for which the sets of descriptors are compared.The introduction of the system of hierarchical features allows to further reduce the calculation time by 2–3 times while ensuring high efficiency of classification.The noise immunity of the method to additive noise has been experimentally studied.According to the results of the research,the marginal degree of the hierarchy of features for reliable classification with the standard deviation of noise less than 30 is the 8-bit distribution.Computing costs increase proportionally with decreasing bit distribution.The method can be used for application tasks where object identification time is critical.展开更多
Total shoulder arthroplasty is a standard restorative procedure practiced by orthopedists to diagnose shoulder arthritis in which a prosthesis replaces the whole joint or a part of the joint.It is often challenging fo...Total shoulder arthroplasty is a standard restorative procedure practiced by orthopedists to diagnose shoulder arthritis in which a prosthesis replaces the whole joint or a part of the joint.It is often challenging for doctors to identify the exact model and manufacturer of the prosthesis when it is unknown.This paper proposes a transfer learning-based class imbalance-aware prosthesis detection method to detect the implant’s manufacturer automatically from shoulder X-ray images.The framework of the method proposes a novel training approach and a new set of batch-normalization,dropout,and fully convolutional layers in the head network.It employs cyclical learning rates and weighting-based loss calculation mechanism.These modifications aid in faster convergence,avoid local-minima stagnation,and remove the training bias caused by imbalanced dataset.The proposed method is evaluated using seven well-known pre-trained models of VGGNet,ResNet,and DenseNet families.Experimentation is performed on a shoulder implant benchmark dataset consisting of 597 shoulder X-ray images.The proposed method improves the classification performance of all pre-trained models by 10–12%.The DenseNet-201-based variant has achieved the highest classification accuracy of 89.5%,which is 10%higher than existing methods.Further,to validate and generalize the proposed method,the existing baseline dataset is supplemented to six classes,including samples of two more implant manufacturers.Experimental results have shown average accuracy of 86.7%for the extended dataset and show the preeminence of the proposed method.展开更多
The problem of image recognition in the computer vision systems is being studied.The results of the development of efficient classification methods,given the figure of processing speed,based on the analysis of the seg...The problem of image recognition in the computer vision systems is being studied.The results of the development of efficient classification methods,given the figure of processing speed,based on the analysis of the segment representation of the structural description in the form of a set of descriptors are provided.We propose three versions of the classifier according to the following principles:“object-etalon”,“object descriptor-etalon”and“vector description of the object-etalon”,which are not similar in level of integration of researched data analysis.The options for constructing clusters over the whole set of descriptions of the etalon database,separately for each of the etalons,as well as the optimal method to compare sets of segment centers for the etalons and object,are implemented.An experimental rating of the efficiency of the created classifiers in terms of productivity,processing time,and classification quality has been realized of the applied.The proposed methods classify the set of etalons without error.We have formed the inference about the efficiency of classification approaches based on segment centers.The time of image processing according to the developedmethods is hundreds of times less than according to the traditional one,without reducing the accuracy.展开更多
In blood or bone marrow,leukemia is a form of cancer.A person with leukemia has an expansion of white blood cells(WBCs).It primarily affects children and rarely affects adults.Treatment depends on the type of leukemia...In blood or bone marrow,leukemia is a form of cancer.A person with leukemia has an expansion of white blood cells(WBCs).It primarily affects children and rarely affects adults.Treatment depends on the type of leukemia and the extent to which cancer has established throughout the body.Identifying leukemia in the initial stage is vital to providing timely patient care.Medical image-analysis-related approaches grant safer,quicker,and less costly solutions while ignoring the difficulties of these invasive processes.It can be simple to generalize Computer vision(CV)-based and image-processing techniques and eradicate human error.Many researchers have implemented computer-aided diagnosticmethods andmachine learning(ML)for laboratory image analysis,hopefully overcoming the limitations of late leukemia detection and determining its subgroups.This study establishes a Marine Predators Algorithm with Deep Learning Leukemia Cancer Classification(MPADL-LCC)algorithm onMedical Images.The projectedMPADL-LCC system uses a bilateral filtering(BF)technique to pre-process medical images.The MPADL-LCC system uses Faster SqueezeNet withMarine Predators Algorithm(MPA)as a hyperparameter optimizer for feature extraction.Lastly,the denoising autoencoder(DAE)methodology can be executed to accurately detect and classify leukemia cancer.The hyperparameter tuning process using MPA helps enhance leukemia cancer classification performance.Simulation results are compared with other recent approaches concerning various measurements and the MPADL-LCC algorithm exhibits the best results over other recent approaches.展开更多
The performance of scene classification of satellite images strongly relies on the discriminative power of the low-level and mid-level feature representation. This paper presents a novel approach, named multi-level ma...The performance of scene classification of satellite images strongly relies on the discriminative power of the low-level and mid-level feature representation. This paper presents a novel approach, named multi-level max-margin analysis (M 3 DA) for semantic classification for high-resolution satellite images. In our M 3 DA model, the maximum entropy discrimination latent Dirichlet allocation (MedLDA) model is applied to learn the topic-level features first, and then based on a bag-of-words repre- sentation of low-level local image features, the large margin nearest neighbor (LMNN) classifier is used to optimize a multiple soft label composed of word-level features (generated by SVM classifier) and topic-level features. The categorization performances on 21-class land-use dataset have demonstrated that the proposed model in multi-level max-margin scheme can distinguish different categories of land-use scenes reasonably.展开更多
Today, many eye diseases jeopardize our everyday lives, such as Diabetic Retinopathy (DR), Age-related Macular Degeneration (AMD), and Glaucoma.Glaucoma is an incurable and unavoidable eye disease that damages the vis...Today, many eye diseases jeopardize our everyday lives, such as Diabetic Retinopathy (DR), Age-related Macular Degeneration (AMD), and Glaucoma.Glaucoma is an incurable and unavoidable eye disease that damages the vision ofoptic nerves and quality of life. Classification of Glaucoma has been an active fieldof research for the past ten years. Several approaches for Glaucoma classification areestablished, beginning with conventional segmentation methods and feature-extraction to deep-learning techniques such as Convolution Neural Networks (CNN). Incontrast, CNN classifies the input images directly using tuned parameters of convolution and pooling layers by extracting features. But, the volume of training datasetsdetermines the performance of the CNN;the model trained with small datasets,overfit issues arise. CNN has therefore developed with transfer learning. The primary aim of this study is to explore the potential of EfficientNet with transfer learning for the classification of Glaucoma. The performance of the current workcompares with other models, namely VGG16, InceptionV3, and Xception usingpublic datasets such as RIM-ONEV2 & V3, ORIGA, DRISHTI-GS1, HRF, andACRIMA. The dataset has split into training, validation, and testing with the ratioof 70:15:15. The assessment of the test dataset shows that the pre-trained EfficientNetB4 has achieved the highest performance value compared to other models listedabove. The proposed method achieved 99.38% accuracy and also better results forother metrics, such as sensitivity, specificity, precision, F1_score, Kappa score, andArea Under Curve (AUC) compared to other models.展开更多
A brain tumor is a mass of abnormal cells in the brain. Brain tumors can be benign (noncancerous) or malignant (cancerous). Conventional diagnosis of a brain tumor by the radiologist is done by examining a set of imag...A brain tumor is a mass of abnormal cells in the brain. Brain tumors can be benign (noncancerous) or malignant (cancerous). Conventional diagnosis of a brain tumor by the radiologist is done by examining a set of images produced by magnetic resonance imaging (MRI). Many computer-aided detection (CAD) systems have been developed in order to help the radiologists reach their goal of correctly classifying the MRI image. Convolutional neural networks (CNNs) have been widely used in the classification of medical images. This paper presents a novel CAD technique for the classification of brain tumors in MRI images. The proposed system extracts features from the brain MRI images by utilizing the strong energy compactness property exhibited by the Discrete Wavelet Transform (DWT). The Wavelet features are then applied to a CNN to classify the input MRI image. Experimental results indicate that the proposed approach outperforms other commonly used methods and gives an overall accuracy of 99.3%.展开更多
In the domain ofmedical imaging,the accurate detection and classification of brain tumors is very important.This study introduces an advanced method for identifying camouflaged brain tumors within images.Our proposed ...In the domain ofmedical imaging,the accurate detection and classification of brain tumors is very important.This study introduces an advanced method for identifying camouflaged brain tumors within images.Our proposed model consists of three steps:Feature extraction,feature fusion,and then classification.The core of this model revolves around a feature extraction framework that combines color-transformed images with deep learning techniques,using the ResNet50 Convolutional Neural Network(CNN)architecture.So the focus is to extract robust feature fromMRI images,particularly emphasizingweighted average features extracted fromthe first convolutional layer renowned for their discriminative power.To enhance model robustness,we introduced a novel feature fusion technique based on the Marine Predator Algorithm(MPA),inspired by the hunting behavior of marine predators and has shown promise in optimizing complex problems.The proposed methodology can accurately classify and detect brain tumors in camouflage images by combining the power of color transformations,deep learning,and feature fusion via MPA,and achieved an accuracy of 98.72%on a more complex dataset surpassing the existing state-of-the-art methods,highlighting the effectiveness of the proposed model.The importance of this research is in its potential to advance the field ofmedical image analysis,particularly in brain tumor diagnosis,where diagnoses early,and accurate classification are critical for improved patient results.展开更多
Machine learning(ML)is increasingly applied for medical image processing with appropriate learning paradigms.These applications include analyzing images of various organs,such as the brain,lung,eye,etc.,to identify sp...Machine learning(ML)is increasingly applied for medical image processing with appropriate learning paradigms.These applications include analyzing images of various organs,such as the brain,lung,eye,etc.,to identify specific flaws/diseases for diagnosis.The primary concern of ML applications is the precise selection of flexible image features for pattern detection and region classification.Most of the extracted image features are irrelevant and lead to an increase in computation time.Therefore,this article uses an analytical learning paradigm to design a Congruent Feature Selection Method to select the most relevant image features.This process trains the learning paradigm using similarity and correlation-based features over different textural intensities and pixel distributions.The similarity between the pixels over the various distribution patterns with high indexes is recommended for disease diagnosis.Later,the correlation based on intensity and distribution is analyzed to improve the feature selection congruency.Therefore,the more congruent pixels are sorted in the descending order of the selection,which identifies better regions than the distribution.Now,the learning paradigm is trained using intensity and region-based similarity to maximize the chances of selection.Therefore,the probability of feature selection,regardless of the textures and medical image patterns,is improved.This process enhances the performance of ML applications for different medical image processing.The proposed method improves the accuracy,precision,and training rate by 13.19%,10.69%,and 11.06%,respectively,compared to other models for the selected dataset.The mean error and selection time is also reduced by 12.56%and 13.56%,respectively,compared to the same models and dataset.展开更多
There is a widespread agreement that lung cancer is one of the deadliest types of cancer,affecting both women and men.As a result,detecting lung cancer at an early stage is crucial to create an accurate treatment plan...There is a widespread agreement that lung cancer is one of the deadliest types of cancer,affecting both women and men.As a result,detecting lung cancer at an early stage is crucial to create an accurate treatment plan and forecasting the reaction of the patient to the adopted treatment.For this reason,the development of convolutional neural networks(CNNs)for the task of lung cancer classification has recently seen a trend in attention.CNNs have great potential,but they need a lot of training data and struggle with input alterations.To address these limitations of CNNs,a novel machine-learning architecture of capsule networks has been presented,and it has the potential to completely transform the areas of deep learning.Capsule networks,which are the focus of this work,are interesting because they can withstand rotation and affine translation with relatively little training data.This research optimizes the performance of CapsNets by designing a new architecture that allows them to perform better on the challenge of lung cancer classification.The findings demonstrate that the proposed capsule network method outperforms CNNs on the lung cancer classification challenge.CapsNet with a single convolution layer and 32 features(CN-1-32),CapsNet with a single convolution layer and 64 features(CN-1-64),and CapsNet with a double convolution layer and 64 features(CN-2-64)are the three capsulel networks developed in this research for lung cancer classification.Lung nodules,both benign and malignant,are classified using these networks using CT images.The LIDC-IDRI database was utilized to assess the performance of those networks.Based on the testing results,CN-2-64 network performed better out of the three networks tested,with a specificity of 98.37%,sensitivity of 97.47%and an accuracy of 97.92%.展开更多
In this paper,we propose hierarchical attention dual network(DNet)for fine-grained image classification.The DNet can randomly select pairs of inputs from the dataset and compare the differences between them through hi...In this paper,we propose hierarchical attention dual network(DNet)for fine-grained image classification.The DNet can randomly select pairs of inputs from the dataset and compare the differences between them through hierarchical attention feature learning,which are used simultaneously to remove noise and retain salient features.In the loss function,it considers the losses of difference in paired images according to the intra-variance and inter-variance.In addition,we also collect the disaster scene dataset from remote sensing images and apply the proposed method to disaster scene classification,which contains complex scenes and multiple types of disasters.Compared to other methods,experimental results show that the DNet with hierarchical attention is robust to different datasets and performs better.展开更多
Accurate cloud classification plays a crucial role in aviation safety,climate monitoring,and localized weather forecasting.Current research has been focusing on machine learning techniques,particularly deep learning b...Accurate cloud classification plays a crucial role in aviation safety,climate monitoring,and localized weather forecasting.Current research has been focusing on machine learning techniques,particularly deep learning based model,for the types identification.However,traditional approaches such as convolutional neural networks(CNNs)encounter difficulties in capturing global contextual information.In addition,they are computationally expensive,which restricts their usability in resource-limited environments.To tackle these issues,we present the Cloud Vision Transformer(CloudViT),a lightweight model that integrates CNNs with Transformers.The integration enables an effective balance between local and global feature extraction.To be specific,CloudViT comprises two innovative modules:Feature Extraction(E_Module)and Downsampling(D_Module).These modules are able to significantly reduce the number of model parameters and computational complexity while maintaining translation invariance and enhancing contextual comprehension.Overall,the CloudViT includes 0.93×10^(6)parameters,which decreases more than ten times compared to the SOTA(State-of-the-Art)model CloudNet.Comprehensive evaluations conducted on the HBMCD and SWIMCAT datasets showcase the outstanding performance of CloudViT.It achieves classification accuracies of 98.45%and 100%,respectively.Moreover,the efficiency and scalability of CloudViT make it an ideal candidate for deployment inmobile cloud observation systems,enabling real-time cloud image classification.The proposed hybrid architecture of CloudViT offers a promising approach for advancing ground-based cloud image classification.It holds significant potential for both optimizing performance and facilitating practical deployment scenarios.展开更多
基金by Makkah Digital Gate Initiative under grant no.(MDP-IRI-3-2020).
文摘Desertification has become a global threat and caused a crisis,especially in Middle Eastern countries,such as Saudi Arabia.Makkah is one of the most important cities in Saudi Arabia that needs to be protected from desertification.The vegetation area in Makkah has been damaged because of desertification through wind,floods,overgrazing,and global climate change.The damage caused by desertification can be recovered provided urgent action is taken to prevent further degradation of the vegetation area.In this paper,we propose an automatic desertification detection system based on Deep Learning techniques.Aerial images are classified using Convolutional Neural Networks(CNN)to detect land state variation in real-time.CNNs have been widely used for computer vision applications,such as image classification,image segmentation,and quality enhancement.The proposed CNN model was trained and evaluated on the Arial Image Dataset(AID).Compared to state-of-the-art methods,the proposed model has better performance while being suitable for embedded implementation.It has achieved high efficiency with 96.47% accuracy.In light of the current research,we assert the appropriateness of the proposed CNN model in detecting desertification from aerial images.
基金Project supported by the Natural Science Foundation of Shandong Province,China (Grant No. ZR2021MF049)the Joint Fund of Natural Science Foundation of Shandong Province (Grant Nos. ZR2022LLZ012 and ZR2021LLZ001)。
文摘We redesign the parameterized quantum circuit in the quantum deep neural network, construct a three-layer structure as the hidden layer, and then use classical optimization algorithms to train the parameterized quantum circuit, thereby propose a novel hybrid quantum deep neural network(HQDNN) used for image classification. After bilinear interpolation reduces the original image to a suitable size, an improved novel enhanced quantum representation(INEQR) is used to encode it into quantum states as the input of the HQDNN. Multi-layer parameterized quantum circuits are used as the main structure to implement feature extraction and classification. The output results of parameterized quantum circuits are converted into classical data through quantum measurements and then optimized on a classical computer. To verify the performance of the HQDNN, we conduct binary classification and three classification experiments on the MNIST(Modified National Institute of Standards and Technology) data set. In the first binary classification, the accuracy of 0 and 4 exceeds98%. Then we compare the performance of three classification with other algorithms, the results on two datasets show that the classification accuracy is higher than that of quantum deep neural network and general quantum convolutional neural network.
基金the National Nature Science Foundation of China (60372057).
文摘To study the scene classification in the Synthetic Aperture Radar (SAR) image, a novel method based on kernel estimate, with the Maxkov context and Dempster-Shafer evidence theory is proposed. Initially, a nonpaxametric Probability Density Function (PDF) estimate method is introduced, to describe the scene of SAR images. And then under the Maxkov context, both the determinate PDF and the kernel estimate method axe adopted respectively, to form a primary classification. Next, the primary classification results are fused using the evidence theory in an unsupervised way to get the scene classification. Finally, a regularization step is used, in which an iterated maximum selecting approach is introduced to control the fragments and modify the errors of the classification. Use of the kernel estimate and evidence theory can describe the complicated scenes with little prior knowledge and eliminate the ambiguities of the primary classification results. Experimental results on real SAR images illustrate a rather impressive performance.
文摘Skin cancer is among the most common malignancies worldwide,but its mortality burden is largely driven by aggressive subtypes such as melanoma,with outcomes varying across regions and healthcare settings.These variations emphasize the importance of reliable diagnostic technologies that support clinicians in detecting skin malignancies with higher accuracy.Traditional diagnostic methods often rely on subjective visual assessments,which can lead to misdiagnosis.This study addresses these challenges by developing HybridFusionNet,a novel model that integrates Convolutional Neural Networks(CNN)with 1D feature extraction techniques to enhance diagnostic accuracy.Utilizing two extensive datasets,BCN20000 and HAM10000,the methodology includes data preprocessing,application of Synthetic Minority Oversampling Technique combined with Edited Nearest Neighbors(SMOTEENN)for data balancing,and optimization of feature selection using the Tree-based Pipeline Optimization Tool(TPOT).The results demonstrate significant performance improvements over traditional CNN models,achieving an accuracy of 0.9693 on the BCN20000 dataset and 0.9909 on the HAM10000 dataset.The HybridFusionNet model not only outperforms conventionalmethods but also effectively addresses class imbalance.To enhance transparency,it integrates post-hoc explanation techniques such as LIME,which highlight the features influencing predictions.These findings highlight the potential of HybridFusionNet to support real-world applications,including physician-assist systems,teledermatology,and large-scale skin cancer screening programs.By improving diagnostic efficiency and enabling access to expert-level analysis,the modelmay enhance patient outcomes and foster greater trust in artificial intelligence(AI)-assisted clinical decision-making.
基金supported by the National Natural Science Foundation of China(Grant Nos.62071315 and 62271336).
文摘The multi-modal characteristics of mineral particles play a pivotal role in enhancing the classification accuracy,which is critical for obtaining a profound understanding of the Earth's composition and ensuring effective exploitation utilization of its resources.However,the existing methods for classifying mineral particles do not fully utilize these multi-modal features,thereby limiting the classification accuracy.Furthermore,when conventional multi-modal image classification methods are applied to planepolarized and cross-polarized sequence images of mineral particles,they encounter issues such as information loss,misaligned features,and challenges in spatiotemporal feature extraction.To address these challenges,we propose a multi-modal mineral particle polarization image classification network(MMGC-Net)for precise mineral particle classification.Initially,MMGC-Net employs a two-dimensional(2D)backbone network with shared parameters to extract features from two types of polarized images to ensure feature alignment.Subsequently,a cross-polarized intra-modal feature fusion module is designed to refine the spatiotemporal features from the extracted features of the cross-polarized sequence images.Ultimately,the inter-modal feature fusion module integrates the two types of modal features to enhance the classification precision.Quantitative and qualitative experimental results indicate that when compared with the current state-of-the-art multi-modal image classification methods,MMGC-Net demonstrates marked superiority in terms of mineral particle multi-modal feature learning and four classification evaluation metrics.It also demonstrates better stability than the existing models.
基金funded by the Deanship of Graduate Studies and Scientific Research at Jouf University under grant No.(DGSSR-2025-02-01296).
文摘Skin diseases affect millions worldwide.Early detection is key to preventing disfigurement,lifelong disability,or death.Dermoscopic images acquired in primary-care settings show high intra-class visual similarity and severe class imbalance,and occasional imaging artifacts can create ambiguity for state-of-the-art convolutional neural networks(CNNs).We frame skin lesion recognition as graph-based reasoning and,to ensure fair evaluation and avoid data leakage,adopt a strict lesion-level partitioning strategy.Each image is first over-segmented using SLIC(Simple Linear Iterative Clustering)to produce perceptually homogeneous superpixels.These superpixels form the nodes of a region-adjacency graph whose edges encode spatial continuity.Node attributes are 1280-dimensional embeddings extracted with a lightweight yet expressive EfficientNet-B0 backbone,providing strong representational power at modest computational cost.The resulting graphs are processed by a five-layer Graph Attention Network(GAT)that learns to weight inter-node relationships dynamically and aggregates multi-hop context before classifying lesions into seven classes with a log-softmax output.Extensive experiments on the DermaMNIST benchmark show the proposed pipeline achieves 88.35%accuracy and 98.04%AUC,outperforming contemporary CNNs,AutoML approaches,and alternative graph neural networks.An ablation study indicates EfficientNet-B0 produces superior node descriptors compared with ResNet-18 and DenseNet,and that roughly five GAT layers strike a good balance between being too shallow and over-deep while avoiding oversmoothing.The method requires no data augmentation or external metadata,making it a drop-in upgrade for clinical computer-aided diagnosis systems.
基金Supported by the National Natural Science Foundation of China(No.61906066)the Zhejiang Provincial Philosophy and Social Science Planning Project(No.21NDJC021Z)+4 种基金Shenzhen Fund for Guangdong Provincial High-level Clinical Key Specialties(No.SZGSP014)Sanming Project of Medicine in Shenzhen(No.SZSM202011015)Shenzhen Science and Technology Planning Project(No.KCXFZ20211020163813019)the Natural Science Foundation of Ningbo City(No.202003N4072)the Postgraduate Research and Innovation Project of Huzhou University(No.2023KYCX52)。
文摘AIM:To conduct a classification study of high myopic maculopathy(HMM)using limited datasets,including tessellated fundus,diffuse chorioretinal atrophy,patchy chorioretinal atrophy,and macular atrophy,and minimize annotation costs,and to optimize the ALFA-Mix active learning algorithm and apply it to HMM classification.METHODS:The optimized ALFA-Mix algorithm(ALFAMix+)was compared with five algorithms,including ALFA-Mix.Four models,including Res Net18,were established.Each algorithm was combined with four models for experiments on the HMM dataset.Each experiment consisted of 20 active learning rounds,with 100 images selected per round.The algorithm was evaluated by comparing the number of rounds in which ALFA-Mix+outperformed other algorithms.Finally,this study employed six models,including Efficient Former,to classify HMM.The best-performing model among these models was selected as the baseline model and combined with the ALFA-Mix+algorithm to achieve satisfactor y classification results with a small dataset.RESULTS:ALFA-Mix+outperforms other algorithms with an average superiority of 16.6,14.75,16.8,and 16.7 rounds in terms of accuracy,sensitivity,specificity,and Kappa value,respectively.This study conducted experiments on classifying HMM using several advanced deep learning models with a complete training set of 4252 images.The Efficient Former achieved the best results with an accuracy,sensitivity,specificity,and Kappa value of 0.8821,0.8334,0.9693,and 0.8339,respectively.Therefore,by combining ALFA-Mix+with Efficient Former,this study achieved results with an accuracy,sensitivity,specificity,and Kappa value of 0.8964,0.8643,0.9721,and 0.8537,respectively.CONCLUSION:The ALFA-Mix+algorithm reduces the required samples without compromising accuracy.Compared to other algorithms,ALFA-Mix+outperforms in more rounds of experiments.It effectively selects valuable samples compared to other algorithms.In HMM classification,combining ALFA-Mix+with Efficient Former enhances model performance,further demonstrating the effectiveness of ALFA-Mix+.
基金Hunan University of Arts and Science provided doctoral research funding for this study (grant number 16BSQD23)Fund of Geography Subject ([2022]351)also provided funding.
文摘Recently,the convolutional neural network(CNN)has been dom-inant in studies on interpreting remote sensing images(RSI).However,it appears that training optimization strategies have received less attention in relevant research.To evaluate this problem,the author proposes a novel algo-rithm named the Fast Training CNN(FST-CNN).To verify the algorithm’s effectiveness,twenty methods,including six classic models and thirty archi-tectures from previous studies,are included in a performance comparison.The overall accuracy(OA)trained by the FST-CNN algorithm on the same model architecture and dataset is treated as an evaluation baseline.Results show that there is a maximal OA gap of 8.35%between the FST-CNN and those methods in the literature,which means a 10%margin in performance.Meanwhile,all those complex roadmaps,e.g.,deep feature fusion,model combination,model ensembles,and human feature engineering,are not as effective as expected.It reveals that there was systemic suboptimal perfor-mance in the previous studies.Most of the CNN-based methods proposed in the previous studies show a consistent mistake,which has made the model’s accuracy lower than its potential value.The most important reasons seem to be the inappropriate training strategy and the shift in data distribution introduced by data augmentation(DA).As a result,most of the performance evaluation was conducted based on an inaccurate,suboptimal,and unfair result.It has made most of the previous research findings questionable to some extent.However,all these confusing results also exactly demonstrate the effectiveness of FST-CNN.This novel algorithm is model-agnostic and can be employed on any image classification model to potentially boost performance.In addition,the results also show that a standardized training strategy is indeed very meaningful for the research tasks of the RSI-SC.
文摘The results of the development of the new fast-speed method of classification images using a structural approach are presented.The method is based on the system of hierarchical features,based on the bitwise data distribution for the set of descriptors of image description.The article also proposes the use of the spatial data processing apparatus,which simplifies and accelerates the classification process.Experiments have shown that the time of calculation of the relevance for two descriptions according to their distributions is about 1000 times less than for the traditional voting procedure,for which the sets of descriptors are compared.The introduction of the system of hierarchical features allows to further reduce the calculation time by 2–3 times while ensuring high efficiency of classification.The noise immunity of the method to additive noise has been experimentally studied.According to the results of the research,the marginal degree of the hierarchy of features for reliable classification with the standard deviation of noise less than 30 is the 8-bit distribution.Computing costs increase proportionally with decreasing bit distribution.The method can be used for application tasks where object identification time is critical.
文摘Total shoulder arthroplasty is a standard restorative procedure practiced by orthopedists to diagnose shoulder arthritis in which a prosthesis replaces the whole joint or a part of the joint.It is often challenging for doctors to identify the exact model and manufacturer of the prosthesis when it is unknown.This paper proposes a transfer learning-based class imbalance-aware prosthesis detection method to detect the implant’s manufacturer automatically from shoulder X-ray images.The framework of the method proposes a novel training approach and a new set of batch-normalization,dropout,and fully convolutional layers in the head network.It employs cyclical learning rates and weighting-based loss calculation mechanism.These modifications aid in faster convergence,avoid local-minima stagnation,and remove the training bias caused by imbalanced dataset.The proposed method is evaluated using seven well-known pre-trained models of VGGNet,ResNet,and DenseNet families.Experimentation is performed on a shoulder implant benchmark dataset consisting of 597 shoulder X-ray images.The proposed method improves the classification performance of all pre-trained models by 10–12%.The DenseNet-201-based variant has achieved the highest classification accuracy of 89.5%,which is 10%higher than existing methods.Further,to validate and generalize the proposed method,the existing baseline dataset is supplemented to six classes,including samples of two more implant manufacturers.Experimental results have shown average accuracy of 86.7%for the extended dataset and show the preeminence of the proposed method.
基金The authors received specific funding for this research-Project Number IF-PSAU-2021/01/18487.
文摘The problem of image recognition in the computer vision systems is being studied.The results of the development of efficient classification methods,given the figure of processing speed,based on the analysis of the segment representation of the structural description in the form of a set of descriptors are provided.We propose three versions of the classifier according to the following principles:“object-etalon”,“object descriptor-etalon”and“vector description of the object-etalon”,which are not similar in level of integration of researched data analysis.The options for constructing clusters over the whole set of descriptions of the etalon database,separately for each of the etalons,as well as the optimal method to compare sets of segment centers for the etalons and object,are implemented.An experimental rating of the efficiency of the created classifiers in terms of productivity,processing time,and classification quality has been realized of the applied.The proposed methods classify the set of etalons without error.We have formed the inference about the efficiency of classification approaches based on segment centers.The time of image processing according to the developedmethods is hundreds of times less than according to the traditional one,without reducing the accuracy.
基金funded by Researchers Supporting Program at King Saud University,(RSPD2024R809).
文摘In blood or bone marrow,leukemia is a form of cancer.A person with leukemia has an expansion of white blood cells(WBCs).It primarily affects children and rarely affects adults.Treatment depends on the type of leukemia and the extent to which cancer has established throughout the body.Identifying leukemia in the initial stage is vital to providing timely patient care.Medical image-analysis-related approaches grant safer,quicker,and less costly solutions while ignoring the difficulties of these invasive processes.It can be simple to generalize Computer vision(CV)-based and image-processing techniques and eradicate human error.Many researchers have implemented computer-aided diagnosticmethods andmachine learning(ML)for laboratory image analysis,hopefully overcoming the limitations of late leukemia detection and determining its subgroups.This study establishes a Marine Predators Algorithm with Deep Learning Leukemia Cancer Classification(MPADL-LCC)algorithm onMedical Images.The projectedMPADL-LCC system uses a bilateral filtering(BF)technique to pre-process medical images.The MPADL-LCC system uses Faster SqueezeNet withMarine Predators Algorithm(MPA)as a hyperparameter optimizer for feature extraction.Lastly,the denoising autoencoder(DAE)methodology can be executed to accurately detect and classify leukemia cancer.The hyperparameter tuning process using MPA helps enhance leukemia cancer classification performance.Simulation results are compared with other recent approaches concerning various measurements and the MPADL-LCC algorithm exhibits the best results over other recent approaches.
基金Supported by the Open Projects Program of National Laboratory of Pattern Recognition and the National Natural Science Foundation of China(91338113)
文摘The performance of scene classification of satellite images strongly relies on the discriminative power of the low-level and mid-level feature representation. This paper presents a novel approach, named multi-level max-margin analysis (M 3 DA) for semantic classification for high-resolution satellite images. In our M 3 DA model, the maximum entropy discrimination latent Dirichlet allocation (MedLDA) model is applied to learn the topic-level features first, and then based on a bag-of-words repre- sentation of low-level local image features, the large margin nearest neighbor (LMNN) classifier is used to optimize a multiple soft label composed of word-level features (generated by SVM classifier) and topic-level features. The categorization performances on 21-class land-use dataset have demonstrated that the proposed model in multi-level max-margin scheme can distinguish different categories of land-use scenes reasonably.
文摘Today, many eye diseases jeopardize our everyday lives, such as Diabetic Retinopathy (DR), Age-related Macular Degeneration (AMD), and Glaucoma.Glaucoma is an incurable and unavoidable eye disease that damages the vision ofoptic nerves and quality of life. Classification of Glaucoma has been an active fieldof research for the past ten years. Several approaches for Glaucoma classification areestablished, beginning with conventional segmentation methods and feature-extraction to deep-learning techniques such as Convolution Neural Networks (CNN). Incontrast, CNN classifies the input images directly using tuned parameters of convolution and pooling layers by extracting features. But, the volume of training datasetsdetermines the performance of the CNN;the model trained with small datasets,overfit issues arise. CNN has therefore developed with transfer learning. The primary aim of this study is to explore the potential of EfficientNet with transfer learning for the classification of Glaucoma. The performance of the current workcompares with other models, namely VGG16, InceptionV3, and Xception usingpublic datasets such as RIM-ONEV2 & V3, ORIGA, DRISHTI-GS1, HRF, andACRIMA. The dataset has split into training, validation, and testing with the ratioof 70:15:15. The assessment of the test dataset shows that the pre-trained EfficientNetB4 has achieved the highest performance value compared to other models listedabove. The proposed method achieved 99.38% accuracy and also better results forother metrics, such as sensitivity, specificity, precision, F1_score, Kappa score, andArea Under Curve (AUC) compared to other models.
文摘A brain tumor is a mass of abnormal cells in the brain. Brain tumors can be benign (noncancerous) or malignant (cancerous). Conventional diagnosis of a brain tumor by the radiologist is done by examining a set of images produced by magnetic resonance imaging (MRI). Many computer-aided detection (CAD) systems have been developed in order to help the radiologists reach their goal of correctly classifying the MRI image. Convolutional neural networks (CNNs) have been widely used in the classification of medical images. This paper presents a novel CAD technique for the classification of brain tumors in MRI images. The proposed system extracts features from the brain MRI images by utilizing the strong energy compactness property exhibited by the Discrete Wavelet Transform (DWT). The Wavelet features are then applied to a CNN to classify the input MRI image. Experimental results indicate that the proposed approach outperforms other commonly used methods and gives an overall accuracy of 99.3%.
基金funding from Prince Sattam bin Abdulaziz University through the Project Number(PSAU/2023/01/24607).
文摘In the domain ofmedical imaging,the accurate detection and classification of brain tumors is very important.This study introduces an advanced method for identifying camouflaged brain tumors within images.Our proposed model consists of three steps:Feature extraction,feature fusion,and then classification.The core of this model revolves around a feature extraction framework that combines color-transformed images with deep learning techniques,using the ResNet50 Convolutional Neural Network(CNN)architecture.So the focus is to extract robust feature fromMRI images,particularly emphasizingweighted average features extracted fromthe first convolutional layer renowned for their discriminative power.To enhance model robustness,we introduced a novel feature fusion technique based on the Marine Predator Algorithm(MPA),inspired by the hunting behavior of marine predators and has shown promise in optimizing complex problems.The proposed methodology can accurately classify and detect brain tumors in camouflage images by combining the power of color transformations,deep learning,and feature fusion via MPA,and achieved an accuracy of 98.72%on a more complex dataset surpassing the existing state-of-the-art methods,highlighting the effectiveness of the proposed model.The importance of this research is in its potential to advance the field ofmedical image analysis,particularly in brain tumor diagnosis,where diagnoses early,and accurate classification are critical for improved patient results.
基金the Deanship of Scientifc Research at King Khalid University for funding this work through large group Research Project under grant number RGP2/421/45supported via funding from Prince Sattam bin Abdulaziz University project number(PSAU/2024/R/1446)+1 种基金supported by theResearchers Supporting Project Number(UM-DSR-IG-2023-07)Almaarefa University,Riyadh,Saudi Arabia.supported by the Basic Science Research Program through the National Research Foundation of Korea(NRF)funded by the Ministry of Education(No.2021R1F1A1055408).
文摘Machine learning(ML)is increasingly applied for medical image processing with appropriate learning paradigms.These applications include analyzing images of various organs,such as the brain,lung,eye,etc.,to identify specific flaws/diseases for diagnosis.The primary concern of ML applications is the precise selection of flexible image features for pattern detection and region classification.Most of the extracted image features are irrelevant and lead to an increase in computation time.Therefore,this article uses an analytical learning paradigm to design a Congruent Feature Selection Method to select the most relevant image features.This process trains the learning paradigm using similarity and correlation-based features over different textural intensities and pixel distributions.The similarity between the pixels over the various distribution patterns with high indexes is recommended for disease diagnosis.Later,the correlation based on intensity and distribution is analyzed to improve the feature selection congruency.Therefore,the more congruent pixels are sorted in the descending order of the selection,which identifies better regions than the distribution.Now,the learning paradigm is trained using intensity and region-based similarity to maximize the chances of selection.Therefore,the probability of feature selection,regardless of the textures and medical image patterns,is improved.This process enhances the performance of ML applications for different medical image processing.The proposed method improves the accuracy,precision,and training rate by 13.19%,10.69%,and 11.06%,respectively,compared to other models for the selected dataset.The mean error and selection time is also reduced by 12.56%and 13.56%,respectively,compared to the same models and dataset.
文摘There is a widespread agreement that lung cancer is one of the deadliest types of cancer,affecting both women and men.As a result,detecting lung cancer at an early stage is crucial to create an accurate treatment plan and forecasting the reaction of the patient to the adopted treatment.For this reason,the development of convolutional neural networks(CNNs)for the task of lung cancer classification has recently seen a trend in attention.CNNs have great potential,but they need a lot of training data and struggle with input alterations.To address these limitations of CNNs,a novel machine-learning architecture of capsule networks has been presented,and it has the potential to completely transform the areas of deep learning.Capsule networks,which are the focus of this work,are interesting because they can withstand rotation and affine translation with relatively little training data.This research optimizes the performance of CapsNets by designing a new architecture that allows them to perform better on the challenge of lung cancer classification.The findings demonstrate that the proposed capsule network method outperforms CNNs on the lung cancer classification challenge.CapsNet with a single convolution layer and 32 features(CN-1-32),CapsNet with a single convolution layer and 64 features(CN-1-64),and CapsNet with a double convolution layer and 64 features(CN-2-64)are the three capsulel networks developed in this research for lung cancer classification.Lung nodules,both benign and malignant,are classified using these networks using CT images.The LIDC-IDRI database was utilized to assess the performance of those networks.Based on the testing results,CN-2-64 network performed better out of the three networks tested,with a specificity of 98.37%,sensitivity of 97.47%and an accuracy of 97.92%.
基金Supported by the National Natural Science Foundation of China(61601176)。
文摘In this paper,we propose hierarchical attention dual network(DNet)for fine-grained image classification.The DNet can randomly select pairs of inputs from the dataset and compare the differences between them through hierarchical attention feature learning,which are used simultaneously to remove noise and retain salient features.In the loss function,it considers the losses of difference in paired images according to the intra-variance and inter-variance.In addition,we also collect the disaster scene dataset from remote sensing images and apply the proposed method to disaster scene classification,which contains complex scenes and multiple types of disasters.Compared to other methods,experimental results show that the DNet with hierarchical attention is robust to different datasets and performs better.
基金funded by Innovation and Development Special Project of China Meteorological Administration(CXFZ2022J038,CXFZ2024J035)Sichuan Science and Technology Program(No.2023YFQ0072)+1 种基金Key Laboratory of Smart Earth(No.KF2023YB03-07)Automatic Software Generation and Intelligent Service Key Laboratory of Sichuan Province(CUIT-SAG202210).
文摘Accurate cloud classification plays a crucial role in aviation safety,climate monitoring,and localized weather forecasting.Current research has been focusing on machine learning techniques,particularly deep learning based model,for the types identification.However,traditional approaches such as convolutional neural networks(CNNs)encounter difficulties in capturing global contextual information.In addition,they are computationally expensive,which restricts their usability in resource-limited environments.To tackle these issues,we present the Cloud Vision Transformer(CloudViT),a lightweight model that integrates CNNs with Transformers.The integration enables an effective balance between local and global feature extraction.To be specific,CloudViT comprises two innovative modules:Feature Extraction(E_Module)and Downsampling(D_Module).These modules are able to significantly reduce the number of model parameters and computational complexity while maintaining translation invariance and enhancing contextual comprehension.Overall,the CloudViT includes 0.93×10^(6)parameters,which decreases more than ten times compared to the SOTA(State-of-the-Art)model CloudNet.Comprehensive evaluations conducted on the HBMCD and SWIMCAT datasets showcase the outstanding performance of CloudViT.It achieves classification accuracies of 98.45%and 100%,respectively.Moreover,the efficiency and scalability of CloudViT make it an ideal candidate for deployment inmobile cloud observation systems,enabling real-time cloud image classification.The proposed hybrid architecture of CloudViT offers a promising approach for advancing ground-based cloud image classification.It holds significant potential for both optimizing performance and facilitating practical deployment scenarios.