critical for guiding treatment and improving patient outcomes.Traditional molecular subtyping via immuno-histochemistry(IHC)test is invasive,time-consuming,and may not fully represent tumor heterogeneity.This study pr...critical for guiding treatment and improving patient outcomes.Traditional molecular subtyping via immuno-histochemistry(IHC)test is invasive,time-consuming,and may not fully represent tumor heterogeneity.This study proposes a non-invasive approach using digital mammography images and deep learning algorithm for classifying breast cancer molecular subtypes.Four pretrained models,including two Convolutional Neural Networks(MobileNet_V3_Large and VGG-16)and two Vision Transformers(ViT_B_16 and ViT_Base_Patch16_Clip_224)were fine-tuned to classify images into HER2-enriched,Luminal,Normal-like,and Triple Negative subtypes.Hyperparameter tuning,including learning rate adjustment and layer freezing strategies,was applied to optimize performance.Among the evaluated models,ViT_Base_Patch16_Clip_224 achieved the highest test accuracy(94.44%),with equally high precision,recall,and F1-score of 0.94,demonstrating excellent generalization.MobileNet_V3_Large achieved the same accuracy but showed less training stability.In contrast,VGG-16 recorded the lowest performance,indicating a limitation in its generalizability for this classification task.The study also highlighted the superior performance of the Vision Transformer models over CNNs,particularly due to their ability to capture global contextual features and the benefit of CLIP-based pretraining in ViT_Base_Patch16_Clip_224.To enhance clinical applicability,a graphical user interface(GUI)named“BCMS Dx”was developed for streamlined subtype prediction.Deep learning applied to mammography has proven effective for accurate and non-invasive molecular subtyping.The proposed Vision Transformer-based model and supporting GUI offer a promising direction for augmenting diagnostic workflows,minimizing the need for invasive procedures,and advancing personalized breast cancer management.展开更多
With network attack technology continuing to develop,traditional anomaly traffic detection methods that rely on feature engineering are increasingly insufficient in efficiency and accuracy.Graph Neural Network(GNN),a ...With network attack technology continuing to develop,traditional anomaly traffic detection methods that rely on feature engineering are increasingly insufficient in efficiency and accuracy.Graph Neural Network(GNN),a promising Deep Learning(DL)approach,has proven to be highly effective in identifying intricate patterns in graph⁃structured data and has already found wide applications in the field of network security.In this paper,we propose a hybrid Graph Convolutional Network(GCN)⁃GraphSAGE model for Anomaly Traffic Detection,namely HGS⁃ATD,which aims to improve the accuracy of anomaly traffic detection by leveraging edge feature learning to better capture the relationships between network entities.We validate the HGS⁃ATD model on four publicly available datasets,including NF⁃UNSW⁃NB15⁃v2.The experimental results show that the enhanced hybrid model is 5.71%to 10.25%higher than the baseline model in terms of accuracy,and the F1⁃score is 5.53%to 11.63%higher than the baseline model,proving that the model can effectively distinguish normal traffic from attack traffic and accurately classify various types of attacks.展开更多
Graph neural networks(GNN)have shown strong performance in node classification tasks,yet most existing models rely on uniform or shared weight aggregation,lacking flexibility in modeling the varying strength of relati...Graph neural networks(GNN)have shown strong performance in node classification tasks,yet most existing models rely on uniform or shared weight aggregation,lacking flexibility in modeling the varying strength of relationships among nodes.This paper proposes a novel graph coupling convolutional model that introduces an adaptive weighting mechanism to assign distinct importance to neighboring nodes based on their similarity to the central node.Unlike traditional methods,the proposed coupling strategy enhances the interpretability of node interactions while maintaining competitive classification performance.The model operates in the spatial domain,utilizing adjacency list structures for efficient convolution and addressing the limitations of weight sharing through a coupling-based similarity computation.Extensive experiments are conducted on five graph-structured datasets,including Cora,Citeseer,PubMed,Reddit,and BlogCatalog,as well as a custom topology dataset constructed from the Open University Learning Analytics Dataset(OULAD)educational platform.Results demonstrate that the proposed model achieves good classification accuracy,while significantly reducing training time through direct second-order neighbor fusion and data preprocessing.Moreover,analysis of neighborhood order reveals that considering third-order neighbors offers limited accuracy gains but introduces considerable computational overhead,confirming the efficiency of first-and second-order convolution in practical applications.Overall,the proposed graph coupling model offers a lightweight,interpretable,and effective framework for multi-label node classification in complex networks.展开更多
The development of brain-computer interfaces(BCI)based on motor imagery(MI)has greatly improved patients’quality of life with movement disorders.The classification of upper limb MI has been widely studied and applied...The development of brain-computer interfaces(BCI)based on motor imagery(MI)has greatly improved patients’quality of life with movement disorders.The classification of upper limb MI has been widely studied and applied in many fields,including rehabilitation.However,the physiological representations of left and right lower limb movements are too close and activated deep in the cerebral cortex,making it difficult to distinguish their features.Therefore,classifying lower limbs motor imagery is more challenging.In this study,we propose a feature extraction method based on functional connectivity,which utilizes phase-locked values to construct a functional connectivity matrix as the features of the left and right legs,which can effectively avoid the problem of physiological representations of the left and right lower limbs being too close to each other during movement.In addition,considering the topology and the temporal characteristics of the electroencephalogram(EEG),we designed a temporal-spatial convolutional network(TSGCN)to capture the spatiotemporal information for classification.Experimental results show that the accuracy of the proposed method is higher than that of existing methods,achieving an average classification accuracy of 73.58%on the internal dataset.Finally,this study explains the network mechanism of left and right foot MI from the perspective of graph theoretic features and demonstrates the feasibility of decoding lower limb MI.展开更多
With the rapid development of the Artificial Intelligence of Things(AIoT),convolutional neural networks(CNNs)have demonstrated potential and remarkable performance in AIoT applications due to their excellent performan...With the rapid development of the Artificial Intelligence of Things(AIoT),convolutional neural networks(CNNs)have demonstrated potential and remarkable performance in AIoT applications due to their excellent performance in various inference tasks.However,the users have concerns about privacy leakage for the use of AI and the performance and efficiency of computing on resource-constrained IoT edge devices.Therefore,this paper proposes an efficient privacy-preserving CNN framework(i.e.,EPPA)based on the Fully Homomorphic Encryption(FHE)scheme for AIoT application scenarios.In the plaintext domain,we verify schemes with different activation structures to determine the actual activation functions applicable to the corresponding ciphertext domain.Within the encryption domain,we integrate batch normalization(BN)into the convolutional layers to simplify the computation process.For nonlinear activation functions,we use composite polynomials for approximate calculation.Regarding the noise accumulation caused by homomorphic multiplication operations,we realize the refreshment of ciphertext noise through minimal“decryption-encryption”interactions,instead of adopting bootstrapping operations.Additionally,in practical implementation,we convert three-dimensional convolution into two-dimensional convolution to reduce the amount of computation in the encryption domain.Finally,we conduct extensive experiments on four IoT datasets,different CNN architectures,and two platforms with different resource configurations to evaluate the performance of EPPA in detail.展开更多
With the increasing complexity of industrial automation,planetary gearboxes play a vital role in largescale equipment transmission systems,directly impacting operational efficiency and safety.Traditional maintenance s...With the increasing complexity of industrial automation,planetary gearboxes play a vital role in largescale equipment transmission systems,directly impacting operational efficiency and safety.Traditional maintenance strategies often struggle to accurately predict the degradation process of equipment,leading to excessive maintenance costs or potential failure risks.However,existing prediction methods based on statistical models are difficult to adapt to nonlinear degradation processes.To address these challenges,this study proposes a novel condition-based maintenance framework for planetary gearboxes.A comprehensive full-lifecycle degradation experiment was conducted to collect raw vibration signals,which were then processed using a temporal convolutional network autoencoder with multi-scale perception capability to extract deep temporal degradation features,enabling the collaborative extraction of longperiod meshing frequencies and short-term impact features from the vibration signals.Kernel principal component analysis was employed to fuse and normalize these features,enhancing the characterization of degradation progression.A nonlinear Wiener process was used to model the degradation trajectory,with a threshold decay function introduced to dynamically adjust maintenance strategies,and model parameters optimized through maximum likelihood estimation.Meanwhile,the maintenance strategy was optimized to minimize costs per unit time,determining the optimal maintenance timing and preventive maintenance threshold.The comprehensive indicator of degradation trends extracted by this method reaches 0.756,which is 41.2%higher than that of traditional time-domain features;the dynamic threshold strategy reduces the maintenance cost per unit time to 55.56,which is 8.9%better than that of the static threshold optimization.Experimental results demonstrate significant reductions in maintenance costs while enhancing system reliability and safety.This study realizes the organic integration of deep learning and reliability theory in the maintenance of planetary gearboxes,provides an interpretable solution for the predictive maintenance of complex mechanical systems,and promotes the development of condition-based maintenance strategies for planetary gearboxes.展开更多
Parkinson’s disease(PD)is a debilitating neurological disorder affecting over 10 million people worldwide.PD classification models using voice signals as input are common in the literature.It is believed that using d...Parkinson’s disease(PD)is a debilitating neurological disorder affecting over 10 million people worldwide.PD classification models using voice signals as input are common in the literature.It is believed that using deep learning algorithms further enhances performance;nevertheless,it is challenging due to the nature of small-scale and imbalanced PD datasets.This paper proposed a convolutional neural network-based deep support vector machine(CNN-DSVM)to automate the feature extraction process using CNN and extend the conventional SVM to a DSVM for better classification performance in small-scale PD datasets.A customized kernel function reduces the impact of biased classification towards the majority class(healthy candidates in our consideration).An improved generative adversarial network(IGAN)was designed to generate additional training data to enhance the model’s performance.For performance evaluation,the proposed algorithm achieves a sensitivity of 97.6%and a specificity of 97.3%.The performance comparison is evaluated from five perspectives,including comparisons with different data generation algorithms,feature extraction techniques,kernel functions,and existing works.Results reveal the effectiveness of the IGAN algorithm,which improves the sensitivity and specificity by 4.05%–4.72%and 4.96%–5.86%,respectively;and the effectiveness of the CNN-DSVM algorithm,which improves the sensitivity by 1.24%–57.4%and specificity by 1.04%–163%and reduces biased detection towards the majority class.The ablation experiments confirm the effectiveness of individual components.Two future research directions have also been suggested.展开更多
With the emphasis on user privacy and communication security, encrypted traffic has increased dramatically, which brings great challenges to traffic classification. The classification method of encrypted traffic based...With the emphasis on user privacy and communication security, encrypted traffic has increased dramatically, which brings great challenges to traffic classification. The classification method of encrypted traffic based on GNN can deal with encrypted traffic well. However, existing GNN-based approaches ignore the relationship between client or server packets. In this paper, we design a network traffic topology based on GCN, called Flow Mapping Graph (FMG). FMG establishes sequential edges between vertexes by the arrival order of packets and establishes jump-order edges between vertexes by connecting packets in different bursts with the same direction. It not only reflects the time characteristics of the packet but also strengthens the relationship between the client or server packets. According to FMG, a Traffic Mapping Classification model (TMC-GCN) is designed, which can automatically capture and learn the characteristics and structure information of the top vertex in FMG. The TMC-GCN model is used to classify the encrypted traffic. The encryption stream classification problem is transformed into a graph classification problem, which can effectively deal with data from different data sources and application scenarios. By comparing the performance of TMC-GCN with other classical models in four public datasets, including CICIOT2023, ISCXVPN2016, CICAAGM2017, and GraphDapp, the effectiveness of the FMG algorithm is verified. The experimental results show that the accuracy rate of the TMC-GCN model is 96.13%, the recall rate is 95.04%, and the F1 rate is 94.54%.展开更多
The integration of image analysis through deep learning(DL)into rock classification represents a significant leap forward in geological research.While traditional methods remain invaluable for their expertise and hist...The integration of image analysis through deep learning(DL)into rock classification represents a significant leap forward in geological research.While traditional methods remain invaluable for their expertise and historical context,DL offers a powerful complement by enhancing the speed,objectivity,and precision of the classification process.This research explores the significance of image data augmentation techniques in optimizing the performance of convolutional neural networks(CNNs)for geological image analysis,particularly in the classification of igneous,metamorphic,and sedimentary rock types from rock thin section(RTS)images.This study primarily focuses on classic image augmentation techniques and evaluates their impact on model accuracy and precision.Results demonstrate that augmentation techniques like Equalize significantly enhance the model's classification capabilities,achieving an F1-Score of 0.9869 for igneous rocks,0.9884 for metamorphic rocks,and 0.9929 for sedimentary rocks,representing improvements compared to the baseline original results.Moreover,the weighted average F1-Score across all classes and techniques is 0.9886,indicating an enhancement.Conversely,methods like Distort lead to decreased accuracy and F1-Score,with an F1-Score of 0.949 for igneous rocks,0.954 for metamorphic rocks,and 0.9416 for sedimentary rocks,exacerbating the performance compared to the baseline.The study underscores the practicality of image data augmentation in geological image classification and advocates for the adoption of DL methods in this domain for automation and improved results.The findings of this study can benefit various fields,including remote sensing,mineral exploration,and environmental monitoring,by enhancing the accuracy of geological image analysis both for scientific research and industrial applications.展开更多
Human disturbance activities is one of the main reasons for inducing geohazards.Ecological impact assessment metrics of roads are inconsistent criteria and multiple.From the perspective of visual observation,the envir...Human disturbance activities is one of the main reasons for inducing geohazards.Ecological impact assessment metrics of roads are inconsistent criteria and multiple.From the perspective of visual observation,the environment damage can be shown through detecting the uncovered area of vegetation in the images along road.To realize this,an end-to-end environment damage detection model based on convolutional neural network is proposed.A 50-layer residual network is used to extract feature map.The initial parameters are optimized by transfer learning.An example is shown by this method.The dataset including cliff and landslide damage are collected by us along road in Shennongjia national forest park.Results show 0.4703 average precision(AP)rating for cliff damage and 0.4809 average precision(AP)rating for landslide damage.Compared with YOLOv3,our model shows a better accuracy in cliff and landslide detection although a certain amount of speed is sacrificed.展开更多
Aspect-oriented sentiment analysis is a meticulous sentiment analysis task that aims to analyse the sentiment polarity of specific aspects. Most of the current research builds graph convolutional networks based on dep...Aspect-oriented sentiment analysis is a meticulous sentiment analysis task that aims to analyse the sentiment polarity of specific aspects. Most of the current research builds graph convolutional networks based on dependent syntactic trees, which improves the classification performance of the models to some extent. However, the technical limitations of dependent syntactic trees can introduce considerable noise into the model. Meanwhile, it is difficult for a single graph convolutional network to aggregate both semantic and syntactic structural information of nodes, which affects the final sentence classification. To cope with the above problems, this paper proposes a bi-channel graph convolutional network model. The model introduces a phrase structure tree and transforms it into a hierarchical phrase matrix. The adjacency matrix of the dependent syntactic tree and the hierarchical phrase matrix are combined as the initial matrix of the graph convolutional network to enhance the syntactic information. The semantic information feature representations of the sentences are obtained by the graph convolutional network with a multi-head attention mechanism and fused to achieve complementary learning of dual-channel features. Experimental results show that the model performs well and improves the accuracy of sentiment classification on three public benchmark datasets, namely Rest14, Lap14 and Twitter.展开更多
Container-based virtualization technology has been more widely used in edge computing environments recently due to its advantages of lighter resource occupation, faster startup capability, and better resource utilizat...Container-based virtualization technology has been more widely used in edge computing environments recently due to its advantages of lighter resource occupation, faster startup capability, and better resource utilization efficiency. To meet the diverse needs of tasks, it usually needs to instantiate multiple network functions in the form of containers interconnect various generated containers to build a Container Cluster(CC). Then CCs will be deployed on edge service nodes with relatively limited resources. However, the increasingly complex and timevarying nature of tasks brings great challenges to optimal placement of CC. This paper regards the charges for various resources occupied by providing services as revenue, the service efficiency and energy consumption as cost, thus formulates a Mixed Integer Programming(MIP) model to describe the optimal placement of CC on edge service nodes. Furthermore, an Actor-Critic based Deep Reinforcement Learning(DRL) incorporating Graph Convolutional Networks(GCN) framework named as RL-GCN is proposed to solve the optimization problem. The framework obtains an optimal placement strategy through self-learning according to the requirements and objectives of the placement of CC. Particularly, through the introduction of GCN, the features of the association relationship between multiple containers in CCs can be effectively extracted to improve the quality of placement.The experiment results show that under different scales of service nodes and task requests, the proposed method can obtain the improved system performance in terms of placement error ratio, time efficiency of solution output and cumulative system revenue compared with other representative baseline methods.展开更多
Breast Cancer(BC)remains a leadingmalignancy among women,resulting in highmortality rates.Early and accurate detection is crucial for improving patient outcomes.Traditional diagnostic tools,while effective,have limita...Breast Cancer(BC)remains a leadingmalignancy among women,resulting in highmortality rates.Early and accurate detection is crucial for improving patient outcomes.Traditional diagnostic tools,while effective,have limitations that reduce their accessibility and accuracy.This study investigates the use ofConvolutionalNeuralNetworks(CNNs)to enhance the diagnostic process of BC histopathology.Utilizing the BreakHis dataset,which contains thousands of histopathological images,we developed a CNN model designed to improve the speed and accuracy of image analysis.Our CNN architecture was designed with multiple convolutional layers,max-pooling layers,and a fully connected network optimized for feature extraction and classification.Hyperparameter tuning was conducted to identify the optimal learning rate,batch size,and number of epochs,ensuring robust model performance.The dataset was divided into training(80%),validation(10%),and testing(10%)subsets,with performance evaluated using accuracy,precision,recall,and F1-score metrics.Our CNN model achieved a magnification-independent accuracy of 97.72%,with specific accuracies of 97.50%at 40×,97.61%at 100×,99.06%at 200×,and 97.25%at 400×magnification levels.These results demonstrate the model’s superior performance relative to existing methods.The integration of CNNs in diagnostic workflows can potentially reduce pathologist workload,minimize interpretation errors,and increase the availability of diagnostic testing,thereby improving BC management and patient survival rates.This study highlights the effectiveness of deep learning in automating BC histopathological classification and underscores the potential for AI-driven diagnostic solutions to improve patient care.展开更多
Gastrointestinal(GI)diseases,including gastric and colorectal cancers,signi-ficantly impact global health,necessitating accurate and efficient diagnostic me-thods.Endoscopic examination is the primary diagnostic tool;...Gastrointestinal(GI)diseases,including gastric and colorectal cancers,signi-ficantly impact global health,necessitating accurate and efficient diagnostic me-thods.Endoscopic examination is the primary diagnostic tool;however,its accu-racy is limited by operator dependency and interobserver variability.Advance-ments in deep learning,particularly convolutional neural networks(CNNs),show great potential for enhancing GI disease detection and classification.This review explores the application of CNNs in endoscopic imaging,focusing on polyp and tumor detection,disease classification,endoscopic ultrasound,and capsule endo-scopy analysis.We discuss the performance of CNN models with traditional dia-gnostic methods,highlighting their advantages in accuracy and real-time decision support.Despite promising results,challenges remain,including data availability,model interpretability,and clinical integration.Future directions include impro-ving model generalization,enhancing explainability,and conducting large-scale clinical trials.With continued advancements,CNN-powered artificial intelligence systems could revolutionize GI endoscopy by enhancing early disease detection,reducing diagnostic errors,and improving patient outcomes.展开更多
Essential proteins are an indispensable part of cells and play an extremely significant role in genetic disease diagnosis and drug development.Therefore,the prediction of essential proteins has received extensive atte...Essential proteins are an indispensable part of cells and play an extremely significant role in genetic disease diagnosis and drug development.Therefore,the prediction of essential proteins has received extensive attention from researchers.Many centrality methods and machine learning algorithms have been proposed to predict essential proteins.Nevertheless,the topological characteristics learned by the centrality method are not comprehensive enough,resulting in low accuracy.In addition,machine learning algorithms need sufficient prior knowledge to select features,and the ability to solve imbalanced classification problems needs to be further strengthened.These two factors greatly affect the performance of predicting essential proteins.In this paper,we propose a deep learning framework based on temporal convolutional networks to predict essential proteins by integrating gene expression data and protein-protein interaction(PPI)network.We make use of the method of network embedding to automatically learn more abundant features of proteins in the PPI network.For gene expression data,we treat it as sequence data,and use temporal convolutional networks to extract sequence features.Finally,the two types of features are integrated and put into the multi-layer neural network to complete the final classification task.The performance of our method is evaluated by comparing with seven centrality methods,six machine learning algorithms,and two deep learning models.The results of the experiment show that our method is more effective than the comparison methods for predicting essential proteins.展开更多
Recommendation systems(RSs)are crucial in personalizing user experiences in digital environments by suggesting relevant content or items.Collaborative filtering(CF)is a widely used personalization technique that lever...Recommendation systems(RSs)are crucial in personalizing user experiences in digital environments by suggesting relevant content or items.Collaborative filtering(CF)is a widely used personalization technique that leverages user-item interactions to generate recommendations.However,it struggles with challenges like the cold-start problem,scalability issues,and data sparsity.To address these limitations,we develop a Graph Convolutional Networks(GCNs)model that captures the complex network of interactions between users and items,identifying subtle patterns that traditional methods may overlook.We integrate this GCNs model into a federated learning(FL)framework,enabling themodel to learn fromdecentralized datasets.This not only significantly enhances user privacy—a significant improvement over conventionalmodels but also reassures users about the safety of their data.Additionally,by securely incorporating demographic information,our approach further personalizes recommendations and mitigates the coldstart issue without compromising user data.We validate our RSs model using the openMovieLens dataset and evaluate its performance across six key metrics:Precision,Recall,Area Under the Receiver Operating Characteristic Curve(ROC-AUC),F1 Score,Normalized Discounted Cumulative Gain(NDCG),and Mean Reciprocal Rank(MRR).The experimental results demonstrate significant enhancements in recommendation quality,underscoring that combining GCNs with CF in a federated setting provides a transformative solution for advanced recommendation systems.展开更多
In the burgeoning field of anomaly detection within attributed networks,traditional methodologies often encounter the intricacies of network complexity,particularly in capturing nonlinearity and sparsity.This study in...In the burgeoning field of anomaly detection within attributed networks,traditional methodologies often encounter the intricacies of network complexity,particularly in capturing nonlinearity and sparsity.This study introduces an innovative approach that synergizes the strengths of graph convolutional networks with advanced deep residual learning and a unique residual-based attention mechanism,thereby creating a more nuanced and efficient method for anomaly detection in complex networks.The heart of our model lies in the integration of graph convolutional networks that capture complex structural relationships within the network data.This is further bolstered by deep residual learning,which is employed to model intricate nonlinear connections directly from input data.A pivotal innovation in our approach is the incorporation of a residual-based attention mech-anism.This mechanism dynamically adjusts the importance of nodes based on their residual information,thereby significantly enhancing the sensitivity of the model to subtle anomalies.Furthermore,we introduce a novel hypersphere mapping technique in the latent space to distinctly separate normal and anomalous data.This mapping is the key to our model’s ability to pinpoint anomalies with greater precision.An extensive experimental setup was used to validate the efficacy of the proposed model.Using attributed social network datasets,we demonstrate that our model not only competes with but also surpasses existing state-of-the-art methods in anomaly detection.The results show the exceptional capability of our model to handle the multifaceted nature of real-world networks.展开更多
The single event effects(SEEs)evaluations caused by atmospheric neutrons were conducted on three different convolutional neural network(CNN)models(Yolov3,MNIST,and ResNet50)in the atmospheric neutron irradiation spect...The single event effects(SEEs)evaluations caused by atmospheric neutrons were conducted on three different convolutional neural network(CNN)models(Yolov3,MNIST,and ResNet50)in the atmospheric neutron irradiation spectrometer(ANIS)at the China Spallation Neutron Source(CSNS).The Yolov3 and MNIST models were implemented on the XILINX28-nm system-on-chip(So C).Meanwhile,the Yolov3 and ResNet50 models were deployed on the XILINX 16-nm Fin FET Ultra Scale+MPSoC.The atmospheric neutron SEEs on the tested CNN systems were comprehensively evaluated from six aspects,including chip type,network architecture,deployment methods,inference time,datasets,and the position of the anchor boxes.The various types of SEE soft errors,SEE cross-sections,and their distribution were analyzed to explore the radiation sensitivities and rules of 28-nm and 16-nm SoC.The current research can provide the technology support of radiation-resistant design of CNN system for developing and applying high-reliability,long-lifespan domestic artificial intelligence chips.展开更多
To enhance the inference efficiency of convolutional neural network(CNN),tensor parallelism is employed to improve the parallelism within operators.However,existing methods are customized to specific networks and hard...To enhance the inference efficiency of convolutional neural network(CNN),tensor parallelism is employed to improve the parallelism within operators.However,existing methods are customized to specific networks and hardware,limiting their generalizability.This paper proposes an approach called resource-adaptive tensor decomposition(RATD)for CNN operators,which aims to achieve an optimal match between computational resources and parallel computing tasks.Firstly,CNN is represented with fine-grained tensors at the lower graph level,thereby decoupling tensors that can be computed in parallel within operators.Secondly,the convolution and pooling operators are fused,and the decoupled tensor blocks are scheduled in parallel.Finally,a cost model is constructed,based on runtime and resource utilization,to iteratively refine the process of tensor block decomposition and automatically determine the optimal tensor decomposition.Experimental results demonstrate that the proposed RATD improves the accuracy of the model by 11%.Compared with CUDA(compute unified device architecture)deep neural network library(cuDNN),RATD achieves an average speedup ratio of 1.21 times in inference time across various convolution kernels,along with a 12%increase in computational resource utilization.展开更多
Developing an accurate and efficient comprehensive water quality prediction model and its assessment method is crucial for the prevention and control of water pollution.Deep learning(DL),as one of the most promising t...Developing an accurate and efficient comprehensive water quality prediction model and its assessment method is crucial for the prevention and control of water pollution.Deep learning(DL),as one of the most promising technologies today,plays a crucial role in the effective assessment of water body health,which is essential for water resource management.This study models using both the original dataset and a dataset augmented with Generative Adversarial Networks(GAN).It integrates optimization algorithms(OA)with Convolutional Neural Networks(CNN)to propose a comprehensive water quality model evaluation method aiming at identifying the optimal models for different pollutants.Specifically,after preprocessing the spectral dataset,data augmentation was conducted to obtain two datasets.Then,six new models were developed on these datasets using particle swarm optimization(PSO),genetic algorithm(GA),and simulated annealing(SA)combined with CNN to simulate and forecast the concentrations of three water pollutants:Chemical Oxygen Demand(COD),Total Nitrogen(TN),and Total Phosphorus(TP).Finally,seven model evaluation methods,including uncertainty analysis,were used to evaluate the constructed models and select the optimal models for the three pollutants.The evaluation results indicate that the GPSCNN model performed best in predicting COD and TP concentrations,while the GGACNN model excelled in TN concentration prediction.Compared to existing technologies,the proposed models and evaluation methods provide a more comprehensive and rapid approach to water body prediction and assessment,offering new insights and methods for water pollution prevention and control.展开更多
基金funded by the Ministry of Higher Education(MoHE)Malaysia through the Fundamental Research Grant Scheme—Early Career Researcher(FRGS-EC),grant number FRGSEC/1/2024/ICT02/UNIMAP/02/8.
文摘critical for guiding treatment and improving patient outcomes.Traditional molecular subtyping via immuno-histochemistry(IHC)test is invasive,time-consuming,and may not fully represent tumor heterogeneity.This study proposes a non-invasive approach using digital mammography images and deep learning algorithm for classifying breast cancer molecular subtypes.Four pretrained models,including two Convolutional Neural Networks(MobileNet_V3_Large and VGG-16)and two Vision Transformers(ViT_B_16 and ViT_Base_Patch16_Clip_224)were fine-tuned to classify images into HER2-enriched,Luminal,Normal-like,and Triple Negative subtypes.Hyperparameter tuning,including learning rate adjustment and layer freezing strategies,was applied to optimize performance.Among the evaluated models,ViT_Base_Patch16_Clip_224 achieved the highest test accuracy(94.44%),with equally high precision,recall,and F1-score of 0.94,demonstrating excellent generalization.MobileNet_V3_Large achieved the same accuracy but showed less training stability.In contrast,VGG-16 recorded the lowest performance,indicating a limitation in its generalizability for this classification task.The study also highlighted the superior performance of the Vision Transformer models over CNNs,particularly due to their ability to capture global contextual features and the benefit of CLIP-based pretraining in ViT_Base_Patch16_Clip_224.To enhance clinical applicability,a graphical user interface(GUI)named“BCMS Dx”was developed for streamlined subtype prediction.Deep learning applied to mammography has proven effective for accurate and non-invasive molecular subtyping.The proposed Vision Transformer-based model and supporting GUI offer a promising direction for augmenting diagnostic workflows,minimizing the need for invasive procedures,and advancing personalized breast cancer management.
基金National Natural Science Foundation of China(Grant No.62103434)National Science Fund for Distinguished Young Scholars(Grant No.62176263).
文摘With network attack technology continuing to develop,traditional anomaly traffic detection methods that rely on feature engineering are increasingly insufficient in efficiency and accuracy.Graph Neural Network(GNN),a promising Deep Learning(DL)approach,has proven to be highly effective in identifying intricate patterns in graph⁃structured data and has already found wide applications in the field of network security.In this paper,we propose a hybrid Graph Convolutional Network(GCN)⁃GraphSAGE model for Anomaly Traffic Detection,namely HGS⁃ATD,which aims to improve the accuracy of anomaly traffic detection by leveraging edge feature learning to better capture the relationships between network entities.We validate the HGS⁃ATD model on four publicly available datasets,including NF⁃UNSW⁃NB15⁃v2.The experimental results show that the enhanced hybrid model is 5.71%to 10.25%higher than the baseline model in terms of accuracy,and the F1⁃score is 5.53%to 11.63%higher than the baseline model,proving that the model can effectively distinguish normal traffic from attack traffic and accurately classify various types of attacks.
基金Support by Sichuan Science and Technology Program[2023YFSY0026,2023YFH0004]Guangzhou Huashang University[2024HSZD01,HS2023JYSZH01].
文摘Graph neural networks(GNN)have shown strong performance in node classification tasks,yet most existing models rely on uniform or shared weight aggregation,lacking flexibility in modeling the varying strength of relationships among nodes.This paper proposes a novel graph coupling convolutional model that introduces an adaptive weighting mechanism to assign distinct importance to neighboring nodes based on their similarity to the central node.Unlike traditional methods,the proposed coupling strategy enhances the interpretability of node interactions while maintaining competitive classification performance.The model operates in the spatial domain,utilizing adjacency list structures for efficient convolution and addressing the limitations of weight sharing through a coupling-based similarity computation.Extensive experiments are conducted on five graph-structured datasets,including Cora,Citeseer,PubMed,Reddit,and BlogCatalog,as well as a custom topology dataset constructed from the Open University Learning Analytics Dataset(OULAD)educational platform.Results demonstrate that the proposed model achieves good classification accuracy,while significantly reducing training time through direct second-order neighbor fusion and data preprocessing.Moreover,analysis of neighborhood order reveals that considering third-order neighbors offers limited accuracy gains but introduces considerable computational overhead,confirming the efficiency of first-and second-order convolution in practical applications.Overall,the proposed graph coupling model offers a lightweight,interpretable,and effective framework for multi-label node classification in complex networks.
基金supported in part by the National Natural Science Foundation of China under Grant 62172368the Natural Science Foundation of Zhejiang Province under Grant LR22F020003.
文摘The development of brain-computer interfaces(BCI)based on motor imagery(MI)has greatly improved patients’quality of life with movement disorders.The classification of upper limb MI has been widely studied and applied in many fields,including rehabilitation.However,the physiological representations of left and right lower limb movements are too close and activated deep in the cerebral cortex,making it difficult to distinguish their features.Therefore,classifying lower limbs motor imagery is more challenging.In this study,we propose a feature extraction method based on functional connectivity,which utilizes phase-locked values to construct a functional connectivity matrix as the features of the left and right legs,which can effectively avoid the problem of physiological representations of the left and right lower limbs being too close to each other during movement.In addition,considering the topology and the temporal characteristics of the electroencephalogram(EEG),we designed a temporal-spatial convolutional network(TSGCN)to capture the spatiotemporal information for classification.Experimental results show that the accuracy of the proposed method is higher than that of existing methods,achieving an average classification accuracy of 73.58%on the internal dataset.Finally,this study explains the network mechanism of left and right foot MI from the perspective of graph theoretic features and demonstrates the feasibility of decoding lower limb MI.
基金supported by the Natural Science Foundation of China No.62362008the Major Scientific and Technological Special Project of Guizhou Province([2024]014).
文摘With the rapid development of the Artificial Intelligence of Things(AIoT),convolutional neural networks(CNNs)have demonstrated potential and remarkable performance in AIoT applications due to their excellent performance in various inference tasks.However,the users have concerns about privacy leakage for the use of AI and the performance and efficiency of computing on resource-constrained IoT edge devices.Therefore,this paper proposes an efficient privacy-preserving CNN framework(i.e.,EPPA)based on the Fully Homomorphic Encryption(FHE)scheme for AIoT application scenarios.In the plaintext domain,we verify schemes with different activation structures to determine the actual activation functions applicable to the corresponding ciphertext domain.Within the encryption domain,we integrate batch normalization(BN)into the convolutional layers to simplify the computation process.For nonlinear activation functions,we use composite polynomials for approximate calculation.Regarding the noise accumulation caused by homomorphic multiplication operations,we realize the refreshment of ciphertext noise through minimal“decryption-encryption”interactions,instead of adopting bootstrapping operations.Additionally,in practical implementation,we convert three-dimensional convolution into two-dimensional convolution to reduce the amount of computation in the encryption domain.Finally,we conduct extensive experiments on four IoT datasets,different CNN architectures,and two platforms with different resource configurations to evaluate the performance of EPPA in detail.
基金funded by scientific research projects under Grant JY2024B011.
文摘With the increasing complexity of industrial automation,planetary gearboxes play a vital role in largescale equipment transmission systems,directly impacting operational efficiency and safety.Traditional maintenance strategies often struggle to accurately predict the degradation process of equipment,leading to excessive maintenance costs or potential failure risks.However,existing prediction methods based on statistical models are difficult to adapt to nonlinear degradation processes.To address these challenges,this study proposes a novel condition-based maintenance framework for planetary gearboxes.A comprehensive full-lifecycle degradation experiment was conducted to collect raw vibration signals,which were then processed using a temporal convolutional network autoencoder with multi-scale perception capability to extract deep temporal degradation features,enabling the collaborative extraction of longperiod meshing frequencies and short-term impact features from the vibration signals.Kernel principal component analysis was employed to fuse and normalize these features,enhancing the characterization of degradation progression.A nonlinear Wiener process was used to model the degradation trajectory,with a threshold decay function introduced to dynamically adjust maintenance strategies,and model parameters optimized through maximum likelihood estimation.Meanwhile,the maintenance strategy was optimized to minimize costs per unit time,determining the optimal maintenance timing and preventive maintenance threshold.The comprehensive indicator of degradation trends extracted by this method reaches 0.756,which is 41.2%higher than that of traditional time-domain features;the dynamic threshold strategy reduces the maintenance cost per unit time to 55.56,which is 8.9%better than that of the static threshold optimization.Experimental results demonstrate significant reductions in maintenance costs while enhancing system reliability and safety.This study realizes the organic integration of deep learning and reliability theory in the maintenance of planetary gearboxes,provides an interpretable solution for the predictive maintenance of complex mechanical systems,and promotes the development of condition-based maintenance strategies for planetary gearboxes.
基金The work described in this paper was fully supported by a grant from Hong Kong Metropolitan University(RIF/2021/05).
文摘Parkinson’s disease(PD)is a debilitating neurological disorder affecting over 10 million people worldwide.PD classification models using voice signals as input are common in the literature.It is believed that using deep learning algorithms further enhances performance;nevertheless,it is challenging due to the nature of small-scale and imbalanced PD datasets.This paper proposed a convolutional neural network-based deep support vector machine(CNN-DSVM)to automate the feature extraction process using CNN and extend the conventional SVM to a DSVM for better classification performance in small-scale PD datasets.A customized kernel function reduces the impact of biased classification towards the majority class(healthy candidates in our consideration).An improved generative adversarial network(IGAN)was designed to generate additional training data to enhance the model’s performance.For performance evaluation,the proposed algorithm achieves a sensitivity of 97.6%and a specificity of 97.3%.The performance comparison is evaluated from five perspectives,including comparisons with different data generation algorithms,feature extraction techniques,kernel functions,and existing works.Results reveal the effectiveness of the IGAN algorithm,which improves the sensitivity and specificity by 4.05%–4.72%and 4.96%–5.86%,respectively;and the effectiveness of the CNN-DSVM algorithm,which improves the sensitivity by 1.24%–57.4%and specificity by 1.04%–163%and reduces biased detection towards the majority class.The ablation experiments confirm the effectiveness of individual components.Two future research directions have also been suggested.
基金supported by the National Key Research and Development Program of China No.2023YFA1009500.
文摘With the emphasis on user privacy and communication security, encrypted traffic has increased dramatically, which brings great challenges to traffic classification. The classification method of encrypted traffic based on GNN can deal with encrypted traffic well. However, existing GNN-based approaches ignore the relationship between client or server packets. In this paper, we design a network traffic topology based on GCN, called Flow Mapping Graph (FMG). FMG establishes sequential edges between vertexes by the arrival order of packets and establishes jump-order edges between vertexes by connecting packets in different bursts with the same direction. It not only reflects the time characteristics of the packet but also strengthens the relationship between the client or server packets. According to FMG, a Traffic Mapping Classification model (TMC-GCN) is designed, which can automatically capture and learn the characteristics and structure information of the top vertex in FMG. The TMC-GCN model is used to classify the encrypted traffic. The encryption stream classification problem is transformed into a graph classification problem, which can effectively deal with data from different data sources and application scenarios. By comparing the performance of TMC-GCN with other classical models in four public datasets, including CICIOT2023, ISCXVPN2016, CICAAGM2017, and GraphDapp, the effectiveness of the FMG algorithm is verified. The experimental results show that the accuracy rate of the TMC-GCN model is 96.13%, the recall rate is 95.04%, and the F1 rate is 94.54%.
文摘The integration of image analysis through deep learning(DL)into rock classification represents a significant leap forward in geological research.While traditional methods remain invaluable for their expertise and historical context,DL offers a powerful complement by enhancing the speed,objectivity,and precision of the classification process.This research explores the significance of image data augmentation techniques in optimizing the performance of convolutional neural networks(CNNs)for geological image analysis,particularly in the classification of igneous,metamorphic,and sedimentary rock types from rock thin section(RTS)images.This study primarily focuses on classic image augmentation techniques and evaluates their impact on model accuracy and precision.Results demonstrate that augmentation techniques like Equalize significantly enhance the model's classification capabilities,achieving an F1-Score of 0.9869 for igneous rocks,0.9884 for metamorphic rocks,and 0.9929 for sedimentary rocks,representing improvements compared to the baseline original results.Moreover,the weighted average F1-Score across all classes and techniques is 0.9886,indicating an enhancement.Conversely,methods like Distort lead to decreased accuracy and F1-Score,with an F1-Score of 0.949 for igneous rocks,0.954 for metamorphic rocks,and 0.9416 for sedimentary rocks,exacerbating the performance compared to the baseline.The study underscores the practicality of image data augmentation in geological image classification and advocates for the adoption of DL methods in this domain for automation and improved results.The findings of this study can benefit various fields,including remote sensing,mineral exploration,and environmental monitoring,by enhancing the accuracy of geological image analysis both for scientific research and industrial applications.
文摘Human disturbance activities is one of the main reasons for inducing geohazards.Ecological impact assessment metrics of roads are inconsistent criteria and multiple.From the perspective of visual observation,the environment damage can be shown through detecting the uncovered area of vegetation in the images along road.To realize this,an end-to-end environment damage detection model based on convolutional neural network is proposed.A 50-layer residual network is used to extract feature map.The initial parameters are optimized by transfer learning.An example is shown by this method.The dataset including cliff and landslide damage are collected by us along road in Shennongjia national forest park.Results show 0.4703 average precision(AP)rating for cliff damage and 0.4809 average precision(AP)rating for landslide damage.Compared with YOLOv3,our model shows a better accuracy in cliff and landslide detection although a certain amount of speed is sacrificed.
文摘Aspect-oriented sentiment analysis is a meticulous sentiment analysis task that aims to analyse the sentiment polarity of specific aspects. Most of the current research builds graph convolutional networks based on dependent syntactic trees, which improves the classification performance of the models to some extent. However, the technical limitations of dependent syntactic trees can introduce considerable noise into the model. Meanwhile, it is difficult for a single graph convolutional network to aggregate both semantic and syntactic structural information of nodes, which affects the final sentence classification. To cope with the above problems, this paper proposes a bi-channel graph convolutional network model. The model introduces a phrase structure tree and transforms it into a hierarchical phrase matrix. The adjacency matrix of the dependent syntactic tree and the hierarchical phrase matrix are combined as the initial matrix of the graph convolutional network to enhance the syntactic information. The semantic information feature representations of the sentences are obtained by the graph convolutional network with a multi-head attention mechanism and fused to achieve complementary learning of dual-channel features. Experimental results show that the model performs well and improves the accuracy of sentiment classification on three public benchmark datasets, namely Rest14, Lap14 and Twitter.
文摘Container-based virtualization technology has been more widely used in edge computing environments recently due to its advantages of lighter resource occupation, faster startup capability, and better resource utilization efficiency. To meet the diverse needs of tasks, it usually needs to instantiate multiple network functions in the form of containers interconnect various generated containers to build a Container Cluster(CC). Then CCs will be deployed on edge service nodes with relatively limited resources. However, the increasingly complex and timevarying nature of tasks brings great challenges to optimal placement of CC. This paper regards the charges for various resources occupied by providing services as revenue, the service efficiency and energy consumption as cost, thus formulates a Mixed Integer Programming(MIP) model to describe the optimal placement of CC on edge service nodes. Furthermore, an Actor-Critic based Deep Reinforcement Learning(DRL) incorporating Graph Convolutional Networks(GCN) framework named as RL-GCN is proposed to solve the optimization problem. The framework obtains an optimal placement strategy through self-learning according to the requirements and objectives of the placement of CC. Particularly, through the introduction of GCN, the features of the association relationship between multiple containers in CCs can be effectively extracted to improve the quality of placement.The experiment results show that under different scales of service nodes and task requests, the proposed method can obtain the improved system performance in terms of placement error ratio, time efficiency of solution output and cumulative system revenue compared with other representative baseline methods.
基金funded by the Deanship of Graduate Studies and Scientific Research at Jouf University under grant No.(DGSSR-2024-02-01096).
文摘Breast Cancer(BC)remains a leadingmalignancy among women,resulting in highmortality rates.Early and accurate detection is crucial for improving patient outcomes.Traditional diagnostic tools,while effective,have limitations that reduce their accessibility and accuracy.This study investigates the use ofConvolutionalNeuralNetworks(CNNs)to enhance the diagnostic process of BC histopathology.Utilizing the BreakHis dataset,which contains thousands of histopathological images,we developed a CNN model designed to improve the speed and accuracy of image analysis.Our CNN architecture was designed with multiple convolutional layers,max-pooling layers,and a fully connected network optimized for feature extraction and classification.Hyperparameter tuning was conducted to identify the optimal learning rate,batch size,and number of epochs,ensuring robust model performance.The dataset was divided into training(80%),validation(10%),and testing(10%)subsets,with performance evaluated using accuracy,precision,recall,and F1-score metrics.Our CNN model achieved a magnification-independent accuracy of 97.72%,with specific accuracies of 97.50%at 40×,97.61%at 100×,99.06%at 200×,and 97.25%at 400×magnification levels.These results demonstrate the model’s superior performance relative to existing methods.The integration of CNNs in diagnostic workflows can potentially reduce pathologist workload,minimize interpretation errors,and increase the availability of diagnostic testing,thereby improving BC management and patient survival rates.This study highlights the effectiveness of deep learning in automating BC histopathological classification and underscores the potential for AI-driven diagnostic solutions to improve patient care.
基金Supported by Open Funds for Shaanxi Provincial Key Laboratory of Infection and Immune Diseases,No.2023-KFMS-1.
文摘Gastrointestinal(GI)diseases,including gastric and colorectal cancers,signi-ficantly impact global health,necessitating accurate and efficient diagnostic me-thods.Endoscopic examination is the primary diagnostic tool;however,its accu-racy is limited by operator dependency and interobserver variability.Advance-ments in deep learning,particularly convolutional neural networks(CNNs),show great potential for enhancing GI disease detection and classification.This review explores the application of CNNs in endoscopic imaging,focusing on polyp and tumor detection,disease classification,endoscopic ultrasound,and capsule endo-scopy analysis.We discuss the performance of CNN models with traditional dia-gnostic methods,highlighting their advantages in accuracy and real-time decision support.Despite promising results,challenges remain,including data availability,model interpretability,and clinical integration.Future directions include impro-ving model generalization,enhancing explainability,and conducting large-scale clinical trials.With continued advancements,CNN-powered artificial intelligence systems could revolutionize GI endoscopy by enhancing early disease detection,reducing diagnostic errors,and improving patient outcomes.
基金the National Natural Science Foundation of China(Nos.11861045 and 62162040)。
文摘Essential proteins are an indispensable part of cells and play an extremely significant role in genetic disease diagnosis and drug development.Therefore,the prediction of essential proteins has received extensive attention from researchers.Many centrality methods and machine learning algorithms have been proposed to predict essential proteins.Nevertheless,the topological characteristics learned by the centrality method are not comprehensive enough,resulting in low accuracy.In addition,machine learning algorithms need sufficient prior knowledge to select features,and the ability to solve imbalanced classification problems needs to be further strengthened.These two factors greatly affect the performance of predicting essential proteins.In this paper,we propose a deep learning framework based on temporal convolutional networks to predict essential proteins by integrating gene expression data and protein-protein interaction(PPI)network.We make use of the method of network embedding to automatically learn more abundant features of proteins in the PPI network.For gene expression data,we treat it as sequence data,and use temporal convolutional networks to extract sequence features.Finally,the two types of features are integrated and put into the multi-layer neural network to complete the final classification task.The performance of our method is evaluated by comparing with seven centrality methods,six machine learning algorithms,and two deep learning models.The results of the experiment show that our method is more effective than the comparison methods for predicting essential proteins.
基金funded by Soonchunhyang University,Grant Numbers 20241422BK21 FOUR(Fostering Outstanding Universities for Research,Grant Number 5199990914048).
文摘Recommendation systems(RSs)are crucial in personalizing user experiences in digital environments by suggesting relevant content or items.Collaborative filtering(CF)is a widely used personalization technique that leverages user-item interactions to generate recommendations.However,it struggles with challenges like the cold-start problem,scalability issues,and data sparsity.To address these limitations,we develop a Graph Convolutional Networks(GCNs)model that captures the complex network of interactions between users and items,identifying subtle patterns that traditional methods may overlook.We integrate this GCNs model into a federated learning(FL)framework,enabling themodel to learn fromdecentralized datasets.This not only significantly enhances user privacy—a significant improvement over conventionalmodels but also reassures users about the safety of their data.Additionally,by securely incorporating demographic information,our approach further personalizes recommendations and mitigates the coldstart issue without compromising user data.We validate our RSs model using the openMovieLens dataset and evaluate its performance across six key metrics:Precision,Recall,Area Under the Receiver Operating Characteristic Curve(ROC-AUC),F1 Score,Normalized Discounted Cumulative Gain(NDCG),and Mean Reciprocal Rank(MRR).The experimental results demonstrate significant enhancements in recommendation quality,underscoring that combining GCNs with CF in a federated setting provides a transformative solution for advanced recommendation systems.
文摘In the burgeoning field of anomaly detection within attributed networks,traditional methodologies often encounter the intricacies of network complexity,particularly in capturing nonlinearity and sparsity.This study introduces an innovative approach that synergizes the strengths of graph convolutional networks with advanced deep residual learning and a unique residual-based attention mechanism,thereby creating a more nuanced and efficient method for anomaly detection in complex networks.The heart of our model lies in the integration of graph convolutional networks that capture complex structural relationships within the network data.This is further bolstered by deep residual learning,which is employed to model intricate nonlinear connections directly from input data.A pivotal innovation in our approach is the incorporation of a residual-based attention mech-anism.This mechanism dynamically adjusts the importance of nodes based on their residual information,thereby significantly enhancing the sensitivity of the model to subtle anomalies.Furthermore,we introduce a novel hypersphere mapping technique in the latent space to distinctly separate normal and anomalous data.This mapping is the key to our model’s ability to pinpoint anomalies with greater precision.An extensive experimental setup was used to validate the efficacy of the proposed model.Using attributed social network datasets,we demonstrate that our model not only competes with but also surpasses existing state-of-the-art methods in anomaly detection.The results show the exceptional capability of our model to handle the multifaceted nature of real-world networks.
基金Project supported by the National Natural Science Foundation of China(Grant No.12305303)the Natural Science Foundation of Hunan Province of China(Grant Nos.2023JJ40520,2024JJ2044,and 2021JJ40444)+3 种基金the Science and Technology Innovation Program of Hunan Province,China(Grant No.2020RC3054)the Postgraduate Scientific Research Innovation Project of Hunan Province,China(Grant No.CX20240831)the Natural Science Basic Research Plan in the Shaanxi Province of China(Grant No.2023-JC-QN0015)the Doctoral Research Fund of University of South China(Grant No.200XQD033)。
文摘The single event effects(SEEs)evaluations caused by atmospheric neutrons were conducted on three different convolutional neural network(CNN)models(Yolov3,MNIST,and ResNet50)in the atmospheric neutron irradiation spectrometer(ANIS)at the China Spallation Neutron Source(CSNS).The Yolov3 and MNIST models were implemented on the XILINX28-nm system-on-chip(So C).Meanwhile,the Yolov3 and ResNet50 models were deployed on the XILINX 16-nm Fin FET Ultra Scale+MPSoC.The atmospheric neutron SEEs on the tested CNN systems were comprehensively evaluated from six aspects,including chip type,network architecture,deployment methods,inference time,datasets,and the position of the anchor boxes.The various types of SEE soft errors,SEE cross-sections,and their distribution were analyzed to explore the radiation sensitivities and rules of 28-nm and 16-nm SoC.The current research can provide the technology support of radiation-resistant design of CNN system for developing and applying high-reliability,long-lifespan domestic artificial intelligence chips.
基金Supported by the National Science and Technology Major Project of China(No.2022ZD0119003)the National Natural Science Foundation of China(No.61834005).
文摘To enhance the inference efficiency of convolutional neural network(CNN),tensor parallelism is employed to improve the parallelism within operators.However,existing methods are customized to specific networks and hardware,limiting their generalizability.This paper proposes an approach called resource-adaptive tensor decomposition(RATD)for CNN operators,which aims to achieve an optimal match between computational resources and parallel computing tasks.Firstly,CNN is represented with fine-grained tensors at the lower graph level,thereby decoupling tensors that can be computed in parallel within operators.Secondly,the convolution and pooling operators are fused,and the decoupled tensor blocks are scheduled in parallel.Finally,a cost model is constructed,based on runtime and resource utilization,to iteratively refine the process of tensor block decomposition and automatically determine the optimal tensor decomposition.Experimental results demonstrate that the proposed RATD improves the accuracy of the model by 11%.Compared with CUDA(compute unified device architecture)deep neural network library(cuDNN),RATD achieves an average speedup ratio of 1.21 times in inference time across various convolution kernels,along with a 12%increase in computational resource utilization.
基金Supported by Natural Science Basic Research Plan in Shaanxi Province of China(Program No.2022JM-396)the Strategic Priority Research Program of the Chinese Academy of Sciences,Grant No.XDA23040101+4 种基金Shaanxi Province Key Research and Development Projects(Program No.2023-YBSF-437)Xi'an Shiyou University Graduate Student Innovation Fund Program(Program No.YCX2412041)State Key Laboratory of Air Traffic Management System and Technology(SKLATM202001)Tianjin Education Commission Research Program Project(2020KJ028)Fundamental Research Funds for the Central Universities(3122019132)。
文摘Developing an accurate and efficient comprehensive water quality prediction model and its assessment method is crucial for the prevention and control of water pollution.Deep learning(DL),as one of the most promising technologies today,plays a crucial role in the effective assessment of water body health,which is essential for water resource management.This study models using both the original dataset and a dataset augmented with Generative Adversarial Networks(GAN).It integrates optimization algorithms(OA)with Convolutional Neural Networks(CNN)to propose a comprehensive water quality model evaluation method aiming at identifying the optimal models for different pollutants.Specifically,after preprocessing the spectral dataset,data augmentation was conducted to obtain two datasets.Then,six new models were developed on these datasets using particle swarm optimization(PSO),genetic algorithm(GA),and simulated annealing(SA)combined with CNN to simulate and forecast the concentrations of three water pollutants:Chemical Oxygen Demand(COD),Total Nitrogen(TN),and Total Phosphorus(TP).Finally,seven model evaluation methods,including uncertainty analysis,were used to evaluate the constructed models and select the optimal models for the three pollutants.The evaluation results indicate that the GPSCNN model performed best in predicting COD and TP concentrations,while the GGACNN model excelled in TN concentration prediction.Compared to existing technologies,the proposed models and evaluation methods provide a more comprehensive and rapid approach to water body prediction and assessment,offering new insights and methods for water pollution prevention and control.