Deep transfer learning has achieved significant success in anomaly detection over the past decade,but data acquisition challenges in practical engineering hinder high-quality feature representation for few-shot learni...Deep transfer learning has achieved significant success in anomaly detection over the past decade,but data acquisition challenges in practical engineering hinder high-quality feature representation for few-shot learning tasks.To address this issue,a novel time-frequency-assisted deep feature enhancement(TFE)mechanism is proposed.Unlike traditional methods that integrate time-frequency analysis with deep neural networks,TFE employs a wavelet scattering transform to establish a parallel time-frequency feature space,where a dual interaction strategy facilitates collaboration between deep feature and time-frequency spaces through two operations:1)Enhancement,where a frequency-importance-driven contrastive learning(FICL)network transfers physically-aware information from wavelet scattering features to deep features,and 2)Feedback,which uses a detection rule adaptation module to minimize bias in wavelet scattering features based on deep feature performance.TFE is applied to a domain-adversarial anomaly detection framework and,through alternating training,significantly enhances both deep feature discriminative power and few-shot anomaly detection.Theoretical analysis confirms that the proposed dual interaction strategy reduces the upper bound of classification error.Experiments on benchmark datasets and a real-world industrial dataset from a large steel factory demonstrate TFE's superior performance and highlight the importance of frequency saliency in transfer learning.Thus,collaboration is shown to outperform integration for few-shot transfer learning in anomaly detection.展开更多
Objective To address the dual challenges of long-tail distribution and feature sparsity in traditional Chinese medicine(TCM)syndrome differentiation within real clinical settings,we propose a data-efficient learning f...Objective To address the dual challenges of long-tail distribution and feature sparsity in traditional Chinese medicine(TCM)syndrome differentiation within real clinical settings,we propose a data-efficient learning framework enhanced by knowledge graphs.Methods We developed Agent-GNN,a three-stage decoupled learning framework,and validated it on the Traditional Chinese Medicine Syndrome Diagnosis(TCM-SD)dataset containing 54152 clinical records across 148 syndrome categories.First,we constructed a comprehensive medical knowledge graph encoding the complete TCM reasoning system.Second,we proposed a Functional Patient Profiling(FPP)method that utilizes large language models(LLMs)combined with Graph Retrieval-Augmented Generation(RAG)to extract structured symptom-etiology-pathogenesis subgraphs from medical records.Third,we employed heterogeneous graph neural networks to learn structured combination patterns explicitly.We compared our method against multiple baselines including BERT,ZY-BERT,ZY-BERT+Know,GAT,and GPT-4 Few-shot,using macro-F1 score as the primary evaluation metric.Additionally,ablation experiments were conducted to validate the contribution of each key component to model performance.Results Agent-GNN achieved an overall macro-F1 score of 72.4%,representing an 8.7 percentage points improvement over ZY-BERT+Know(63.7%),the strongest baseline among traditional methods.For long-tail syndromes with fewer than 10 samples,Agent-GNN reached a macro-F1 score of 58.6%,compared with 39.3%for ZY-BERT+Know and 41.2%for GPT-4 Few-shot,representing relative improvements of 49.2%and 42.2%,respectively.Ablation experiments confirmed that the explicit modeling of etiology-pathogenesis nodes contributed 12.4 percentage points to this enhanced long-tail syndrome performance.Conclusion This study proposes Agent-GNN,a knowledge graph-enhanced framework that effectively addresses the long-tail distribution challenge in TCM syndrome differentiation.By explicitly modeling manifestation-mechanism-essence patterns through structured knowledge graphs,our approach achieves superior performance in data-scarce scenarios while providing interpretable reasoning paths for TCM intelligent diagnosis.展开更多
To integrate traditional culture and modern technology,Shandong University’s School of Software has promoted an interdisciplinary teaching project called IYAN&ITAN,the I Ching Knowledge Graph.The project,driven b...To integrate traditional culture and modern technology,Shandong University’s School of Software has promoted an interdisciplinary teaching project called IYAN&ITAN,the I Ching Knowledge Graph.The project,driven by I Ching texts,guides students to practice natural language processing(NLP)and knowledge graph technology in a task-oriented curriculum,based on constructivism,situated learning,and inquiry-based pedagogy,with a progressive and task-oriented teaching model.The platform established enables the retrieval of knowledge,parsing of text,symbolic-numeric analysis,and historical commentary integration,making possible multidimensional,structured representation of I Ching knowledge,and offering an extensible reference for interdisciplinary learning in the context of New Engineering Education.展开更多
Mental-health risk detection seeks early signs of distress from social media posts and clinical transcripts to enable timely intervention before crises.When such risks go undetected,consequences can escalate to self-h...Mental-health risk detection seeks early signs of distress from social media posts and clinical transcripts to enable timely intervention before crises.When such risks go undetected,consequences can escalate to self-harm,long-term disability,reduced productivity,and significant societal and economic burden.Despite recent advances,detecting risk from online text remains challenging due to heterogeneous language,evolving semantics,and the sequential emergence of new datasets.Effective solutions must encode clinically meaningful cues,reason about causal relations,and adapt to new domains without forgetting prior knowledge.To address these challenges,this paper presents a Continual Neuro-Symbolic Graph Learning(CNSGL)framework that unifies symbolic reasoning,causal inference,and continual learning within a single architecture.Each post is represented as a symbolic graph linking clinically relevant tags to textual content,enriched with causal edges derived from directional Point-wise Mutual Information(PMI).A two-layer Graph Convolutional Network(GCN)encodes these graphs,and a Transformer-based attention pooler aggregates node embeddings while providing interpretable tag-level importances.Continual adaptation across datasets is achieved through the Multi-Head Freeze(MH-Freeze)strategy,which freezes a shared encoder and incrementally trains lightweight task-specific heads(small classifiers attached to the shared embedding).Experimental evaluations across six diverse mental-health datasets ranging from Reddit discourse to clinical interviews,demonstrate that MH-Freeze consistently outperforms existing continual-learning baselines in both discriminative accuracy and calibration reliability.Across six datasets,MH-Freeze achieves up to 0.925 accuracy and 0.923 F1-Score,with AUPRC≥0.934 and AUROC≥0.942,consistently surpassing all continual-learning baselines.The results confirm the framework’s ability to preserve prior knowledge,adapt to domain shifts,and maintain causal interpretability,establishing CNSGL as a promising step toward robust,explainable,and lifelong mental-health risk assessment.展开更多
Graph Federated Learning(GFL)has shown great potential in privacy protection and distributed intelligence through distributed collaborative training of graph-structured data without sharing raw information.However,exi...Graph Federated Learning(GFL)has shown great potential in privacy protection and distributed intelligence through distributed collaborative training of graph-structured data without sharing raw information.However,existing GFL approaches often lack the capability for comprehensive feature extraction and adaptive optimization,particularly in non-independent and identically distributed(NON-IID)scenarios where balancing global structural understanding and local node-level detail remains a challenge.To this end,this paper proposes a novel framework called GFL-SAR(Graph Federated Collaborative Learning Framework Based on Structural Amplification and Attention Refinement),which enhances the representation learning capability of graph data through a dual-branch collaborative design.Specifically,we propose the Structural Insight Amplifier(SIA),which utilizes an improved Graph Convolutional Network(GCN)to strengthen structural awareness and improve modeling of topological patterns.In parallel,we propose the Attentive Relational Refiner(ARR),which employs an enhanced Graph Attention Network(GAT)to perform fine-grained modeling of node relationships and neighborhood features,thereby improving the expressiveness of local interactions and preserving critical contextual information.GFL-SAR effectively integrates multi-scale features from every branch via feature fusion and federated optimization,thereby addressing existing GFL limitations in structural modeling and feature representation.Experiments on standard benchmark datasets including Cora,Citeseer,Polblogs,and Cora_ML demonstrate that GFL-SAR achieves superior performance in classification accuracy,convergence speed,and robustness compared to existing methods,confirming its effectiveness and generalizability in GFL tasks.展开更多
Multivariate anomaly detection plays a critical role in maintaining the stable operation of information systems.However,in existing research,multivariate data are often influenced by various factors during the data co...Multivariate anomaly detection plays a critical role in maintaining the stable operation of information systems.However,in existing research,multivariate data are often influenced by various factors during the data collection process,resulting in temporal misalignment or displacement.Due to these factors,the node representations carry substantial noise,which reduces the adaptability of the multivariate coupled network structure and subsequently degrades anomaly detection performance.Accordingly,this study proposes a novel multivariate anomaly detection model grounded in graph structure learning.Firstly,a recommendation strategy is employed to identify strongly coupled variable pairs,which are then used to construct a recommendation-driven multivariate coupling network.Secondly,a multi-channel graph encoding layer is used to dynamically optimize the structural properties of the multivariate coupling network,while a multi-head attention mechanism enhances the spatial characteristics of the multivariate data.Finally,unsupervised anomaly detection is conducted using a dynamic threshold selection algorithm.Experimental results demonstrate that effectively integrating the structural and spatial features of multivariate data significantly mitigates anomalies caused by temporal dependency misalignment.展开更多
Deep graph contrastive clustering has attracted widespread attentions due to its self-supervised representation learning paradigm and superior clustering performance.Although,two challenges emerge and result in high c...Deep graph contrastive clustering has attracted widespread attentions due to its self-supervised representation learning paradigm and superior clustering performance.Although,two challenges emerge and result in high computational costs.Most existing contrastive methods adopt the data augmentation and then representation learning strategy,where representation learning with trainable graph convolution is coupled with complex and fixed data augmentation,inevitably limiting the efficiency and flexibility.The similarity metric between positive-negative sample pairs is complex and contrastive objective is partial,limiting the discriminability of representation learning.To solve these challenges,a novel wide graph clustering network(WGCN)adhering to representation and then augmentation framework is proposed,which mainly consists of multiorder filter fusion(MFF)and double-level contrastive learning(DCL)modules.Specifically,the MFF module integrates multiorder low-pass filters to extract smooth and multi-scale topological features,utilizing self-attention fusion to reduce redundancy and obtain comprehensive embedding representation.Further,the DCL module constructs two augmented views by the parallel parameter-unshared Siamese encoders rather than complex augmentations on graph.To achieve simple yet effective self-supervised learning,representation self-supervision and structural consistency oriented double-level contrastive loss is designed,where representation self-supervision maximizes the agreement between pairwise augmented embedding representations and structural consistency promotes the mutual information correlation between appending neighborhoods with similar semantics.Extensive experiments on six benchmark datasets demonstrate the superiority of the proposed WGCN,especially highlighting its time-saving characteristic.The code could be available in the https://github.com/Tianxiang Zhao0474/WGCN.展开更多
Multimodal spatiotemporal data from smart city consumer electronics present critical challenges including cross-modal temporal misalignment,unreliable data quality,limited joint modeling of spatial and temporal depend...Multimodal spatiotemporal data from smart city consumer electronics present critical challenges including cross-modal temporal misalignment,unreliable data quality,limited joint modeling of spatial and temporal dependencies,and weak resilience to adversarial updates.To address these limitations,EdgeST-Fusion is introduced as a cross-modal federated graph transformer framework for context-aware smart city analytics.The architecture integrates cross-modal embedding networks for modality alignment,graph transformer encoders for spatial dependency modeling,temporal self-attention for dynamic pattern learning,and adaptive anomaly detection to ensure data quality and security during aggregation.A privacy-preserving federated learning protocol with differential privacy guarantees enables collaborative model training without centralizing sensitive data.The framework employs data-quality-aware weighted aggregation to enhance robustness against noisy and malicious client updates.Experimental evaluation on the GeoLife,PeMS-Bay,and SmartHome+datasets demonstrates that EdgeST-Fusion achieves 21.8%improvement in prediction accuracy,35.7%reduction in communication overhead,and 29.4%enhancement in security resilience compared to recent baselines.Real-world deployment across three smart city testbeds validates practical viability with 90.0%average accuracy and sub-250 ms inference latency.The proposed framework remains feasible for deployment on heterogeneous and resource-constrained consumer electronics devices whilemaintaining strong privacy guarantees and scalability for large-scale urban environments.展开更多
Predicting the behavior of renewable energy systems requires models capable of generating accurate forecasts from limited historical data,a challenge that becomes especially pronounced when commissioning new facil-iti...Predicting the behavior of renewable energy systems requires models capable of generating accurate forecasts from limited historical data,a challenge that becomes especially pronounced when commissioning new facil-ities where operational records are scarce.This review aims to synthesize recent progress in data-efficient deep learning approaches for addressing such“cold-start”forecasting problems.It primarily covers three interrelated domains—solar photovoltaic(PV),wind power,and electrical load forecasting—where data scarcity and operational variability are most critical,while also including representative studies on hydropower and carbon emission prediction to provide a broader systems perspective.To this end,we examined trends from over 150 predominantly peer-reviewed studies published between 2019 and mid-2025,highlighting advances in zero-shot and few-shot meta-learning frameworks that enable rapid model adaptation with minimal labeled data.Moreover,transfer learning approaches combined with spatiotemporal graph neural networks have been employed to transfer knowledge from existing energy assets to new,data-sparse environments,effectively capturing hidden dependencies among geographic features,meteorological dynamics,and grid structures.Synthetic data generation has further proven valuable for expanding training samples and mitigating overfitting in cold-start scenarios.In addition,large language models and explainable artificial intelligence(XAI)—notably conversational XAI systems—have been used to interpret and communicate complex model behaviors in accessible terms,fostering operator trust from the earliest deployment stages.By consolidating methodological advances,unresolved challenges,and open-source resources,this review provides a coherent overview of deep learning strategies that can shorten the data-sparse ramp-up period of new energy infrastructures and accelerate the transition toward resilient,low-carbon electricity grids.展开更多
Predicting the productivity of multistage fractured horizontal wells plays an important role in exploiting unconventional resources.In recent years,machine learning(ML)models have emerged as a new approach for such st...Predicting the productivity of multistage fractured horizontal wells plays an important role in exploiting unconventional resources.In recent years,machine learning(ML)models have emerged as a new approach for such studies.However,the scarcity of sufficient real data for model training often leads to imprecise predictions,even though the models trained with real data better characterize geological and engineering features.To tackle this issue,we propose an ML model that can obtain reliable results even with a small amount of data samples.Our model integrates the synthetic minority oversampling technique(SMOTE)to expand the data volume,the support vector machine(SVM)for model training,and the particle swarm optimization(PSO)algorithm for optimizing hyperparameters.To enhance the model performance,we conduct feature fusion and dimensionality reduction.Additionally,we examine the influences of different sample sizes and ML models for training.The proposed model demonstrates higher prediction accuracy and generalization ability,achieving a predicted R^(2)value of up to 0.9 for the test set,compared to the traditional ML techniques with an R^(2)of 0.13.This model accurately predicts the production of fractured horizontal wells even with limited samples,supplying an efficient tool for optimizing the production of unconventional resources.Importantly,the model holds the potential applicability to address similar challenges in other fields constrained by scarce data samples.展开更多
The rapid growth of biomedical data,particularly multi-omics data including genomes,transcriptomics,proteomics,metabolomics,and epigenomics,medical research and clinical decision-making confront both new opportunities...The rapid growth of biomedical data,particularly multi-omics data including genomes,transcriptomics,proteomics,metabolomics,and epigenomics,medical research and clinical decision-making confront both new opportunities and obstacles.The huge and diversified nature of these datasets cannot always be managed using traditional data analysis methods.As a consequence,deep learning has emerged as a strong tool for analysing numerous omics data due to its ability to handle complex and non-linear relationships.This paper explores the fundamental concepts of deep learning and how they are used in multi-omics medical data mining.We demonstrate how autoencoders,variational autoencoders,multimodal models,attention mechanisms,transformers,and graph neural networks enable pattern analysis and recognition across all omics data.Deep learning has been found to be effective in illness classification,biomarker identification,gene network learning,and therapeutic efficacy prediction.We also consider critical problems like as data quality,model explainability,whether findings can be repeated,and computational power requirements.We now consider future elements of combining omics with clinical and imaging data,explainable AI,federated learning,and real-time diagnostics.Overall,this study emphasises the need of collaborating across disciplines to advance deep learning-based multi-omics research for precision medicine and comprehending complicated disorders.展开更多
This paper investigates the challenges associated with Unmanned Aerial Vehicle (UAV) collaborative search and target tracking in dynamic and unknown environments characterized by limited field of view. The primary obj...This paper investigates the challenges associated with Unmanned Aerial Vehicle (UAV) collaborative search and target tracking in dynamic and unknown environments characterized by limited field of view. The primary objective is to explore the unknown environments to locate and track targets effectively. To address this problem, we propose a novel Multi-Agent Reinforcement Learning (MARL) method based on Graph Neural Network (GNN). Firstly, a method is introduced for encoding continuous-space multi-UAV problem data into spatial graphs which establish essential relationships among agents, obstacles, and targets. Secondly, a Graph AttenTion network (GAT) model is presented, which focuses exclusively on adjacent nodes, learns attention weights adaptively and allows agents to better process information in dynamic environments. Reward functions are specifically designed to tackle exploration challenges in environments with sparse rewards. By introducing a framework that integrates centralized training and distributed execution, the advancement of models is facilitated. Simulation results show that the proposed method outperforms the existing MARL method in search rate and tracking performance with less collisions. The experiments show that the proposed method can be extended to applications with a larger number of agents, which provides a potential solution to the challenging problem of multi-UAV autonomous tracking in dynamic unknown environments.展开更多
Image classification is crucial for various applications,including digital construction,smart manu-facturing,and medical imaging.Focusing on the inadequate model generalization and data privacy concerns in few-shot im...Image classification is crucial for various applications,including digital construction,smart manu-facturing,and medical imaging.Focusing on the inadequate model generalization and data privacy concerns in few-shot image classification,in this paper,we propose a federated learning approach that incorporates privacy-preserving techniques.First,we utilize contrastive learning to train on local few-shot image data and apply various data augmentation methods to expand the sample size,thereby enhancing the model’s generalization capabilities in few-shot contexts.Second,we introduce local differential privacy techniques and weight pruning methods to safeguard model parameters,perturbing the transmitted parameters to ensure user data privacy.Finally,numerical simulations are conducted to demonstrate the effectiveness of our proposed method.The results indicate that our approach significantly enhances model generalization and test accuracy compared to several popular federated learning algorithms while maintaining data privacy,highlighting its effectiveness and practicality in addressing the challenges of model generalization and data privacy in few-shot image scenarios.展开更多
Although conventional object detection methods achieve high accuracy through extensively annotated datasets,acquiring such large-scale labeled data remains challenging and cost-prohibitive in numerous real-world appli...Although conventional object detection methods achieve high accuracy through extensively annotated datasets,acquiring such large-scale labeled data remains challenging and cost-prohibitive in numerous real-world applications.Few-shot object detection presents a new research idea that aims to localize and classify objects in images using only limited annotated examples.However,the inherent challenge in few-shot object detection lies in the insufficient sample diversity to fully characterize the sample feature distribution,which consequently impacts model performance.Inspired by contrastive learning principles,we propose an Implicit Feature Contrastive Learning(IFCL)module to address this limitation and augment feature diversity for more robust representational learning.This module generates augmented support sample features in a mixed feature space and implicitly contrasts them with query Region of Interest(RoI)features.This approach facilitates more comprehensive learning of both intra-class feature similarity and inter-class feature diversity,thereby enhancing the model’s object classification and localization capabilities.Extensive experiments on PASCAL VOC show that our method achieves a respective improvement of 3.2%,1.8%,and 2.3%on 10-shot of three Novel Sets compared to the baseline model FPD.展开更多
Federated Graph Learning (FGL) enables model training without requiring each client to share local graph data, effectively breaking data silos by aggregating the training parameters from each terminal while safeguardi...Federated Graph Learning (FGL) enables model training without requiring each client to share local graph data, effectively breaking data silos by aggregating the training parameters from each terminal while safeguarding data privacy. Traditional FGL relies on a centralized server for model aggregation;however, this central server presents challenges such as a single point of failure and high communication overhead. Additionally, efficiently training a robust personalized local model for each client remains a significant objective in federated graph learning. To address these issues, we propose a decentralized Federated Graph Learning framework with efficient communication, termed Decentralized Federated Graph Learning via Surrogate Model (SD_FGL). In SD_FGL, each client is required to maintain two models: a private model and a surrogate model. The surrogate model is publicly shared and can exchange and update information directly with any client, eliminating the need for a central server and reducing communication overhead. The private model is independently trained by each client, allowing it to calculate similarity with other clients based on local data as well as information shared through the surrogate model. This enables the private model to better adjust its training strategy and selectively update its parameters. Additionally, local differential privacy is incorporated into the surrogate model training process to enhance privacy protection. Testing on three real-world graph datasets demonstrates that the proposed framework improves accuracy while achieving decentralized Federated Graph Learning with lower communication overhead and stronger privacy safeguards.展开更多
Air quality estimation assesses the pollution level in the air,supports public health warnings,and is a valuable tool in environmental management.Although air sensors have proven helpful in this task,sensors are often...Air quality estimation assesses the pollution level in the air,supports public health warnings,and is a valuable tool in environmental management.Although air sensors have proven helpful in this task,sensors are often expensive and difficult to install,while cameras are becoming more popular and accessible,from which images can be collected as data for deep learning models to solve the above task.This leads to another problem:several labeled images are needed to achieve high accuracy when deep-learningmodels predict air quality.In this research,we have threemain contributions:(1)Collect and publish an air quality estimation dataset,namely PTIT_AQED,including environmental image data and air quality;(2)Propose a deep learning model to predict air quality with few data,called PTIT_FAQE(PTIT Few-shot air quality estimation).We build PTIT_FAQE based on EfficientNet-a CNN architecture that ensures high performance in deep learning applications and Few-shot Learning with Prototypical Networks.This helps the model use only a fewtraining data but still achieve high accuracy in air quality estimation.And(3)conduct experiments to prove the superiority of PTIT_FAQE compared to other studies on both PTIT_AQED and APIN datasets.The results show that our model achieves an accuracy of 0.9278 and an F1-Score of 0.9139 on the PTIT_AQED dataset and an accuracy of 0.9467 and an F1-Score of 0.9371 on the APIN dataset,which demonstrate a significant performance improvement compared to previous studies.We also conduct detailed experiments to evaluate the impact of each component on model performance.展开更多
Computational analysis can accurately detect drug-gene interactions(DGIs)cost-effectively.However,transductive learning models are the hotspot to reveal the promising performance for unknown DGIs(both drugs and genes ...Computational analysis can accurately detect drug-gene interactions(DGIs)cost-effectively.However,transductive learning models are the hotspot to reveal the promising performance for unknown DGIs(both drugs and genes are present in the training model),without special attention to the unseen DGIs(both drugs and genes are absent in the training model).In view of this,this study,for the first time,proposed an inductive learning-based model for the precise identification of unseen DGIs.In our study,by integrating disease nodes to avoid data sparsity,a multi-relational drug-disease-gene(DDG)graph was constructed to achieve effective fusion of data on DDG intro-relationships and inter-actions.Following the extraction of graph features by utilizing graph embedding algorithms,our next step was the retrieval of the attributes of individual gene and drug nodes.In this way,a hybrid feature characterization was represented by integrating graph features and node attributes.Machine learning(ML)models were built,enabling the fulfillment of transductive predictions of unknown DGIs.To realize inductive learning,this study generated an innovative idea of transforming known node vectors derived from the DDG graph into representations of unseen nodes using node similarities as weights,enabling inductive predictions for the unseen DGIs.Consequently,the final model was superior to existing models,with significant improvement in predicting both external unknown and unseen DGIs.The practical feasibility of our model was further confirmed through case study and molecular docking.In summary,this study establishes an efficient data-driven approach through the proposed modeling,suggesting its value as a promising tool for accelerating drug discovery and repurposing.展开更多
Few-shot point cloud 3D object detection(FS3D)aims to identify and locate objects of novel classes within point clouds using knowledge acquired from annotated base classes and a minimal number of samples from the nove...Few-shot point cloud 3D object detection(FS3D)aims to identify and locate objects of novel classes within point clouds using knowledge acquired from annotated base classes and a minimal number of samples from the novel classes.Due to imbalanced training data,existing FS3D methods based on fully supervised learning can lead to overfitting toward base classes,which impairs the network’s ability to generalize knowledge learned from base classes to novel classes and also prevents the network from extracting distinctive foreground and background representations for novel class objects.To address these issues,this thesis proposes a category-agnostic contrastive learning approach,enhancing the generalization and identification abilities for almost unseen categories through the construction of pseudo-labels and positive-negative sample pairs unrelated to specific classes.Firstly,this thesis designs a proposal-wise context contrastive module(CCM).By reducing the distance between foreground point features and increasing the distance between foreground and background point features within a region proposal,CCM aids the network in extracting more discriminative foreground and background feature representations without reliance on categorical annotations.Secondly,this thesis utilizes a geometric contrastive module(GCM),which enhances the network’s geometric perception capability by employing contrastive learning on the foreground point features associated with various basic geometric components,such as edges,corners,and surfaces,thereby enabling these geometric components to exhibit more distinguishable representations.This thesis also combines category-aware contrastive learning with former modules to maintain categorical distinctiveness.Extensive experimental results on FS-SUNRGBD and FS-ScanNet datasets demonstrate the effectiveness of this method with average precision exceeding the baseline by up to 8%.展开更多
Graph similarity learning aims to calculate the similarity between pairs of graphs.Existing unsupervised graph similarity learning methods based on contrastive learning encounter challenges related to random graph aug...Graph similarity learning aims to calculate the similarity between pairs of graphs.Existing unsupervised graph similarity learning methods based on contrastive learning encounter challenges related to random graph augmentation strategies,which can harm the semantic and structural information of graphs and overlook the rich structural information present in subgraphs.To address these issues,we propose a graph similarity learning model based on learnable augmentation and multi-level contrastive learning.First,to tackle the problem of random augmentation disrupting the semantics and structure of the graph,we design a learnable augmentation method to selectively choose nodes and edges within the graph.To enhance contrastive levels,we employ a biased random walk method to generate corresponding subgraphs,enriching the contrastive hierarchy.Second,to solve the issue of previous work not considering multi-level contrastive learning,we utilize graph convolutional networks to learn node representations of augmented views and the original graph and calculate the interaction information between the attribute-augmented and structure-augmented views and the original graph.The goal is to maximize node consistency between different views and learn node matching between different graphs,resulting in node-level representations for each graph.Subgraph representations are then obtained through pooling operations,and we conduct contrastive learning utilizing both node and subgraph representations.Finally,the graph similarity score is computed according to different downstream tasks.We conducted three sets of experiments across eight datasets,and the results demonstrate that the proposed model effectively mitigates the issues of random augmentation damaging the original graph’s semantics and structure,as well as the insufficiency of contrastive levels.Additionally,the model achieves the best overall performance.展开更多
Full ceramic bearings are mission-critical components in oil-free environments,such as food processing,semiconductor manufacturing,and medical applications.Developing effective fault diagnosis methods for these bearin...Full ceramic bearings are mission-critical components in oil-free environments,such as food processing,semiconductor manufacturing,and medical applications.Developing effective fault diagnosis methods for these bearings is essential to ensuring operational reliability and preventing costly failures.Traditional supervised deep learning approaches have demonstrated promise in fault detection,but their dependence on large labeled datasets poses significant challenges in industrial settings where fault-labeled data is scarce.This paper introduces a few-shot learning approach for full ceramic bearing fault diagnosis by leveraging the pre-trained GPT-2 model.Large language models(LLMs)like GPT-2,pre-trained on diverse textual data,exhibit remarkable transfer learning and few-shot learning capabilities,making them ideal for applications with limited labeled data.In this study,acoustic emission(AE)signals from bearings were processed using empirical mode decomposition(EMD),and the extracted AE features were converted into structured text for fine-tuning GPT-2 as a fault classifier.To enhance its performance,we incorporated a modified loss function and softmax activation with cosine similarity,ensuring better generalization in fault identification.Experimental evaluations on a laboratory-collected full ceramic bearing dataset demonstrated that the proposed approach achieved high diagnostic accuracy with as few as five labeled samples,outperforming conventional methods such as k-nearest neighbor(KNN),large memory storage and retrieval(LAMSTAR)neural network,deep neural network(DNN),recurrent neural network(RNN),long short-term memory(LSTM)network,and model-agnostic meta-learning(MAML).The results highlight LLMs’potential to revolutionize fault diagnosis,enabling faster deployment,reduced reliance on extensive labeled datasets,and improved adaptability in industrial monitoring systems.展开更多
基金supported in part by the National Natural Science Foundation of China(62472146)the Key Technologies Research Development Joint Foundation of Henan Province of China(225101610001)。
文摘Deep transfer learning has achieved significant success in anomaly detection over the past decade,but data acquisition challenges in practical engineering hinder high-quality feature representation for few-shot learning tasks.To address this issue,a novel time-frequency-assisted deep feature enhancement(TFE)mechanism is proposed.Unlike traditional methods that integrate time-frequency analysis with deep neural networks,TFE employs a wavelet scattering transform to establish a parallel time-frequency feature space,where a dual interaction strategy facilitates collaboration between deep feature and time-frequency spaces through two operations:1)Enhancement,where a frequency-importance-driven contrastive learning(FICL)network transfers physically-aware information from wavelet scattering features to deep features,and 2)Feedback,which uses a detection rule adaptation module to minimize bias in wavelet scattering features based on deep feature performance.TFE is applied to a domain-adversarial anomaly detection framework and,through alternating training,significantly enhances both deep feature discriminative power and few-shot anomaly detection.Theoretical analysis confirms that the proposed dual interaction strategy reduces the upper bound of classification error.Experiments on benchmark datasets and a real-world industrial dataset from a large steel factory demonstrate TFE's superior performance and highlight the importance of frequency saliency in transfer learning.Thus,collaboration is shown to outperform integration for few-shot transfer learning in anomaly detection.
基金Sichuan TCM Culture Coordinated Development Research Center Project(2023XT131)National Key Science and Technology Project of China(2023ZD0509405)National Natural Science Foundation of China(82174236).
文摘Objective To address the dual challenges of long-tail distribution and feature sparsity in traditional Chinese medicine(TCM)syndrome differentiation within real clinical settings,we propose a data-efficient learning framework enhanced by knowledge graphs.Methods We developed Agent-GNN,a three-stage decoupled learning framework,and validated it on the Traditional Chinese Medicine Syndrome Diagnosis(TCM-SD)dataset containing 54152 clinical records across 148 syndrome categories.First,we constructed a comprehensive medical knowledge graph encoding the complete TCM reasoning system.Second,we proposed a Functional Patient Profiling(FPP)method that utilizes large language models(LLMs)combined with Graph Retrieval-Augmented Generation(RAG)to extract structured symptom-etiology-pathogenesis subgraphs from medical records.Third,we employed heterogeneous graph neural networks to learn structured combination patterns explicitly.We compared our method against multiple baselines including BERT,ZY-BERT,ZY-BERT+Know,GAT,and GPT-4 Few-shot,using macro-F1 score as the primary evaluation metric.Additionally,ablation experiments were conducted to validate the contribution of each key component to model performance.Results Agent-GNN achieved an overall macro-F1 score of 72.4%,representing an 8.7 percentage points improvement over ZY-BERT+Know(63.7%),the strongest baseline among traditional methods.For long-tail syndromes with fewer than 10 samples,Agent-GNN reached a macro-F1 score of 58.6%,compared with 39.3%for ZY-BERT+Know and 41.2%for GPT-4 Few-shot,representing relative improvements of 49.2%and 42.2%,respectively.Ablation experiments confirmed that the explicit modeling of etiology-pathogenesis nodes contributed 12.4 percentage points to this enhanced long-tail syndrome performance.Conclusion This study proposes Agent-GNN,a knowledge graph-enhanced framework that effectively addresses the long-tail distribution challenge in TCM syndrome differentiation.By explicitly modeling manifestation-mechanism-essence patterns through structured knowledge graphs,our approach achieves superior performance in data-scarce scenarios while providing interpretable reasoning paths for TCM intelligent diagnosis.
基金support provided by the Shandong University Education and Teaching Reform Research Project(2024Y232)the“New 20 Regulations for Universities”funding program of Jinan(202228089).
文摘To integrate traditional culture and modern technology,Shandong University’s School of Software has promoted an interdisciplinary teaching project called IYAN&ITAN,the I Ching Knowledge Graph.The project,driven by I Ching texts,guides students to practice natural language processing(NLP)and knowledge graph technology in a task-oriented curriculum,based on constructivism,situated learning,and inquiry-based pedagogy,with a progressive and task-oriented teaching model.The platform established enables the retrieval of knowledge,parsing of text,symbolic-numeric analysis,and historical commentary integration,making possible multidimensional,structured representation of I Ching knowledge,and offering an extensible reference for interdisciplinary learning in the context of New Engineering Education.
基金supported by the National Research Foundation of Korea(NRF)grant funded by the Korea government(MSIT)(RS-2025-00518960)in part by the National Research Foundation of Korea(NRF)grant funded by the Korea government(MSIT)(RS-2025-00563192).
文摘Mental-health risk detection seeks early signs of distress from social media posts and clinical transcripts to enable timely intervention before crises.When such risks go undetected,consequences can escalate to self-harm,long-term disability,reduced productivity,and significant societal and economic burden.Despite recent advances,detecting risk from online text remains challenging due to heterogeneous language,evolving semantics,and the sequential emergence of new datasets.Effective solutions must encode clinically meaningful cues,reason about causal relations,and adapt to new domains without forgetting prior knowledge.To address these challenges,this paper presents a Continual Neuro-Symbolic Graph Learning(CNSGL)framework that unifies symbolic reasoning,causal inference,and continual learning within a single architecture.Each post is represented as a symbolic graph linking clinically relevant tags to textual content,enriched with causal edges derived from directional Point-wise Mutual Information(PMI).A two-layer Graph Convolutional Network(GCN)encodes these graphs,and a Transformer-based attention pooler aggregates node embeddings while providing interpretable tag-level importances.Continual adaptation across datasets is achieved through the Multi-Head Freeze(MH-Freeze)strategy,which freezes a shared encoder and incrementally trains lightweight task-specific heads(small classifiers attached to the shared embedding).Experimental evaluations across six diverse mental-health datasets ranging from Reddit discourse to clinical interviews,demonstrate that MH-Freeze consistently outperforms existing continual-learning baselines in both discriminative accuracy and calibration reliability.Across six datasets,MH-Freeze achieves up to 0.925 accuracy and 0.923 F1-Score,with AUPRC≥0.934 and AUROC≥0.942,consistently surpassing all continual-learning baselines.The results confirm the framework’s ability to preserve prior knowledge,adapt to domain shifts,and maintain causal interpretability,establishing CNSGL as a promising step toward robust,explainable,and lifelong mental-health risk assessment.
基金supported by National Natural Science Foundation of China(62466045)Inner Mongolia Natural Science Foundation Project(2021LHMS06003)Inner Mongolia University Basic Research Business Fee Project(114).
文摘Graph Federated Learning(GFL)has shown great potential in privacy protection and distributed intelligence through distributed collaborative training of graph-structured data without sharing raw information.However,existing GFL approaches often lack the capability for comprehensive feature extraction and adaptive optimization,particularly in non-independent and identically distributed(NON-IID)scenarios where balancing global structural understanding and local node-level detail remains a challenge.To this end,this paper proposes a novel framework called GFL-SAR(Graph Federated Collaborative Learning Framework Based on Structural Amplification and Attention Refinement),which enhances the representation learning capability of graph data through a dual-branch collaborative design.Specifically,we propose the Structural Insight Amplifier(SIA),which utilizes an improved Graph Convolutional Network(GCN)to strengthen structural awareness and improve modeling of topological patterns.In parallel,we propose the Attentive Relational Refiner(ARR),which employs an enhanced Graph Attention Network(GAT)to perform fine-grained modeling of node relationships and neighborhood features,thereby improving the expressiveness of local interactions and preserving critical contextual information.GFL-SAR effectively integrates multi-scale features from every branch via feature fusion and federated optimization,thereby addressing existing GFL limitations in structural modeling and feature representation.Experiments on standard benchmark datasets including Cora,Citeseer,Polblogs,and Cora_ML demonstrate that GFL-SAR achieves superior performance in classification accuracy,convergence speed,and robustness compared to existing methods,confirming its effectiveness and generalizability in GFL tasks.
基金supported by Natural Science Foundation of Qinghai Province(2025-ZJ-994M)Scientific Research Innovation Capability Support Project for Young Faculty(SRICSPYF-BS2025007)National Natural Science Foundation of China(62566050).
文摘Multivariate anomaly detection plays a critical role in maintaining the stable operation of information systems.However,in existing research,multivariate data are often influenced by various factors during the data collection process,resulting in temporal misalignment or displacement.Due to these factors,the node representations carry substantial noise,which reduces the adaptability of the multivariate coupled network structure and subsequently degrades anomaly detection performance.Accordingly,this study proposes a novel multivariate anomaly detection model grounded in graph structure learning.Firstly,a recommendation strategy is employed to identify strongly coupled variable pairs,which are then used to construct a recommendation-driven multivariate coupling network.Secondly,a multi-channel graph encoding layer is used to dynamically optimize the structural properties of the multivariate coupling network,while a multi-head attention mechanism enhances the spatial characteristics of the multivariate data.Finally,unsupervised anomaly detection is conducted using a dynamic threshold selection algorithm.Experimental results demonstrate that effectively integrating the structural and spatial features of multivariate data significantly mitigates anomalies caused by temporal dependency misalignment.
基金supported by the National Natural Science Foundation of China(62225303,62403043,62433004)the Beijing Natural Science Foundation(4244085)+1 种基金the Postdoctoral Fellowship Program of China Postdoctoral Science Foundation(GZC20230203)the China Postdoctoral Science Foundation(2023M740201)。
文摘Deep graph contrastive clustering has attracted widespread attentions due to its self-supervised representation learning paradigm and superior clustering performance.Although,two challenges emerge and result in high computational costs.Most existing contrastive methods adopt the data augmentation and then representation learning strategy,where representation learning with trainable graph convolution is coupled with complex and fixed data augmentation,inevitably limiting the efficiency and flexibility.The similarity metric between positive-negative sample pairs is complex and contrastive objective is partial,limiting the discriminability of representation learning.To solve these challenges,a novel wide graph clustering network(WGCN)adhering to representation and then augmentation framework is proposed,which mainly consists of multiorder filter fusion(MFF)and double-level contrastive learning(DCL)modules.Specifically,the MFF module integrates multiorder low-pass filters to extract smooth and multi-scale topological features,utilizing self-attention fusion to reduce redundancy and obtain comprehensive embedding representation.Further,the DCL module constructs two augmented views by the parallel parameter-unshared Siamese encoders rather than complex augmentations on graph.To achieve simple yet effective self-supervised learning,representation self-supervision and structural consistency oriented double-level contrastive loss is designed,where representation self-supervision maximizes the agreement between pairwise augmented embedding representations and structural consistency promotes the mutual information correlation between appending neighborhoods with similar semantics.Extensive experiments on six benchmark datasets demonstrate the superiority of the proposed WGCN,especially highlighting its time-saving characteristic.The code could be available in the https://github.com/Tianxiang Zhao0474/WGCN.
基金supported by the University of Tabuk,Saudi Arabia。
文摘Multimodal spatiotemporal data from smart city consumer electronics present critical challenges including cross-modal temporal misalignment,unreliable data quality,limited joint modeling of spatial and temporal dependencies,and weak resilience to adversarial updates.To address these limitations,EdgeST-Fusion is introduced as a cross-modal federated graph transformer framework for context-aware smart city analytics.The architecture integrates cross-modal embedding networks for modality alignment,graph transformer encoders for spatial dependency modeling,temporal self-attention for dynamic pattern learning,and adaptive anomaly detection to ensure data quality and security during aggregation.A privacy-preserving federated learning protocol with differential privacy guarantees enables collaborative model training without centralizing sensitive data.The framework employs data-quality-aware weighted aggregation to enhance robustness against noisy and malicious client updates.Experimental evaluation on the GeoLife,PeMS-Bay,and SmartHome+datasets demonstrates that EdgeST-Fusion achieves 21.8%improvement in prediction accuracy,35.7%reduction in communication overhead,and 29.4%enhancement in security resilience compared to recent baselines.Real-world deployment across three smart city testbeds validates practical viability with 90.0%average accuracy and sub-250 ms inference latency.The proposed framework remains feasible for deployment on heterogeneous and resource-constrained consumer electronics devices whilemaintaining strong privacy guarantees and scalability for large-scale urban environments.
文摘Predicting the behavior of renewable energy systems requires models capable of generating accurate forecasts from limited historical data,a challenge that becomes especially pronounced when commissioning new facil-ities where operational records are scarce.This review aims to synthesize recent progress in data-efficient deep learning approaches for addressing such“cold-start”forecasting problems.It primarily covers three interrelated domains—solar photovoltaic(PV),wind power,and electrical load forecasting—where data scarcity and operational variability are most critical,while also including representative studies on hydropower and carbon emission prediction to provide a broader systems perspective.To this end,we examined trends from over 150 predominantly peer-reviewed studies published between 2019 and mid-2025,highlighting advances in zero-shot and few-shot meta-learning frameworks that enable rapid model adaptation with minimal labeled data.Moreover,transfer learning approaches combined with spatiotemporal graph neural networks have been employed to transfer knowledge from existing energy assets to new,data-sparse environments,effectively capturing hidden dependencies among geographic features,meteorological dynamics,and grid structures.Synthetic data generation has further proven valuable for expanding training samples and mitigating overfitting in cold-start scenarios.In addition,large language models and explainable artificial intelligence(XAI)—notably conversational XAI systems—have been used to interpret and communicate complex model behaviors in accessible terms,fostering operator trust from the earliest deployment stages.By consolidating methodological advances,unresolved challenges,and open-source resources,this review provides a coherent overview of deep learning strategies that can shorten the data-sparse ramp-up period of new energy infrastructures and accelerate the transition toward resilient,low-carbon electricity grids.
基金supported by the National Natural Science Foundation of China(52274055)the Shandong Provincial Natural Science Foundation(ZR2022YQ50)the Taishan Scholar Program of Shandong Province(tsqn202408088)。
文摘Predicting the productivity of multistage fractured horizontal wells plays an important role in exploiting unconventional resources.In recent years,machine learning(ML)models have emerged as a new approach for such studies.However,the scarcity of sufficient real data for model training often leads to imprecise predictions,even though the models trained with real data better characterize geological and engineering features.To tackle this issue,we propose an ML model that can obtain reliable results even with a small amount of data samples.Our model integrates the synthetic minority oversampling technique(SMOTE)to expand the data volume,the support vector machine(SVM)for model training,and the particle swarm optimization(PSO)algorithm for optimizing hyperparameters.To enhance the model performance,we conduct feature fusion and dimensionality reduction.Additionally,we examine the influences of different sample sizes and ML models for training.The proposed model demonstrates higher prediction accuracy and generalization ability,achieving a predicted R^(2)value of up to 0.9 for the test set,compared to the traditional ML techniques with an R^(2)of 0.13.This model accurately predicts the production of fractured horizontal wells even with limited samples,supplying an efficient tool for optimizing the production of unconventional resources.Importantly,the model holds the potential applicability to address similar challenges in other fields constrained by scarce data samples.
文摘The rapid growth of biomedical data,particularly multi-omics data including genomes,transcriptomics,proteomics,metabolomics,and epigenomics,medical research and clinical decision-making confront both new opportunities and obstacles.The huge and diversified nature of these datasets cannot always be managed using traditional data analysis methods.As a consequence,deep learning has emerged as a strong tool for analysing numerous omics data due to its ability to handle complex and non-linear relationships.This paper explores the fundamental concepts of deep learning and how they are used in multi-omics medical data mining.We demonstrate how autoencoders,variational autoencoders,multimodal models,attention mechanisms,transformers,and graph neural networks enable pattern analysis and recognition across all omics data.Deep learning has been found to be effective in illness classification,biomarker identification,gene network learning,and therapeutic efficacy prediction.We also consider critical problems like as data quality,model explainability,whether findings can be repeated,and computational power requirements.We now consider future elements of combining omics with clinical and imaging data,explainable AI,federated learning,and real-time diagnostics.Overall,this study emphasises the need of collaborating across disciplines to advance deep learning-based multi-omics research for precision medicine and comprehending complicated disorders.
基金supported by the National Natural Science Foundation of China(Nos.12272104,U22B2013).
文摘This paper investigates the challenges associated with Unmanned Aerial Vehicle (UAV) collaborative search and target tracking in dynamic and unknown environments characterized by limited field of view. The primary objective is to explore the unknown environments to locate and track targets effectively. To address this problem, we propose a novel Multi-Agent Reinforcement Learning (MARL) method based on Graph Neural Network (GNN). Firstly, a method is introduced for encoding continuous-space multi-UAV problem data into spatial graphs which establish essential relationships among agents, obstacles, and targets. Secondly, a Graph AttenTion network (GAT) model is presented, which focuses exclusively on adjacent nodes, learns attention weights adaptively and allows agents to better process information in dynamic environments. Reward functions are specifically designed to tackle exploration challenges in environments with sparse rewards. By introducing a framework that integrates centralized training and distributed execution, the advancement of models is facilitated. Simulation results show that the proposed method outperforms the existing MARL method in search rate and tracking performance with less collisions. The experiments show that the proposed method can be extended to applications with a larger number of agents, which provides a potential solution to the challenging problem of multi-UAV autonomous tracking in dynamic unknown environments.
基金supported by Suzhou Science and Technology Plan(Basic Research)Project under Grant SJC2023002Postgraduate Research&Practice Innovation Program of Jiangsu Province under Grant KYCX23_3322.
文摘Image classification is crucial for various applications,including digital construction,smart manu-facturing,and medical imaging.Focusing on the inadequate model generalization and data privacy concerns in few-shot image classification,in this paper,we propose a federated learning approach that incorporates privacy-preserving techniques.First,we utilize contrastive learning to train on local few-shot image data and apply various data augmentation methods to expand the sample size,thereby enhancing the model’s generalization capabilities in few-shot contexts.Second,we introduce local differential privacy techniques and weight pruning methods to safeguard model parameters,perturbing the transmitted parameters to ensure user data privacy.Finally,numerical simulations are conducted to demonstrate the effectiveness of our proposed method.The results indicate that our approach significantly enhances model generalization and test accuracy compared to several popular federated learning algorithms while maintaining data privacy,highlighting its effectiveness and practicality in addressing the challenges of model generalization and data privacy in few-shot image scenarios.
基金funded by the China Chongqing Municipal Science and Technology Bureau,grant numbers CSTB2024TIAD-CYKJCXX0009,CSTB2024NSCQ-LZX0043,CSTB2022NSCQ-MSX0288Chongqing Municipal Commission of Housing and Urban-Rural Development,grant number CKZ2024-87+3 种基金the Chongqing University of Technology Graduate Education High-Quality Development Project,grant number gzlsz202401the Chongqing University of Technology—Chongqing LINGLUE Technology Co.,Ltd.Electronic Information(Artificial Intelligence)Graduate Joint Training Basethe Postgraduate Education and Teaching Reform Research Project in Chongqing,grant number yjg213116the Chongqing University of Technology-CISDI Chongqing Information Technology Co.,Ltd.Computer Technology Graduate Joint Training Base.
文摘Although conventional object detection methods achieve high accuracy through extensively annotated datasets,acquiring such large-scale labeled data remains challenging and cost-prohibitive in numerous real-world applications.Few-shot object detection presents a new research idea that aims to localize and classify objects in images using only limited annotated examples.However,the inherent challenge in few-shot object detection lies in the insufficient sample diversity to fully characterize the sample feature distribution,which consequently impacts model performance.Inspired by contrastive learning principles,we propose an Implicit Feature Contrastive Learning(IFCL)module to address this limitation and augment feature diversity for more robust representational learning.This module generates augmented support sample features in a mixed feature space and implicitly contrasts them with query Region of Interest(RoI)features.This approach facilitates more comprehensive learning of both intra-class feature similarity and inter-class feature diversity,thereby enhancing the model’s object classification and localization capabilities.Extensive experiments on PASCAL VOC show that our method achieves a respective improvement of 3.2%,1.8%,and 2.3%on 10-shot of three Novel Sets compared to the baseline model FPD.
基金supported by InnerMongolia Natural Science Foundation Project(2021LHMS06003)Inner Mongolia University Basic Research Business Fee Project(114).
文摘Federated Graph Learning (FGL) enables model training without requiring each client to share local graph data, effectively breaking data silos by aggregating the training parameters from each terminal while safeguarding data privacy. Traditional FGL relies on a centralized server for model aggregation;however, this central server presents challenges such as a single point of failure and high communication overhead. Additionally, efficiently training a robust personalized local model for each client remains a significant objective in federated graph learning. To address these issues, we propose a decentralized Federated Graph Learning framework with efficient communication, termed Decentralized Federated Graph Learning via Surrogate Model (SD_FGL). In SD_FGL, each client is required to maintain two models: a private model and a surrogate model. The surrogate model is publicly shared and can exchange and update information directly with any client, eliminating the need for a central server and reducing communication overhead. The private model is independently trained by each client, allowing it to calculate similarity with other clients based on local data as well as information shared through the surrogate model. This enables the private model to better adjust its training strategy and selectively update its parameters. Additionally, local differential privacy is incorporated into the surrogate model training process to enhance privacy protection. Testing on three real-world graph datasets demonstrates that the proposed framework improves accuracy while achieving decentralized Federated Graph Learning with lower communication overhead and stronger privacy safeguards.
文摘Air quality estimation assesses the pollution level in the air,supports public health warnings,and is a valuable tool in environmental management.Although air sensors have proven helpful in this task,sensors are often expensive and difficult to install,while cameras are becoming more popular and accessible,from which images can be collected as data for deep learning models to solve the above task.This leads to another problem:several labeled images are needed to achieve high accuracy when deep-learningmodels predict air quality.In this research,we have threemain contributions:(1)Collect and publish an air quality estimation dataset,namely PTIT_AQED,including environmental image data and air quality;(2)Propose a deep learning model to predict air quality with few data,called PTIT_FAQE(PTIT Few-shot air quality estimation).We build PTIT_FAQE based on EfficientNet-a CNN architecture that ensures high performance in deep learning applications and Few-shot Learning with Prototypical Networks.This helps the model use only a fewtraining data but still achieve high accuracy in air quality estimation.And(3)conduct experiments to prove the superiority of PTIT_FAQE compared to other studies on both PTIT_AQED and APIN datasets.The results show that our model achieves an accuracy of 0.9278 and an F1-Score of 0.9139 on the PTIT_AQED dataset and an accuracy of 0.9467 and an F1-Score of 0.9371 on the APIN dataset,which demonstrate a significant performance improvement compared to previous studies.We also conduct detailed experiments to evaluate the impact of each component on model performance.
基金funded by the National Natural Science Foundation of China(Grant No.:22173065)the Sichuan International Science and Technology Innovation Cooperation Project,China(Grant No.:24GJHZ0431).
文摘Computational analysis can accurately detect drug-gene interactions(DGIs)cost-effectively.However,transductive learning models are the hotspot to reveal the promising performance for unknown DGIs(both drugs and genes are present in the training model),without special attention to the unseen DGIs(both drugs and genes are absent in the training model).In view of this,this study,for the first time,proposed an inductive learning-based model for the precise identification of unseen DGIs.In our study,by integrating disease nodes to avoid data sparsity,a multi-relational drug-disease-gene(DDG)graph was constructed to achieve effective fusion of data on DDG intro-relationships and inter-actions.Following the extraction of graph features by utilizing graph embedding algorithms,our next step was the retrieval of the attributes of individual gene and drug nodes.In this way,a hybrid feature characterization was represented by integrating graph features and node attributes.Machine learning(ML)models were built,enabling the fulfillment of transductive predictions of unknown DGIs.To realize inductive learning,this study generated an innovative idea of transforming known node vectors derived from the DDG graph into representations of unseen nodes using node similarities as weights,enabling inductive predictions for the unseen DGIs.Consequently,the final model was superior to existing models,with significant improvement in predicting both external unknown and unseen DGIs.The practical feasibility of our model was further confirmed through case study and molecular docking.In summary,this study establishes an efficient data-driven approach through the proposed modeling,suggesting its value as a promising tool for accelerating drug discovery and repurposing.
文摘Few-shot point cloud 3D object detection(FS3D)aims to identify and locate objects of novel classes within point clouds using knowledge acquired from annotated base classes and a minimal number of samples from the novel classes.Due to imbalanced training data,existing FS3D methods based on fully supervised learning can lead to overfitting toward base classes,which impairs the network’s ability to generalize knowledge learned from base classes to novel classes and also prevents the network from extracting distinctive foreground and background representations for novel class objects.To address these issues,this thesis proposes a category-agnostic contrastive learning approach,enhancing the generalization and identification abilities for almost unseen categories through the construction of pseudo-labels and positive-negative sample pairs unrelated to specific classes.Firstly,this thesis designs a proposal-wise context contrastive module(CCM).By reducing the distance between foreground point features and increasing the distance between foreground and background point features within a region proposal,CCM aids the network in extracting more discriminative foreground and background feature representations without reliance on categorical annotations.Secondly,this thesis utilizes a geometric contrastive module(GCM),which enhances the network’s geometric perception capability by employing contrastive learning on the foreground point features associated with various basic geometric components,such as edges,corners,and surfaces,thereby enabling these geometric components to exhibit more distinguishable representations.This thesis also combines category-aware contrastive learning with former modules to maintain categorical distinctiveness.Extensive experimental results on FS-SUNRGBD and FS-ScanNet datasets demonstrate the effectiveness of this method with average precision exceeding the baseline by up to 8%.
文摘Graph similarity learning aims to calculate the similarity between pairs of graphs.Existing unsupervised graph similarity learning methods based on contrastive learning encounter challenges related to random graph augmentation strategies,which can harm the semantic and structural information of graphs and overlook the rich structural information present in subgraphs.To address these issues,we propose a graph similarity learning model based on learnable augmentation and multi-level contrastive learning.First,to tackle the problem of random augmentation disrupting the semantics and structure of the graph,we design a learnable augmentation method to selectively choose nodes and edges within the graph.To enhance contrastive levels,we employ a biased random walk method to generate corresponding subgraphs,enriching the contrastive hierarchy.Second,to solve the issue of previous work not considering multi-level contrastive learning,we utilize graph convolutional networks to learn node representations of augmented views and the original graph and calculate the interaction information between the attribute-augmented and structure-augmented views and the original graph.The goal is to maximize node consistency between different views and learn node matching between different graphs,resulting in node-level representations for each graph.Subgraph representations are then obtained through pooling operations,and we conduct contrastive learning utilizing both node and subgraph representations.Finally,the graph similarity score is computed according to different downstream tasks.We conducted three sets of experiments across eight datasets,and the results demonstrate that the proposed model effectively mitigates the issues of random augmentation damaging the original graph’s semantics and structure,as well as the insufficiency of contrastive levels.Additionally,the model achieves the best overall performance.
文摘Full ceramic bearings are mission-critical components in oil-free environments,such as food processing,semiconductor manufacturing,and medical applications.Developing effective fault diagnosis methods for these bearings is essential to ensuring operational reliability and preventing costly failures.Traditional supervised deep learning approaches have demonstrated promise in fault detection,but their dependence on large labeled datasets poses significant challenges in industrial settings where fault-labeled data is scarce.This paper introduces a few-shot learning approach for full ceramic bearing fault diagnosis by leveraging the pre-trained GPT-2 model.Large language models(LLMs)like GPT-2,pre-trained on diverse textual data,exhibit remarkable transfer learning and few-shot learning capabilities,making them ideal for applications with limited labeled data.In this study,acoustic emission(AE)signals from bearings were processed using empirical mode decomposition(EMD),and the extracted AE features were converted into structured text for fine-tuning GPT-2 as a fault classifier.To enhance its performance,we incorporated a modified loss function and softmax activation with cosine similarity,ensuring better generalization in fault identification.Experimental evaluations on a laboratory-collected full ceramic bearing dataset demonstrated that the proposed approach achieved high diagnostic accuracy with as few as five labeled samples,outperforming conventional methods such as k-nearest neighbor(KNN),large memory storage and retrieval(LAMSTAR)neural network,deep neural network(DNN),recurrent neural network(RNN),long short-term memory(LSTM)network,and model-agnostic meta-learning(MAML).The results highlight LLMs’potential to revolutionize fault diagnosis,enabling faster deployment,reduced reliance on extensive labeled datasets,and improved adaptability in industrial monitoring systems.